Recent comments in /f/MachineLearning
[deleted] t1_j7gfa1b wrote
Reply to comment by supersoldierboy94 in [D] Yann Lecun seems to be very petty against ChatGPT by supersoldierboy94
[removed]
NoLifeGamer2 t1_j7geyw5 wrote
I love how bloom was just like "F*ck it let's one-up openAI"
po-handz t1_j7gemeh wrote
Reply to comment by blablanonymous in [N] GitHub CEO on why open source developers should be exempt from the EU’s AI Act by EmbarrassedHelp
I'm curious how much you've interacts with the homeless? Any soup kitchens or charity events? There's maybe a 1 out of 50 chance you come across some one who's well put together, education, has a job, but is just a few bucks short each months
those aren't the people waiting in line at the shelter
noone_relevant t1_j7gamsz wrote
Reply to High-speed cameras and deep learning [Research] by A15L
You said it yourself in the question. What is the potential application? Also does it have to be online? If not even though the camera is high speed it is similar to other camera for deep learning
maxip89 t1_j7g95yt wrote
Reply to comment by BrisklyBrusque in [N] "I got access to Google LaMDA, the Chatbot that was so realistic that one Google engineer thought it was conscious. First impressions" by That_Violinist_18
Does this proof something? If I fire you does that proof that you worked on atomic missiles?
BrisklyBrusque t1_j7g8v7a wrote
Reply to comment by maxip89 in [N] "I got access to Google LaMDA, the Chatbot that was so realistic that one Google engineer thought it was conscious. First impressions" by That_Violinist_18
I don’t think it was marketing. The individual who claimed the AI was sentient was dismissed for his actions.
MysteryInc152 t1_j7g83pw wrote
Reply to comment by Cheap_Meeting in [D] List of Large Language Models to play with. by sinavski
GLM-130B is really really good. https://crfm.stanford.edu/helm/latest/?group=core_scenarios
I think some instruction tuning is all it needs to match the text-davinci models
_d0s_ t1_j7g250y wrote
Reply to High-speed cameras and deep learning [Research] by A15L
Not the same, but I'd suggest to look into event cameras
mrpogiface t1_j7g03gj wrote
Do we actually know that chatGPT is the full 175B? With codex being 13B and still enormously powerful, and previous instruction tuned models (in the paper) being 6.7B it seems likely that they have it working on a much smaller parameter count
CriticalTemperature1 t1_j7fzacn wrote
Google has their AI Test Kitchen for LaMDA
MysteryInc152 t1_j7fymgb wrote
Reply to comment by Cheap_Meeting in [D] List of Large Language Models to play with. by sinavski
don't think so
mostlyhydrogen OP t1_j7fydvb wrote
Reply to comment by YOLOBOT666 in [D] Querying with multiple vectors during embedding nearest neighbor search? by mostlyhydrogen
The goal is to harvest training data for ML. If there is a difficult edge case the model is struggling with, the best way to improve model performance is to harvest additional training data for that edge case. You stop when the model performance meets your requirements.
mostlyhydrogen OP t1_j7fxwyx wrote
Reply to comment by RingoCatKeeper in [D] Querying with multiple vectors during embedding nearest neighbor search? by mostlyhydrogen
>ScaNN interface features
Nope. Notice that the results have shape (10000, 20) instead of (20,). That is just doing a batched query i.e. "for each of these 10k input vectors, find me 20 neighbors". What I need is a joint query, i.e. "given these 10k positive examples, give me an additional 20 candidate samples".
gopher9 t1_j7fq4kw wrote
Reply to comment by m98789 in [D] List of Large Language Models to play with. by sinavski
With RWKV-4-Pile-14B-20230204-7324.pth released 2 hours ago, as you can see at https://huggingface.co/BlinkDL/rwkv-4-pile-14b/tree/main.
But yeah, it's still WIP.
DrHaz0r t1_j7fmzlx wrote
I think you are all missing the point that Schmidhuber has basically invented Chat-GPT in the 90s already. Just with smaller networks, smaller datasets and less compute power. Although even Schmidhuber basically just stole from Gauss, but no one talks about that.
supersoldierboy94 OP t1_j7fjwzd wrote
Reply to comment by PredictorX1 in [D] Yann Lecun seems to be very petty against ChatGPT by supersoldierboy94
i mean, you can see others using argument of authority for sure
supersoldierboy94 OP t1_j7fhasq wrote
Reply to comment by PredictorX1 in [D] Yann Lecun seems to be very petty against ChatGPT by supersoldierboy94
i mean, you can see argument of authority from other people
PredictorX1 t1_j7ffj8j wrote
>I get that he is one of the godfathers of AI.
What does that even mean? Very many people have contributed to this field.
m98789 t1_j7ffidp wrote
Reply to comment by gopher9 in [D] List of Large Language Models to play with. by sinavski
(final release around Feb-15-2023):
supersoldierboy94 OP t1_j7fcbh3 wrote
Reply to comment by bacon_boat in [D] Yann Lecun seems to be very petty against ChatGPT by supersoldierboy94
Meta is a leader in the research community alongside Google as top contributors. The funny thing is that he started posting that graph of AI related paper contributions to show supremacy and to undermine OpenAI and DeepMind as merely consumers of research. But Meta hasnt provided any product from their research that has reached the public. When they tried, they immediately shut it down.
He also kinda blames the public perception as to why Meta cannot publish products without scrutiny pointing the thing that people are still overly criticizing Facebook/Meta for obviously great reasons in the past.
It is indeed a massive milestone maybe a bit above Stable Diffusion. I'd still argue that Github Copilot was bigger but since its mainly for devs, it didnt get the publicity that it wanted. It's a massive milestone because common folks pondered the idea of AI takeover which have shifted every one else's perspective on the domain. It's the culmination of decades of R&D that the public can interact to -- a gateway to AI and its complexities.
Common folks and the public do not really care about sophisticated algos that never see the light of day.
beezlebub33 t1_j7fbjii wrote
"Henry Ford did nothing revolutionary, the engineering work in making a car isn't particularly difficult, it's just perceived that way by the public. There will be a half dozen other car manufacturers in 6 months."
LeCun is going too far the opposite way. I would not be surprised if he has access to systems at FAIR that could do something similar, so dismisses the whole thing or misses the main point. But, like Ford, what OpenAI has done with Dalle2 and ChatGPT is make AI useable and available to us benighted common folk.
It doesn't matter whether Google and Meta not releasing something like this is due to a can't or a won't. It's all the same to the rest of humanity who can't use it in either case.
bacon_boat t1_j7fbjcg wrote
I think his view reflects his disappointment as a researcher that it's not novel ideas and algorithms that lead to success. It's scale + engineering.
But anyone with a broader view sees that ChatGPT represents a massive milestone for AI.
Who really cares how novel the algorithms are, openAI built a killer product, and deserve the recognition.
Lecun is maybe also salty because Deepmind / OpenAI are perceived as leaders, and Meta isn't.
[deleted] t1_j7f9vm6 wrote
The perfect language would be Mentalese.
[deleted] t1_j7f845z wrote
Reply to comment by 42gauge in [N] OpenAI starts selling subscriptions to its ChatGPT bot by bikeskata
[removed]
dataslacker t1_j7gfa6c wrote
Reply to [D] Yann Lecun seems to be very petty against ChatGPT by supersoldierboy94
There’s probably some resentment that google and meta could have released something similar over a year ago but chose not to because they didn’t think it would be responsible. Now the company that was founded on being “responsible” released it to the world it a way that hasn’t satisfied a lot of researchers.