Recent comments in /f/MachineLearning
chief167 t1_j9ku5mq wrote
Reply to comment by activatedgeek in [D] "Deep learning is the only thing that currently works at scale" by GraciousReformer
I don't think it implies that all datasets are equally likely. I think it only implies that given all possible datasets, there is no best approach to modelling them. All possible != All are equally likely
But I don't have my book with me, and I do t trust the internet since it seems to lead to random blogposts instead of the original paper (Wikipedia gave a 404 in the footnotes)
chief167 t1_j9kt5ho wrote
Reply to comment by Featureless_Bug in [D] "Deep learning is the only thing that currently works at scale" by GraciousReformer
We use gradient boosting at quite a big scale. Not LLM big, but still big. It's just not NLP or CV at all. It's for fraud detection in large transactional tabular datasets. And it outperforms basically all neural network, shallow or deep, approaches.
IdentifiableParam t1_j9ksw3a wrote
Really interesting work. Might be worth doing even if courts decide this isn't going to work as a legal defense.
[deleted] t1_j9ksmp8 wrote
Reply to [P] MIT Introduction to Data-Centric AI by anishathalye
[removed]
vyasnikhil96 OP t1_j9ksj4v wrote
Reply to comment by bluemason in [R] Provable Copyright Protection for Generative Models by vyasnikhil96
We already handle that as our notion is not based on reproduction but is rather a information-theoretic notion. We also have a parameter that measures how much information we have "reproduced" vs adapted which can be set depending on the underlying models and the use case.
bluemason t1_j9kqrdg wrote
Reply to comment by vyasnikhil96 in [R] Provable Copyright Protection for Generative Models by vyasnikhil96
Not a lawyer, but yes.
Copyright infringement goes beyond reproduction of a work. An original adaptation can also violate copyright.
kermunnist t1_j9kpp3s wrote
Reply to comment by __lawless in [R] Multimodal Chain-of-Thought Reasoning in Language Models - Amazon Web Services Zhuosheng Zhang et al - Outperforms GPT-3.5 by 16% (75%->91%) and surpasses human performance on ScienceQA while having less than 1B params! by Singularian2501
I wonder how flamingo would compare
howtorewriteaname t1_j9kmbrv wrote
There's no mathematical formulation of that statement because there's no mathematical reasoning behind that statement. It's just an opinion (which I believe it isn't true btw)
hpstring t1_j9kkkr0 wrote
Interesting research! Let's try to convince artists with this kind of work
relevantmeemayhere t1_j9kin48 wrote
Reply to comment by adventuringraw in [D] "Deep learning is the only thing that currently works at scale" by GraciousReformer
I agree with you. I was just pointing out that to say they are the only solution is foolish, as the quote implied
This quote could have just been used without much context, so grain of salt.
throwaway2676 t1_j9kilst wrote
Reply to [D] Simple Questions Thread by AutoModerator
Are there any developments in the ASIC/analog computing space that people are really excited about? I think most people know about google's TPUs by now, but is there anything else with the potential to threaten the dominance of GPUs in the next few years?
-vertigo-- t1_j9kih7k wrote
Reply to comment by activatedgeek in [D] "Deep learning is the only thing that currently works at scale" by GraciousReformer
hmm for some reason the arxiv links are giving 403 forbidden
relevantmeemayhere t1_j9kifhu wrote
Reply to comment by VirtualHat in [D] "Deep learning is the only thing that currently works at scale" by GraciousReformer
Linear models are often preferred for the reasons you mentioned. Under fitting is almost always preferred to overfitting.
relevantmeemayhere t1_j9ki2x1 wrote
Reply to comment by Featureless_Bug in [D] "Deep learning is the only thing that currently works at scale" by GraciousReformer
Because they are useful for some problems and not others, like every algorithm? Nowhere in my statement did I say they are monolithic in their use across all subdomains of ml
The statement was that deep learning is the only thing that works at scale. It’s not lol. Deep learning struggles in a lot of situations.
relevantmeemayhere t1_j9khp8m wrote
Reply to comment by VirtualHat in [D] "Deep learning is the only thing that currently works at scale" by GraciousReformer
As you mentioned, this is highly dependent on the functional relationship of the data.
You wouldn’t not use domain knowledge to determine that.
Additionally, non linear models tend to have their own drawbacks. Lack of interpretability, high variability being some of them
hpstring t1_j9kf6te wrote
Reply to comment by activatedgeek in [D] "Deep learning is the only thing that currently works at scale" by GraciousReformer
This is a very good answer! I want to add that apart from generalization, the fact that we have efficient optimization algorithms that can find quite good minima also contributes a lot to the deep learning magic.
LetterRip t1_j9ker51 wrote
Reply to [D] Faster Flan-T5 inference by _learn_faster_
See this tutorial - converts to ONXX CPU, then to tensor-RT for a 3-6x speedup.
https://developer.nvidia.com/blog/optimizing-t5-and-gpt-2-for-real-time-inference-with-tensorrt/
[deleted] t1_j9kdvxd wrote
Reply to comment by GraciousReformer in [D] "Deep learning is the only thing that currently works at scale" by GraciousReformer
[deleted]
[deleted] t1_j9kbzaz wrote
[removed]
thomasahle OP t1_j9kapw7 wrote
Reply to comment by ChuckSeven in Unit Normalization instead of Cross-Entropy Loss [Discussion] by thomasahle
Even with angles you can still have exponentially many vectors that are nearly orthogonal to each other, if that's what you mean...
I agree the representations will be different. Indeed one issue may be that large negative entries will be penalized as much as large positive ones, which is not the case for logsumexp...
But on the other hand more "geometric" representations like this, based on angles, may make the vectors more suitable for stuff like LSH.
Disastrous_Nose_1299 OP t1_j9kak0o wrote
Reply to comment by Blakut in [Discussion] Exploring the Black Box Theory and Its Implications for AI, God, and Ethics by Disastrous_Nose_1299
thank you.
Disastrous_Nose_1299 OP t1_j9kai85 wrote
Reply to comment by Blakut in [Discussion] Exploring the Black Box Theory and Its Implications for AI, God, and Ethics by Disastrous_Nose_1299
"Ok, but you are going about it the wrong way. The possibility of god existing or not is irrelevant to what we can know about his existence. If something is unkowable, then any categorical statements about it are invalid. Yes, we can consider the possibility, but if you can't ever tell if it's true or not, this approach makes no sense.
we do not have enough evidence to suggest that god exists or does not exist in a black hole.
And what i'm saying is that since we will never have that evidence, no matter what, it is irrelevant to approach the problem from this angle."
I respect your position, but this is where we are going to have to agree to disagre, Ive got to shovel some ones drive way. It was nice talking to you.
Blakut t1_j9kagtv wrote
Reply to comment by Disastrous_Nose_1299 in [Discussion] Exploring the Black Box Theory and Its Implications for AI, God, and Ethics by Disastrous_Nose_1299
alright have fun
Blakut t1_j9ka6sn wrote
Reply to comment by Disastrous_Nose_1299 in [Discussion] Exploring the Black Box Theory and Its Implications for AI, God, and Ethics by Disastrous_Nose_1299
>But the point i am trying to make is that it is a possibility that God exists.
Ok, but you are going about it the wrong way. The possibility of god existing or not is irrelevant to what we can know about his existence. If something is unkowable, then any categorical statements about it are invalid. Yes, we can consider the possibility, but if you can't ever tell if it's true or not, this approach makes no sense.
>we do not have enough evidence to suggest that god exists or does not exist in a black hole.
And what i'm saying is that since we will never have that evidence, no matter what, it is pointless to approach the problem from this angle.
Featureless_Bug t1_j9kuu22 wrote
Reply to comment by chief167 in [D] "Deep learning is the only thing that currently works at scale" by GraciousReformer
Large scale is somewhere to the north of 1-2 TB of data. Even if you had that much data, in absolutely most cases tabular data has such a simplistic structure that you wouldn't need that much data to achieve the same performance - so I wouldn't call any kind of tabular data large scale to be frank