Recent comments in /f/MachineLearning
NotActual t1_j916pb7 wrote
Reply to [D] Please stop by [deleted]
Ironically, ChatGPT might make a decent automod!
dojoteef t1_j916kx3 wrote
Reply to [D] Please stop by [deleted]
See previous discussion: https://old.reddit.com/r/MachineLearning/comments/110swn2/d_quality_of_posts_in_this_sub_going_down
LcuBeatsWorking t1_j916fye wrote
Reply to [D] Please stop by [deleted]
I agree.
All subs related to AI or ML appear to get flooded with this stuff right now.
schwagggg t1_j916cc5 wrote
Reply to comment by Oripy in [D] Simple Questions Thread by AutoModerator
so actor critic without critic is just policy gradient/reinforce/score function gradient, first two names used in RL, last one used in stats/OR.
short answer is policy gradient tends to have high variances empirically, so people use control variates to control its variance, and the critic is simply the control variate.
high variance methods usually converge to worse local minimas than low variance ones. u can verify this by taking or the critic function entirely. try it itself with that tutorial
CacheMeUp t1_j915ffp wrote
Breaking the security-by-required-effort assumption of various human interactions, especially among strangers.
It used to take effort to voice opinions on social media and other mass-communication platform, making the public trust that these are authentic messages representing real people. The scalability of this technology breaks that assumption. This has started before, and LLMs take it to a whole new level.
abstractcontrol t1_j915epu wrote
Reply to comment by I_will_delete_myself in [N] Google is increasing the price of every Colab Pro tier by 10X! Pro is 95 Euro and Pro+ is 433 Euro per month! Without notifying users! by FreePenalties
It is best to use a script in order to get a Paperspace notebook. Otherwise, yeah, you are going to have a hard time sometimes. The availability does depend on the timezone from what I've heard.
Cherubin0 t1_j915ayg wrote
That only the people in power are allowed to use AI while the rest is not. Like some kind if AI aristocrats. But this will probably happen when the regulations come.
schwagggg t1_j914xct wrote
Reply to comment by aCuRiOuSguuy in [D] Simple Questions Thread by AutoModerator
can you share the syllabus and some of the early assignments?
Berzerka t1_j913vha wrote
Reply to comment by Kitchen_Tower2800 in [D] Lion , An Optimizer That Outperforms Adam - Symbolic Discovery of Optimization Algorithms by ExponentialCookie
Ever heard of large models?
Ulfgardleo t1_j912luv wrote
computer vision is a much broader problem domain than text to image or text to video. AFAIK 3D pose estimation under occlusions is an unsolved problem, still.
[deleted] t1_j911942 wrote
Reply to comment by baffo32 in [D] HuggingFace considered harmful to the community. /rant by drinkingsomuchcoffee
[deleted]
WarAndGeese t1_j90z7bb wrote
Shoutout to /r/huggingface/
GlobalMammoth t1_j90xs6c wrote
Reply to [D] Short survey of optimization methods by medwatt
Here is a playlist of Constantine Caramanis at UT Austin that covers optimization theory in depth:
https://youtube.com/playlist?list=PLXsmhnDvpjORzPelSDs0LSDrfJcqyLlZc
baffo32 t1_j90uucx wrote
Reply to comment by hpstring in [D] HuggingFace considered harmful to the community. /rant by drinkingsomuchcoffee
it’s important if you’re publishing large software packages of course lots of hobbyists also learn in the field
ml-research t1_j90suwh wrote
Finding open problems
East-Beginning9987 t1_j90q25r wrote
Reply to [D] Coauthor Paper? by [deleted]
From my experience in interviewing for pre doc programs, people down weigh your controbution to the paper if youre not the first author and often you have to explicitly make sure to let them know what you did exactly , etc.however I think having a good contribution from your side would eventually lead to you being able to talk more clearly about what you did and show the interviewer that you know stuff. Since you’re in second year, i guess you will be able to work on first author papers which would then strengthen having second or later author papers.
PrestigiousCloud9 t1_j90q0hq wrote
Reply to [D] Simple Questions Thread by AutoModerator
I am currently a student learning ML. I have done some courses on Machine Learning and I know the theoretical part of it i.e. about algorithms like Random Forest , Decision Tree , SVM , KNN etc. But I want to work of some ML projects how should I start ? How can I gain practical knowledge of ML ? How can I make projects to improve my resume ? I particularly struggle to identify which ML algorithm is suitable for a particular problem statment?
ilovethrills t1_j90noyx wrote
Reply to comment by currentscurrents in [D] What are the worst ethical considerations of large language models? by BronzeArcher
But that can be said on paper for thousands of things. Not sure if it actually translates in real life. Although there might be some push to label such content as AI generated, similar to how "Ad" and "promoted" are labelled in results.
bjergerk1ng OP t1_j90mnb7 wrote
Reply to comment by anonymousTestPoster in [D] Formalising information flow in NN by bjergerk1ng
He linked https://arxiv.org/abs/1905.04271, not sure what is happening lol.
filipposML t1_j90m8x0 wrote
Reply to [D] Formalising information flow in NN by bjergerk1ng
Maybe you are interested in Tishby's rate distortion. E.g. in this paper they do an analysis of the behaviour of mutual information in the hidden layers as a neural network is trained to convergence.
farmingvillein t1_j90m0ab wrote
Reply to comment by adt in [D] Compare open source LLMs by President_Xi_
> For models, see my up-to-date list of models:
Which tab is germane to OP's request?
> but I am specifically refering to performance after finetuning.
So far as I can tell, there is nothing here that is responsive to OP's query. But there is a lot here--perhaps I read too quickly.
a1_jakesauce_ t1_j90k4h6 wrote
Reply to comment by NotARedditUser3 in [D] What are the worst ethical considerations of large language models? by BronzeArcher
All I found was this https://twitter.com/kliu128/status/1623472922374574080?s=21
NotARedditUser3 t1_j90j0er wrote
Reply to comment by a1_jakesauce_ in [D] What are the worst ethical considerations of large language models? by BronzeArcher
If you spend some time looking up how microsoft's gpt integrated chat / ai works, it does this. Lookup the thread of tweets for the hacker that exposed its internal codename 'Syndey'; it scrapes his twitter profile, realizes he exposed its secrets in prior convo's after social engineering hacking it with a few conversations, and then turns hostile to him.
a1_jakesauce_ t1_j90iv0q wrote
Reply to comment by NotARedditUser3 in [D] What are the worst ethical considerations of large language models? by BronzeArcher
This describes a LLM + reinforcement learning hybrid that has been trained to navigate webpages for arbitrary tasks. I’m not sure how far away this is, or if it already exists. Someone below mentioned an action transformer which may be related
[deleted] t1_j917vlr wrote
Reply to comment by NotARedditUser3 in [D] What are the worst ethical considerations of large language models? by BronzeArcher
[deleted]