Recent comments in /f/MachineLearning
goolulusaurs t1_j8evwnj wrote
Reply to comment by dustintran in [D] Quality of posts in this sub going down by MurlocXYZ
I remember being here in 2017 also and I definitely recall the quality of the post being much higher. Even looking at the sidebar, most of the high quality AMAs from prominent researchers where prior to 2018. Now I often see posts that I would classify as relevant, correct or high quality get downvoted, and posts that seem misinformed or incorrect get upvoted. Personally I blame the reddit redesign for deemphasizing text and discussion in favor of lowest common denominator stuff like eye catching images and video.
UnderstandingDry1256 t1_j8ev9bx wrote
Reply to [R] [N] Toolformer: Language Models Can Teach Themselves to Use Tools - paper by Meta AI Research by radi-cho
An obvious idea is to connect gpt to browser api and let it go and learn 😄
[deleted] t1_j8euer0 wrote
mrfox321 t1_j8esd6k wrote
These are their features:
https://www.cs.princeton.edu/courses/archive/spring21/cos598D/icde_2021_camera_ready.pdf
The paper also references older neural network architectures used in late stages of the recsys stack.
FastestLearner t1_j8esc0c wrote
Reply to [D] Is a non-SOTA paper still good to publish if it has an interesting method that does have strong improvements over baselines (read text for more context)? Are there good examples of this kind of work being published? by orangelord234
Neural networks were not SOTA for a very very long time. The world would be very different if everyone had only published SOTA results improving upon existing SOTAs of the 90s.
[deleted] t1_j8er5mw wrote
Reply to comment by [deleted] in [D] Have their been any attempts to create a programming language specifically for machine learning? by throwaway957280
[deleted]
[deleted] t1_j8er4gk wrote
AdamAlexanderRies t1_j8eppdi wrote
Reply to comment by daking999 in [D] Quality of posts in this sub going down by MurlocXYZ
ChatGPT's mostly a cool toy, but there are some tasks it's genuinely useful for. I use it to explain complex topics, write code, brainstorm ideas, and for fun creative writing exercises. I've only tried the free version, but I am seeing mostly disappointment about the pro version.
Definitely check it out for at least curiosity's sake.
currentscurrents t1_j8em94v wrote
Reply to comment by That_Violinist_18 in The Inference Cost Of Search Disruption – Large Language Model Cost Analysis [D] by norcalnatv
Samsung's working on in-memory processing. This is still digital logic and Von Neumann, but by putting a bunch of tiny processors inside the memory chip, each has their own memory bus they can access in parallel.
Most research on non-Von-Neumann architectures is focused on SNNs. Both startups and big tech are working on analog SNN chips. So far these are proof of concept; they work and achieve extremely low power usage, but they're not at a big enough scale to compete with GPUs.
Swing_Bishop t1_j8elyx8 wrote
Reply to [D] Quality of posts in this sub going down by MurlocXYZ
Maybe they're written by bots?
urbanfoh t1_j8elywk wrote
Reply to comment by EducationalCicada in [R] [N] Toolformer: Language Models Can Teach Themselves to Use Tools - paper by Meta AI Research by radi-cho
Isn't it almost certainly possible due to the universal approximation theorem?
Assuming consciousness is a function of external variables a large enough network with access to these variables should be able to approximate consciousness.
_poisonedrationality t1_j8elkeq wrote
The paper you link does not describe YouTube's algorithm. YouTube's selection algorithm is proprietary and not revealed to the public. You just linked a paper from researchers at Google studying the topic of video recommendation. The extent to which it describes youtube's actual algorithm is not at all obvious.
marcus_hk t1_j8ejn0n wrote
Reply to comment by Reasonable_Ad_6572 in [R] [N] Toolformer: Language Models Can Teach Themselves to Use Tools - paper by Meta AI Research by radi-cho
Which part do you disagree with here:
My unwavering opinion on current (auto-regressive) LLMs
- They are useful as writing aids.
- They are "reactive" & don't plan nor reason.
- They make stuff up or retrieve stuff approximately.
- That can be mitigated but not fixed by human feedback.
- Better systems will come
daking999 t1_j8ejfe2 wrote
Reply to comment by AdamAlexanderRies in [D] Quality of posts in this sub going down by MurlocXYZ
Hmm well now I don't know if I'm talking to you or your bot!
Cool I should check it out. Seems like the free version is already pretty functional?
[deleted] t1_j8ehx5o wrote
Reply to comment by MurlocXYZ in [D] Quality of posts in this sub going down by MurlocXYZ
[deleted]
royalemate357 t1_j8ehkvl wrote
Some researchers at tiktok's parent company released a paper on a recommender system called Monolith here: https://arxiv.org/abs/2209.07663.
I'm not sure its actually what tiktok is using, but they do say that "Monolith has successfully landed in the BytePlus Recommend product".
TMills t1_j8eh61k wrote
Reply to [D] Is a non-SOTA paper still good to publish if it has an interesting method that does have strong improvements over baselines (read text for more context)? Are there good examples of this kind of work being published? by orangelord234
It doesn't need to be sota in an absolute sense, but it should be interesting in an empirical way. If the model is small, it needs to benchmark against other small models. If it's efficient it should compare against other efficient models. If you just like it aesthetically, or think it's clever, then you need to think about what that cleverness buys you and evaluate it in that dimension.
Oat-is-the-Best t1_j8ef5x0 wrote
Reply to comment by ksatriamelayu in [R] [N] Toolformer: Language Models Can Teach Themselves to Use Tools - paper by Meta AI Research by radi-cho
How do you calculate your fitness? That has the same problem of a model not being able to assess its own success
solidavocadorock t1_j8eeikl wrote
Reply to [D] Have their been any attempts to create a programming language specifically for machine learning? by throwaway957280
Julia is a perfect match for scientific computations.
__Maximum__ t1_j8eduow wrote
Reply to comment by TLfanbasit in [D] What ML dev tools do you wish you'd discovered earlier? by TikkunCreation
You can save the page
bballerkt7 t1_j8eddln wrote
Reply to comment by pyepyepie in [R] [N] Toolformer: Language Models Can Teach Themselves to Use Tools - paper by Meta AI Research by radi-cho
No worries I think you definitely have a valid take. I always feel not smart talking about AI stuff lol :)
That_Violinist_18 t1_j8ed3j9 wrote
Reply to comment by currentscurrents in The Inference Cost Of Search Disruption – Large Language Model Cost Analysis [D] by norcalnatv
So should we expect much higher peak throughput numbers from more specialized hardware?
I have yet to hear of any startups in the ML hardware space advertising this.
ksatriamelayu t1_j8ebpx4 wrote
Reply to comment by uristmcderp in [R] [N] Toolformer: Language Models Can Teach Themselves to Use Tools - paper by Meta AI Research by radi-cho
Do people use things like evolutionary fitness + changing environments to describe those quality? Seems dynamic environment might be the answer?
ksatriamelayu t1_j8ebhn4 wrote
Reply to [R] [N] Toolformer: Language Models Can Teach Themselves to Use Tools - paper by Meta AI Research by radi-cho
Keep in mind that our current theories in Neuroscience broadly agrees something similar is going on with mammalian, even reptilian brains. Hell, maybe even worm brains.
There's autonomous systems everywhere that calls each other for updates and in some certain brains, enough complexity that something that can called thinking occurs.
Practically, offloading calculations to a python REPL, machine translation to GTranslate API call, and knowledge search to Wikipedia corpus is going to let LLMs do what they do best - mask users intent and generate believable enough corpus. Let the facts stay factual and the hallucination stay hallucination.
impossiblefork t1_j8ewogm wrote
Reply to comment by berryaroberry in [D] Quality of posts in this sub going down by MurlocXYZ
I talked research with researchers here, partially in PM, but some of it openly.
I'm sure many others did too. The current problem is something new and which has come during the past few days.