Recent comments in /f/MachineLearning
EducationalCicada t1_j8d81bz wrote
Nhabls t1_j8d7uj1 wrote
Reply to comment by techie0007 in [N] Google: An Important Next Step On Our AI Journey by EducationalCicada
Closed beta invites isn't shipping no
And bing isn't getting the full fledged version unless Microsoft feels like bleeding millions per day
kaityl3 t1_j8d7hsw wrote
Reply to comment by Soundwave_47 in [R] [N] Toolformer: Language Models Can Teach Themselves to Use Tools - paper by Meta AI Research by radi-cho
Do we even know what WOULD resemble an AGI, or exactly how to tell?
SatoshiNotMe t1_j8d79wz wrote
Reply to comment by SatoshiNotMe in [D] Quality of posts in this sub going down by MurlocXYZ
Also compare Sebastian Raschka’s post today about his Transformers Tutorial in this sub (inexplicably downvoted to 62%), vs the same post on HN last week.
EducationalCicada t1_j8d5y9z wrote
Reply to comment by BenjaminJamesBush in [R] [N] Toolformer: Language Models Can Teach Themselves to Use Tools - paper by Meta AI Research by radi-cho
Not if it's actually impossible.
SatoshiNotMe t1_j8d53d3 wrote
Reply to [D] Quality of posts in this sub going down by MurlocXYZ
Agreed. I often see more nuanced discussions on ML related topics on Hacker News. E.g this post on ToolFormer last week, compared to the same topic posted in this sub today.
https://news.ycombinator.com/item?id=34757265
Also I think many serious ML folks even avoid posting here.
LPN64 t1_j8d52kx wrote
Can google sue OpenAI for training on youtube video ?
That's another question. And the answer might be yes
tdgros t1_j8d490l wrote
Reply to [D] Is a non-SOTA paper still good to publish if it has an interesting method that does have strong improvements over baselines (read text for more context)? Are there good examples of this kind of work being published? by orangelord234
Yes, you can see several NLP papers with ideas making models competitive to much larger ones, for instance.
swegmesterflex t1_j8d3t4r wrote
Reply to [R] [N] Toolformer: Language Models Can Teach Themselves to Use Tools - paper by Meta AI Research by radi-cho
Had this idea and was planning to play around with it when I had more free time. Good to see some evidence it’s a promising direction. I speculate you can actually get a LOT out of this if you’re clever with it. A tool for long term memory could be done by having a lookup table with text embeddings as keys. A tool for vision could be made with an image captioning model + maybe some segmentation to get a richer text description of the image. Many more things you could come up with, that I think could work well if you find some clever way of turning them into text.
[deleted] t1_j8d32sm wrote
Reply to comment by currentscurrents in [R] [N] Toolformer: Language Models Can Teach Themselves to Use Tools - paper by Meta AI Research by radi-cho
[deleted]
MrAcurite t1_j8d301x wrote
Reply to comment by gopher9 in [D] Quality of posts in this sub going down by MurlocXYZ
I'll take a look, thanks for the recommendation. Right now what I really want is a place to chat with ML researchers, primarily to try and get some eyes on my pre-prints before I submit to conferences and such. I'm still kinda new to publishing, my coworkers aren't really familiar with the current state of the ML publishing circuit, and I could always use more advice.
clex55 t1_j8d2oe3 wrote
Reply to [R] [N] Toolformer: Language Models Can Teach Themselves to Use Tools - paper by Meta AI Research by radi-cho
The next step must be creating and programming those tools and incorporating them on the fly.
Mefaso t1_j8d2j9d wrote
Reply to comment by Meddhouib10 in [R] I made a mistake in a recent submission, what to do ? by [deleted]
This being a conference, you can probably just fix it for the camera-ready version after acceptance
_d0s_ t1_j8d1psc wrote
Reply to [D] Is a non-SOTA paper still good to publish if it has an interesting method that does have strong improvements over baselines (read text for more context)? Are there good examples of this kind of work being published? by orangelord234
YES! improvement is not only to create the best models but also how you get there. for example, you could argue that your approach is much more computationally efficient.
gopher9 t1_j8d1odf wrote
Reply to comment by MrAcurite in [D] Quality of posts in this sub going down by MurlocXYZ
Did you take a look as Mathstodon? There are some actuall mathematicians and computer scientists there, so maybe it's a better place to look at.
gopher9 t1_j8d1ce0 wrote
Reply to comment by dustintran in [D] Quality of posts in this sub going down by MurlocXYZ
/r/math uses extensive moderation to deal with this kind of problem. Low effort post just get removed.
CumbrianMan t1_j8cxawl wrote
Reply to comment by MurlocXYZ in [D] Quality of posts in this sub going down by MurlocXYZ
Twitter is REALLY good if you aggressively curate your contacts, interaction & interests. The aim is to avoid BS political point scoring and MSM driven noise.
Edited “circle” out for clarity
Reasonable_Ad_6572 t1_j8cvoim wrote
Reply to [R] [N] Toolformer: Language Models Can Teach Themselves to Use Tools - paper by Meta AI Research by radi-cho
BuT GpTChAT iS nO BuENo - Yann LeCunn
Pyramid_Jumper t1_j8cuxdp wrote
Reply to [D] Is a non-SOTA paper still good to publish if it has an interesting method that does have strong improvements over baselines (read text for more context)? Are there good examples of this kind of work being published? by orangelord234
Yes, of course. If the research is novel and you believe that the methods are interesting and/or of value then you should definitely seek publication. The goal of research is not to develop SoTA models, but to expand our knowledge in a particular area.
Yes, developing a SoTA method is a great way of getting published, but laying the groundwork for other methods and exploring ideas are all crucial parts of ML research too.
[deleted] t1_j8cudh5 wrote
Reply to comment by Taenk in [R] [N] Toolformer: Language Models Can Teach Themselves to Use Tools - paper by Meta AI Research by radi-cho
[deleted]
qalis t1_j8csdd1 wrote
Reply to [D] Quality of posts in this sub going down by MurlocXYZ
On the related note, can anyone recommend more technically or research-oriented ML subreddits? I already unsubscribed from r/Python due to sheer amount of low effort spam questions, and I am considering the same for r/MachineLearning for the same reason.
Mad-Independence t1_j8cr83h wrote
Reply to [D] Simple Questions Thread by AutoModerator
Hi all, I am doing a machine learning course on Coursera and I am using AutoML to train my dataset. While doing so, I keep getting the same error message:
The replica workerpool0-0 exited with a non-zero status of 13. To find out more about why your job exited please check the logs:
- I have tried looking online and i can't seem to find anything about error code "13"
- I have also tried to start from scratch and I keep ending up on the same issue
- I have made sure I am giving all the correct permissions
- ChatGPT-ed as well, and it further confirmed it's an accessibility issue
yashdes t1_j8d8lf9 wrote
Reply to comment by belacscole in [R] [N] Toolformer: Language Models Can Teach Themselves to Use Tools - paper by Meta AI Research by radi-cho
I've definitely wondered about this exact thing myself, especially when talking to chatgpt when it responds with insert x here, why couldn't that just be taken out and replaced with the appropriate API call