Recent comments in /f/MachineLearning
mycall t1_j8bjo05 wrote
Reply to comment by bballerkt7 in [R] [N] Toolformer: Language Models Can Teach Themselves to Use Tools - paper by Meta AI Research by radi-cho
Progress comes in a multitude of mysterious ways.
thundergolfer t1_j8bjgpu wrote
Reply to comment by PHEEEEELLLLLEEEEP in [D] What ML dev tools do you wish you'd discovered earlier? by TikkunCreation
If you don't have issues then definitely don't bother migrating! Something like Metaflow or Modal is much more built for purpose. Airflow was designed for the Hadoop era of data engineering; it's straining under changes that have happened in the Python, container, and ML ecosystems.
big_gondola t1_j8biqv4 wrote
Reply to comment by belacscole in [R] [N] Toolformer: Language Models Can Teach Themselves to Use Tools - paper by Meta AI Research by radi-cho
I might say we gain general intelligence by creating different models for different tasks and gain experience on when to call which. This has the when to call which, but not the creation of new models.
bballerkt7 t1_j8bimv8 wrote
Reply to [R] [N] Toolformer: Language Models Can Teach Themselves to Use Tools - paper by Meta AI Research by radi-cho
AGI getting closer everyday
JustSomeMemelord t1_j8bi118 wrote
Reply to [D] Quality of posts in this sub going down by MurlocXYZ
Posts like these are no better
master3243 t1_j8bg2wl wrote
It would be amazing if this supports the whisper package directly instead of whisper in transformers from huggingface. (I know this isn't just for whisper but I do really need to speedup whisper)
alayaMatrix t1_j8bg0mk wrote
Reply to comment by DonRybron in [D] Can Google sue OpenAI for using the Transformer in their products? by t0t0t4t4
Wow, I'm really interested at this history, do you have more detailed literature about this?
I_will_delete_myself t1_j8bfq3e wrote
Reply to comment by konrradozuse in [D] Can Google sue OpenAI for using the Transformer in their products? by t0t0t4t4
I disagree about that. Imagine you invest millions of dollars then someone makes millions of it and you lose millions of dollars.
belacscole t1_j8bf6ol wrote
Reply to [R] [N] Toolformer: Language Models Can Teach Themselves to Use Tools - paper by Meta AI Research by radi-cho
I wonder if this is the ultimate path to reaching general intelligence. After all, humans evolved by learning to master tools.
colugo t1_j8be3q6 wrote
Reply to [D] Quality of posts in this sub going down by MurlocXYZ
It's ChatGPT writing about ChatGPT
MurlocXYZ OP t1_j8bdr4q wrote
Reply to comment by dustintran in [D] Quality of posts in this sub going down by MurlocXYZ
Dang it. I was hoping I can get away with not having a Twitter account
dustintran t1_j8bdcv6 wrote
Reply to [D] Quality of posts in this sub going down by MurlocXYZ
r/MachineLearning today has 2.6 million subscribers. The more influx of newcomers the more beginner-friendly posts get upvoted. This is OK—don't get me wrong—it's just a different setting.
Academic discussions were popular back when there were only 50-100K. In fact, I remember in 2017 being in OpenAI offices and every morning, seeing a row of researchers with reddit on their monitor. Discussions mostly happen now on Twitter.
rust_dfdx t1_j8bd9bu wrote
Reply to comment by SwayStar123 in [D] Have their been any attempts to create a programming language specifically for machine learning? by throwaway957280
Hah this was a nice surprise - yes the whole point of dfdx is to do just this!
PHEEEEELLLLLEEEEP t1_j8bac1k wrote
Reply to comment by thundergolfer in [D] What ML dev tools do you wish you'd discovered earlier? by TikkunCreation
Why is airflow bad and can you suggest alternatives? I'm using airflow now with no issues
ferndoll6677 t1_j8b4kz9 wrote
Reply to comment by vzq in [R] DIGIFACE-1M — synthetic dataset with one million images for face recognition by t0ns0fph0t0ns
Most epic game how many questions would be required?
EnjoyableGamer t1_j8b36s2 wrote
Reply to [D] Quality of posts in this sub going down by MurlocXYZ
Not just you, it pivoted with the narrative that existing models will scale and stand the test of time with more data and bigger models.
Chamrockk t1_j8b2gwe wrote
thundergolfer t1_j8b2g6w wrote
> If you're deploying models to production
Airflow is not a good tool for ML development. Leave Airflow back in 2018. Also Modal can do prod model deployment, model pipelines, and inference.
machineko t1_j8b0zyv wrote
Reply to [D] Speed up HuggingFace Inference Pipeline by [deleted]
Are you interested in reducing the latency or just cutting down the cost? Can you run the workload on GPUs instead?
For BERT-type models, doing some compression and using inference libraries can easily get you 5-10x speedup. If interested, I'd be happy to share more resources on this.
leeliop t1_j8azsq4 wrote
Reply to [R] DIGIFACE-1M — synthetic dataset with one million images for face recognition by t0ns0fph0t0ns
When is the NFT drop
ArnoF7 t1_j8azbzj wrote
Reply to [D] Quality of posts in this sub going down by MurlocXYZ
Discussion in this subreddit is always a bit hit and miss. After all, reddit as a community has almost no gate keeping. While this could be a good thing, there are of course downsides to it.
If you look at this post about batch norm, you see that there are people who brought up interesting insights, and there are a good chunk of people who clearly have never even read the paper carefully. And this post is 5 years ago.
[deleted] t1_j8ax03l wrote
Reply to comment by DoxxThis1 in [D] Have their been any attempts to create a programming language specifically for machine learning? by throwaway957280
[removed]
MiuraDude t1_j8aw11t wrote
Qdrant for the vector database and Kern AI refinery for data labeling!
berryaroberry t1_j8aveyl wrote
Reply to [D] Quality of posts in this sub going down by MurlocXYZ
The following is my opinion; so bias is there. My feeling is the sub was never about academic discussions per se. The papers and academic discussions acted like vessels to carry people towards "(deep learning hype + money flow+ industry jobs)" island. In most of the earlier discussions ,if you follow them closely, you will see that there was never really a push for genuine understanding, rather people looking for easy way to earn "publication currency". Initial impression was having some kinda project or publication could land people a high-paying job. Probably later people realized that actually they don't need to worry about papers and stuff, rather doing some kinda quick LLM based project will help to land high-paying jobs even faster. I mean LLMs are currently at the peak of hype. Thus we have more random looking posts.
aicharades OP t1_j8bjsxr wrote
Reply to comment by ShermanTSE in [P] ChatGPT without size limits: upload any pdf and apply any prompt to it by aicharades
You can find map at http://www.wrotescan.com/Map_Only. It will break your document into 12000 character chunks then run the map prompt on each chunk.