Recent comments in /f/MachineLearning
I_will_delete_myself OP t1_j95u7e7 wrote
Reply to comment by RideOrDieRemember in [D] Things you wish you knew before you started training on the cloud? by I_will_delete_myself
Aws isn't the only one doing spot instances
Lust4Me t1_j95u5zl wrote
Reply to [R] neural cloth simulation by LegendOfHiddnTempl
relevant to post from earlier this week:
MysteryInc152 OP t1_j95u3t2 wrote
Reply to comment by Professor_Entropy in [D] Toolformer implementation using only few-shot prompting by MysteryInc152
Seems like something a chain of thought example in the pre prompt would fix more than any deficiency in the approach.
Also eliminating arithmetic errors doesn't mean you'd eliminate logical/reasoning errors.
Professor_Entropy t1_j95txkb wrote
You can't still rely on its results. "What's the volume of 1000 KG of Ice?" doesn't work, the model asks "1000 * 0.919" instead of "1000 / 0.919"
MysteryInc152 OP t1_j95rp8c wrote
Reply to comment by Taenk in [D] Toolformer implementation using only few-shot prompting by MysteryInc152
Taenk t1_j95rfg2 wrote
Can you please link the demo without going through twitter? It won’t load for me.
MysteryInc152 OP t1_j95r8ni wrote
Reply to comment by ilovethrills in [D] Toolformer implementation using only few-shot prompting by MysteryInc152
Much simpler approach compared to langchain ( and this is self supervised) but they attempt to do the same thing.
tdls_to t1_j95pqh9 wrote
Reply to comment by iacolippo in [D] Does langchain upload all user’s data to Openai? by westeast1000
according to open AI TOS you can send them an email to opt out of them using your prompts for model training because the right to the "content" you send to their API is perpetually yours. they also license the outputs to you to use as see fit (since you're paying them to use the service). so, on paper, you can use it for "serious" purposes without an issue. that said, the legal aspects of this whole thing is still work in progress and I strongly suggest you discuss the implications with your internal legal team before sending any sensitive company info
ilovethrills t1_j95p6p7 wrote
Is this like langchain?
LegendOfHiddnTempl OP t1_j95ok8m wrote
Reply to [R] neural cloth simulation by LegendOfHiddnTempl
>We present a general framework for the garment animation problem through unsupervised deep learning inspired in physically based simulation. Existing trends in the literature already explore this possibility. Nonetheless, these approaches do not handle cloth dynamics. Here, we propose the first methodology able to learn realistic cloth dynamics unsupervisedly, and henceforth, a general formulation for neural cloth simulation. The key to achieve this is to adapt an existing optimization scheme for motion from simulation based methodologies to deep learning. Then, analyzing the nature of the problem, we devise an architecture able to automatically disentangle static and dynamic cloth subspaces by design. We will show how this improves model performance. Additionally, this opens the possibility of a novel motion augmentation technique that greatly improves generalization. Finally, we show it also allows to control the level of motion in the predictions. This is a useful, never seen before, tool for artists. We provide of detailed analysis of the problem to establish the bases of neural cloth simulation and guide future research into the specifics of this domain. arxiv.org
>
>github.com/hbertiche/NeuralClothSim
overactor t1_j95oem0 wrote
Reply to comment by KPTN25 in [D] Please stop by [deleted]
Your ridiculous hyperbole is not helping your argument. It's entirely possible that sentience is an instrumental goal for achieving a certain level of text prediction. And I don't see why a sufficiently large LLM definitely couldn't achieve it. It could be that another few paradigm shifts will be needed, but it could also be an we need to do is scaling up. I think anyone who claims to know if LLMs can achieve sentience is either ignorant or lying.
guaranteednotabot t1_j95m1yr wrote
Reply to [D] Simple Questions Thread by AutoModerator
How is the cost of queries to AI tools such as ChatGPT determined?
Sorry for the beginner question, but I keep seeing numbers such as 2 cents per query being quoted for a ChatGPT query.
How much processing power is required to complete a query? Does it scale with the number of parameters - or does number of parameters only affect memory usage?
nanashi500 t1_j95l3xx wrote
Reply to [D] Please stop by [deleted]
You can’t ask for this to stop because:
- Not everybody is knowledgeable
- Not everybody is smart
That being said, the questions can become a bother to answer over time, so I just pick and choose if and when I want to respond.
KPTN25 t1_j95kx5j wrote
Reply to comment by overactor in [D] Please stop by [deleted]
Reproducing language is a very different problem than true thought or self-awareness, is why.
LLMs are no more likely to become sentient than a linear regression or random forest model. Frankly, they're no more likely than a peanut butter sandwich to achieve sentience.
Is it possible that we've bungled our study of peanut butter sandwiches so badly that we may have missed some incredible sentience-granting mechanism? I guess, but it's so absurd and infinitesimal it's not worth considering or entertaining practically.
The black box argument is intellectually lazy. We have a better understanding of what is happening in LLMs and other models than most clickbaity headlines imply.
badabummbadabing t1_j95kmxk wrote
Reply to comment by MysteryInc152 in [D] Toolformer implementation using only few-shot prompting by MysteryInc152
This is absolutely wild.
westeast1000 OP t1_j95j8by wrote
Reply to comment by iacolippo in [D] Does langchain upload all user’s data to Openai? by westeast1000
I see. So i guess its unusable for any serious use cases in companies due to regulations and policies. Thanks for the response I thought i was missing something because I never see people question this
__lawless t1_j95ixov wrote
Reply to [D] Things you wish you knew before you started training on the cloud? by I_will_delete_myself
Use code-server (VS code in browser) it is amazing
overactor t1_j95hrop wrote
Reply to comment by KPTN25 in [D] Please stop by [deleted]
I really don't think you can say that with such confidence. If you were saying they no existing LLMs have achieved sentience and they can't at the scale we're working today, I'd agree, but I really don't see how you can be so sure that increasing the size and training data couldn't result in sentience somewhere down the line.
Mefaso t1_j95hl4n wrote
Reply to comment by RideOrDieRemember in [D] Things you wish you knew before you started training on the cloud? by I_will_delete_myself
Maybe try different regions?
Mefaso t1_j95hjkm wrote
Reply to comment by Demortus in [D] Things you wish you knew before you started training on the cloud? by I_will_delete_myself
>Running Linux on your desktop/laptop makes it significantly easier to run projects on the cloud
Just as a note, this can easily be done in a docker consider on windows as well.
gopher9 t1_j95hafv wrote
Reply to [D] Lack of influence in modern AI by I_like_sources
Neural networks are by design black boxes. You get great performance in exchange of explainability. This does not mean though that you have no control over the result.
> Example Stable Diffusion. You don't like what the eyes look like, yet you don't know how to make them more realistic.
ControlNet allows to guide image generation: https://github.com/lllyasviel/ControlNet.
> Example NLP. The chatbot does not give you logical answers? Try another random model.
Or give it some examples and ask to reason step by step. Alternatively, finetune it on examples. You can also teach LLM to use external tools, thus avoiding using LLM for reasoning.
sneakpeekbot t1_j95gyjw wrote
Reply to comment by DigThatData in [D] Please stop by [deleted]
Here's a sneak peek of /r/MLQuestions using the top posts of the year!
#1: I've recorded over 1500 farts to train a model to recognize farts. Who or how do I share the dataset with to be more available to anyone who may find it useful for audio tasks?
#2: Does anyone here use newer or custom frameworks aside from TensorFlow, Keras and PyTorch?
#3: I 25f want to get into AI research/Engineering - but I’m a administrative assistant w a theatre/philosophy degree
^^I'm ^^a ^^bot, ^^beep ^^boop ^^| ^^Downvote ^^to ^^remove ^^| ^^Contact ^^| ^^Info ^^| ^^Opt-out ^^| ^^GitHub
DigThatData t1_j95gxlf wrote
Reply to comment by maxToTheJ in [D] Please stop by [deleted]
i think something changed in the past week though. /r/MLQuestions has recently been getting a lot of "can you recommend a free AI app that does <generic thing>?". I'm wondering if there was a news piece that went viral or something that turned a new flood of people on to what's been happening in AI or something like that.
iacolippo t1_j95ffj7 wrote
If you use the OpenAI models inside langchain, then yes.
milleniumsentry t1_j95v74v wrote
Reply to comment by I_like_sources in [D] Lack of influence in modern AI by I_like_sources
I disagree. They are completely related, and directly to the black box problem.
I wish I found this article a month ago, because it sums up a lot of the 'ai's are unknowable' nonsense.
Being a blackbox, is not an inherent quality of an AI. It's an inherent quality of a badly designed AI. Eventually, we will have methods that allow us to query why a particular result was given.
They are unknowable, because we have not designed them to be. The tech is in it's infancy. Give it time.