Recent comments in /f/MachineLearning
zazzersmel t1_jankg28 wrote
Reply to [D] Is there an ML project out there that recommends movies based on more than the usual features? by of_a_varsity_athlete
it just depends on the available data. someone has to tag those films with the appropriate info youre interested in. the modeling isnt necessarily any different.
avialex t1_janjx6r wrote
Reply to comment by mmmniple in [D] Are Genetic Algorithms Dead? by TobusFire
Appears to be here: https://openreview.net/forum?id=ibNr25jJrf
edit: actually after reading it, I don't think this is the referenced publication, but it's still interesting
lifesthateasy t1_janig00 wrote
Reply to comment by currentscurrents in [D] Blake Lemoine: I Worked on Google's AI. My Fears Are Coming True. by blabboy
They're black boxes in a sense that it's hard to oversee all of the activations together. But it's very easy to understand what each neuron does, and you can even check outputs at each layer to see what's happening inside.
Look you sound like you went to an online course and have a basic understanding of basic buzzwords but have never studied the topic in depth.
Lol if you think rocks might be sentient, there's no way I can make you understand why LLMs are not.
You're even wrong on sentience and consciousness, for once you keep mixing these two concepts together which makes it harder to converse, as you keep changing what you're discussing. And then again, we do have a definition for sentience, and there have been studies that have proven for example in multiple animal species that they are in fact sentient, and zero studies that have shown the same on rocks. Even the notion is idiotic.
LetterRip t1_jani50o wrote
Reply to comment by jinnyjuice in [D] OpenAI introduces ChatGPT and Whisper APIs (ChatGPT API is 1/10th the cost of GPT-3 API) by minimaxir
We don't know the supply demand curve, so we can't know for sure that the revenue increased.
currentscurrents t1_jangzvf wrote
Reply to comment by lifesthateasy in [D] Blake Lemoine: I Worked on Google's AI. My Fears Are Coming True. by blabboy
Hah! Not even close, they're almost black boxes.
But even if we did, that wouldn't help us tell whether or not they're sentient, because we'd still need understand to sentience. For all we know everything down to dumb rocks could be sentient. Or maybe I'm the only conscious entity in the universe - there's just no data.
Timdegreat t1_jangbi7 wrote
Reply to comment by sebzim4500 in [D] OpenAI introduces ChatGPT and Whisper APIs (ChatGPT API is 1/10th the cost of GPT-3 API) by minimaxir
You got a point there! I haven't given it too much thought really -- I def need to check out ada.
But wouldn't the ChatGPT embeddings still be better? Given that they're cheap, why not use the better option?
ajt9000 t1_janfx47 wrote
Reply to comment by sugar_scoot in [D] Are Genetic Algorithms Dead? by TobusFire
This comment make me wonder if the same rules about using one-hot encoding instead of ordinal encoding for classifiers still apply to a neural net trained with a gradient-less search algorithm like a GA instead of backprop.
ajt9000 t1_janfj0c wrote
Reply to [D] Are Genetic Algorithms Dead? by TobusFire
Who says genetic algorithms are dead? They're pretty much dead for training neural nets absolutely, but there are tons of other more general optimization problems that GAs (or more generally evolutionary algorithms) are well suited for.
Not to mention they still have plenty of utility as a search algorithm for hyperparameters so they aren't even dead for neural applications.
[deleted] t1_janfihy wrote
Reply to comment by PassionatePossum in [D] Are Genetic Algorithms Dead? by TobusFire
[deleted]
noeda t1_janf9cr wrote
Reply to [D] Are Genetic Algorithms Dead? by TobusFire
I use CMA-ES (a type of evolutionary algorithm) for training neural networks for finance stuff. The neural networks involved are not superhuge so it works out (IIRC the number of parameters is around ~500-1000).
The fitness function is pretty complicated and written in Rust and I put a lot of effort to making it fast because these algorithms need to evaluate it many many times. I feel using evolutionary algorithms makes coding simpler because you do not need to care that whatever you are writing is differentiable or that some backprop/gradient descent library needs to be able to "see" inside your function.
I do think my use case is a bit more niche. I live in hope that some breakthrough happened that made evolutionary algorithms practically usable for large neural networks.
proton-man t1_janca53 wrote
Reply to comment by [deleted] in [D] Are Genetic Algorithms Dead? by TobusFire
It was. Dumb too. Because of the limitations of memory and computing power at the time you had to constantly tweak parameters to optimize learning speed, avoid overfitting, avoid local optimums, etc. Only to find that the best performing model was the one generated by your 2 AM code with the fundamental flaw and the random parameters you chose while high.
Pikalima t1_janc14v wrote
Reply to comment by [deleted] in [D] OpenAI introduces ChatGPT and Whisper APIs (ChatGPT API is 1/10th the cost of GPT-3 API) by minimaxir
Iād say we need an /r/VXJunkies equivalent for statistical learning theory, but the real deal is close enough.
dragosconst t1_janbuui wrote
Reply to [D] Are Genetic Algorithms Dead? by TobusFire
I think there are few problems were a couple extra assumptions that could make much more efficient methods work (not NNs necessarily of course) don't hold. I'm not sure there exist problems where genetic algos outperform other methods, disregarding problems where only genetic algos work.
scawsome t1_janbtva wrote
Reply to comment by rm-rf_ in [D] Are Genetic Algorithms Dead? by TobusFire
Not necessarily. Bayesian methods work great when you have expensive objective function evaluations that can only be evaluated in serial (or limited parallel evaluations). Bayesian methods aren't ideal in massively parallelizable evaluations (evaluating >100 points at a time) or when evaluations are relatively cheap. It depends on the cost of optimizing the acquisition function. I've actually played around with combining BO with evolutionary algorithms to extend BO towards massively parallelizable evaluations and have seen some promising results.
dansmonrer t1_janb38v wrote
Reply to [D] Podcasts about ML research? by Tight-Vacation-9410
TWIML as others mentioned, Eye On AI also
fasttosmile t1_janaaex wrote
Reply to comment by harharveryfunny in [D] OpenAI introduces ChatGPT and Whisper APIs (ChatGPT API is 1/10th the cost of GPT-3 API) by minimaxir
GCP, speechmatics, rev, otter.ai, assemblyai etc. etc. offer similar or better performance, as well as streaming and a much more rich output.
marcus_hk t1_jan8rmh wrote
Reply to [D] Are Genetic Algorithms Dead? by TobusFire
They might see a resurgence in dynamic multi-agent environments.
sebzim4500 t1_jan85s7 wrote
Reply to comment by Timdegreat in [D] OpenAI introduces ChatGPT and Whisper APIs (ChatGPT API is 1/10th the cost of GPT-3 API) by minimaxir
Yeah, I get that's that embeddings are used for semantic search but would you really want to use a model as big as ChatGPT to compute the embeddings? (Given how cheap and effective Ada is)
lost_fodder6947 t1_jan7zat wrote
Nice... š
Timdegreat t1_jan7sel wrote
Reply to comment by sebzim4500 in [D] OpenAI introduces ChatGPT and Whisper APIs (ChatGPT API is 1/10th the cost of GPT-3 API) by minimaxir
You can use the embeddings to search through documents. First, create embeddings of your documents. Then create an embedding of your search query. Do a similarity measurement between the document embeddings and the search embedding. Surface the top N documents.
[deleted] t1_jan7rz4 wrote
Reply to [D] Are Genetic Algorithms Dead? by TobusFire
[deleted]
mmmniple t1_jan7i1a wrote
Reply to comment by filipposML in [D] Are Genetic Algorithms Dead? by TobusFire
It sounds very interesting. Is it available to read? Thanks
mikonvergence OP t1_jan5fnj wrote
Reply to comment by SnooMarzipans1345 in [P] A minimal framework for image diffusion (including high-resolution) by mikonvergence
Hi! Sure, here it goes:
It's a course about making AI models that can create images. These models can that by learning from a dataset of example images. "Diffusion" is a new type of AI model that works very well for this task.
The course will work best for those familiar with training deep neural networks for generative tasks, so I would advise catching up on topics like VAEs or GANs. However, the video course material is quite short (about 1,5 hrs) so you can just play it and see if it works for you or not!
extracensorypower t1_jan4yel wrote
Reply to [D] Are Genetic Algorithms Dead? by TobusFire
I think they're still useful for "no information at all" scenarios where attempting a solution is just too time consuming or not possible using other methods (e.g. traveling salesman problem).
As a practical matter, I think they're best integrated with other methods as "first cut" solutions that get you closer to something you can work out with a neural net or rule based system.
That said, I'm unaware of any NN or rule based solution better than a GA for solving the traveling salesman problem even now. So, maybe some P-NP problems will always be best attacked with GAs.
LetterRip t1_janljeo wrote
Reply to comment by lucidraisin in [D] OpenAI introduces ChatGPT and Whisper APIs (ChatGPT API is 1/10th the cost of GPT-3 API) by minimaxir
Ah, I'd not seen the Block Recurrent Transformers paper before, interesting.