Recent comments in /f/MachineLearning
Kaleidophon t1_jalv0qs wrote
Reply to comment by Emergency_Apricot_77 in [D] backprop through beam sampling ? by SaltyStackSmasher
>Why go with BLEU though ? OP didn't particularly mention optimizing sequence level metrics.
From OPs post above:
>is this possible at all ? I think if we can backprop through beam sampling, we can directly optimise for bleu ?
drplan t1_jalud65 wrote
Reply to [D] Are Genetic Algorithms Dead? by TobusFire
Genetic algorithms are still useful for strange objective functions that defy analytical approaches, such as anything based on complex simulations. But it somehow has always been this way.
Nowadays things have changed by generative models for code generation. A few years ago Genetic Programming (and it's many variants) was the only approach to do this, now some problem can just be solved by asking a language model to write the code for xyz.
keepthepace t1_jalu94j wrote
Reply to comment by red75prime in [D] Blake Lemoine: I Worked on Google's AI. My Fears Are Coming True. by blabboy
The key thing we need is agency. The current chatbots lack the long-term coherency we expect from an agent, because they do not plan towards specific goals, so they just jump from one thing to another.
1bir t1_jalti3p wrote
Reply to [D] Podcasts about ML research? by Tight-Vacation-9410
Be careful what you listen to while driving; don't let this month's Tesla become next month's robot overlord...
Hunterhal t1_jalt078 wrote
Reply to [D] Are Genetic Algorithms Dead? by TobusFire
Hyperparam optimization?
[deleted] t1_jalsgjd wrote
Reply to [D] Are Genetic Algorithms Dead? by TobusFire
[removed]
xEdwin23x t1_jals78j wrote
Reply to comment by bjergerk1ng in [D] What are the most known architectures of Text To Image models ? by AImSamy
I'm guessing he refers to this one: https://parti.research.google/
frequenttimetraveler t1_jalqh95 wrote
Well in a way they are 'coming true'. He is now a professional fearmonger for pay
[deleted] t1_jalq5f9 wrote
Reply to comment by M_Alani in [D] Are Genetic Algorithms Dead? by TobusFire
[deleted]
WarAndGeese t1_jalq339 wrote
Reply to comment by Educational-Net303 in [D] OpenAI introduces ChatGPT and Whisper APIs (ChatGPT API is 1/10th the cost of GPT-3 API) by minimaxir
Don't let it demotivate competitors. They are making money somehow, and planning to make massive amounts more. Hence the space is ripe for tons of competition, and those other companies would also be on track to make tons of money. Hence, jump in competitors, the market is waiting for you.
Hsemar t1_jalp8as wrote
Reply to comment by lucidraisin in [D] OpenAI introduces ChatGPT and Whisper APIs (ChatGPT API is 1/10th the cost of GPT-3 API) by minimaxir
but does flash attention help with auto-regressive generation? My understanding was that it prevents materializing the large kv dot product during training. At inference (one token at a time) with kv caching this shouldn't be that relevant right?
PassionatePossum t1_jalnb1q wrote
Reply to [D] Are Genetic Algorithms Dead? by TobusFire
I think, my professor summarized it very well: "Genetic algorithms is what you do when everything else fails."
What he meant by that is, that they are very inefficient optimizers. You need to evaluate lots and lots of configurations because you are stepping around more of less blindly in the parameter space and you are only relying on luck and a few heuristics to improve your fitness. But their advantage is that they will always work as long as you can define some sort of fitness function.
If you can get a gradient, you are immediately more efficient because you already know in which direction you need to step to get a better solution.
But of course there is room for all algorithms. Even when you can do gradient descent, there are problems where it quickly gets stuck in a local optimum. There are approaches how to "restart" the algorithm to find a better local optimum. I'm not that familiar with that kind of optimization but it is not inconceivable that genetic algorithms might have a role to play in such a scenario.
serge_cell t1_jalnarf wrote
Reply to [D] Are Genetic Algorithms Dead? by TobusFire
The notable diffrence between GA and other random searches is cross-over operator, and in it's theory "building blocks" hypothesis. Neither were confirmed during years (dozens of years) of attemted use of GA.
bo_peng OP t1_jalmszp wrote
Reply to comment by ID4gotten in [P] ChatRWKV v2 (can run RWKV 14B with 3G VRAM), RWKV pip package, and finetuning to ctx16K by bo_peng
It's actually quite good at Q&A if you use my prompt templates:
+gen \nExpert Questions & Helpful Answers\nAsk Research Experts\nQuestion:\nXXXXXXXXXXXXXXX?\n\nFull Answer:\n
+gen \nAsk Expert\n\nQuestion:\nXXXXXXXXXXXXXXXX?\n\nExpert Full Answer:\n
+gen \nQ & A\n\nQuestion:\nXXXXXXXXXXXXXXXXX?\n\nDetailed Expert Answer:\n
sobe86 t1_jalldpg wrote
Reply to comment by sugar_scoot in [D] Are Genetic Algorithms Dead? by TobusFire
Plus also there needs to be a learnable, nontrivial 'strategy' to take advantage of, otherwise it's not going to beat simulated annealing except on speed. The couple of times I've used it in practice, SA was about as good as we could get performance-wise.
[deleted] t1_jall6xi wrote
Reply to comment by jinnyjuice in [D] OpenAI introduces ChatGPT and Whisper APIs (ChatGPT API is 1/10th the cost of GPT-3 API) by minimaxir
[deleted]
jinnyjuice t1_jalkbvu wrote
Reply to comment by LetterRip in [D] OpenAI introduces ChatGPT and Whisper APIs (ChatGPT API is 1/10th the cost of GPT-3 API) by minimaxir
How do we know these technical improvements result in 90% extra revenue? I feel I'm missing some link here.
Trotskyist t1_jalk4j5 wrote
Reply to comment by Lychee7 in [D] OpenAI introduces ChatGPT and Whisper APIs (ChatGPT API is 1/10th the cost of GPT-3 API) by minimaxir
A token is (roughly) 4 characters. Both prompt and result are counted.
FinancialElephant t1_jaliqsh wrote
Reply to [D] Are Genetic Algorithms Dead? by TobusFire
Genetic optimization might be dead in most cases. I think a lot of the ideas aside from optimization algorithms are still relevant.
I've found GP techniques can yield parsimonious models. A lot of the big research these days is on big models, but GP seems good for small, parsimonious, and elegant models. Good for low data regimes, specialized problems, and problems where you have expert knowledge you can encode. Generally speaking I like working with GP becuase you end up with a parsimonious and interpretable model (opposite of a lot of NN research).
In practice I've found importance sampling methods to work about as good as genetic optimization for optimizing GP trees/grammars for the small amount of work I did with them. I haven't found either method to edge out by much, but it could depend on the problem.
I don't know if this is considered GP (or GA) without a genetic optimization method. However I think we can say that the notion of optimizing a symbolic tree or grammar was heavily developed within GP, even if today you may use some monte carlo optimization method in practice.
nomorerainpls t1_jalhznm wrote
I assume Google stands behind their decision after reading this article
visarga t1_jalh1r1 wrote
Reply to comment by Dendriform1491 in [D] Are Genetic Algorithms Dead? by TobusFire
You don't always need a population of neural networks, it could be a population of prompts or even a population of problem solutions.
If you're using GA to solve specific coding problems, then there is one paper where they use LLM to generate diffs for code. The LLM was the mutation operator, and they even fine-tune it iteratively.
lifesthateasy t1_jalgvq6 wrote
Reply to comment by currentscurrents in [D] Blake Lemoine: I Worked on Google's AI. My Fears Are Coming True. by blabboy
Exactly, but we completely understand how neural networks work down to a tee.
visarga t1_jalgrla wrote
Reply to [D] Are Genetic Algorithms Dead? by TobusFire
visarga t1_jalg9iu wrote
Reply to comment by fmai in [D] OpenAI introduces ChatGPT and Whisper APIs (ChatGPT API is 1/10th the cost of GPT-3 API) by minimaxir
I think the main pain point was memory usage.
curiousshortguy t1_jalwn6p wrote
Reply to [D] Podcasts about ML research? by Tight-Vacation-9410
Talking machines has a great hook to machine learning researchers in the NeurIPS community, but it's also very much an ivory tower view with very little relevance for research outside the privilege bubble.