Recent comments in /f/MachineLearning
DataGOGO t1_j813dui wrote
It is legal until a court says otherwise.
OkCandle6431 t1_j811ol6 wrote
Reply to [D] Critique of statistics research from machine learning perspectives (and vice versa)? by fromnighttilldawn
Where I'm at 'statistics' is what me and my co-workers call what we do, and 'machine learning' is what goes in the grant application. I'm sure this differs across regions/faculty/industry/whatever.
goj-145 t1_j80xlao wrote
Reply to comment by Tlaloc-Es in [D] Is it legal to use images or videos with copyright to train a model? by Tlaloc-Es
Not really hard when the model is spitting out watermarked images.
Tlaloc-Es OP t1_j80xdxu wrote
Reply to comment by goj-145 in [D] Is it legal to use images or videos with copyright to train a model? by Tlaloc-Es
But anyway, is hard to demonstrate which is the dataset of a model right? in the case of Getty you can probably get images that look like Getty image dataset, but for a predictor? and if this case for example where "there wasn't any law" or predecessor case can lose the lawsuit having to pay?
goj-145 t1_j80ufu1 wrote
We're going to find out soon with the Getty lawsuit. Until then, gray area.
ellemoe-is-elleva t1_j80pr75 wrote
Reply to [D] Locally-runnable text to speech AI? by gruevy
Pyttsx, mbrola, mimic 3. I like the mimic 3. Which is lightweight. And can run on docker or just native.
I started out with mycroft which has mimic 3 build in. But you can run it just stand alone as well and quite easy to set up. https://mycroft.ai/mimic-3/
If you want to go down the rabbithole of speech synthesis and analsys check out praat praat.org it's a quiet impressive piece of software.
Royal-Landscape9353 t1_j80oma3 wrote
Reply to [D] Locally-runnable text to speech AI? by gruevy
Try TortoiseTTS on the highest quality setting
gruevy OP t1_j80oc7f wrote
Reply to comment by [deleted] in [D] Locally-runnable text to speech AI? by gruevy
ah my bad then, must have misread. i'll take another look
[deleted] t1_j80o2s9 wrote
Reply to comment by gruevy in [D] Locally-runnable text to speech AI? by gruevy
It does both buddy
gruevy OP t1_j80mwkq wrote
Reply to comment by [deleted] in [D] Locally-runnable text to speech AI? by gruevy
Yeah that's speech to text. I want text to speech. Thanks tho
[deleted] t1_j80mgtf wrote
Reply to comment by gruevy in [D] Locally-runnable text to speech AI? by gruevy
Available_Lion_652 OP t1_j80lcuv wrote
Reply to comment by ehlen in [D] RTX 3090 with i7 7700k, training bottleneck by Available_Lion_652
I would really appreciate it if you can try to finetune a T5-XXL Flan model from Huggingface on your hardware. I am curious if it works and if there is a big bottleneck. Thank you
gruevy OP t1_j80kuq5 wrote
Reply to comment by [deleted] in [D] Locally-runnable text to speech AI? by gruevy
google search didn't get me much, can you be more specific?
[deleted] t1_j80jvp3 wrote
Reply to [D] Locally-runnable text to speech AI? by gruevy
Speech WebKit
eigenham t1_j80ijfv wrote
Reply to comment by jimmymvp in [D] Constrained Optimization in Deep Learning by d0cmorris
Link(s)? I can find more from examples, just need a thread to pull. Thanks!
slashdave t1_j80hs32 wrote
Reply to [D] Critique of statistics research from machine learning perspectives (and vice versa)? by fromnighttilldawn
Different goals and different tools
ehlen t1_j80bfkq wrote
I have this exact setup (7700k & 3090). If you want me to try something out, I am happy to try running it.
jimmymvp t1_j807b94 wrote
There's a bunch of cool work on using constrained optimization as a layer in neural nets, differentiation through argmin. I'm not sure if this answers your question.
AdFew4357 t1_j806plm wrote
Reply to comment by jimmymvp in [D] Critique of statistics research from machine learning perspectives (and vice versa)? by fromnighttilldawn
The whole landscape of ML research is a hunt to chase SOTA by tweaking an architecture here or using a different optimizer there and then squeezing out 0.2% accuracy on some well known imaging dataset in an attempt to churn out papers. That’s not science if you ask me.
jimmymvp t1_j806dx2 wrote
Reply to comment by AdFew4357 in [D] Critique of statistics research from machine learning perspectives (and vice versa)? by fromnighttilldawn
Just communicating what I've heard. Nevertheless, I think the whole interpretable ML community (at the very least) would disagree with you on this one :). Reducing ML to "plug and chug" is well... Speaks for itself :D
Ready-Acanthaceae970 OP t1_j803d96 wrote
Reply to comment by bjr1973 in [D]Image Recognition ability of machine learning in financial markets questions by Ready-Acanthaceae970
this is great. thanks
SwitchOrganic t1_j801prt wrote
Reply to comment by themusicdude1997 in [D] Critique of statistics research from machine learning perspectives (and vice versa)? by fromnighttilldawn
My guess is ML scientists generally care less about statistical rigor which can lead to poor outcomes due to not properly understanding the data, assumptions, risk involved, etc
Ex: Zillow
_eminorhan_ t1_j7zwlu9 wrote
People should be more skeptical of "emergent abilities" in big models: 1) Papers claiming such abilities generally use undertrained small models as per chinchilla scaling (compute is not controlled + suboptimal hyperparam choices for small models) and 2) these papers generally use a semilogx plot to demonstrate "emergence" but even a linear relationship will look exponential in such a plot. I'm not sure if I'd want to call a simple linear relationship "emergent".
AdFew4357 t1_j7ztran wrote
Reply to comment by jimmymvp in [D] Critique of statistics research from machine learning perspectives (and vice versa)? by fromnighttilldawn
Stats is finding interpretable ways to look at and mode data that ML plug and chug cs people don’t do
ehlen t1_j818wj1 wrote
Reply to comment by Available_Lion_652 in [D] RTX 3090 with i7 7700k, training bottleneck by Available_Lion_652
Ok I’ll try it out. Might be a few days as I am out of town ATM.