Recent comments in /f/MachineLearning

personnealienee t1_j82ygry wrote

messing with target sound extraction by adding just barebones masknet architechture on top of samplernn. I want to apply this architecture to extracting different layers in electronic misic. for example, pick out just the snare drum track from the full drum machine mix. It is easy to generate datasets using DawDreamer (generating random drum patterns using a sampler currently). considering adding conditioning by the output of a differentiable filter bank

2

askingforhelp1111 t1_j81ggm0 wrote

Sure, I have a few links. All of them have an inference speed of 4-9 seconds.

https://huggingface.co/poom-sci/WangchanBERTa-finetuned-sentiment

https://huggingface.co/ayameRushia/bert-base-indonesian-1.5G-sentiment-analysis-smsa

I call each checkpoint like this:

nlp = pipeline('sentiment-analysis',
            model=checkpoint, 
            tokenizer=checkpoint)

Thank you!

1

ZBMakesSongs t1_j81dm5r wrote

If you want ML TTS, there are a lot of open source models out there, problem is most of them are trained on the same data, so your going to get similar voice options for the most part. You can definitely train your own text to speech, and pretty easily as well, but I'm assuming you don't want to go that route. Maybe try starting with Coqui TTS, but for reading long documents it definitely has its fair share of issues.

2

d0cmorris OP t1_j819chm wrote

Exactly. I mean I can easily define L2-constraints for the weights of my network and then do constrained optimization, which would at least theoretically be equivalent to L2-regularization/weight decay. But this is not quite useful, I am wondering whether there are applications of constraints where it actually makes sense.

1