Recent comments in /f/MachineLearning
lonelyrascal t1_j7vwy0c wrote
Reply to [P] Creating an embedding from a CNN by zanzagaes2
PCA has O(n^3) time complexity. Instead of doing that, why don't you pass the embedding through an autoencoder?
lonelyrascal t1_j7vwf3f wrote
Reply to comment by lonelyrascal in [D] Similarity b/w two vectors by TKMater
Well it also depends on whether you have sparse vectors or not... Some metrics work best for sparse vectors and other work for dense vectors. Check out scikit learn's pairwise distances metrics to understand better
lonelyrascal t1_j7vw6f4 wrote
Reply to [D] Similarity b/w two vectors by TKMater
Try cosine similarly?
notdelet t1_j7vv9pi wrote
Reply to comment by tdgros in [D] Constrained Optimization in Deep Learning by d0cmorris
You can get constrained optimization in general for unconstrained nonlinear problems (see the work N Sahinidis has done on BARON). The feasible sets are defined in the course of solving the problem and introducing branches. But that is both slow, doesn't scale to NN sizes, and doesn't really answer the question ML folks are asking (see the talk at the IAS on "Is Optimization the Right Language for ML").
the_new_scientist t1_j7vu5fk wrote
Yes, the DINO paper showed that the ability to perform segmentation emerges from self-supervised vision transformers.
https://arxiv.org/abs/2104.14294
Edit: oops, didn't realize you said image generation models, thought you asked for just vision models.
Jucuco t1_j7vsc6p wrote
Reply to [D] Similarity b/w two vectors by TKMater
Inner product between metrics tells you the projection of one onto the other. Whatever the space you're building reperesents
zanzagaes2 OP t1_j7vpols wrote
Reply to comment by schludy in [P] Creating an embedding from a CNN by zanzagaes2
You are right, both tSNE and UMAP documentation recommend going to 30-50 features before using them. In this case the result is quite similar to the one I found, though.
zanzagaes2 OP t1_j7vpd89 wrote
Reply to comment by schludy in [P] Creating an embedding from a CNN by zanzagaes2
Yes, I think that's the case because I am getting far more reasonable values comparing the projection to 2d/3d of the embedding rather than the full 500 feature vector.
Is there a better way to do this than projecting into a smaller space (using reduction dimensionality techniques or encoder-decoder approach) and using L2 there?
CeFurkan OP t1_j7vouon wrote
Reply to comment by logsinh in [D] Are there any AI model that I can use to improve very bad quality sound recording? Removing noise and improving overall quality by CeFurkan
thanks
no idea where to put this code in nuwave2
CeFurkan OP t1_j7volvh wrote
Reply to comment by vivehelpme in [D] Are there any AI model that I can use to improve very bad quality sound recording? Removing noise and improving overall quality by CeFurkan
but what about synching? how to solve synching problem?
i haven't found any way to re-voice with proper synchronization
i can prepare a perfect .vtt file but how to sync it with video?
CeFurkan OP t1_j7vof8i wrote
Reply to comment by pronunciaai in [D] Are there any AI model that I can use to improve very bad quality sound recording? Removing noise and improving overall quality by CeFurkan
thank you
DACUS1995 t1_j7vltrw wrote
As you said most deep-learning models use some sort of regularization at training so there is some implicit constraint on the actual values of the weights, even more so when the number of parameters goes in the range of billions where you will have an inherent statistical distribution of the feature importance. On the more explicit and fixed side, there are a couple of papers and efforts in the area of quantization where parameter outliers in various layers affect the precision of quantized representation, so you would want a reduced variance in the block or layers values. For example, you can check this: https://arxiv.org/abs/1901.09504.
pronunciaai t1_j7vlt9i wrote
Reply to comment by CeFurkan in [D] Are there any AI model that I can use to improve very bad quality sound recording? Removing noise and improving overall quality by CeFurkan
Yes dns64 is better/larger than their other pretrained (48 I think)
[deleted] t1_j7vl6xh wrote
Reply to comment by logsinh in [D] Are there any AI model that I can use to improve very bad quality sound recording? Removing noise and improving overall quality by CeFurkan
[removed]
vivehelpme t1_j7vfol7 wrote
Reply to comment by CeFurkan in [D] Are there any AI model that I can use to improve very bad quality sound recording? Removing noise and improving overall quality by CeFurkan
Instead of trying to salvage the original recording why not recreate it by putting the text transcript into a text-to-speech model?
As you have it transcribed you don't even need to do any advanced speech recognition that filters the noise, just paste the text into something a bit more advanced than Microsoft Sam
CeFurkan OP t1_j7vf77g wrote
Reply to comment by vivehelpme in [D] Are there any AI model that I can use to improve very bad quality sound recording? Removing noise and improving overall quality by CeFurkan
what you mean by that? i have transcripts but then what to do? thank you
CeFurkan OP t1_j7vf5fu wrote
Reply to comment by jeanfeydy in [D] Are there any AI model that I can use to improve very bad quality sound recording? Removing noise and improving overall quality by CeFurkan
thank you for trying to help
sonofmath t1_j7vehdk wrote
Reply to comment by mr_house7 in [D] List of RL Papers by C_l3b
Not really, I think the main strength of the library is that it is designed to be easy to understand how the algorithms are implemnted. At the time, the main alternative was OpenAI/Stable baselines, which was quite obscure to understand how the algorithms are implemented. On the other hand, the algorithms do not use some more advanced tricks that enhance performance
However, there are better libraries now. In the same spirit, there is CleanRL, that is clean (with algorithms in one file) , but also performent. If you are looking for a modular easy-to-use library, I would recommend Stable Baselines3
tdgros t1_j7vdocr wrote
With constrained optimization, you usually have a feasible set for the variables you optimize, but in a NN training you optimize millions of weights that aren't directly meaningful, so in general, it's not clear if you can define a feasible set for each of them.
mLalush t1_j7vcqph wrote
Love your write ups /u/pommedeterresautee . Especially the fact that they're written with human beings in mind. I mean that as a compliment, seeing as the vast majority of stuff concerning cuda and low level optimization is impenetrable.
I periodically check kernl.ai to see whether the documentation and tutorial sections have been expanded. My advice is put some real effort and focus in to examples and tutorials. It is key for an optimization/acceleration library. 10x-ing the users of a library like this is much more likely to come from spending 10 out of every 100 developer hours writing tutorials, as opposed to spending 8 or 9 of those tutorial-writing hours on developing new features that only a small minority understand how to use and apply.
vivehelpme t1_j7vcbx5 wrote
Reply to [D] Are there any AI model that I can use to improve very bad quality sound recording? Removing noise and improving overall quality by CeFurkan
Transcribe them and put the transcripts in TTS
TKMater OP t1_j7vayeo wrote
Reply to comment by BenoitParis in [D] Similarity b/w two vectors by TKMater
I have classified some data points as anomaly in timeseries data and calculated feature importance vectors for them. Now I want to calculate similarity between two anomalous data points based on their feature importance vector.
Character_Internet_3 t1_j7vaukr wrote
Reply to [D]Image Recognition ability of machine learning in financial markets questions by Ready-Acanthaceae970
Is possible, but if you have the data to generate those images, will be easier to treat the data not as an image (chart)
BenoitParis t1_j7v9wue wrote
Reply to [D] Similarity b/w two vectors by TKMater
Lots to choose from:
https://docs.scipy.org/doc/scipy/reference/spatial.distance.html
How do your vectors look like? What do you intend to do with them? Will you be clustering them? Indexing them? How many are there? How did you obtain them? What do they represent? What is their type?
big_ol_tender t1_j7vx88w wrote
Reply to [D]Image Recognition ability of machine learning in financial markets questions by Ready-Acanthaceae970
Hundreds of firms have already tried this years ago- there is no alpha left in this type of information. Source: work at a hedge fund and have friends at two sigma, citadel, and D. E. Shaw. If you want to do it as an academic exercise go ahead, but this is not how you make money.