Recent comments in /f/MachineLearning
CeFurkan OP t1_j7wsy5f wrote
Reply to comment by express_mode_420 in [D] Are there any AI model that I can use to improve very bad quality sound recording? Removing noise and improving overall quality by CeFurkan
so it is a logical layout
any software that can do it?
Locomule t1_j7wsobl wrote
Reply to comment by CeFurkan in [D] Are there any AI model that I can use to improve very bad quality sound recording? Removing noise and improving overall quality by CeFurkan
MonsterMMORPG eh? Very interesting!! :D
danielgafni t1_j7wsnlw wrote
Reply to [D]Image Recognition ability of machine learning in financial markets questions by Ready-Acanthaceae970
The approach you are describing isn’t the best.
-
There is no sense in rendering these images as OHLCV data is timeseries, not 2D images. Most of the data would just be white pixels. Which is not really wrong but is greatly inefficient. Instead of using 2D convolutions 1D convolutions can be used on the timeseries directly (which is called a Wavenet) which would remove rendering from your pipeline and greatly speedup training and inference.
-
OHLCV data won’t give you enough information to neither predict the future or backtest your trading algorithm accurately due to loss of data after aggregations.
fasttosmile t1_j7ws729 wrote
Reply to comment by DigThatData in [D] Using LLMs as decision engines by These-Assignment-936
So cool!
[deleted] t1_j7wrqyp wrote
Reply to comment by DigThatData in [D] Using LLMs as decision engines by These-Assignment-936
[removed]
YOLOBOT666 t1_j7wrm1z wrote
What about saving the dataset into batches as individual files, then use the data loader to load the files as batches for transformers? Keeping the batch size reasonable for the GPU memory.
For any preprocessing/scaling, this could be done on the CPU side and would not consume much memory^
lonelyrascal t1_j7wp5yv wrote
Reply to comment by zanzagaes2 in [P] Creating an embedding from a CNN by zanzagaes2
Ok cool. Yeah keras has basic encoder decoder architecture in its documentation. If that's not something you like, you can always ask chatGPT ;)
CeFurkan OP t1_j7wleql wrote
Reply to comment by evanthebouncy in [D] Are there any AI model that I can use to improve very bad quality sound recording? Removing noise and improving overall quality by CeFurkan
it is fine this is my public email
​
thank you for warning
currentscurrents t1_j7wk84r wrote
Reply to comment by andreichiffa in [D] Are there emergent abilities of image models? by These-Assignment-936
While those are on the same topic, they're very different papers. The Anthropic paper spends most of its time going on about safety/bias/toxicity, while the Google paper is focused on more useful things like the technical abilities of the models.
evanthebouncy t1_j7wjt36 wrote
Reply to comment by CeFurkan in [D] Are there any AI model that I can use to improve very bad quality sound recording? Removing noise and improving overall quality by CeFurkan
don't put your email in public like this. dm the guy. remove the email while you still can.
EQ and Compression are good techniques to try, reaper is free. I'm sure your friend can show you.
DigThatData t1_j7wjpp4 wrote
express_mode_420 t1_j7wizoa wrote
Reply to comment by CeFurkan in [D] Are there any AI model that I can use to improve very bad quality sound recording? Removing noise and improving overall quality by CeFurkan
I'm not sure how I'd go about syncing it, but would this be an adequate workaround:
- break apart your script in small chunks by time stamp
- generate different tts recordings off of each time stamp
- generate an audio file that inserts each of the produced recordings at their respective time-stamped location
- replace the audio of the recording with your newly produced recording
CeFurkan OP t1_j7wi9v3 wrote
Reply to comment by Locomule in [D] Are there any AI model that I can use to improve very bad quality sound recording? Removing noise and improving overall quality by CeFurkan
> audio editor Reaper to apply EQ and Compression
if you make a video i would watch it and show me how to do :D
CeFurkan OP t1_j7whknq wrote
Reply to comment by Locomule in [D] Are there any AI model that I can use to improve very bad quality sound recording? Removing noise and improving overall quality by CeFurkan
wow this is amazing
how can I contact to you?
my email : monstermmorpg@gmail.com
my discord : MonsterMMORPG#2198
CeFurkan OP t1_j7whf7d wrote
Reply to comment by express_mode_420 in [D] Are there any AI model that I can use to improve very bad quality sound recording? Removing noise and improving overall quality by CeFurkan
i have vtt file you know the subtitles we use for movies
but i haven't found and text to speech that can generate speech with that timing
do you know any?
​
about your suggested approach, any way to automatically do it? i mean we generate speech then we sync but how?
VaxxBetrayal t1_j7wfxkg wrote
Reply to comment by nielsrolf in [D] Are there emergent abilities of image models? by These-Assignment-936
Hmmmm. Hmmmmmmmm
___luigi t1_j7wf94b wrote
I guess that this is similar to trend that we have seen with “Decision Transformers” and how they have been proposed as a practical alternative for RL.
currentscurrents t1_j7wf3u0 wrote
>What is the standard modeling approach to these kinds of problems?
The standard approach is reinforcement learning. It works, but it's not very sample-efficient and takes many iterations to train.
LLMs are probably so good at this because of their strong meta-learning abilities; during the process of pretraining they not only learn the task but also learn good strategies for learning new tasks.
This has some really interesting implications. Pretraining seems to drastically improve sample efficiency even if the pretraining was on a very different task. Maybe we could pretrain on a very large amount of synthetic, generated data before doing our real training on our finitely-sized real datasets.
7366241494 t1_j7wb6ok wrote
Reply to comment by imaginethezmell in [N] "I got access to Google LaMDA, the Chatbot that was so realistic that one Google engineer thought it was conscious. First impressions" by That_Violinist_18
ChatGPT has the same problems with astronomy data. Blame the marketing team not the technology.
ggf31416 t1_j7waxlu wrote
It will depend on how much preprocessing and augmentation is needed. I don't think text needs much preprocessing or augmentation, but for example image classification or detection training needs to create a different augmented image on each iteration and will benefit from a more powerful processor.
Note that you can also use cloud services. If you aren't dealing with confidential data vast ai often is one of the cheapest, otherwise you can use Lambda Labs, Google Engine, AWS or other services. At least in the case of Google Engine and AWS you have to request access to GPU instances, which may take some time.
zanzagaes2 OP t1_j7w5sr1 wrote
Reply to comment by lonelyrascal in [P] Creating an embedding from a CNN by zanzagaes2
I will try encoder-decoder architecture, mainly to try to improve the embedding. Right now asymptotics of PCA have not proven a problem, sklearn implementation performs PCA on ~1.000 features vectors almost immediately.
Do you have any reference on any encoder-decoder architecture I can use?
vannak139 t1_j7w5otz wrote
Reply to comment by Optoplasm in [D] Image object detection, but for 1 dimensional data? by Optoplasm
So, the simple strategy here, which kind of ignores your variable length objects, is to simply classify CNN receptive fields directly, and then Max Pool the multiple classification frames.
So, lets say that your sequence is 1024. You build a CNN that has a receptive field of 32, and a stride of 16. This network applied to the sequence will offer something like 63 "frames". Typically, the CNN would expand this network representation up with a large number of channels, take the GlobalMaxPooling to merge these frame's information, and then classify the sample.
Instead, you should classify the frames directly, meaning your output looks like 63 separate sigmoid classifications associated with regions of the signal. Then, you simply take the maximum of each classification likelihood, and use this for your image-level classification.
After training, you can remove the GlobalMaxPooling layer, and look at the segment classifications directly.
[deleted] t1_j7w44k2 wrote
[removed]
express_mode_420 t1_j7w3mrm wrote
Reply to comment by CeFurkan in [D] Are there any AI model that I can use to improve very bad quality sound recording? Removing noise and improving overall quality by CeFurkan
Could you speech-to-text your lecture, collecting timestamps, do the same with TTS and automagically sync that way?
CeFurkan OP t1_j7wt0as wrote
Reply to comment by Locomule in [D] Are there any AI model that I can use to improve very bad quality sound recording? Removing noise and improving overall quality by CeFurkan
ye that is the game ,i develop : https://www.monstermmorpg.com