Recent comments in /f/MachineLearning
katadh t1_j7s68c6 wrote
Reply to comment by wintermute93 in [Discussion] Cognitive science inspired AI research by theanswerisnt42
There has been a lot of progress in the last 2 - 3 years. They're still not quite at the level of ANNs in general but have been gaining ground quickly and do outperform ANNs on some specific tasks -- usually things with a temporal component but low data dimensionality per time-step. Another area with comparable results to ANNs would be object detection.
dancingnightly t1_j7s355b wrote
In a sense, you can communicate between semantic text embeddings and LM models through this method(would operate differently to multi modal embeddings): https://www.lesswrong.com/posts/mkbGjzxD8d8XqKHzA/the-singular-value-decompositions-of-transformer-weight
This method, which is only practical for toy problems really right now, would allow you to use semantic embeddings to find what to look for when doing SVD on an (autoregressive) LM. You could depend this on the input, for example, transforming your embedding into the keys to apply the abduction with in that process, and impacting the generation of logits. I'm not sure this would behave much differently to altering the logit_bias of tokens, but it would be interesting to hear if it was.
Optoplasm OP t1_j7rkrc3 wrote
Reply to comment by vannak139 in [D] Image object detection, but for 1 dimensional data? by Optoplasm
I would ideally be able to go in and select a start time and end time for each event within the longer timeseries and assign it a class, like you do for YOLO training labels, but in 1D. I can assemble and label the dataset, but the labeled segments will be extremely variable in length.
Chamrockk t1_j7rk2i7 wrote
Reply to [D] List of RL Papers by C_l3b
.
MisterBadger t1_j7rj1bf wrote
Reply to comment by klop2031 in [N] Getty Images sues AI art generator Stable Diffusion in the US for copyright infringement by Wiskkey
Machine learning algorithms are not even "intelligent" enough to filter out Getty watermarks.
They do not have minds or experiences, any more than zbrush or cinema4D or any other complicated software do.
Furthermore, they do not produce outputs like humans do - the speed and scale are more akin to automated car factories than human tinkers.
Fair use laws were not designed with them in mind.
theLanguageSprite t1_j7rgvme wrote
Reply to comment by geeky_username in [N] Google: An Important Next Step On Our AI Journey by EducationalCicada
Artificial general intelligence at this time of human development at this level of hardware, localized entirely within your warehouse?
theLanguageSprite t1_j7relsn wrote
Reply to [D] Which is the fastest and lightweight ultra realistic TTS for real-time voice cloning? by akshaysri0001
You have to pay to use the api and it’s completely closed source but resemble.ai works pretty well
[deleted] t1_j7redf1 wrote
like_a_tensor t1_j7rbdno wrote
Reply to comment by MrOfficialCandy in [D] Is English the optimal language to train NLP models on? by MrOfficialCandy
Sounds like you want something like a logical representation of sentences. Reducing sentences to first order logic might be what you're looking for. There's also AMRs (Abstract Meaning Representations). The problem with AMRs is that they need to be built, which is non-trivial for machines and time-consuming for humans.
MrSurname t1_j7ralcv wrote
Reply to Wouldn’t it be a good idea to bring a more energy efficient language into the ML world to reduce the insane costs a bit?[D] by thedarklord176
You can minimize the energy costs by hosting the primary interface on a COBOL mainframe, with eleventeen node-processors.
mjaltthrowaway t1_j7r9zon wrote
Reply to comment by SnuggleWuggleSleep in [D] What techniques can I use to tell if a problem is likely enough to be solved by ML so as to justify compiling the dataset? by SnuggleWuggleSleep
No it's real human. For some reason it didn't take in my first paragraph. What I was saying was that I own a high value domain related to vaccines and I was also thinking something along the similar lines of what OP posted.
And then the rest of my post was about starting from first principles and asking the simple question what problem does this machine learning website or app solve.
MrOfficialCandy OP t1_j7r85li wrote
Reply to comment by like_a_tensor in [D] Is English the optimal language to train NLP models on? by MrOfficialCandy
That doesn't help at all. Reading tokens at the byte level does not stop the word "they" or "it" from being vague in the context of a sentence.
MrOfficialCandy OP t1_j7r7xg3 wrote
Reply to comment by uoftsuxalot in [D] Is English the optimal language to train NLP models on? by MrOfficialCandy
It's not just about swapping tokens for other tokens. It's that grammatical structure (of any language) which can convey ambiguous meaning.
aicharades OP t1_j7r594n wrote
Reply to comment by szabokb in [P] ChatGPT without size limits: upload any pdf and apply any prompt to it by aicharades
I’m able to access it from https://www.wrotescan.com. Are you getting an error message?
like_a_tensor t1_j7r4mdx wrote
There's been some work on getting models to work at the byte level. An example: https://arxiv.org/abs/2105.13626
szabokb t1_j7r3yaz wrote
Is the website down?
[deleted] t1_j7r1mxo wrote
SnuggleWuggleSleep OP t1_j7qyhhs wrote
Reply to comment by mjaltthrowaway in [D] What techniques can I use to tell if a problem is likely enough to be solved by ML so as to justify compiling the dataset? by SnuggleWuggleSleep
Reads like an ai that needs more training wrote this.
mr_house7 t1_j7qpnc4 wrote
Reply to comment by sonofmath in [D] List of RL Papers by C_l3b
Thank you so much for your reply.
I did take a look at your suggestions. The Sergey Levine course seems really awesome, It is definitely on the list to do. But I'm a little bit conflicted regarding the David Silver course. I have a few follow-up questions, if you don't mind:
I'm currently doing Hugging Face Deep RL course. It is free and at the end I get a certification of completion, not that I care much about the certification, but it is always nice to get a certificate.
Also, isn't the David Silver course a little outdated? It seems that the video lectures were made 7 years ago. I guess the basics don't change, but I was wondering what is the best course to take. I think they are around the same lvl of difficulty. What would you choose?
ksblur t1_j7qozkl wrote
Reply to comment by infinity in [N] Microsoft announces new "next-generation" LLM, will be integrated with Bing and Edge by currentscurrents
> you.com
First time hearing about that search engine. I gave it a go, and man is it bad. I don’t understand how they think people will be loyal users of their “AI” search engine when nothing is intelligent about it.
I asked a simple query: “should I wear a jacket tomorrow?”, expecting it to interpret my query as “will it rain/be cold tomorrow” and this was the answer:
> It depends on the weather and the occasion. If it is mild or warm outside, then wearing a jacket may not be necessary. Ultimately, it is up to you to decide what to wear based on the weather, the occasion, and your personal style.
MAVAAMUSICMACHINE t1_j7ql8pt wrote
Reply to [D] List of RL Papers by C_l3b
You can follow along CMU’s deep RL course, they post which research papers they go over here: https://cmudeeprl.github.io/703website_f22/lectures/
JackBlemming t1_j7ql45e wrote
Reply to comment by Iunaml in [N] New Book on Synthetic Data: Version 3.0 Just Released by MLRecipes
There's nothing wrong with relevant self promotion, especially if it's high quality material. Obviously bad/irrelevant stuff should be removed, but that's up to the mods discretion.
I personally bookmarked this for later as it's very interesting to me.
uoftsuxalot t1_j7qkcq7 wrote
No, information is not reduced by just using another code
starstruckmon t1_j7qfcsl wrote
Reply to comment by jturp-sc in [N] Microsoft announces new "next-generation" LLM, will be integrated with Bing and Edge by currentscurrents
Looks so dated..
katadh t1_j7s73hw wrote
Reply to comment by currentscurrents in [Discussion] Cognitive science inspired AI research by theanswerisnt42
SNN - ANN conversion and surrogate gradient methods can both get good results these days, so training has become a lot more comparable to ANNs than it was in the past. I would agree though that there is a disconnect between the hardware and software still which is preventing SNNs from reaching the dream of super low power models.