Recent comments in /f/MachineLearning
ilyakuzovkin t1_jazwvuh wrote
Reply to To RL or Not to RL? [D] by vidul7498
I think RL is a niche by definition, but that's not a bad thing. If the problem you want to solve is about agents operating in interactive environments and maximizing some kind of utility function along the way - surely RL is your workhorse here.
Over the course of the last years we have seen successful applications of RL outside that narrow field of problems, where a problem that is seemingly not about agents and environments can still be formulated as an MDP and then solved with an RL approach. Because of these examples there seems to be a looming sentiment that RL is somehow "instead of" supervised, and questions like "which is better RL or supervised" arise.
My take on this would he that both are applicable in their appropriate spaces of problem formulations. Some problems are made to be solved with SL, some other ones with RL. And while it is feasible to twist an SL problem into RL framework, or even vice versa, it does not imply that one or the other is the ultimate tool.
Same way as one wouldn't use RL to multiply two numbers (except for academic interest), one should not use RL if it is not the right framework for the problem at hand. But for some other problems RL will definitely be (and already is, like in Go, Chess, Startcraft) the future.
currentscurrents t1_jazwqft wrote
Reply to To RL or Not to RL? [D] by vidul7498
The reason you want to do RL is that there's problem scenarios where RL is the only way to learn the problem.
Unsupervised learning can teach a model to understand the world, and supervised learning can teach a model to complete a human-defined task. But reinforcement learning can teach a model to choose its own tasks to complete arbitrary goals.
Trouble is, the training signal in reinforcement learning is a lot smaller, so you need ridiculous amounts of training data. Current thinking is that you need to use unsupervised learning to learn a world model + RL to learn how to achieve goals inside that model. This combination has worked very well for things like DreamerV3.
radi-cho OP t1_jazvwtm wrote
Reply to comment by Fickle_Dragonfly3090 in [P] diffground - A simplistic Android UI to access ControlNet and instruct-pix2pix. by radi-cho
More models will be added continuously. Depending on Apple's review timing and app usage, the iOS version might be coming in 1-2 weeks.
Bentastico t1_jazub0z wrote
Reply to comment by Art10001 in [P] LazyShell - GPT based autocomplete for zsh by rumovoice
bro relax ðŸ˜
[deleted] t1_jaztupq wrote
I also received the access link with .ac.kr domain.
Fickle_Dragonfly3090 t1_jazrqt5 wrote
Reply to comment by radi-cho in [P] diffground - A simplistic Android UI to access ControlNet and instruct-pix2pix. by radi-cho
How soon?)
ke7cfn t1_jazlzk6 wrote
Reply to comment by rumovoice in [P] LazyShell - GPT based autocomplete for zsh by rumovoice
Looks like here's what needs to be done:
https://www.reddit.com/r/macapps/comments/lea865/how_to_install_sshfs_on_big_sur/
rumovoice OP t1_jazl88i wrote
Reply to comment by ke7cfn in [P] LazyShell - GPT based autocomplete for zsh by rumovoice
> download and build sshfs for an m1 mac
its answer: git clone https://github.com/osxfuse/sshfs.git && cd sshfs && ./autogen.sh && ./configure && make && sudo make install
it doesn't do well in cases where it needs some recent knowledge like m1 issues
ke7cfn t1_jazjfwv wrote
Can it download and build sshfs for an m1 mac by query ?
chungexcy t1_jaz8tm4 wrote
Reply to [R] High-resolution image reconstruction with latent diffusion models from human brain activity by SleekEagle
>Figure 3 shows the results of visual reconstruction for one subject (subj01). We generated five images for each test image and selected the generated images with highest PSMs.
Something is not quite right. When they select the generated image, they use PSM score to select the best in 5. To calculate the PSM, I believe you need the original image (target, ground truth). It's like the LDM gives you five choices and you use your target pick the most similar one and then claim that this one is similar your target?
Accomplished-Fly-96 t1_jaz8gc8 wrote
Reply to [R] High-resolution image reconstruction with latent diffusion models from human brain activity by SleekEagle
In Sections 3.3 and 3.4, the authors mention linear models for the mapping between the text embeddings and the fMRI. I looked at their repository, but it does not have any code yet. Does anyone have a better idea about these linear models the authors talk about?
BrotherAmazing t1_jaz5fnx wrote
Reply to comment by rpnewc in [D] The Sentences Computers Can't Understand, But Humans Can by New_Computer3619
This is what I came here to say.
If one just reads about how ChatGPT was trained and understands some basics of machine learning, it’s quite obvious what you say has to be true. you
rumovoice OP t1_jaz2inw wrote
Reply to comment by kekinor in [P] LazyShell - GPT based autocomplete for zsh by rumovoice
shell_gpt seems more mature than that
https://github.com/TheR1D/shell_gpt
Both of those are separate scripts rather than hotkey bindings that work inline in your shell.
kekinor t1_jayxe3j wrote
Reply to comment by rumovoice in [P] LazyShell - GPT based autocomplete for zsh by rumovoice
There is also plz, which seems more mature:
Art10001 t1_jayg0yg wrote
Reply to comment by rumovoice in [P] LazyShell - GPT based autocomplete for zsh by rumovoice
SLA level of 99 % uptime/availability results in the following periods of allowed downtime/unavailability:
Daily: 14m 24s
Weekly: 1h 40m 48s
Monthly: 7h 14m 41s
Quarterly: 21h 44m 4.4s
Yearly: 3d 14h 56m 18s
Art10001 t1_jayfsfe wrote
Reply to comment by BeautifulLurker in [P] LazyShell - GPT based autocomplete for zsh by rumovoice
I've seen DM like 5 times in recent memory...
In reddit we use PM, Private Message. Not DM, Direct Message, which is a Discordism.
Sirisian t1_jayfok9 wrote
Reply to comment by Quazar_omega in [P] LazyShell - GPT based autocomplete for zsh by rumovoice
They bought https://ai.com the other day if you missed that. It directs to ChatGPT for now.
Art10001 t1_jayflg9 wrote
Reply to comment by maxToTheJ in [P] LazyShell - GPT based autocomplete for zsh by rumovoice
They have billions of dollars.
Art10001 t1_jayfgsc wrote
Reply to comment by crayphor in [P] LazyShell - GPT based autocomplete for zsh by rumovoice
I suggest another model such as OPT or even Flan-T5, because they're much easier to setup than OAI's outdated instructions that use outdated package versions that effectively demand a for-purpose VM or Docker.
ninjasaid13 t1_jay68c4 wrote
Where's the option that says, "I didn't get access"?
PM_ME_ENFP_MEMES t1_jay58f5 wrote
Superschlenz t1_jay1s3d wrote
Just wondering how many clicked on a random topic to see the 208 votes instead of waiting 2 days and 14 hours for the result...
Edit: topic → option
DAlmighty t1_jaxzv4n wrote
Reply to comment by Trotskyist in [P] LazyShell - GPT based autocomplete for zsh by rumovoice
That’s how I feel about https://regexr.com/. Without it, I’d be lost as well so don’t feel too bad.
Trotskyist t1_jaxyg2b wrote
Reply to comment by DAlmighty in [P] LazyShell - GPT based autocomplete for zsh by rumovoice
Been using regex's regularly for like 10 years now...still don't know how the hell to write them. Shameful, I know.
Shout out to some random guy named Olaf Neumann, without whom I'd be screwed: https://regex-generator.olafneumann.org/
2blazen t1_jazyryq wrote
Reply to comment by rpnewc in [D] The Sentences Computers Can't Understand, But Humans Can by New_Computer3619
Do you think an LLM can be taught to recognize when a question would require advanced reasoning to answer, or is it inherently impossible?