Recent comments in /f/MachineLearning
CellWithoutCulture t1_javhjpc wrote
Reply to comment by LetterRip in [D] OpenAI introduces ChatGPT and Whisper APIs (ChatGPT API is 1/10th the cost of GPT-3 API) by minimaxir
I mean... why were they not doing this already? They would have to code it but it seems like low hanging fruit
> memory efficient attention. 10x-20x increase in batch size.
That seems large, which paper has that?
rumovoice OP t1_javfy8c wrote
Reply to comment by omgpop in [P] LazyShell - GPT based autocomplete for zsh by rumovoice
it inserts $(pwd)
omgpop t1_javey7y wrote
Reply to comment by rumovoice in [P] LazyShell - GPT based autocomplete for zsh by rumovoice
How does it get the current dir in your example?
LeanderKu t1_javeqbn wrote
Reply to comment by A_HumblePotato in [R] High-resolution image reconstruction with latent diffusion models from human brain activity by SleekEagle
They probably don’t generalize. I bet they tried it
Thin_Sky t1_jav7a6e wrote
Reply to comment by harharveryfunny in [D] OpenAI introduces ChatGPT and Whisper APIs (ChatGPT API is 1/10th the cost of GPT-3 API) by minimaxir
Where do you find info on these 8k and 32k token prices? Is this listed on their page or is it leaked from consultations?
[deleted] t1_jav5pu1 wrote
What dataset did you use to train the model? I'm creating something similar for an app and looking for a dataset.
Edit: NVM, you are using OpenAI API.
Coyote-Sweaty t1_jav4ul6 wrote
Reply to comment by oneandonly13579 in [R] High-resolution image reconstruction with latent diffusion models from human brain activity by SleekEagle
Actually, its from 2022
karius85 t1_jav4q78 wrote
Reply to [R] High-resolution image reconstruction with latent diffusion models from human brain activity by SleekEagle
Any relation to the team in Kyoto who popped up in vsauce MindField episode a couple of years back?
rumovoice OP t1_jav292n wrote
Reply to comment by omgpop in [P] LazyShell - GPT based autocomplete for zsh by rumovoice
It sends only OS name and current command. I tried to avoid sending anything else for privacy reasons.
omgpop t1_jav1omw wrote
Does it/could it send your directory/file tree as part of the prompt?
drplan t1_jav01pf wrote
Reply to comment by TobusFire in [D] Are Genetic Algorithms Dead? by TobusFire
I think the best approach for this is thinking about the search space and the fitness landscape. If different components of the solution vector can independently improve the fitness crossover operators will have a positive impact.
Another aspect is the search space itself. Is it real-valued, is it binary, is it a tree-like structure,..?
Traditionally genetic algorithms are operating on binary encodings, and they often work ok problem which have binary solutions (a fixed-size vector of bits). These problem do not have gradient to start with. However one should investigate beforehand if there are combinatorial approaches to solve the problem.
For real-valued problems with no gradient: evolution strategies with a smart mutation operation like CMA (covariance matrix adaption) would be a good choice.
rumovoice OP t1_jauxt0h wrote
https://github.com/not-poma/lazyshell
A smart autocomplete script invoked with ALT+G. It can modify the existing command as well.
of_a_varsity_athlete OP t1_jauqrc5 wrote
Reply to comment by mrtransisteur in [D] Is there an ML project out there that recommends movies based on more than the usual features? by of_a_varsity_athlete
What prompts are you using?
currentscurrents t1_jatvmtm wrote
Reply to comment by OrangeYouGlad100 in [R] High-resolution image reconstruction with latent diffusion models from human brain activity by SleekEagle
You're right, I misread it. I thought they held out 4 patients for tests. But upon rereading, their dataset only had 4 patients total and they held out the set of images that were seen by all of them.
>NSD provides data acquired from a 7-Tesla fMRI scanner over 30–40 sessions during which each subject viewed three repetitions of 10,000 images. We analyzed data for four of the eight subjects who completed all imaging sessions (subj01, subj02, subj05, and subj07).
...
>We used 27,750 trials from NSD for each subject (2,250 trials out of the total 30,000 trials were not publicly released by NSD). For a subset of those trials (N=2,770 trials), 982 images were viewed by all four subjects. Those trials were used as the test dataset, while the remaining trials (N=24,980) were used as the training dataset.
4 patients is small by ML standards, but with medical data you gotta make do with what you can get.
I think my second question is still valid though. How much of the image comes from the brain data vs from the StableDiffusion pretraining? Pretraining isn't inherently bad - and if your dataset is 4 patients, you're gonna need it - but it makes the results hard to interpret.
OrangeYouGlad100 t1_jatt83m wrote
Reply to comment by currentscurrents in [R] High-resolution image reconstruction with latent diffusion models from human brain activity by SleekEagle
This is what they wrote:
"For a subset of those trials (N=2,770 trials), 982 images were viewed by all four subjects. Those trials were used as the test dataset, while the remaining trials (N=24,980) were used as the training dataset."
That makes it sound like 982 images were not used for training
OrangeYouGlad100 t1_jatsw72 wrote
Reply to comment by currentscurrents in [R] High-resolution image reconstruction with latent diffusion models from human brain activity by SleekEagle
> so their MRI->latent space model has seen every one of the 10,000 images in the dataset
Are you sure about that? I wasn't able to understand their test method from the paper, but it sounds like they held out some images from training
A_HumblePotato t1_jati59p wrote
Reply to [R] High-resolution image reconstruction with latent diffusion models from human brain activity by SleekEagle
Looks interesting, but as another user pointed out not particularly novel (aside from the decoder model being used). One thing I wish these studies did is to test these models on subjects that weren’t used for training of the model, to see if these methods generalize to several people (or at least a few-shot training/testing on new subjects). I do actually like the idea of using latent diffusion models for these tasks, as long-term our brain does not store perfect reconstruction of images.
Visible-Moment-8974 t1_jath1zi wrote
Reply to [R] High-resolution image reconstruction with latent diffusion models from human brain activity by SleekEagle
exciting! one day we will be able to visualize whats goin on in this amazing creature: https://www.youtube.com/watch?v=0vKCLJZbytU&ab_channel=NatureonPBS
currentscurrents t1_jat9lvg wrote
Reply to comment by WarAndGeese in [N] EleutherAI has formed a non-profit by StellaAthena
It's a joke. OpenAI was supposed to be a nonprofit too, now they look more like a Microsoft subsidiary.
SleekEagle OP t1_jaszawj wrote
Reply to comment by Zestyclose-Debt-4712 in [R] High-resolution image reconstruction with latent diffusion models from human brain activity by SleekEagle
It looks like, rather than conditioning on text they condition on the fMRI, but it's unclear to me exactly how they map between the two and why this would even work without finetuning. TBH I haven't had time to read the paper so I don't know the details, but figured I'd drop the paper in case anyone was interested!
Zestyclose-Debt-4712 t1_jasycf3 wrote
Reply to [R] High-resolution image reconstruction with latent diffusion models from human brain activity by SleekEagle
Does this research make any real sense? Creating a low resolution image from brain activity has been done before and is amazing. But using a pretrained denoising network on the noisy image will add just details that have nothing to do with the brain activity. Just like those ai-„enlarge/zoom“ models imagine/add details that never were in the original picture.
Or am I missing something here and they address the issue?
currentscurrents t1_jasxijr wrote
Reply to [R] High-resolution image reconstruction with latent diffusion models from human brain activity by SleekEagle
I'm a wee bit cautious.
Their test set is a set of patients, not images, so their MRI->latent space model has seen every one of the 10,000 images in the dataset. Couldn't it simply have learned to classify them? Previous work has very successfully classified objects based on brain activity.
How much information are they actually getting out of the brain? They're using StableDiffusion to create the images, which has a lot of world knowledge about images pretrained into it. I wish there was a way to measure how of the output image is coming from the MRI scan vs from StableDiffusion's world knowledge.
xGovernor t1_jasx7r9 wrote
Reply to comment by Im2bored17 in [D] OpenAI introduces ChatGPT and Whisper APIs (ChatGPT API is 1/10th the cost of GPT-3 API) by minimaxir
You needed the secret api key, included with the plus edition. Prior to Whispers I don't believe you could obtain a secret key. Also gave early access to new features and provides me turbo day one. Also I've used to much more and got turbo to work with my plus subscription.
Had to find a workaround. Don't feel scammed. Plus I've been having too much fun with it.
helliun t1_jasuo76 wrote
Reply to [R] High-resolution image reconstruction with latent diffusion models from human brain activity by SleekEagle
Wow the implications here are kinda insane
possibilistic t1_javklac wrote
Reply to [P] LazyShell - GPT based autocomplete for zsh by rumovoice
Is this a Super Mario RPG reference?