Recent comments in /f/MachineLearning
theidiotrocketeer t1_j9e9iw9 wrote
Reply to [D] Simple Questions Thread by AutoModerator
Is it psychotic to use a GPT based model for what could be treated as image segmentation?
For my task, I trained a GPT model to predict a mask for an Input Integer Matrix with certain rows being entirely a spurious value. Where the mask is replacing the spurious integers with X's. It is a text based model for what could be considered an image task.
Snoo9704 t1_j9e8k2w wrote
Reply to [D] Large Language Models feasible to run on 32GB RAM / 8 GB VRAM / 24GB VRAM by head_robotics
I'm a super learning noob, but is there a reason you can't substitute large amounts of VRAM with large amounts of DRAM?
I know RAM bandwidth is important, but does it make that much of a difference if I got 256GB of quad channel DRAM and only 8GB VRAM? Compared to a more typical 32GB DRAM and 24GB VRAM?
halixness t1_j9e80y1 wrote
Reply to [D] Large Language Models feasible to run on 32GB RAM / 8 GB VRAM / 24GB VRAM by head_robotics
So far I have tried BLOOM Petals (a distributed LLM), inference took me around 30s for a single prompt on a 8GB VRAM gpu, but not bad!
Baeocystin t1_j9e6s12 wrote
Reply to comment by Last-Belt-4010 in [D] Large Language Models feasible to run on 32GB RAM / 8 GB VRAM / 24GB VRAM by head_robotics
The tl;dr for all GPU questions is that CUDA is the answer. There are no other even 'kinda' contenders.
I'm not happy about the monopoly, but that's where we're at, and there is nothing on the horizon pointing otherwise, either.
ajt9000 t1_j9e4jah wrote
Reply to comment by Borrowedshorts in [D] On papers forcing the use of GANs where it is not relevant by AlmightySnoo
Two opposite concepts
[deleted] t1_j9e1uoq wrote
Reply to [D] Large Language Models feasible to run on 32GB RAM / 8 GB VRAM / 24GB VRAM by head_robotics
[deleted]
catch23 t1_j9dxlze wrote
Reply to comment by smallfried in [D] Large Language Models feasible to run on 32GB RAM / 8 GB VRAM / 24GB VRAM by head_robotics
Their benchmark was done on a 16GB T4 which is anything but beefy. The T4 maxes out at 80W power consumption, and was primarily marketed toward model inference. The T4 is the cheapest GPU offered by google cloud.
Sandy_dude OP t1_j9duz1x wrote
Reply to comment by IIIBlueberry in [R] Looking for papers which are modified variational autoencoder (VAE) by Sandy_dude
Thank you!
smallfried t1_j9dtyf7 wrote
Reply to comment by catch23 in [D] Large Language Models feasible to run on 32GB RAM / 8 GB VRAM / 24GB VRAM by head_robotics
That is very interesting!
The paper is not yet on GitHub, but I'm assuming the hardware requirements are as mentioned one beefy consumer GPU (3090) and a whole bunch of DRAM (>210GB) ?
I've played with opt-175b and with a bit of twiddling it can actually generate some Python code :)
This is very exciting as it gets these models into the prosumer range hardware!
YinYang-Mills t1_j9dtwjh wrote
Reply to [D] Large Language Models feasible to run on 32GB RAM / 8 GB VRAM / 24GB VRAM by head_robotics
Is there a way to do it with single precision?
xrailgun t1_j9dtp9c wrote
Reply to comment by Emergency_Apricot_77 in [D] Large Language Models feasible to run on 32GB RAM / 8 GB VRAM / 24GB VRAM by head_robotics
It might not be unreasonable to think maybe OP primarily wants the functionality of current LLMs, and if something can provide that more efficiently (or has promise to in the near future), s/he may want to know about it too.
pyonsu2 t1_j9ds6j5 wrote
Reply to [D] Large Language Models feasible to run on 32GB RAM / 8 GB VRAM / 24GB VRAM by head_robotics
Depends on what you’re trying to do but just use OpenAI APIs. Your effort/time is also expensive.
ClothesRight t1_j9drvwt wrote
Reply to comment by Valachio in [D] What's the best way to capture a person's 3D likeness right now? by Valachio
I do not, sorry.
Valachio OP t1_j9dqpnt wrote
Reply to comment by xorbinant_ranchu in [D] What's the best way to capture a person's 3D likeness right now? by Valachio
Interesting stuff. The research is actually a couple of years old which is surprising, I expected it to be like months old and not years old. Are you involved with the project?
Valachio OP t1_j9dqmsh wrote
Reply to comment by parabellum630 in [D] What's the best way to capture a person's 3D likeness right now? by Valachio
DECA looks cool and it looks viable for my project. I'll look into it a bit deeper. Thanks for the suggestion.
Valachio OP t1_j9dqisx wrote
Reply to comment by ClothesRight in [D] What's the best way to capture a person's 3D likeness right now? by Valachio
That's super cool. Do you know what kind of techniques they are using?
Superschlenz t1_j9doev2 wrote
Reply to comment by Valachio in [D] What's the best way to capture a person's 3D likeness right now? by Valachio
Anytwo would be cheaper. One person is standing still and the other person is moving around them making photos or a video with a phone. Turntables large enough to hold a person and actuators for moving the phone up and down sound expensive.
ClothesRight t1_j9dn7ud wrote
Try luma AI https://lumalabs.ai/
johnnydaggers t1_j9dmhhv wrote
I have been doing my best to beat them back in peer-review, but I can only do so much...
IIIBlueberry t1_j9dkgon wrote
Agreeable-Run-9152 t1_j9dh1pu wrote
Reply to comment by Mefaso in [D] On papers forcing the use of GANs where it is not relevant by AlmightySnoo
I would assume that someone who is capable of programming a GAN and go through all the steps of Parameter Tuning at some Point should realize that the randomness shouldnt do anything.
Ferocious_Armadillo t1_j9dgsw0 wrote
Reply to comment by Stellar_____ in [D] Simple Questions Thread by AutoModerator
The “line down the middle” is where the “predicted” shark is the true shark (the type of shark was guessed correctly. The color of the square represents the certainty of the guess. *How confident is the software that it made the “right” guess? In these cases, it’s 100% confident since it is the right one!)
Pretty cool if you ask me!
You can match them up to see for yourself.
Valachio OP t1_j9dgmvq wrote
Reply to comment by mansumi_ in [D] What's the best way to capture a person's 3D likeness right now? by Valachio
I want to let users upload a 3D model of themselves onto an app and try out different clothing on their model. I can tell you more over message if you're interested in hearing more
machineko t1_j9dgblo wrote
You can use langchain with open source models like Flan-T5 or GPT-J as well. Just need to deploy them as an API endpoint and point to it.
tyras_ t1_j9e9kp0 wrote
Reply to comment by nikola-b in [D] Large Language Models feasible to run on 32GB RAM / 8 GB VRAM / 24GB VRAM by head_robotics
Free for now or free for an hour as the pricing tab indicates?