Recent comments in /f/MachineLearning
mrwafflezzz OP t1_j9b414n wrote
Reply to comment by squidward2022 in [D] Relu + sigmoid output activation by mrwafflezzz
Very interesting. Thanks!
gpt-doktor-6b t1_j9b3u79 wrote
Reply to [D] Large Language Models feasible to run on 32GB RAM / 8 GB VRAM / 24GB VRAM by head_robotics
You might be interested in this tutorial on loading large models. They promise you the ability to inference model as long as you have enough disk space.
harharveryfunny t1_j9b30et wrote
Reply to comment by harharveryfunny in [D] Something basic I don't understand about Nerfs by alik31239
Not sure why this got downvoted given that it's correct. ChatGPT is also well capable of explaining how this mapping is learnt (using a view-consistency loss mapping from the 3D voxels back to a 2D view and comparing to image).
wywywywy t1_j9b2kqu wrote
Reply to comment by xrailgun in [D] Large Language Models feasible to run on 32GB RAM / 8 GB VRAM / 24GB VRAM by head_robotics
So, not scientific at all, but I've noticed that checkpoint file size * 0.6 is pretty close to actual VRAM requirement for LLM.
But you're right it'd be nice to have a table handy.
CommunismDoesntWork t1_j9b1qjb wrote
Reply to [D] Large Language Models feasible to run on 32GB RAM / 8 GB VRAM / 24GB VRAM by head_robotics
I'm surprised pytorch doesn't have an option to load models partially in a just in time basis yet. That way even an infinitely large model can be infered on.
marixer t1_j9b0x65 wrote
The step you're missing there is finding the cameras positions and angles with something like COLMAP, predicting them by extracting features from the images, pairing and triangulating. That data is then used alongside the RGB images to train the nerf
Pyramid_Jumper t1_j9ayed5 wrote
Been a while since I’ve read the paper but I don’t think you’re missing anything - apart from data in the correct format that is. You’ll need the aforementioned 5D vectors to be able to train/use this model.
If you can’t get that data then I’d suggest you look at other work that cite NeRF that maybe have data in a similar format to the data you do have
harharveryfunny t1_j9aydo9 wrote
Here's the key, thanks to CHatGPT:
Data preparation: First, the training data is preprocessed to convert the 2D images and camera poses into a set of 3D points and corresponding colors. Each 2D image is projected onto a 3D point cloud using the corresponding camera pose, resulting in a set of 3D points with associated colors.
go2carter t1_j9ay3ix wrote
This is very doable. What domain is your data? e.g. tabular, images, videos?
If it's vision, are you able to share a bit on what quality control metrics you have and whether you need to detect or classify anything in the images / video?
[deleted] t1_j9axwrk wrote
[deleted]
CatalyzeX_code_bot t1_j9avn4h wrote
Found relevant code at https://github.com/yenchenlin/nerf-pytorch + all code implementations here
--
To opt out from receiving code links, DM me
TruthAndDiscipline t1_j9avex4 wrote
Reply to comment by DevarshTare in [D] What matters while running models? by DevarshTare
I'm using a 3060 (no ti) with 12GB VRAM and train locally as well. Performance is fine, too.
xrailgun t1_j9avboh wrote
Reply to comment by wywywywy in [D] Large Language Models feasible to run on 32GB RAM / 8 GB VRAM / 24GB VRAM by head_robotics
Thanks!
I wish model publishers would indicate rough (V)RAM requirements...
squidward2022 t1_j9au3bg wrote
Reply to comment by mrwafflezzz in [D] Relu + sigmoid output activation by mrwafflezzz
Yup! If you look at the graph of tanh you will see relu(tanh) will smush the left half of the graph to 0. The right half of the graph on (0,infty) ranges in value from 0 and 1 but you can see saturation towards 1 starts to occur around 2-2.5. Since relu leaves this half unchanged you’ll be able to approach 1 very effectively with reasonable finite values.
[deleted] t1_j9atzu6 wrote
Reply to comment by squidward2022 in [D] Relu + sigmoid output activation by mrwafflezzz
[deleted]
[deleted] t1_j9ati5p wrote
Reply to comment by ArmagedonAshhole in [D] Large Language Models feasible to run on 32GB RAM / 8 GB VRAM / 24GB VRAM by head_robotics
[deleted]
[deleted] t1_j9at634 wrote
Reply to comment by Rockingtits in [D] Large Language Models feasible to run on 32GB RAM / 8 GB VRAM / 24GB VRAM by head_robotics
[deleted]
AtomicNixon t1_j9ar9hw wrote
Reply to [P] I've been commissioned to make 1000+ variations of my unique geometric art, while retaining its essential characteristics. It's been suggested that I use GAN to create permutations of my art. Any advice/directions? by eternalvisions
There's not enough here for a network to latch onto. I've trained nets on a variety of geometric patterns of differing styles so I know the minimum needed. I think banme's suggestion is the way to go. Figure out what your personal algo is and go with that.
wywywywy t1_j9ar2tk wrote
Reply to comment by xrailgun in [D] Large Language Models feasible to run on 32GB RAM / 8 GB VRAM / 24GB VRAM by head_robotics
I did test larger but it didn't run. I can't remember which ones, probably GPT-J. I recently got a 3090 so I can load larger models now.
As for quality, my use case is simple (writing prompt to help with writing stories & articles) and nothing sophisticated, and they worked well. Until ChatGPT came along. I use ChatGPT instead now.
xrailgun t1_j9aq903 wrote
Reply to comment by wywywywy in [D] Large Language Models feasible to run on 32GB RAM / 8 GB VRAM / 24GB VRAM by head_robotics
Did you test any larger and it wouldn't run?
Also, any comments so far among those? Good? Bad? Easy? Etc?
banmeyoucoward t1_j9apt54 wrote
Reply to [P] I've been commissioned to make 1000+ variations of my unique geometric art, while retaining its essential characteristics. It's been suggested that I use GAN to create permutations of my art. Any advice/directions? by eternalvisions
What tool did you use to make the art on your website?
Your style relies heavily on recursion and similarities between scales, which conv nets are not good at, but programatic descriptions of images like LOGO are very good at. My strategy would be to manually write simple LOGO, python (or whatever tool you initially used) programs that generate each of the images on your site, and then prompt Chat-GPT with “write a program that generates an image combining ideas from <Program A> and <Program B>
wywywywy t1_j9apjs3 wrote
Reply to [D] Large Language Models feasible to run on 32GB RAM / 8 GB VRAM / 24GB VRAM by head_robotics
I had a 3070 with 8GB and I managed to run these locally through KoboldAI.
Meta OPT 2.7B
EleutherAI GPT-Neo 2.7B
BigScience Bloom 1.7B
violet_zamboni t1_j9ak3ba wrote
Reply to [P] I've been commissioned to make 1000+ variations of my unique geometric art, while retaining its essential characteristics. It's been suggested that I use GAN to create permutations of my art. Any advice/directions? by eternalvisions
I don’t thing machine learning is going to get you good results. Have you asked this on r/generative ? It’s like something from there.
AdFew4357 t1_j9ai5ra wrote
Reply to [D] Simple Questions Thread by AutoModerator
Any MS level statisticians who have moved into ML research?
tdgros t1_j9b43pe wrote
Reply to comment by harharveryfunny in [D] Something basic I don't understand about Nerfs by alik31239
it's downvoted because it doesn't add anything to the conversation, OP has already stated that they know what info is input, they just don't know where to get it from. Someone already answered correctly at the top.