Recent comments in /f/MachineLearning

sneakpeekbot t1_j9j6ey9 wrote

Here's a sneak peek of /r/StableDiffusion using the top posts of all time!

#1: 🐢Turtleybug🐞 | 125 comments
#2: "Can an AI draw hands?" | 109 comments
#3: Stelfie Log #4 : Ulysses and the Trojan horse | 127 comments


^^I'm ^^a ^^bot, ^^beep ^^boop ^^| ^^Downvote ^^to ^^remove ^^| ^^Contact ^^| ^^Info ^^| ^^Opt-out ^^| ^^GitHub

1

No-Belt7582 t1_j9j6edy wrote

Ok it will try to share the resources here soon. I saved them somewhere yesterday

[1] J. Howard, “fast.ai - 1st Two Lessons of From Deep Learning Foundations to Stable Diffusion,” Oct. 19, 2022. https://www.fast.ai/posts/part2-2022-preview.html (accessed Feb. 20, 2023). [2] use_excalidraw, “Well-Researched Comparison of Training Techniques (Lora, Inversion, Dreambooth, Hypernetworks),” r/StableDiffusion, Jan. 15, 2023. www.reddit.com/r/StableDiffusion/comments/10cgxrx/wellresearched_comparison_of_training_techniques/ (accessed Feb. 20, 2023). [3] terrariyum, “Advanced advice for model training / fine-tuning and captioning,” r/StableDiffusion, Feb. 17, 2023. www.reddit.com/r/StableDiffusion/comments/114dxgl/advanced_advice_for_model_training_finetuning_and/ (accessed Feb. 20, 2023). [4] BarTraditional6305, “Alternative tools to fine tune stable diffusion models?,” r/StableDiffusion, Feb. 08, 2023. www.reddit.com/r/StableDiffusion/comments/10wqgok/alternative_tools_to_fine_tune_stable_diffusion/ (accessed Feb. 13, 2023). [5] Thick_Journalist_348, “LORA vs dream booth?,” r/StableDiffusion, Feb. 02, 2023. www.reddit.com/r/StableDiffusion/comments/10rihf7/lora_vs_dream_booth/ (accessed Feb. 13, 2023). [6] “AUTOMATIC1111/stable-diffusion-webui: Stable Diffusion web UI.” https://github.com/AUTOMATIC1111/stable-diffusion-webui (accessed Feb. 13, 2023). [7] AbdBarho, “Stable Diffusion WebUI Docker.” Feb. 13, 2023. Accessed: Feb. 13, 2023. [Online]. Available: https://github.com/AbdBarho/stable-diffusion-webui-docker [8] “stochasticai/x-stable-diffusion.” Stochastic, Feb. 13, 2023. Accessed: Feb. 13, 2023. [Online]. Available: https://github.com/stochasticai/x-stable-diffusion [9] N. W. Foong, “How to Fine-tune Stable Diffusion using Textual Inversion,” Medium, Oct. 24, 2022. https://towardsdatascience.com/how-to-fine-tune-stable-diffusion-using-textual-inversion-b995d7ecc095 (accessed Feb. 09, 2023). [10] N. W. Foong, “How to Fine-tune Stable Diffusion using Dreambooth,” Medium, Feb. 04, 2023. https://towardsdatascience.com/how-to-fine-tune-stable-diffusion-using-dreambooth-dfa6694524ae (accessed Feb. 09, 2023). [11] “The Annotated Diffusion Model.” https://huggingface.co/blog/annotated-diffusion (accessed Jan. 31, 2023). [12] J. Alammar, “The Illustrated Stable Diffusion.” https://jalammar.github.io/illustrated-stable-diffusion/ (accessed Jan. 31, 2023). [13] “Understanding Stable Diffusion from ‘Scratch.’” https://scholar.harvard.edu/binxuw/classes/machine-learning-scratch/materials/stable-diffusion-scratch (accessed Jan. 31, 2023).

3

Nill444 t1_j9j68sv wrote

>Today, science has explanations for many of those natural phenomena

Depends what you mean by explanations. You can always keep asking "why?" and at some point you won't be able to answer so you can just put god in there just like the early humans did. If you think what they did was reasonable then the same principle applies here, it's just that we have a deeper level of understanding

2

Raaaaaav t1_j9j21oh wrote

You could look at a YOLO tutorial, that's how I started back whenn YOLO was still new .

Maybe this repo will help you:

https://github.com/ultralytics/yolov5

But keep in mind that you will have to learn AI techniques to really be able to control the outcome of your model and that the training process can become unnecessarily expensive if you don't know when and how to optimize it.

1

currentscurrents t1_j9j0gt7 wrote

According to their paper, the LLM is doing task decomposition. You're able to give it high-level instructions like "go to the kitchen and make an omelette", and it breaks it down into actions like get eggs, get pan, get oil, put oil in pan, put eggs in pan, etc.

You could use something like this to give high-level instructions to a robot in plain English.

2

ChuckSeven t1_j9iyuc2 wrote

hmm not sure, but I think if you don't exponentiate you cannot fit n targets into a d-dimensional space if n > d and you want there to exist a vector v for each target such that the outcome is a one-hot distribution (or 0 loss).

Basically, if you have 10 targets but only a 2-dimensional space you need to have enough non-linearity in the projection to your target space such that there exists a 2d vector which gives 0 loss for each target.

edit: MNIST only has 10 classes so you are probably fine. Furthermore, softmax of the dot product "care exponentially more" about the angle of the prediction vector than the scale. If you use norm, I'd think that you only care about angle which likely leads to different representations. The fact that those may improve performance highly depends how your model may rely on scale to learn certain predictions. Maybe in case of mnist, relying on scale worsens performance (e.g. if you want a wild guess, because it maybe makes "predictions more certain" simply if it has more pixels set to 1).

3

vurt72 OP t1_j9iytee wrote

https://www.reddit.com/r/stablediffusion/wiki/tutorials/ ?

there's not a lot, and not a lot about machine learning tbh.

also, what i'm doing is fine-tuning? i've googled that as well and it's not totally clear to me what the differences would be. for fine tuning i did read (discord, stable tuner) that Class is not needed for doing that.

1