Recent comments in /f/MachineLearning

Rubberdiver t1_j99txcd wrote

I noticed ChatGPT can show me some example code but it's far from working (eg. Variables not defined...).

My project: I try to track fishes in a pond filmed from above and calculate their speed to see health issues if their movementspeed changes. For training I have videos of different daytimes.

ChatGPT gave me code but never told me really how to train a model on the PC that will work good enough on a Raspberry Pi 3 or 4. Is there any "known to work" code or tutorial that I can use to start my project from? I did some stuff in Python on the Pi, but I'm far from a programmer. Help?

1

pyepyepie t1_j99prs0 wrote

Reply to comment by TeamRocketsSecretary in [D] Please stop by [deleted]

LOL, I don't know what to say. I personally don't have anything smart to say about this question currently, it's as if you ask me if there is external life. Sure, I would watch it on Netflix if I have time, but generally speaking, it's way out of my field of interest. When you say snake oil, do you mean AI ExPeRtS? Why would you care about it? I think it's good that ML becomes mainstream.

1

CurrentlyJoblessFML t1_j99mphb wrote

I definitely think diffusion based generative ai models are a great idea. And whole heartedly agree that training GANs can be very painful. Head over to the hugging face diffusers library and you should be able to find a few models that are able to do unconditional image generation. They also have cookie cutter scripts that you can just execute to start training your model from the get go. They also have detailed instructions for how you can set up your own training data.

Although I have been working with these models for a while and I think training diffusion models can be very computationally intensive. Do you have access to a GPU cluster? If not, I’d recommend a U-Net based approach which you could train on GPU/TPUs on Google colab.

I have been using these class of models for my masters thesis and I would be happy to help in case you have any questions. Good luck! :)

2

GeorgLegato t1_j99k0rt wrote

and since it is simple black and white drawings you could use my vectorise extension to produce thousands of svg. so no scaling issues for large canvas or use at any smaller resolutions on mobiles etc

2

icelahtte t1_j99jzil wrote

Engineers, Mechanics, and Contractors of Thailand, I need your help!

I am currently outsourcing to the best Drive Company in Asia for my Water and Heavy industry business. I am gathering reviews and feedback about specific brands to decide what is the best Drive Company in the Thailand market. 

You can click the link to access the survey form: https://forms.gle/qr7h7Z6HhcDjmyHH7

Please feel free to leave a comment if you're interested or if you know someone who's related to these fields. I would greatly appreciate your response and time. Thank you so much!

1

Dovermore t1_j99j8ns wrote

I am trying to find tasks that use discrete tokens as inputs to do classification. E.g. some NLP classification tasks operate on a set of tokens (words, characters, special token sets, etc), and malware classification can operate on raw bytes. Is there any other domain that uses discrete sequences of tokens as inputs?

1

sam__izdat t1_j99j0iu wrote

You're not likely to get much help there, unfortunately. With SD, your best bet would probably be Dreambooth, which you can get with the Huggingface diffusers library. It might be overcomplicating matters, if the site is representative of your training data, though. GANs can be notoriously difficult to train but it's probably worth a shot here -- it's a pretty basic use case. You might look into data augmentation and try a u-net with a single-channel output.

A slightly more advanced option might be ProGAN. Here's a good video tutorial if that's your thing.

6

radi-cho OP t1_j99fh5s wrote

About the intuition that it would produce responses further from the human ones (in fact, we see that for this variant, the BLEU is lower) - in a way, it could work as a regularization to produce more diverse responses and prevent some overfitting. That loss mostly affects the additional head's weights which are removed during inference, but we also multiply it by an optimal constant to be sure it doesn't affect the whole architecture too much. I've sent you a PM if you wish to receive some more details or empirical insights.

2

impossiblefork t1_j99edtf wrote

17

Taenk t1_j99bo8q wrote

Maybe ask over at /r/stablediffusion and check out aesthetic gradients over there. Might be able to replicate your art style and scale it to the thousands of images you'll need to generate.

5

Red-Portal t1_j994qi0 wrote

AISTATS tend to be more popular these days, probably due to the conference timing. If you don't want to submit to AAAI, AISTATS is the other option. Also, the review process is much less noisy due to the better focus, and you get 5 reviews in general. In terms of content, they have slightly different flavors. Traditionally, people doing Bayesian nonparametrics have favored UAI, and it still somewhat seems to be the case.

2