Recent comments in /f/MachineLearning

tysam_and_co OP t1_j6gzgf9 wrote

I've created a writeup for each release in https://github.com/tysam-code/hlb-CIFAR10/releases

Maybe I could do more at some point (like the original "bag of tricks" blogpost) once the smoke all clears, but I spent ~8-10 hours straight manually tuning hyperparameters yesterday night so I am smoked! Though usually I am pretty beat on release days due to trying to keep up with everything in the dopamine rush. :D

17

MrAcurite t1_j6gr1n8 wrote

I would argue that EE is actually a better major than CS for ML. It beefs up your Math and Statistics chops with DiffEQ, Quantum, and the like, and also includes enough Linear Algebra and Statistics to get you sorted. As a Math major doing ML research, I'm kind of embarrassed by how weak my background in Signal Processing is, and am working through a textbook on DSP in my spare time to fix that.

5

PretendiFendi t1_j6goq91 wrote

No probably not. If you have money to travel and your work can be done at home, you could probably arrange to fly back and forth later in your phd. No one is going to know where you are.

I did an experimental PhD, and there was a guy in my group who had a baby with his phd student wife across the country. He flew back and forth without any of us knowing. He would typically do weeks of work in a single week by working literally 20 hours a day and then go stay with her for two weeks and fly back before his next meeting so no one knew. We all didn’t know bc he was co-advised, so we all thought when he wasn’t in our lab he was in the other.

What I’m saying is that as you get older and more independent you can get away with a lot, people don’t check how you’re doing your work.

1

tysam_and_co OP t1_j6gjzzf wrote

Many thanks! I've found that the 'speed hunger' for me is truly insatiable -- we're almost at half the time it takes to train as when we started, and I find myself just as hungry to make it faster and faster. The Sisyphean hill is real, though I suppose it is more easily justified with a goal in mind! 😄😁

9

dineNshine t1_j6gikpr wrote

Children and pets are not the same as adults. Guns are also different from language models and image generators. A gun is a weapon, but a language model isn't.

Adding certain protections might be necessary for objects that can otherwise cause bodily harm to the user (e.g. gun safeties), but if you think that people must be prevented from accessing information because they are too stupid to properly evaluate it, then you might as well abolish democracy.

I am not doubting that people can evaluate information incorrectly. The issue is that nobody can do it in an unbiased way. The people doing the censorship don't know all that much better and often don't have the right intentions, as is often demonstrated.

It has been shown that ChatGPT has strong political biases as a result of the tampering applied to make it "safe". I find this concerning.

1