Recent comments in /f/MachineLearning
C_l3b OP t1_j7qc0cr wrote
Reply to comment by fnbr in [D] List of RL Papers by C_l3b
That looks really cool, thx
jmmcd t1_j7qb9i1 wrote
Reply to comment by SimonJDPrince in [D] Understanding Vision Transformer (ViT) - What are the prerequisites? by SAbdusSamad
This book is really excellent! I'm working through it and collecting a few typos. I'll pass them on when done. I'm going to recommend it to my students this semester.
---AI--- t1_j7qa9ec wrote
Reply to comment by LeftToSketch in [N] Google: An Important Next Step On Our AI Journey by EducationalCicada
Thanks
fnbr t1_j7q9v1m wrote
Reply to [D] List of RL Papers by C_l3b
Sutton and Barto is obligatory if you want to learn RL, imo. Even as an experienced researcher, I read it every year or so. It's very approachable.
currentscurrents t1_j7q8q5v wrote
Reply to comment by wintermute93 in [Discussion] Cognitive science inspired AI research by theanswerisnt42
So far nobody's figured out a good way to train them.
You can't easily do backprop, but you wouldn't want to anyway - the goal of SNNs is to run on ultra-low-power analog computers. For this you need local learning, where neurons can learn by communicating only with adjacent neurons. There's some ideas (forward-forward learning, predictive coding, etc) but so far nothing is as good as backprop.
There's a bit of a chicken-and-egg problem too. Without a good way to train SNNs, there's little interest in the specialized hardware - and without the hardware, there's little interest in good ways to train them. You can emulate them on regular computers but that removes all their benefits.
answersareallyouneed t1_j7q83qv wrote
Reply to [D] What do you think about this 16 week curriculum for existing software engineers who want to pursue AI and ML? by Imaginary-General687
I’d add lecture(s) talking about MAP/MLE, bias-variance trade off, and model interpretability, common pitfalls (Eg. Concept drift), and (maybe) building ml systems.
I’d skip the lectures on reinforcement learning and gans and maybe add a lecture on recommender systems. I’d say you need quite a bit of knowledge on both of these topics before you can actually solve real/practical problems.
Honestly, 16 weeks isn’t a lot of time to learn/digest all of this material in depth. I’d focus a lot more on the practical.
Iunaml t1_j7q38uc wrote
So that's an ad.
I don't like this "subtle" style of marketing. We're talking about a $63 book and yet the first sentence is puzzling.
z_fi t1_j7q0h1g wrote
Reply to [D] What do you think about this 16 week curriculum for existing software engineers who want to pursue AI and ML? by Imaginary-General687
A typical machine learning curriculum should cover the following topics:
Introduction to machine learning
Linear Regression
Logistic Regression
Decision Trees and Random Forests
Naive Bayes
k-Nearest Neighbors (k-NN)
Support Vector Machines (SVMs)
Neural Networks
Convolutional Neural Networks (CNNs)
Recurrent Neural Networks (RNNs)
Generative Adversarial Networks (GANs)
Clustering (K-means, Hierarchical)
Dimensionality Reduction (PCA, t-SNE)
Ensemble Methods
Model evaluation and selection
Hyperparameter tuning
Regularization
Bias-Variance Trade-off
Overfitting and Underfitting
Model interpretability and explainability
[deleted] t1_j7q0a2c wrote
sonofmath t1_j7q03db wrote
Reply to comment by mr_house7 in [D] List of RL Papers by C_l3b
Well.. kind of. Now for courses I would recommend Silver's course, followed by Levine's course, which are both available on youtube (besides reading the Sutton-Barto book). But besides the reading list, it also provides a detailed explaination of the most important model-free algorithms, as well as code implementations that are supposed to be as easy to understand as possible. Now if you want performent code for research/personal projects, I would not recommend SpinningUp, but it is a great way to learn how they are implemented.
ThrillHouseofMirth t1_j7q0053 wrote
Reply to comment by jobeta in [N] Getty Images sues AI art generator Stable Diffusion in the US for copyright infringement by Wiskkey
The dalle2 thing is an interesting wrinkle but a judge will need to be convinced that a photograph and generated-image that looks like a photograph are the same thing. They aren’t imo but it doesn’t matter what I think.
iron_proxy t1_j7pyvou wrote
Reply to [D] List of RL Papers by C_l3b
I woukd start with kaggle's RL course. Its a good into and has links to david silver's lecture series and sutton and barto's text book. Both are excellent intoductions to rl theory
VelveteenAmbush t1_j7pxngk wrote
Reply to comment by currentscurrents in [N] Microsoft announces new "next-generation" LLM, will be integrated with Bing and Edge by currentscurrents
Yes, 100% agree. This "can we coerce the model into saying something bad" is just a game that journalists play to catastrophize new technology and juice their engagement metrics. There's bad stuff on the internet, too, and you can find it with search engines. We still use search engines because they're incredibly useful.
The embarrassing part is that Google was so afraid of these BS stories that they kept LaMDA stuck in a warehouse for over two years while OpenAI and Microsoft lapped them.
wintermute93 t1_j7pxlsj wrote
Reply to comment by katadh in [Discussion] Cognitive science inspired AI research by theanswerisnt42
Have spiking networks actually produced any meaningful results? Granted, the last time I looked into the field was like 5 years ago, but back then the answer was definitely "no, these are just a toy".
rand3289 t1_j7pwnh3 wrote
Using spikes (points in time) instead of symbols (ex: [0,1]) to avoid the "Chinese Room argument". Here is my take on it: https://github.com/rand3289/PerceptionTime#readme
bocatty31 t1_j7pvp2d wrote
katadh t1_j7pv6f1 wrote
Look into spiking neural networks if you're not aware of them already
vannak139 t1_j7puon5 wrote
How you would approach this really depends on a few things. The most important question is, do you have the target data you want to get out of the network? It is possible, in some cases, to highlight regions of interest using only sample-level classification data. However, this usually is very context specific. If you have target data where these regions are already specified, a normal supervised learning method for wave forms should be perfectly workable, and will likely use 1D CNNs.
ElectroNight t1_j7poggj wrote
Reply to comment by Dr_Love2-14 in [Discussion] Is ChatGPT and/or OpenAI really the leader in the space? by wonderingandthinking
Meh, size of research team does not strongly correlate outcome quality and innovation. Furthermore bulky teams can reinforce momentum on a certain approach that turns into a dead end long term. Meanwhile small teams elsewhere start from a completely orthogonal approach and sometimes truly innovate. I'm not convinced Google has the right approach for the long term, organizationally or technically. Not saying ChatGPT is a Google killer either, yet.
Phoneaccount25732 t1_j7pncu3 wrote
Reply to [D] What do you think about this 16 week curriculum for existing software engineers who want to pursue AI and ML? by Imaginary-General687
You messed up the text in box 12, it's a duplicate of box 11.
AutomaticAccount6832 t1_j7pj318 wrote
Reply to [N] Microsoft announces new "next-generation" LLM, will be integrated with Bing and Edge by currentscurrents
I hope they don’t forget to make it compatible to Sharepoint and Teams as everything they do. Why would we need performance if we can have compatibility?
chrvt t1_j7pizuw wrote
Reply to [D] Normalizing Flows in 2023? by wellfriedbeans
We used NFs to estimate the ID of data, achieving SOTA results for very high-dimensional data where classical nearest neighbor methods fail:
GitGudOrGetGot t1_j7phuiz wrote
Reply to comment by PK_thundr in [N] Microsoft announces new "next-generation" LLM, will be integrated with Bing and Edge by currentscurrents
Boobies
mr_house7 t1_j7phbx1 wrote
Reply to comment by sonofmath in [D] List of RL Papers by C_l3b
Is this also kind of a course?
klop2031 t1_j7qcak1 wrote
Reply to comment by MisterBadger in [N] Getty Images sues AI art generator Stable Diffusion in the US for copyright infringement by Wiskkey
Take a gander here: https://youtu.be/G08hY8dSrUY At min 8 and 9 sec Seems like no one knows how scotus will deal with it but a good argument is that an AI is experiencing are like humans and generates new work by mixing in its skill.
Further, it seems like the law may only differentiate it by the intelligences' physical makeup.
And to be honest, it seems like the only ppl mad about generative networks producing art are the artists about to lose their jobs.
Who cares if an AI can create art, if one only cares about the creative aspect then the human can make art too, no one is stopping them. But really its about money.