Recent comments in /f/MachineLearning

YOLOBOT666 t1_j6ziz4m wrote

Yeah, this would be a course in RL, most likely using RL bible as main reference textbook. Agree with the other comment, these lectures are all available online.

What I found valuable in attending a course in person was the prof, lots of insights and intuitions explained in person/office hours was the most valuable part for me. While I was taking the RL course in person, I also referenced online lectures and notes.

In terms of data science interviews and jobs, Bayesian would be more useful, at least more than RL unless you found yourself in robotics or some very niche industry.

1

Professional_Poet489 t1_j6zgk6h wrote

You can find good lectures on all of these topics on youtube, coursera, etc, but that's also true about Bayesian methods. RL is more fun IMO, but less employable for now. RL is used all over the place for things like recommender engines, ad promotion, etc. The concepts are super valuable. Bayesian methods are a bit more generic and common, and tbh are going out of vogue in most of robotics.

1

LeumasInkwater t1_j6zdg60 wrote

Honestly all the GPT stuff they are introducing seems pretty useful.

I like the idea of having automatic tasks generated after a meeting. I usually jot down 'follow-up' items while in meetings, and send them out to relevant coworkers afterward. It would only save me 5 minutes or so after every call, but could maybe help me focus more on what's being said rather than writing everything down 🤷‍♂️.

Also flagging parts of a meeting that you missed, auto-chapters, and tagging sections by the speaker all seem genuinely helpful.

That being said, my company doesn't use Microsoft products, so I hope to see features like this come to other platforms.

9

visarga t1_j6zb9em wrote

Many AI teams are scrambling now to label data with GPT-3 and train their small efficient models from GPT-3 predictions. This makes the hard part of data labelling much easier, speeds up development 10 times. In the end you get your cheap & fast models that work about as good as GPT-3 but only on a narrow task.

7

fuscarili OP t1_j6zb7y9 wrote

This is the syllabus:

Reinforcement learning in non-sequential problems:

  • Non-contextual multi-armed bandits
  • Contextual Multi-Armed Bandits

Reinforcement learning in sequential problems:

  • Dynamic planning. Bellman's formula
  • Value-Based Methods
  • Policy-based methods
  • Actor-critic methods
  • Model-Based Methods

Would you say it's within the basic stuff? I honestly have no clue

2