Recent comments in /f/MachineLearning

gsvclass OP t1_j9xttgi wrote

While it may seem that way correct answers are always expected but never delivered everything works within a margin of error with humans it's pretty large and not easy to fix. Also "correct" is subjective. LLMs are language models use the knowlede embedded in their wieghts combined with the context provided by the prompt to do their best. The positive thing here is that that the margin of error is actively being reduced withn LLMs and not so with however we did this before.

1

Linear-- OP t1_j9xt0nh wrote

I've now done some further research and read the comments.

By far, my conclusion is that, SSL is indeed, a type of SL. It contains features and corresponding label(s). From wikipedia:

>Supervised learning (SL) is a machine learning paradigm for problems where the available data consists of labeled examples, meaning that each data point contains features (covariates) and an associated label.

Since this is not a debate, I do not want to dwell on the definition. And indeed, *self-*supervised means that it does not require extra resource-consuming labelling from human, making training with huge datasets possible, like GPT-3.

And I disagree that seeing SSL as a kind of SL is the "wrong level" as a comment suggestted. What I originally intended to confirm was that, language modeling, which gives rise to GPT-3/ChatGPT... Is a kind of supervised learning with a large quantity (and sometimes good quality) of data. Strong model with simple, old methods.

−1

cthorrez t1_j9xstlw wrote

People are rushing to deploy LLMs in search, summarization, virtual assistants, question answering and countless other applications where correct answers are expected.

The reason they want to get to the latent space close to the answer is because they want the LLM to output the correct answer.

1

KingsmanVince t1_j9xr8oe wrote

>Not so constructive.

It's not much I am aware. However, what I mean that names of both training paradigm already told you a part of the answer. The last paragraph of mine is to refer two other comments to create a more sufficient answer.

Moreover, the names of both already pointed it's somewhat related. Therefore, this line

>So I think classifying them as disjoint is somewhat misleading.

is obvious. I don't know who have said "classifying them as disjoint" to you. Clearly they didn't pay attention to the names.

4

Linear-- OP t1_j9xqtsx wrote

It's clear that human and other animals must learn with reinforcement -- requiring the agent to act and recevive feedback/reward. This is an important part and I don't think it's proper to classify it as SSL. Moreover, psychology on learning points out that problem-solving and immediate feedback is very important for learning outcomes -- these feedbacks are typically human labels or reward signal.

1

KingsmanVince t1_j9xk6rf wrote

>Isn't self-supervised learning(SSL) simply a kind of SL?

Don't their names already tell that? Self-supervised learning... supervised learning...

>So I think classifying them as disjoint is somewhat misleading.

Who said this?

The ways of determining labels of both paradigms are different (as u/cthorrez said). Moreover, the objectives are different (as u/currentscurrents said).

7

Desticheq t1_j9xiv9l wrote

Well, in terms of "out-of-the-box," I'm not sure what else could be better. AWS, Azure or Google provide empty units basically, and you'd have to configure all the "Ops" stuff like network, security, load balancing, etc. That's not that difficult if you do it once in a while, but for a "test-it-and-forget-it" project it might be too difficult.

2

currentscurrents t1_j9xg9kn wrote

You're looking at the wrong level. SSL is a different training objective. Everything else about the model and optimizer is the same, but you're training it on a different problem.

Also SSL has other advantages beyond being cheaper. SL can only teach you ideas humans already know, while SSL learns from the data directly. It would be fundamentally impossible to create labels for every single concept a large model like GPT-3 knows.

Yann Lecun is almost certainly right that most human learning is SSL. Very little of our input data is labeled - and for animals, possibly none.

9

gt33m t1_j9xfxid wrote

Not certain where banning AI came into the discussion. It’s just not going to happen and I don’t see anyone proposing it. However, it shouldn’t be the other extreme either where everyone is running a nuclear plant in their backyard.

To draw parallels from your example, AI needs a lot of regulation, industry standards and careful handling. The current technology is still immature but if the right structures are not put in place now, it will be too late to put the genie back in the bottle later.

3

gt33m t1_j9xapzz wrote

Like I said this is similar to the guns argument. Banning guns does not stop people from Killing each other but easy access to guns amplifies the problem.

AI as a tool of automation is a force multiplier that is going to be indistinguishable from human action.

3