Recent comments in /f/MachineLearning

currentscurrents t1_j9zwkw3 wrote

If I'm reading it right, it only works for LLMs that call an external source. Like Toolformer or Bing Chat. There's no way to inject it into ChatGPT or Github Copilot, it isn't a training data poisoning attack.

I think I remember somebody doing something like this against bing chat. They would give it a link to their blog, which contained the full prompt.

8

[deleted] t1_j9ztf95 wrote

  • Region of proposal NNs:A Region Proposal Network, or RPN, is a fully convolutional network that simultaneously predicts object bounds and objectness scores at each position.
  • YOLO uses the Pr(Object) which I am not sure whether it is a prediction from the model, but it seems so
  • And Region of Interest. That is widely used as well, but does not exactly tell "objectness".
2

[deleted] t1_j9zrouh wrote

I dont know what GM is, but for the second part, you can imagine that training a model is like finding a set of optimal conditions. However, some models find good conditions but not the best ones, this normally depends in the Loss function and other characteristics.

2

Scyther99 t1_j9zomj7 wrote

First point is like saying phishing was nonexistent before we invented computers and internet, so we dont have to worry about it once we invent them. There have been no AGI. There have been no comparable events. Basing it on fact that asteroid killing all life on earth is unlikely does not make sense.

6

should_go_work t1_j9zclbn wrote

Pattern Recognition and Machine Learning (PRML) and Elements of Statistical Learning (ESL) are two of the standard references that will give you what you're looking for with regards to the more classical topics you allude to (linear models, kernels, boosting, etc.).

1

currentscurrents t1_j9yxr37 wrote

Look up predictive coding; neuroscientists came up with it in the 80s and 90s.

A good portion of learning works by trying to predict the future and updating your brain's internal model when you're wrong. This is especially involved in perception and world modeling tasks, like vision processing or commonsense physics.

You would have a very hard time learning this from RL. Rewards are sparse in the real world, and if you observe something that doesn't affect your reward function, RL can't learn from it. But predictive coding/self-supervised learning can learn from every bit of data you observe.

You do also use RL, because there are some things you can only learn through RL. But this becomes much easier once you already have a rich mental model of the world. Getting good at predicting the future makes you very good at predicting what will maximize your reward.

6

qalis t1_j9y4c1m wrote

Yes, absolutely, for any size of the dataset and model this is strictly necessary. You can use cross-validation, Leave-One-Out CV, or bootstrap techniques (e.g. 0.632+ bootstrap). You don't need to validate if you don't have any hyperparameters, but this is very rarely the case; the only examples I can think of is Random Forest and Extremely Randomized Trees, where sufficiently large number of trees is typically enough.

1

Hyper1on t1_j9y3vz1 wrote

I mean, I don't see how you get a plausible explanation of BingGPT from underfitting either. As you say, models are underfit on some types of data, but I think the key here is the finetuning procedure, either normal supervised, or RLHF, which is optimising for a particular type of dialogue data in which the model is asked to act as an "Assistant" to a human user.

Part of the reason I suspect my explanation is right is that ChatGPT and BingGPT were almost certainly finetuned on large amounts of dialogue data, collected from interactions with users, and yet most of the failure modes of BingGPT that made the media are not stuff like "we asked it to solve this complex reasoning problem and it failed horribly", they are instead coming from prompts which are very much in distribution for dialogue data, such as asking the model what it thinks about X, or asking the model to pretend it is Y and you would expect the model to have seen dialogues which start similarly before. I find underfitting on this data to be quite unlikely as an explanation.

3

GandhisLittleHelper t1_j9xwca7 wrote

Has anyone made spectrogram 2 spectrogram models for music analysis, specifically demixing such as isolating vocals? I’m currently using a Mel spectrogram for the input and output but struggling to get good results. Using a hop length=512,n_fft=2048, no_mels=128. My model is currently a bi directional GRU model with 3 layers and a 256 hidden size.Does anyone know a good model type to use and/or good audio transformations for this project?

Cheers

1

Linear-- OP t1_j9xu7pn wrote

You can not just confidently infer meaning from the name. Is "Light Year" a unit of time?

By your logic, "unsupervised learning" is not supervised learning, while SSL is sometimes classified as part of unsupervised learning, so now SSL isn't SL as well!

So "I think classifying them as disjoint is somewhat misleading."

is obvious.

My fault, deleted. Satisfied now?

−2