Recent comments in /f/MachineLearning
[deleted] t1_j9ztf95 wrote
Reply to comment by slickvaguely in [D] Simple Questions Thread by AutoModerator
- Region of proposal NNs:A Region Proposal Network, or RPN, is a fully convolutional network that simultaneously predicts object bounds and objectness scores at each position.
- YOLO uses the Pr(Object) which I am not sure whether it is a prediction from the model, but it seems so
- And Region of Interest. That is widely used as well, but does not exactly tell "objectness".
[deleted] t1_j9zrouh wrote
Reply to comment by FrostedFlake212 in [D] Simple Questions Thread by AutoModerator
I dont know what GM is, but for the second part, you can imagine that training a model is like finding a set of optimal conditions. However, some models find good conditions but not the best ones, this normally depends in the Loss function and other characteristics.
Scyther99 t1_j9zomj7 wrote
Reply to comment by VirtualHat in [D] To the ML researchers and practitioners here, do you worry about AI safety/alignment of the type Eliezer Yudkowsky describes? by SchmidhuberDidIt
First point is like saying phishing was nonexistent before we invented computers and internet, so we dont have to worry about it once we invent them. There have been no AGI. There have been no comparable events. Basing it on fact that asteroid killing all life on earth is unlikely does not make sense.
taken_every_username OP t1_j9zo17x wrote
Reply to comment by currentscurrents in [R] [P] New ways of breaking app-integrated LLMs with prompt injection by taken_every_username
Doesn't seem like there are any good mitigations right now and it affects pretty much all the useful use-cases for LLMs, even code completion...
should_go_work t1_j9zclbn wrote
Reply to comment by TinkerAndThinker in [D] Simple Questions Thread by AutoModerator
Pattern Recognition and Machine Learning (PRML) and Elements of Statistical Learning (ESL) are two of the standard references that will give you what you're looking for with regards to the more classical topics you allude to (linear models, kernels, boosting, etc.).
[deleted] t1_j9zaenf wrote
Reply to [D] Simple Questions Thread by AutoModerator
[removed]
currentscurrents t1_j9z82po wrote
Reply to [R] [P] New ways of breaking app-integrated LLMs with prompt injection by taken_every_username
Interesting. LLMs really need a better way to understand what instructions they should follow and what instructions they should ignore.
Neural network security is getting to be a whole subfield at this point. Adversarial attacks, training data poisoning, etc.
PhoibusApollo t1_j9z0hes wrote
Reply to comment by memberjan6 in [D] A funny story from my interview by nobody0014
That’s just weird
currentscurrents t1_j9yxr37 wrote
Reply to comment by Linear-- in [D] Isn't self-supervised learning(SSL) simply a kind of SL? by Linear--
Look up predictive coding; neuroscientists came up with it in the 80s and 90s.
A good portion of learning works by trying to predict the future and updating your brain's internal model when you're wrong. This is especially involved in perception and world modeling tasks, like vision processing or commonsense physics.
You would have a very hard time learning this from RL. Rewards are sparse in the real world, and if you observe something that doesn't affect your reward function, RL can't learn from it. But predictive coding/self-supervised learning can learn from every bit of data you observe.
You do also use RL, because there are some things you can only learn through RL. But this becomes much easier once you already have a rich mental model of the world. Getting good at predicting the future makes you very good at predicting what will maximize your reward.
DanielHendrycks t1_j9ytp0j wrote
Reply to [D] To the ML researchers and practitioners here, do you worry about AI safety/alignment of the type Eliezer Yudkowsky describes? by SchmidhuberDidIt
Here is a course for ML researchers about research areas that help reduce risks from AI (including today's risks as well as more extreme forms of them):
[deleted] t1_j9y7uzf wrote
[deleted]
davidmezzetti OP t1_j9y7tmq wrote
Reply to comment by visarga in [P] Introducing txtchat, next-generation conversational search and workflows by davidmezzetti
With the current version, yes it runs an embeddings query for each message. I plan to handle threaded conversations shortly. In that scenario, the chat history will be provided to the prompt.
visarga t1_j9y7fro wrote
Reply to comment by Linear-- in [D] Isn't self-supervised learning(SSL) simply a kind of SL? by Linear--
Words in language are both observations and actions. So language modelling is also a kind of supervised policy learning?
So... Self Supervised Learning is Unsupervised & Supervised & Reinforcement Learning.
visarga t1_j9y79v0 wrote
Reply to comment by cthorrez in [D] Isn't self-supervised learning(SSL) simply a kind of SL? by Linear--
But the text coming from a human should be considered "manually" labelled, right?
visarga t1_j9y7624 wrote
Reply to [P] Introducing txtchat, next-generation conversational search and workflows by davidmezzetti
Does it do only one round of retrieval?
qalis t1_j9y4c1m wrote
Yes, absolutely, for any size of the dataset and model this is strictly necessary. You can use cross-validation, Leave-One-Out CV, or bootstrap techniques (e.g. 0.632+ bootstrap). You don't need to validate if you don't have any hyperparameters, but this is very rarely the case; the only examples I can think of is Random Forest and Extremely Randomized Trees, where sufficiently large number of trees is typically enough.
Hyper1on t1_j9y3vz1 wrote
Reply to comment by Imnimo in [D] To the ML researchers and practitioners here, do you worry about AI safety/alignment of the type Eliezer Yudkowsky describes? by SchmidhuberDidIt
I mean, I don't see how you get a plausible explanation of BingGPT from underfitting either. As you say, models are underfit on some types of data, but I think the key here is the finetuning procedure, either normal supervised, or RLHF, which is optimising for a particular type of dialogue data in which the model is asked to act as an "Assistant" to a human user.
Part of the reason I suspect my explanation is right is that ChatGPT and BingGPT were almost certainly finetuned on large amounts of dialogue data, collected from interactions with users, and yet most of the failure modes of BingGPT that made the media are not stuff like "we asked it to solve this complex reasoning problem and it failed horribly", they are instead coming from prompts which are very much in distribution for dialogue data, such as asking the model what it thinks about X, or asking the model to pretend it is Y and you would expect the model to have seen dialogues which start similarly before. I find underfitting on this data to be quite unlikely as an explanation.
28Smiles t1_j9y2fw4 wrote
At least use an Ensemble and cross-validate, this way u get at least some meaningful results, but you are still in danger of overfitting
[deleted] t1_j9y1g6o wrote
Reply to [D] Simple Questions Thread by AutoModerator
[deleted]
paradigmai t1_j9xyrcy wrote
IMO, although the optimization techniques are the same, it is important to make this distinction because SSL does not require curated labels. And in some use cases SSL is not an option at all.
Siltala t1_j9xwvyh wrote
Reply to comment by KingsmanVince in [D] Isn't self-supervised learning(SSL) simply a kind of SL? by Linear--
Why does it not stand for Sexual Learning? I see a business opportunity…
GandhisLittleHelper t1_j9xwlz9 wrote
Reply to comment by JGoodle in [D] Simple Questions Thread by AutoModerator
The same but feeding frames from the videos into model like a CNN-RNN model which keeps memory of previous frames, but will obviously be a much bigger dataset.
GandhisLittleHelper t1_j9xwca7 wrote
Reply to [D] Simple Questions Thread by AutoModerator
Has anyone made spectrogram 2 spectrogram models for music analysis, specifically demixing such as isolating vocals? I’m currently using a Mel spectrogram for the input and output but struggling to get good results. Using a hop length=512,n_fft=2048, no_mels=128. My model is currently a bi directional GRU model with 3 layers and a 256 hidden size.Does anyone know a good model type to use and/or good audio transformations for this project?
Cheers
Linear-- OP t1_j9xu7pn wrote
Reply to comment by KingsmanVince in [D] Isn't self-supervised learning(SSL) simply a kind of SL? by Linear--
You can not just confidently infer meaning from the name. Is "Light Year" a unit of time?
By your logic, "unsupervised learning" is not supervised learning, while SSL is sometimes classified as part of unsupervised learning, so now SSL isn't SL as well!
So "I think classifying them as disjoint is somewhat misleading."
is obvious.
My fault, deleted. Satisfied now?
currentscurrents t1_j9zwkw3 wrote
Reply to comment by taken_every_username in [R] [P] New ways of breaking app-integrated LLMs with prompt injection by taken_every_username
If I'm reading it right, it only works for LLMs that call an external source. Like Toolformer or Bing Chat. There's no way to inject it into ChatGPT or Github Copilot, it isn't a training data poisoning attack.
I think I remember somebody doing something like this against bing chat. They would give it a link to their blog, which contained the full prompt.