Recent comments in /f/MachineLearning

_Arsenie_Boca_ OP t1_j9gix7q wrote

If I understand you correctly, that would mean that bottlenecks only interesting when

a) you further use the lower dimensional features as output like in autoencoders b) you are interested in knowing if your features have lower intrinsic dimension

Both are not met in many cases such as normal ResNets. Could you elaborate how you believe bottlenecks act as regularizers?

2

marcus_hk t1_j9gij1a wrote

Looks great. Might not be intelligible to those who don't know what they're looking at, though. Maybe include labels of, say, filters, what each slice of input represents, etc.?

Would like to see the same for normalization layers. And RNNs. And transformers. Keep it up!

61

Professional_Poet489 t1_j9gh652 wrote

The theory is that bottlenecks are a compression / regularization mechanism. If you have a smaller number of parameters in the bottleneck than overall in the net, and you get high quality results from the output, then the bottleneck layer must be capturing the information required to drive the output to the correct results. The fact that these intermediate layers are often used for embeddings indicates that this is a real phenomenon.

32

yaosio t1_j9gfypb wrote

This has very limited use as they already have the tools to deal with it. There's a second bot of some kind that reads the chat and deletes things if it doesn't like what it sees. Adding the ability to detect when commands are giving through a webpage would close it off. Then you would need some extra clever methods of working around it, such as putting the page in a format Sydney can read but the bot can't read.

1

aMericanEthnic t1_j9gf0l3 wrote

Bottlenecks are typically a point that is outside of control, purposeful implementation of a bottleneck can only be explained as an attempt at ambiguity in the sense that it’s an attempt to appear of create the feel of a real world issue’ , they “bottlenecks” are unnecessary and should be removed…

−14

notdelet t1_j9g627c wrote

> Assuming Gaussianity and then using maximum likelihood gives yields an L2 error minimization problem.

Incorrect, only true if you fix the scale parameter. I normally wouldn't nitpick like this but your unnecessary usage of bold made me.

> (if you interpret training as maximum likelihood estimation)

> a squared loss does not "hide a Gaussian assumption".

It does... if you interpret training as (conditional) MLE. Give me a non-Gaussian distribution with an MLE estimator that yields MSE loss. Also, residuals are explicitly not orthogonal projections whenever the variables are dependent.

0

OpeningVariable t1_j9g5llt wrote

I don't think it can make a "real" research paper, but it surely is interesting to know. I think, writing it up in a short workshop paper could work. I also think, if you continue working on this and have multiple instances of observations and injections made over time, it could maybe become an overview article and something that could go in a journal.

1

Ok-Assignment7469 t1_j9g51o4 wrote

These models are mainly based on reinforcement learning and the goal is to give you an answer which makes u happy the most. If you keep bugging it , eventually it will tell you the password at some point, because you are asking for it , and the bot s main goal is to satisfy your questions with probability and not reasoning because it was not designed to have a reasonable behavior

−6

BrandonBilliard t1_j9g3111 wrote

Hey,

Many of the proposed legal regulations for systems such as autonomous vehicles mention the need for explainability or transparency in the decision making processes of said vehicles. My understanding however was that due to their deep-learning processes, this is either extremely hard or impossible to do?

Is my understanding correct? Or is explainability possible in deep-learning systems?

1