Recent comments in /f/MachineLearning
goldemerald t1_j5psrue wrote
Reply to [D] CVPR Reviews are out by banmeyoucoward
3 borderlines?!?!?
I've been in a state of shock for an hour.
AdFew4357 t1_j5pqkqb wrote
I have one minor gripe about deep learning textbooks. I think they are great references, but should not be used as a way for beginners to get into the field. I genuinely feel like time is better spent on the student going down a rabbit hole of actual papers of maybe one of the chapters of those books, say, a student reads the chapter on graph neural networks and the proceeds to read everything in graph neural networks, rather than read the whole book on different subsections.
SatoshiNotMe t1_j5ponoh wrote
Reply to comment by SimonJDPrince in [P] New textbook: Understanding Deep Learning by SimonJDPrince
Looks like a great book so far. I think it is definitely valuable to focus on giving a clear understanding of some topics rather than covering everything while compromising depth of understanding
SimonJDPrince OP t1_j5pdwwr wrote
Reply to comment by bythenumbers10 in [P] New textbook: Understanding Deep Learning by SimonJDPrince
That's not a bad idea actually!
bythenumbers10 t1_j5pa9iz wrote
When to reach for deep learning over older, simpler methods. Just an executive summary to keep folks from sandblasting soda crackers, or being forced to.
tfburns OP t1_j5p4owf wrote
Reply to comment by terranop in [N] Call for Tiny Papers @ ICLR, a DEI initiative by tfburns
We've updated the working def :)
Cyclone4096 t1_j5owtmi wrote
Reply to [D] Simple Questions Thread by AutoModerator
I don’t have too much background on ML. I want to build a fairly small neural network that has only one input which comes from a time series data and has to give only one output for that data. My loss function aggregates the entire time series output to get a single scalar value. I’m using PyTorch and when I call “.backward()” on the loss function it takes a long time (understandably). Is there an easier way to do this rather than doing backward gradient calculation on a loss function that itself is a result if 100s of millions values? Note that the neural network itself is tiny, maybe less than 100 weights, but my issue is that I don’t have any golden target, but I want to minimize a complex function calculated from the entire time series output.
ArnoF7 t1_j5omrfh wrote
Reply to comment by FastestLearner in [D] Multiple Different GPUs? by Maxerature
Great insight. Appreciate it
SimonJDPrince OP t1_j5olz4s wrote
Reply to comment by TheMachineTookShape in [P] New textbook: Understanding Deep Learning by SimonJDPrince
Thanks! If you send your real name to the e-mail on the front page of the book, then I'll add you to the acknowledgements.
SimonJDPrince OP t1_j5olux3 wrote
Reply to comment by new_name_who_dis_ in [P] New textbook: Understanding Deep Learning by SimonJDPrince
That was kind of my impression. And I do discuss this in the chapters on transformers and regularization. Was wondering if there is more to it.
zoontechnicon t1_j5oizdf wrote
Reply to comment by kernel_KP in [D] Simple Questions Thread by AutoModerator
You could build an autoencoder using CNNs and use the latent vectors as input to a clustering algorithm.
new_name_who_dis_ t1_j5oix1c wrote
Reply to comment by arsenyinfo in [P] New textbook: Understanding Deep Learning by SimonJDPrince
Fine tuning isn’t any different than just training…? You just don’t start with random network, but fine tuning doesn’t really have anything different besides that and the size of the dataset
zoontechnicon t1_j5oiraa wrote
Reply to [D] Simple Questions Thread by AutoModerator
I'm trying to use this model to summarize text: https://huggingface.co/bigscience/mt0-large Text generation seems to end after the special end token </s> however. I wonder how I would coax it to generate longer texts. Any ideas?
[deleted] t1_j5oiopw wrote
Reply to comment by [deleted] in [D] Simple Questions Thread by AutoModerator
[deleted]
Apprehensive-Grade81 t1_j5oijh0 wrote
Reply to comment by arsenyinfo in [P] New textbook: Understanding Deep Learning by SimonJDPrince
Definitely would like something like this. Maybe SOTA benchmarks as well.
TheMachineTookShape t1_j5ohns9 wrote
Reply to comment by SimonJDPrince in [P] New textbook: Understanding Deep Learning by SimonJDPrince
There's another on page 349 in section "Combination with other models":
>...will ensure that the the aggregated posterior...
TheMachineTookShape t1_j5ofkzb wrote
Reply to comment by TheMachineTookShape in [P] New textbook: Understanding Deep Learning by SimonJDPrince
Sorry, you've written the instructions right there on the Web page! Just ignore me...
TheMachineTookShape t1_j5ofex4 wrote
What is the most efficient way for someone to tell you about typos, or provide suggestions? I'll try to have a read over the weekend.
tfburns OP t1_j5ofatr wrote
Reply to comment by terranop in [N] Call for Tiny Papers @ ICLR, a DEI initiative by tfburns
Very valid point. Working on it (and also thinking to add some others, too).
bacocololo t1_j5of34s wrote
Reply to comment by SimonJDPrince in [P] New textbook: Understanding Deep Learning by SimonJDPrince
just above 3.1.2 sum of slopes from the the regions
terranop t1_j5of10a wrote
Is there a reason why trans people (specifically trans men, since trans people of other genders would qualify under the "gender" criterion) are not included in the URM criteria? It seems kinda odd to include all these other minority groups but not trans people. Are transgender people not actually underrepresented at ICLR?
tfburns OP t1_j5oe3nw wrote
Reply to comment by certain_entropy in [N] Call for Tiny Papers @ ICLR, a DEI initiative by tfburns
Changed. Thanks for the suggestion!
certain_entropy t1_j5od4pw wrote
Any chance you might relax the criteria for volunteering? The form calls for a website and open review account which many prospective volunteers might not have.
SimonJDPrince OP t1_j5ocrdo wrote
Reply to comment by [deleted] in [P] New textbook: Understanding Deep Learning by SimonJDPrince
I'd say that mine is more internally consistent -- all the notation is consistent across all equations and figures. I have made 275 new figures, whereas he has curated existing figures from papers. Mine is more in depth on the topics that it covers (only deep learning), but his has much greater breadth. His is more of a reference work, whereas mine is intended mainly for people learning this for the first time.
Full credit to Kevin Murphy -- writing book is much more work than people think, and so completing that monster is quite an achievement.
Thanks for tip about Hacker News -- that's a good idea.
KrakenInAJar t1_j5pvfuw wrote
Reply to [D] CVPR Reviews are out by banmeyoucoward
3 borderline (3), 1 weak accept (4), better than feared but I guess I will run on adrenaline for the next month...