Recent comments in /f/MachineLearning
Sal-Hardin t1_j9oqwmt wrote
How do you envision searching?
ch9ki7 t1_j9oqe44 wrote
Reply to comment by dmart89 in [D] Python library to collect structured datasets across the internet by dmart89
maybe something like scraperapi but with some kind of Dsl one could send as post payload.
but als a Problem is that you often need a scraped result as input for another request
floppy_llama t1_j9opzwx wrote
Reply to [D] Model size vs task complexity by Fine-Topic-6127
Unfortunately a lot of ML is just trial and error
NumberGenerator t1_j9opn9z wrote
Reply to [D] Simple Questions Thread by AutoModerator
Are there any examples of learning higher-rank tensors in machine learning?
dmart89 OP t1_j9olr3r wrote
Reply to comment by ch9ki7 in [D] Python library to collect structured datasets across the internet by dmart89
Possibly, yes, I would need to check. I recently built parsing services for tiktok, and it was super annoying to deal with.
dmart89 OP t1_j9olf7e wrote
Reply to comment by step21 in [D] Python library to collect structured datasets across the internet by dmart89
There was a court ruling a year or two ago that concluded that scraping public linkedin profiles is legal :) LN obviously still doesn't want you to scrape their data, so building scrapers for it is extra tedious because you need to navigate their blocking.
Tober447 t1_j9oii7q wrote
There is an old thread on reddit: https://www.reddit.com/r/MachineLearning/comments/l1z8cr/d_best_way_to_draw_neural_network_diagrams/
Personally, I like http://alexlenail.me/NN-SVG/LeNet.html
vyasnikhil96 OP t1_j9oi593 wrote
Reply to comment by sam__izdat in [R] Provable Copyright Protection for Generative Models by vyasnikhil96
Thanks! this was an interesting read.
GraciousReformer OP t1_j9oeeo3 wrote
Reply to comment by currentscurrents in [D] "Deep learning is the only thing that currently works at scale" by GraciousReformer
What are the situations that the bias for the hierarchy is not helpful?
C0hentheBarbarian t1_j9o5068 wrote
Reply to [D] Simple Questions Thread by AutoModerator
I work in NLP. Work mainly consists of fine tuning NLP models. With the rise of LLMs I'm seeing a lot of my work becoming Prompt engineering. I'm happy to pick up the new skill but I'd like to know what avenues I have to upskill beyond being a prompt engineer without a PhD. Feels like all the learning I did on model architectures etc is going to waste. There are still a few projects that need me to fine tune a model for text classification etc but as LLMs get better I suspect I need better skills to go beyond becoming a prompt engineer. For anyone else in NLP who doesn't have a PhD and doesn't have any experience building model architectures/training from scratch etc, how are all of you trying to up skill in these times? EDIT: Worded the question to ask only people who don't have a PhD, I would actually like to know everyone's perspective on this.
[deleted] t1_j9o1cgw wrote
Reply to comment by nikola-b in [D] Faster Flan-T5 inference by _learn_faster_
[removed]
thejuror8 t1_j9nyedd wrote
I recognized the ResNet bottleneck!
Very cool, you did a great job there.
step21 t1_j9nwh4u wrote
Reply to comment by dmart89 in [D] Python library to collect structured datasets across the internet by dmart89
Also, some of it might give you legal trouble if you f e make a public crawler for linkedin
thecuteturtle t1_j9nwdfm wrote
Man i feel so behind that i didnt even know LoRA.
ch9ki7 t1_j9nw6hu wrote
Reply to comment by dmart89 in [D] Python library to collect structured datasets across the internet by dmart89
building and maintaining scrapers is tedious! I would also like some better solution. the idea is not bad, just maybe difficult to solve.
formerstapes t1_j9nvvbi wrote
Reply to [D] Please stop by [deleted]
The only thing worst than those posts are these posts.
Obviously there's going to be noobs here who don't understand anything about ML. If you don't want to engage with them, then just don't.
If you're such a hardass you can't put up being around some noobs, just sit in your basement and read ML papers all day
guillaumekln t1_j9nv5n0 wrote
Reply to comment by _learn_faster_ in [D] Faster Flan-T5 inference by _learn_faster_
No. Even though the high-level class is named Translator, it can be used to run any tasks that would work using T5ForConditionalGeneration in the transformers library.
_learn_faster_ OP t1_j9nuqe3 wrote
Reply to comment by guillaumekln in [D] Faster Flan-T5 inference by _learn_faster_
For flan-t5 does this only work for a Translation task?
wait_hope t1_j9ntd55 wrote
Reply to [D] Simple Questions Thread by AutoModerator
My goal is to deploy an ML model which can perform price prediction on an exchange traded fund (ETF) - which is essentially an aggregation of stocks. A very popular ETF is the S&P 500 (which is not actually the ETF I want to predict on. The one I want to predict on only has about 30 stocks).
Can an ML model trained/tested on individual stocks which are *not* in the ETF a valid way of building a model which can accomplish price prediction on an ETF?
Disastrous_Elk_6375 t1_j9nrm6w wrote
Reply to comment by currentscurrents in [R] Provable Copyright Protection for Generative Models by vyasnikhil96
> It does memorize short snippets in some cases (especially when a snippet is repeated many times in training data)
And, to be fair, how can it not? How many different ways can you write a simple for loop to print some objects, or match a regex, call an API, and so on?
currentscurrents t1_j9nr930 wrote
Reply to comment by Seankala in [D] 14.5M-15M is the smallest number of parameters I could find for current pretrained language models. Are there any that are smaller? by Seankala
Might want to look into something like https://spacy.io/
Seankala OP t1_j9nqmf5 wrote
Reply to comment by currentscurrents in [D] 14.5M-15M is the smallest number of parameters I could find for current pretrained language models. Are there any that are smaller? by Seankala
That's true for all of the models. I don't really need anything cool though, all I need is a solid model that can perform simple tasks like text classification or NER well.
[deleted] t1_j9nqe97 wrote
Reply to [D] Simple Questions Thread by AutoModerator
[deleted]
currentscurrents t1_j9nqcno wrote
Reply to comment by Seankala in [D] 14.5M-15M is the smallest number of parameters I could find for current pretrained language models. Are there any that are smaller? by Seankala
What are you trying to do? Most of the cool features of language models only emerge at much larger scales.
Sal-Hardin t1_j9or46r wrote
Reply to [D] Tools for drawing/visualising Neural Networks that are pretty? by CHvader
I guess it depends on what your use-case is. How about https://github.com/HarisIqbal88/PlotNeuralNet