Recent comments in /f/MachineLearning

niclas_wue t1_j842pyo wrote

Hey, great idea, looks very interesting. Do you use the abstract as an input or do you actually parse the paper? I built something quite similar: http://www.arxiv-summary.com which summarizes trending AI papers as bullet points. However, I think a chrome extension allows for a much more flexible paper choice, which is really great.

37

cajmorgans t1_j8416i1 wrote

Even if it will become illegal, the democracy of Machine Learning depends on it being legal. If Getty wins this, it would mean that a few pretty large companies would be the only ones that can build large models because they “own” most of the data. Facebook for example does a lot of stuff to prevent people scrape public data from their apps.

3

jimmymvp t1_j83v503 wrote

I'm not sure if you have a good overview of ML research if this is your claim. Sounds like you've read too many blog posts on transformers. I'd suggest going through some conference proceedings to get a good overview, there's some pretty rigorous (not just stats) stuff out there. I agree though that there is a substantial subset of research in ML that works towards tweaking and pushing the boundaries of what can be achieved with existing methods, which is for me personally exciting to see! A lot of cool stuff came out of scaling up and tweaking the architectures.

2

_sshin_ OP t1_j83t0er wrote

https://chrome.google.com/webstore/detail/arxivgpt/fbbfpcjhnnklhmncjickdipdlhoddjoh

To use this extension, simply install it and visit a link to an arXived paper. It will generate a summary of the paper, including a one sentence summary, 3-5 questions for the authors, and 3-5 suggestions for related topics. The query prompt can be customized to fit your specific needs and preferences

6

thiru_2718 t1_j83kkbh wrote

Poker depends on looking far enough ahead to be able to play game theory optimal (GTO) moves that maximize the expected value over a long run of hands. You can train a transformer on a ton of data, and get it to predict context-specific plays, but if the number of possible decision-branches is growing exponentially, is this enough?

But honestly, I don't know much about these types of RL-type problems. How is AlphaGo structured?

3

TaxSuspicious8708 t1_j83hqiw wrote

I'm working in game dev, and on side project I'm currently building a little (newbie) ml framework in c# to discover FFNN, CNN and probly RNN. I'm currently struggling on the backpropagation in convolutionnal layer but that's a matter of time before it works (I hope) 😂

I'm very curious to see the possible applications in game AI. I already did some testing projects before, simple ML agents with small fully connected networks.. But I want to go further and probly try to mix the utility based ai pattern with reinforcement learning methods or genetic algorithm.

I also think Convolutionnal network could maybe be used to input some 'spatialized' data to an ai agent and allow him to take decision about movement or so..

3

goj-145 t1_j831dqg wrote

The question is can you use copyrighted info to train a model. The answer is we don't know yet.

The current lawsuit that will define precedent on this is for image generation using copyrighted Getty images in a training model. It's proven that Getty images are used because the watermark shows up in the output of the model many times which is the answer to "how can they prove it".

Once that is defined, then we will know if it is legal or not in those jurisdictions. And then we will get to the "do we do it anyways even though it's illegal?"

3