Recent comments in /f/MachineLearning

Downchuck t1_j8500e1 wrote

Perhaps the number of unique queries is overstated: through vector similarity search and result caching, the vast majority of lookups would be duplicate searches already materialized. OpenAI has now introduced a "premium" option suggesting a market for premium search - suggesting room for more cash inflows. This may change their spend strategy, perhaps spending less on marketing and more on hardware.

2

norcalnatv OP t1_j84wfs7 wrote

"Our model is built from the ground up on a per-inference basis, but it lines up with Sam Altman’s tweet and an interview he did recently. We assume that OpenAI used a GPT-3 dense model architecture with a size of175 billion parameters, hidden dimension of 16k, sequence length of 4k,average tokens per response of 2k, 15 responses per user, 13 million daily active users, FLOPS utilization rates 2x higher than FasterTransformer at <2000ms latency, int8 quantization, 50% hardware utilization rates due to purely idle time, and $1 cost per GPU hour. Please challenge our assumptions"

4

bubudumbdumb t1_j84w7r2 wrote

Reply to comment by lmtog in [D] Transformers for poker bot by lmtog

Correct but the goal is not to train but to infer. I am not saying it wouldn't work just that I don't see why the priors of a transformer model would work better than RNNs or LSTMs in modeling the rewards of each play. Maybe there is something that I don't get about pocker that maps the game to graphs that can be learned through self attention.

2

lmtog OP t1_j84vc3x wrote

Reply to comment by thiru_2718 in [D] Transformers for poker bot by lmtog

Thats what I'am not quite sure about. I assume the result would not be close to the nash equilibrium.

But I don't know since I have not worked with transformers before.

I think it comes down to, can we train a transformer with feedback on what hands were good and which ones were not. Looking at other responses it seems like that is not possible.

1

lmtog OP t1_j84uk0n wrote

I think the training part is what I was missing.

I thought you would train a transformer like a normal neural net in the sense that you tell it what output you like and what is wrong.

Looking into it a bit more I assume you could get an output but nothing close to the nash equilibrium.

Thank you for the feedback.

1

IronRabbit69 t1_j84njph wrote

Tabular CFR can be approximated with a neural network, as Noam Brown (1st author of Pluribus) and co-authors show in follow-up work: https://arxiv.org/abs/1811.00164

But you're comparing apples to oranges a bit asking if transformers can replace CFR. Transformers are a neural net architecture. You could of course encode poker stuff in text and feed that to a transformer which predicts the right move to play. But how do you train that network? CFR is a self-play learning algorithm (sort of like Alphago's MCTS) which learns good policies.

2

bubudumbdumb t1_j84mygn wrote

The strength of transformers lies in the transfer of representations learned over large corpuses of text or images. Those are less likely to bring capabilities that generalise to pocker so traditional RL and Monte Carlo approaches are likely to have the upper hand. Pocker's challenges are not linguistic or visual perspective challenges.

1

Ulfgardleo t1_j84fokp wrote

legally the data is not public and the fact that facebook is actively trying to prevent scraping is making it very difficult to argue otherwise.

Legally, the data cnanot be public. The users give facebook a non-exclusive license with limited rights to store and process the data. From this does not follow the right that anyone who sees the shared images (for example) has a right to process them as well. If that wasthe case, the terms (https://www.facebook.com/terms.php 3.1) would have to state under which license the works are redistributed by facebook.

2

Ulfgardleo t1_j84fdfl wrote

if it is illegal now it would be super illegal then, because removing watermarks on its own typically violates the license of the material.

&#x200B;

The question is 100% the same as "can i include GPLv3 code in my commercial closed source repository if i remove the license headers and ensure that the code ris never published?"

0