Recent comments in /f/MachineLearning

jess-plays-games t1_j5fy01j wrote

U could sli a 1080 with a previous card easily enough but they don't share ram they just use one cards vram.

The 2000 series and later don't support sli anymore and instead use nvlink witch does share vram between cards.

There a handy lil program called any sli I used to use

3

hey_look_its_shiny t1_j5fu53q wrote

You don't need to implement a full-scale LLM in order to degrade watermarks at scale or even mix-and-match watermarked inputs. People who aren't even trying get halfway there now with crappy synonym engines.

And before you ask, no, I'm not going to technically spec it for you. Instead I suggest using the upvote pattern from this expert community to run backprop on your beliefs. ;)

5

ardula99 t1_j5fsdsw wrote

That is what adversarial data points are -- people have discovered that it is possible to "confuse" image models using attacks such as these. Take a normal picture of cat, feed it into an image model and it'll label it correctly and say - hey, I'm 97% sure there's a cat in this. Change a small number of pixels using some algorithm (say <1% of the entire image) - to a human, it will still look like a cat, but an image model now thinks it's a stop sign (or something equally unlikely) with 90%+ probability.

2

doIneedtohaveone1 t1_j5fqzkf wrote

Does any one know how to solve the PDE for it in python? Any kind of reference material would be appreciated!

It's been long since I came across any PDEs and have forgotten everything related to it.

1

hey_look_its_shiny t1_j5fnyt6 wrote

I'm not OP, but the words "won't work in the long term" from their original statement are not synonymous with "useless".

Your original comment was disrespectful, and while you have raised some valid points along the way, they're collectively misaligned with the original statement you were responding to. You've been fighting a strawman, and it shows in how the community received your comments.

16

GalaxyGoldMiner t1_j5flc0v wrote

Thinking out loud, instead of watermarking you could just look at each tokens conditional probability of being sampled based on the prior tokens; if the probabilities are high in aggregate it is likely to hang come from low temperature GPT. This assumes that transformer models trained by different companies (on presumably overlapping data) will have different enough predictions in long sequences.

1