Recent comments in /f/MachineLearning
Ye1488 t1_jawh7km wrote
Reply to comment by possibilistic in [P] LazyShell - GPT based autocomplete for zsh by rumovoice
boomer
rumovoice OP t1_jawfmk1 wrote
Reply to comment by BezoomnyBrat in [P] LazyShell - GPT based autocomplete for zsh by rumovoice
I'm not sure if bash has autocomplete capabilities like that (like asking for a query under the current command line)
MaoamWins t1_jawb7b3 wrote
My Uni doesn't use .edu (German Uni), I wonder if that's a factor
Disastrous_Elk_6375 t1_jawatbb wrote
Reply to comment by WittyBananaPeel in Did you get access to Meta AI's LLAMA? [Discussion] by WittyBananaPeel
As a large language model I can't comment, but my friends say it works :D
visarga t1_jaw9d2e wrote
Reply to comment by [deleted] in [P] LazyShell - GPT based autocomplete for zsh by rumovoice
ALL the data
MuonManLaserJab t1_jaw7lmj wrote
Reply to comment by possibilistic in [P] LazyShell - GPT based autocomplete for zsh by rumovoice
You're thinking of that other library, Keraskeras Cola.
BezoomnyBrat t1_jaw7k23 wrote
Looks great, definitely going to try it out. Pity it works only with zsh though and not with bash
New_Computer3619 OP t1_jaw68xf wrote
Reply to comment by currentscurrents in [D] The Sentences Computers Can't Understand, But Humans Can by New_Computer3619
Nice. Now I have a paper to read. :) Thank you.
DAlmighty t1_jaw5m9v wrote
Reply to comment by rumovoice in [P] LazyShell - GPT based autocomplete for zsh by rumovoice
Regexes are definitely the bane of my existence.
[deleted] t1_jaw4rqp wrote
Reply to comment by DAlmighty in [P] LazyShell - GPT based autocomplete for zsh by rumovoice
[deleted]
maxToTheJ t1_jaw2nk0 wrote
Reply to comment by DAlmighty in [P] LazyShell - GPT based autocomplete for zsh by rumovoice
Especially given the compute used on OpenAI end. This isn’t sustainable
hiptobecubic t1_jaw1yhv wrote
Reply to comment by DAlmighty in [P] LazyShell - GPT based autocomplete for zsh by rumovoice
The ffmpeg example is worth it alone
WittyBananaPeel OP t1_jaw0e38 wrote
Reply to comment by Disastrous_Elk_6375 in Did you get access to Meta AI's LLAMA? [Discussion] by WittyBananaPeel
Does that link "actually" works? 😂
rumovoice OP t1_javxdca wrote
Reply to comment by DAlmighty in [P] LazyShell - GPT based autocomplete for zsh by rumovoice
If you know and remember the command - yes, if you need to google or read the man first this could be faster. Or sometimes faster for complex commands with subshells and regexes.
currentscurrents t1_javx4pw wrote
The Winograd Schema is a test of commonsense reasoning. It's hard because it requires not just knowledge of english, but also knowledge of the real world.
But as you found, it's pretty much solved now. As of 2019 LLMs could complete it with better than 90% accuracy, which means it was actually already solved when Tom Scott made his video.
Disastrous_Elk_6375 t1_javwm8d wrote
laughs in magnet link
RoninUTA t1_javs4zg wrote
Impressive!
DAlmighty t1_javrbwq wrote
This is definitely impressive but it also feels like more work than to just bang out the commands.
CellWithoutCulture t1_javqw9s wrote
Reply to comment by LetterRip in [D] OpenAI introduces ChatGPT and Whisper APIs (ChatGPT API is 1/10th the cost of GPT-3 API) by minimaxir
Fantastic reply, it's great to see all those concrete advances thst made it intro prod. Thanks for sharing.
New_Computer3619 OP t1_javqiuu wrote
Reply to comment by DSM-6 in [D] The Sentences Computers Can't Understand, But Humans Can by New_Computer3619
I tried the same questions in separate chats as in the edited post. ChatGPT gave incorrect/unsatisfying answers this time. May be without context from previous Q&A, it can only infer using grammar rule? What do you think?
LetterRip t1_javpxbv wrote
Reply to comment by CellWithoutCulture in [D] OpenAI introduces ChatGPT and Whisper APIs (ChatGPT API is 1/10th the cost of GPT-3 API) by minimaxir
> I mean... why were they not doing this already? They would have to code it but it seems like low hanging fruit
GPT-3 came out in 2020 (they had their initial price then a modest price drop early on).
Flash attention is June of 2022.
Quantization we've only figured out how to do it fairly lossless recently (especially int4). Tim Dettmers LLM int8 is from August 2022.
https://arxiv.org/abs/2208.07339
> That seems large, which paper has that?
See
https://github.com/HazyResearch/flash-attention/raw/main/assets/flashattn_memory.jpg
>We show memory savings in this graph (note that memory footprint is the same no matter if you use dropout or masking). Memory savings are proportional to sequence length -- since standard attention has memory quadratic in sequence length, whereas FlashAttention has memory linear in sequence length. We see 10X memory savings at sequence length 2K, and 20X at 4K. As a result, FlashAttention can scale to much longer sequence lengths.
DSM-6 t1_javnmz2 wrote
Personally, I think the answer is existing bias in the training data.
I don’t know enough about chatgpt to state this as fact, but I think it’s safe to assume that chatgpt understands or adheres to grammar rules. I.e. nowhere in the code does it state “antecedent pronouns should refer to the subject of a sentence”
Instead I assume chatgpt grammar comes repeated convention in the training data. Enough data in which the antecedent refers to something other then the sentence object means that the “they” can refer to any of the preceding nouns. In that case “councilmen fear voilence” is a far more common sentence in the training than “protesters fear violence”
Then again your example was passive tense, so I dunno 🤷♀️.
notaninvestor633 t1_javmt3c wrote
Wow this is awesome. Can’t wait to tinker around with it
harharveryfunny t1_javmsab wrote
Reply to comment by Thin_Sky in [D] OpenAI introduces ChatGPT and Whisper APIs (ChatGPT API is 1/10th the cost of GPT-3 API) by minimaxir
It's a leak, but seems to be legitimate.
https://twitter.com/transitive_bs/status/1628118163874516992
svantevid t1_jawk06h wrote
Reply to comment by MaoamWins in Did you get access to Meta AI's LLAMA? [Discussion] by WittyBananaPeel
Worked with .ac.uk e-mail for me.