Recent comments in /f/MachineLearning
kandalete OP t1_j8i674e wrote
Reply to comment by Fast-for-a-starfish in [R] [P] LUCAS: LUng CAncer Screening dataset by kandalete
Thank you, I will try it
fullouterjoin t1_j8i5kci wrote
Reply to [N] "I got access to Google LaMDA, the Chatbot that was so realistic that one Google engineer thought it was conscious. First impressions" by That_Violinist_18
I think it is important to hear directly from Blake Lemoine his reasoning by raising the issue of LaMDA, https://www.youtube.com/watch?v=kgCUn4fQTsc
2blazen t1_j8i5fyx wrote
Reply to comment by NoLifeGamer2 in [Discussion] The need for noise in stable diffusion by AdministrationOk2735
That was my understanding as well, noise ensures "randomness"
Thin-Shirt6688 OP t1_j8i53bg wrote
Reply to comment by ImZanga in [R] What are some papers that describe TikTok's algorithm? by Thin-Shirt6688
Thanks! Much appreciated.
AnotsuKagehisa t1_j8i3ttc wrote
Its a lot easier to create a variety of shapes this way, instead of being stuck with a predetermined shape.
Zondartul t1_j8i1bdg wrote
Reply to comment by Disastrous_Elk_6375 in [R] [P] OpenAssistant is a fully open-source chat-based assistant that understands tasks, can interact with third-party systems, and retrieve information dynamically to do so. by radi-cho
Sorry, I just casually watch Yannic Kilcher's YT videos, so I don't know much else.
Tyson1405 t1_j8i09j7 wrote
Reply to [D] Simple Questions Thread by AutoModerator
Hello,
Most of the time I only have my old laptop available without a dGPU and a 5 year old I7 dual core.
Training on the thing takes lots of time. What could you suggest for training models online? My datasets are often in the 2-10gb Range. I don’t have a problem to pay like 30-50 Euros monthly.
I heard colab pro was super good but since they changed to the compute units model it got pretty meh? Or is it still good? Otherwise I heard about paperclip.
What else can you recommend? I only want to train models online and then export them using joblib. I am also a Student just in Case there are some nice discounts.
Appreciate any help!
thedude0425 t1_j8hzozd wrote
Reply to comment by belacscole in [R] [N] Toolformer: Language Models Can Teach Themselves to Use Tools - paper by Meta AI Research by radi-cho
Intelligence and physical traits evolved in humans through random mutation that eventually allowed humans to use tools.
lwl t1_j8hoxpg wrote
Reply to comment by astonzhang in [R] Multimodal Chain-of-Thought Reasoning in Language Models - Amazon Web Services Zhuosheng Zhang et al - Outperforms GPT-3.5 by 16% (75%->91%) and surpasses human performance on ScienceQA while having less than 1B params! by Singularian2501
Super interesting work, thank you for sharing! If you are still active on reddit - we noticed that the pdf is no longer available on arxiv, are you able to say why that is?
tdgros t1_j8hn77o wrote
Reply to comment by gopher9 in [Discussion] The need for noise in stable diffusion by AdministrationOk2735
This one as well: https://openreview.net/pdf?id=QsVditUhXR
NoLifeGamer2 t1_j8hmag2 wrote
To my understanding, if you use noise, then you can generate different images using the same algorithm, just by changing the noise. If you have a blank canvas, there is only 1 initial starting position (blank), so there would be only 1 output image.
gopher9 t1_j8hl55i wrote
There's a paper that does that and also other transformations as well: https://arxiv.org/pdf/2208.09392.pdf
kvg_one t1_j8hke19 wrote
Reply to High-speed cameras and deep learning [Research] by A15L
You may find SPADs and Quanta image sensors interesting.
boadie t1_j8hhtwi wrote
Reply to [D] Have their been any attempts to create a programming language specifically for machine learning? by throwaway957280
In the opposite direction from your question is a very interesting project, TinyNN all implemented as close to the metal as possible and very fast: https://github.com/NVlabs/tiny-cuda-nn
Also in the vague neighbourhood of your question is the Triton compiler, while on the surface being a Python jit compiler is language coverage is much smaller than Python and you can view it as a small dsl, all the interesting bits are way below that level: https://openai.com/blog/triton/
VP4770 t1_j8hh2pn wrote
andreichiffa t1_j8hf2th wrote
Reply to comment by ateqio in [D] Looking for recommendations for an affordable API service to classify AI-generated text by ateqio
10% is what OpenAI considered as "good enough" for theirs, but the problem is with the fact that the detection is not uniform. Most neurodivergent folks will be misclassified as generative models, just as for people with social anxiety who tend to be wordy. Non-native and non-fluent English speakers are the other big false-positive triggers.
Fast-for-a-starfish t1_j8heo8j wrote
I am able to access the files under http://157.253.243.19/LUCAS/ maybe you can try an VPN?
Disastrous_Elk_6375 t1_j8hdb2r wrote
Reply to comment by Zondartul in [R] [P] OpenAssistant is a fully open-source chat-based assistant that understands tasks, can interact with third-party systems, and retrieve information dynamically to do so. by radi-cho
Do you know if distilling will be possible after instruct finetuning and the RLHF steps? I know it works on "vanilla" models, but I haven't searched anything regarding distillation of instruct trained models.
Zondartul t1_j8hd29v wrote
Reply to comment by Disastrous_Elk_6375 in [R] [P] OpenAssistant is a fully open-source chat-based assistant that understands tasks, can interact with third-party systems, and retrieve information dynamically to do so. by radi-cho
The plan is to make it kinda good and train in (on industrial hardware) and then distill it down to a smaller model that ideally can fit in a consumer GPU. It's going to be big at first but they do want to make it small eventually.
ateqio OP t1_j8hcsz6 wrote
Reply to comment by andreichiffa in [D] Looking for recommendations for an affordable API service to classify AI-generated text by ateqio
What's the ratio of false positives? honestly curious
trnka t1_j8hcpwt wrote
Reply to [D] Simple Questions Thread by AutoModerator
I've been learning more about multilingual neural machine translation models lately such as the one in Google's recent paper:
Bapna, A., Caswell, I., Kreutzer, J., Firat, O., van Esch, D., Siddhant, A., Niu, M., Baljekar, P., Garcia, X., Macherey, W., Breiner, T., Axelrod, V., Riesa, J., Cao, Y., Chen, M. X., Macherey, K., Krikun, M., Wang, P., Gutkin, A., … Hughes, M. (2022). BUILDING MACHINE TRANSLATION SYSTEMS FOR THE NEXT THOUSAND LANGUAGES
I'm not sure I understand why it works for languages with no parallel data with any language though.... for instance Latinized Hindi doesn't have any parallel data. Why would the encoder or decoder representations of Latinized Hindi be compatible with any other language?
Is it because byte-pair encoding is done across languages, and that Latinized Hindi will have some word overlap with languages that DO have parallel data? So then it's encouraging the learning algorithm to represent those languages in the same latent space?
andreichiffa t1_j8hawd4 wrote
Reply to comment by ateqio in [D] Looking for recommendations for an affordable API service to classify AI-generated text by ateqio
I have reported to Huggingface what its detector was used for and its failure modes (hint:false positives are worse). In the first days of December. They decided to keep it up. It’s on their consciousness.
Same thing with API providers. Those willing to sell you one are selling you snake oil. It’s on their consciousness.
Same thing for you. You want to build an app that sells snake oil that can be harmful in a lot of scenarios? It’s on your consciousness.
But at that point you even don’t need an API to build it.
[deleted] t1_j8h99q4 wrote
Reply to comment by drcopus in [R] [N] Toolformer: Language Models Can Teach Themselves to Use Tools - paper by Meta AI Research by radi-cho
[removed]
[deleted] t1_j8h8mxq wrote
Reply to comment by [deleted] in [R] [N] Toolformer: Language Models Can Teach Themselves to Use Tools - paper by Meta AI Research by radi-cho
[removed]
kandalete OP t1_j8i7g7o wrote
Reply to comment by Fast-for-a-starfish in [R] [P] LUCAS: LUng CAncer Screening dataset by kandalete
I've used 1.1.1.1 and changed my IP to Phillipines. But I still can't access it. Could you tell me how heavy is the dataset?