Recent comments in /f/MachineLearning
draconicmoniker t1_j5v76ho wrote
Reply to [D] Self-Supervised Contrastive Approaches that don’t use large batch size. by shingekichan1996
Non-contrastive approaches e.g. SWaV can handle lower batch sizes https://arxiv.org/abs/2006.09882
keisukegoda3804 t1_j5v6c7q wrote
Reply to comment by Kacper-Lukawski in [D] Efficient retrieval of research information for graduate research by [deleted]
How does qdrant compare to other offerings with regards to filtered search?
dancingnightly t1_j5v5zwe wrote
The internet isn't accessed live by most of these models, as others have said.
You can finetune language models, but you don't add knowledge as such to them; you bias them to output more words in similar order to your sample data; it won't add facts as such if you do this fine tuning.
One approach you can do though is semantic search through your notes for a given topic/search query. You basically collect the relevant notes with meanings similar to your topic/search query. Then you can populate a prompt with that text. The answer will use that information and any facts, if the model is big enough and RLHF tuned (like ChatGPT/Instruct/text-00x models from OpenAI).
An open source module for this is GPTIndex, I also work on a commercial solution which encompasses videos etc too and has some optimisations. It is possible you can add data/facts from the internet to the prompt(context) at time of generation too; you can use an approach like WebGPT.
olegranmo OP t1_j5v4gh8 wrote
DogeMD t1_j5v41n2 wrote
Reply to comment by olegranmo in [R] Tsetlin Machine in Medical Research - Striking Differences Between Tsetlin Machine Interpretability and Deep Learning Attention by olegranmo
This sounds like a great idea! I will pm you my e-mail to take the conversation further
satireplusplus t1_j5v24u2 wrote
Reply to comment by Paedor in [D] Self-Supervised Contrastive Approaches that don’t use large batch size. by shingekichan1996
If you don't have 8 GPUs you can always run the same computation 8x in series on one GPU. Then you merge the results the same way the parallel implementation would do it. In most cases that's probably gonna end up being a form of gradient accumulation. Think of it this way: you basically compute your distances on a subset of n, but since there are much fewer pairs of distances, the gradient would be noisy. So you just run it a couple of times and average the result to get an approximation of the real thing. Very likely that this is what the parallel implementation does too.
olegranmo OP t1_j5v1xsq wrote
Reply to comment by DogeMD in [R] Tsetlin Machine in Medical Research - Striking Differences Between Tsetlin Machine Interpretability and Deep Learning Attention by olegranmo
Hi DogeMD,
Thanks for the questions! I introduced the Tsetlin machine in 2018 as an interpretable and transparent alternative to deep learning, and it is getting increasingly popular, showing promising results in several domains. The paper reports the first approach to using Tsetlin machines for ECG classification, and it is fantastic that you see potential opportunities in myocardial infarction prediction. If you like, I can do an online tutorial on Tsetlin machines with you and your team to give you a headstart?
avd4292 t1_j5v0hud wrote
Reply to [D] CVPR Reviews are out by banmeyoucoward
4,3,4. What are the chances?
blueSGL t1_j5uzwkq wrote
Reply to comment by CasaDeCastello in [D]Are there any known AI systems today that are significantly more advanced than chatGPT ? by Xeiristotle
why did you put that evil in my mind.
The willpower needed not to use that at every possible juncture might be a bit much for me.
pm_me_your_pay_slips t1_j5uxih0 wrote
Reply to comment by BinodBoppa in [D]Are there any known AI systems today that are significantly more advanced than chatGPT ? by Xeiristotle
He's a very famous AI pioneer from rance.
KrakenInAJar t1_j5uwbp0 wrote
Reply to comment by Expensive-Track in [D] CVPR Reviews are out by banmeyoucoward
I think its one to five, but it is relativly arbitrary, the real confidence score is between the lines of the actual review. If a reviewer is inclined to change his mind it usually shows by the way the review is written.
DogeMD t1_j5uvxiw wrote
Reply to [R] Tsetlin Machine in Medical Research - Striking Differences Between Tsetlin Machine Interpretability and Deep Learning Attention by olegranmo
Ole, I haven’t heard about the Tsetlin machine before. My group is doing some ML research using CNN architectures to predict myocardial infarctions. Would love to explore the use of Tsetlin machines for showing ECG signs of infarction to users (doctors) since EU legislation mandates explainability. Have you tried anything like this before and if so, do you think the Tsetlin machine would be a good candidate? We are based in Lund, southern Sweden
MysteryInc152 t1_j5uvo3i wrote
Reply to comment by Kamimashita in [D]Are there any known AI systems today that are significantly more advanced than chatGPT ? by Xeiristotle
Nothing that would beat Open AI's stuff (Google's stuff) is open for inference or finetuning from the public.
I think the best Open source alternative is this
https://github.com/THUDM/GLM-130B
https://huggingface.co/spaces/THUDM/GLM-130B
But it's not finetuned for instruction so you have to prompt/approach it like a text completer. And also you'll need a 4x3090 to get it running locally.
The best open source instruction finetuned models are the flan t5 models
https://huggingface.co/google/flan-t5-xxl
If you're not necessarily looking for open source but still actual alternatives that aren't just an API wraparound of GPT, you can try cohere
Good thing is that it's completely free for non commercial or non production use
or alephalpha
Not free but the pricing is decent and they have a visual language model as well. Something like flamingo
https://www.deepmind.com/blog/tackling-multiple-tasks-with-a-single-visual-language-model
emad_eldeen t1_j5uvkvw wrote
Reply to comment by FastestLearner in [R] Best service for scientific paper correction by Meddhouib10
Grammerly is great!
currentscurrents t1_j5uvgk8 wrote
Reply to comment by Kamimashita in [D]Are there any known AI systems today that are significantly more advanced than chatGPT ? by Xeiristotle
Not at this time. Google says they're going to release some kind of LLM-based product this year though.
FlavoredQuark t1_j5uv2t9 wrote
Reply to comment by bigbird1996 in [D] CVPR Reviews are out by banmeyoucoward
Patience and belief in yourself is key
farmingvillein t1_j5utusn wrote
Reply to comment by MysteryInc152 in [D]Are there any known AI systems today that are significantly more advanced than chatGPT ? by Xeiristotle
You're probably right, but has anyone built an updated set of benchmarks to compare chatgpt with Google's publicly released numbers? (Maybe yes? Maybe I'm out of the loop?) Chatgpt is sufficiently different than gpt3.5 that I think we'd need to rerun benchmarks to compare.
(And, of course, even if we did, there are open questions of potential data leakage--always a concern, but maybe an extra concern here, since it is unclear whether OpenAI would have prioritized that issue in chatgpt build out. Certainly would have been low on my list, personally.)
Kamimashita t1_j5usxq7 wrote
Reply to comment by MysteryInc152 in [D]Are there any known AI systems today that are significantly more advanced than chatGPT ? by Xeiristotle
Do you know if any of those other models have APIs for queries/inference or finetuning? I've looked into GPT alternatives and most seem closed off.
Paedor t1_j5ur6tx wrote
Reply to comment by altmly in [D] Self-Supervised Contrastive Approaches that don’t use large batch size. by shingekichan1996
The trouble is that contrastive methods often compare elements from the same batch, instead of treating elements as independent like pretty much all other ML (except batchnorm).
As a simple example with a really weird version of contrastive learning: with a batch of 2N, contrastive learning might use the 4N^2 distances between batch elements to calculate a loss, while with two accumulated batches of N, contrastive learning could only use 2N^2 pairs for loss.
[deleted] t1_j5uooap wrote
Reply to [D] Self-Supervised Contrastive Approaches that don’t use large batch size. by shingekichan1996
[deleted]
machine_learning7 t1_j5uoo02 wrote
Reply to [D] CVPR Reviews are out by banmeyoucoward
2 strong R's, 1 weak accept. Probably a very low chance of success.
Can I add more references to the paper between now? I.e. I wish to answer some reviewers by saying I will add the references they want me to add and talk about the differences between my work and the reference.
Nmanga90 t1_j5uolx8 wrote
Reply to [D]Are there any known AI systems today that are significantly more advanced than chatGPT ? by Xeiristotle
Google definitely does. Chat GPT is based off of GPT3-175B, and google has put out several models that outperform it. Like 4 or 5 I think and each of those significantly outperforms it’s predecessor
theunixman t1_j5unatg wrote
Meddhouib10 OP t1_j5umy78 wrote
Reply to comment by zyl1024 in [R] Best service for scientific paper correction by Meddhouib10
None of us is a native english speaker. We tried our best to not make mistake and be clear. But I only want an external opinion only for this paper as it is my first.
Ok-Yogurtcloset-4508 t1_j5v9bbi wrote
Reply to [D] CVPR Reviews are out by banmeyoucoward
Are there any stats available for the review scores?