Recent comments in /f/MachineLearning

IntelArtiGen t1_j5tmia0 wrote

chatGPT is probably a very good model on the task it had to solve (to be a great conversational agent based on openAI data), but there are better models regarding the broad task of language understanding. You could adapt these models to be conversational agents, and they could probably beat chatGPT if they had access to the same dataset. But it would still be this specific task of being a great conversational agent. It's not the task of "thinking by itself like humans".

So it depends on what "more advanced" means. There are probably more "advanced" tasks towards AGI. But towards being a great conversational agent perhaps openAI has the best task-dataset combo today. At least I'm quite sure that there aren't systems which would be "significantly" more advanced than that, because I think the current limit is that it's "just" a very good conversational agent.

18

IntelArtiGen t1_j5tijjx wrote

I managed to use SwAV on 1 GPU (8GB), batch size 240, 224x224 images, FP16, ResNet18.

Of course it works, the problem isn't just the batch size but the accuracy - batchsize trade-off, and the accuracy was quite bad (still usable for my task though). If 50% top5 on imagenet is ok for you, you can do it. But I'm not sure there are many tasks where it makes sense.

Perhaps contrastive learning isn't the best for single GPU. I'm not sure about the current SOTA on this task.

3