Recent comments in /f/MachineLearning
a_khalid1999 OP t1_j6efynn wrote
Reply to comment by Red-Portal in [D] AI Theory - Signal Processing? by a_khalid1999
Sounds fun. During my Bachelors I took ML courses along with Signal Processing and other EE stuff wanting to somehow specialize in some intersection, now ending up kinda in this identity crisis of being a CS or a EE. Make EE great again! :D
YoutubeStruggle OP t1_j6efuzb wrote
Reply to comment by mkzoucha in [P] AI Content Detector by YoutubeStruggle
I agree, but the point is AI, and e.g. chatGPT, will always have one way to generate content. Whereas humans may have diverse ways of writing and suppose if we consider an essay or an article, the way of writing by a human would vary with every single sentence but it would remain the same for AI throughout. That's how AI-generated content can be detected. If we do para-wise analysis, we would get better results and a clearer picture but it won't be the same for sentence-wise analysis. And there should not be any possible way that for a particular human, all the generated paragraphs come out to be detected as AI-generated.
Red-Portal t1_j6efht6 wrote
Reply to comment by a_khalid1999 in [D] AI Theory - Signal Processing? by a_khalid1999
That's a more recent trend. Until the late 2000s, computer vision was basically combining machine learning techniques with image processing: Design filters to extract features, and slap them into a classifier. Naturally, lots of Fourier, wavelets, and other weird bases. Very different times.
a_khalid1999 OP t1_j6efhay wrote
Reply to comment by mo6phr in [D] AI Theory - Signal Processing? by a_khalid1999
At this point, I feel dumb for writing this post. If I ak not mistaken, Dr. Lecunn has a ECE background before his PhD, also evident from the invention of CNNs, that have convolution, fundamentally a EE thing
a_khalid1999 OP t1_j6eev77 wrote
Reply to comment by Red-Portal in [D] AI Theory - Signal Processing? by a_khalid1999
Interesting perspective. I did not know Computer Vision was a EE-dominant field at one point, I mean I knew Image Processing is a EE thing, but Vision just gave the ... CS vibe, I mean when I took it in my Bachelor's it was labelled as a CS course.
So basically one way of looking at things could be, and as a EE I'm obviously biased, it's not the Signal Process Engineers moving into ML, it's the CS guys starting to use Signal Processing, cuz where I've been the impression is always that AI is completely a CS thing and the EE's coming in this field are coming due to the lack of job opportunities
[deleted] t1_j6eehil wrote
Reply to comment by YoutubeStruggle in [P] AI Content Detector by YoutubeStruggle
[deleted]
[deleted] t1_j6eebca wrote
Reply to comment by A_HumblePotato in [D] AI Theory - Signal Processing? by a_khalid1999
[deleted]
[deleted] t1_j6edxfy wrote
Reply to [D] Simple Questions Thread by AutoModerator
[deleted]
mkzoucha t1_j6edujv wrote
Reply to comment by YoutubeStruggle in [P] AI Content Detector by YoutubeStruggle
Also, one more thing, at the end of the day there is no way to prove either way without having students record their screens and entire rooms (or only do in person) when writing papers
Red-Portal t1_j6edkjk wrote
Reply to [D] AI Theory - Signal Processing? by a_khalid1999
One of the bull's eye contributions of signal processing to deep learning was this paper. From a signal processing perspective, naive pooling is obviously problematic because you're decimating without limiting the signal bandwidth. That paper showed that in 2019. Shows how much computer vision has changed from an EE-dominant field to a CS field, where signal processing is not common knowledge.
HermanCainsGhost t1_j6edce9 wrote
Reply to comment by IshKebab in [R] InstructPix2Pix: Learning to Follow Image Editing Instructions by Illustrious_Row_9971
It's already in a free app, Draw Things
Note: not mine, just like it a lot
mkzoucha t1_j6ed1z9 wrote
Reply to comment by YoutubeStruggle in [P] AI Content Detector by YoutubeStruggle
I did not have time to try this specific one but I have tried at least 10 others. Sorry, not trying to be negative or anything. They’re are just tons of different models, each of which would need a separate detection model. The model was trained on human writing, so it’s bound to have humanistic sound, and some humans are bound to have a writing voice similar to the output of AI content creators. There is also no real standard ‘human’ way of writing to clearly separate the two. Combine that with the difference in results based on the prompt and it quickly becomes an insurmountable task in my opinion.
At the end of the day, I applaud your efforts, truly but realistically I think your model is significantly overfit to a very small percentage of possible samples, both AI and human generated.
mo6phr t1_j6ecafp wrote
Reply to [D] AI Theory - Signal Processing? by a_khalid1999
Computer vision is built on top of image processing, which is basically just 2d signal processing
YoutubeStruggle OP t1_j6ec8vq wrote
Reply to comment by MrEloi in [P] AI Content Detector by YoutubeStruggle
The use of AI tools should definitely be appreciated. It is saving a lot of time and as a fellow developer, I would highly encourage it. But the classification of human-generated content is necessary as AI-generated content could be misleading, making it important to distinguish it from human-generated content. Also detecting AI-generated content can help ensure the quality of information being shared and consumed, especially in sensitive domains such as news and medicine.
albertzeyer t1_j6ebian wrote
Reply to comment by JustOneAvailableName in [D] Why are there no End2End Speech Recognition models using the same Encoder-Decoder learning process as BART (no CTC) ? by KarmaCut132
It's a bit strange indeed that the GCP or Azure results are not so great. As said, I do actually research on speech recognition, and Google is probably the biggest player in this field, and usually always with the very best results.
My explanation is, they don't really use such good and big models for GCP. Maybe they want to reduce the computational cost as much as possible.
But you also anyway have to be a bit careful in what you compare. Your results might be flawed when your finetuning data is close to your validation set (e.g. similar domain, similar sound conditions). Because in case of GCP, they have very generic models, working for all kinds of domains, all kinds of conditions.
YoutubeStruggle OP t1_j6ebgjv wrote
Reply to comment by mkzoucha in [P] AI Content Detector by YoutubeStruggle
Did you give it a try? It shouldn't be easy to fool this tool. Can you give an example of when it gives a false positive? Your feedback is appreciated.
MrEloi t1_j6ebcks wrote
Reply to [P] AI Content Detector by YoutubeStruggle
Students should declare use of AI tools.
Educators should accept - ideally encourage - AI tool use.
design_ai_bot_human t1_j6eb3wx wrote
Reply to comment by JohnConquest in [R] InstructPix2Pix: Learning to Follow Image Editing Instructions by Illustrious_Row_9971
second your Whisper request.
mkzoucha t1_j6ealr7 wrote
Reply to [P] AI Content Detector by YoutubeStruggle
Too many false positives and ways to trick it, this has been proven over and over again
Kerbal634 t1_j6ead9n wrote
Reply to comment by [deleted] in [R] InstructPix2Pix: Learning to Follow Image Editing Instructions by Illustrious_Row_9971
Definitely don't blame the capitalists going with the cheaper option at the expense of quality for sure
Vegetable-Skill-9700 OP t1_j6e8txc wrote
Vegetable-Skill-9700 OP t1_j6e8o99 wrote
Reply to comment by StoicBatman in [P] Launching my first ever open-source project and it might make your ChatGPT answers better by Vegetable-Skill-9700
Firstly, by measuring data drift and analyzing user behavior, UpTrain identifies which prompts/questions were unseen by the model or the cases where the user was unsatisfied with the model output. It automatically collects those cases for the model to retrain upon.
Secondly, you can use the package to define a custom rule and filter out relevant data sets to retrain ChatGPT for your use case.
Say you want to use LLM to write product descriptions for Nike shoes and have a database of Nike customer chats:
a) Rachel - I don't like these shoes. I want to return them. How do I do that?
b) Ross - These shoes are great! I love them. I wear them every day while practicing unagi.
c) Chandler - Are there any better shoes than Nike? 👟 😍
You probably want to filter out cases with positive sentiments or cases with lots of emojis. With UpTrain, you can easily define such rules as a python function and collect those cases.
I am working on an example highlighting how all the above can be done. It should be done in a week. Stay tuned!
Ch1nada OP t1_j6e5ns3 wrote
Reply to comment by nTro314 in [P] Automating a Youtube Shorts channel with Huggingface Transformers and After Effects by Ch1nada
Thank you! I'm just happy I got to deploy a side-project, the pile of "projects that sound great but lost steam half-way through" was getting too big haha
Yeitgeist t1_j6e5hzy wrote
Cursed woody
ninjawick t1_j6egkao wrote
Reply to [R] InstructPix2Pix: Learning to Follow Image Editing Instructions by Illustrious_Row_9971
The balance between image and text cgf is awkward. Doesn't give consistent results. Creates totally different images but with given prompt. Hope they find something to fix it.