orange and yellow neon lights

The possible proliferation of misinformation because of AI

As much as I love all these AI tools popping into the scene left and right, one thing has always left me somewhat dreading of what it could mean for the future.

I’ve been tinkering around with AI tools like ChatGPT and Midjourney the past few months, and admittedly found myself amazed at how freaking good these are even when they’re just relatively new to public access. I have even put ChatGPT to use as part of my workflow and day-to-day – coding assistance, copywriting ideas, etc. I’m also witnessing what feels like FOMO with companies scrambling to integrate AI into their products and services.

But, something eventually dawned on me: Am I crazy to think that AI – if left unregulated and kept being used irresponsibly – could eventually distort most of facts and truths that we get online?

I mean, if you think about it and as far as I know, currently these tools take the information it uses from user inputs and interactions, and what’s currently publicly available on the entire web. If we keep on creating content using AI without even bothering to review and correct supposed facts and accepted truths, then use AI to further “rehash” those AI-generated content for more AI-generated content, doesn’t that lead to inaccuracy eventually?

Even the creators of these tools admit and put it out there that info generated by their AI “may produce inaccurate information about people, places, or facts.”

What do you think? Are we moving into the future of AI too fast? Does it need to be regulated as soon as possible? Should there be immediate laws against its misuse?

Published by

Bibiano Wenceslao

SaaS customer success. Workflow automation. QA testing. Work-from-home dad.

Leave a Reply

Your email address will not be published. Required fields are marked *

This site uses Akismet to reduce spam. Learn how your comment data is processed.