OpenAI has an AI text detector but doesn't want to publish it

OpenAI has developed some new tools to detect content generated by ChatGPT and its AI models, but it is not going to roll them out yet. The company has come up with a way to overlay a type of watermark onto AI-produced text. This built-in indicator could accomplish the goal of guessing when AI has written some content. However, OpenAI is hesitant to offer this feature when it could harm those who use its models for benign purposes.

OpenAI’s new method would employ algorithms capable of embedding subtle markers in the text generated by ChatGPT. Though invisible to the naked eye, the tool would use a specific format of words and phrases that signal the origin of ChatGPT’s text. There are obvious reasons why this could be a boon for generative AI as an industry, as OpenAI notes. Watermarking could play a critical role in combating misinformation, ensuring transparency in content creation and preserving the integrity of digital communications. It’s also similar to a tactic OpenAI already employs for its AI-generated images. The DALL-E 3 text-to-image model produces visuals with metadata explaining their AI origin, including invisible digital watermarks that can even survive any attempts to remove them through editing.

scroll to top