chatgpt
OpenAI, which released the viral ChatGPT chatbot last year, unveiled a tool that’s intended to help show if text has been authored by an artificial intelligence program and passed off as human.
The tool will flag content written by OpenAI’s products as well as other AI authoring software. However, the company said “it still has a number of limitations — so it should be used as a complement to other methods of determining the source of text instead of being the primary decision-making tool.”
In the Microsoft Corp.-backed company’s evaluations, only 26% of AI-written text was correctly identified. It also flagged 9% of human-written text as being composed by AI.
The tool, called a classifier, will be available as a web app, along with some resources for teachers, the company said in a statement Tuesday. The popularity of ChatGPT has given rise to authorship concerns as students and workers use the bot to create reports and content and pass it off as their own. It’s also spurred worries about the ease of auto-generated misinformation campaigns.
“While it is impossible to reliably detect all AI-written text, we believe good classifiers can inform mitigations for false claims that AI-generated text was written by a human: for example, running automated misinformation campaigns, using AI tools for academic dishonesty, and positioning an AI chatbot as a human,” OpenAI said in a blog post.
Sécurité des mots de passe : bonnes pratiques pour éviter les failles La sécurité des…
Ransomware : comment prévenir et réagir face à une attaque Le ransomware est l’une des…
Cybersécurité et e-commerce : protéger vos clients et vos ventes En 2025, les sites e-commerce…
Les ransomwares : comprendre et se défendre contre cette menace En 2025, les ransomwares représentent…
RGPD et cybersécurité : comment rester conforme en 2025 Depuis sa mise en application en…
VPN : un outil indispensable pour protéger vos données Le VPN, ou « Virtual Private…
This website uses cookies.