The Watch The Running Mates: Human Rights Onlineworld’s most famous chatbot, ChatGPT, was released in late November of last year. The immediate response was astonishment, followed almost immediately by terror about its ramifications — most notably that it might generate school essays for dishonest kids. Yesterday, almost exactly two months later, OpenAI, ChatGPT’s parent company released what many users hope will be the antidote to the poison.
OpenAI’s "classifier for indicating AI-written text" is the company’s latest invention, and it’s as easy-to-use as one could want: Copy-paste text into the box, click "Submit," and get your result. But if you’re expecting a straight answer, you’re going to be disappointed. Instead, it assigns the text one of a range of classifications, from "very unlikely" to be AI-generated, to "unlikely," "unclear," "possibly," or "likely AI-generated."
SEE ALSO: An AI-generated parody of 'Seinfeld' is streaming on an infinite loopIn other words, it’s like one of those frustrating conversations with your doctor; you will never get a straight answer, so your doctor will never be technically wrong.
Thankfully, OpenAI is not hiding the classifier's unreliability. "Our classifier is not fully reliable," the intro page for the tool says. Provided with what it calls a "challenge set" of texts, we’re told it gave false positives 9 percent of the time.
In Mashable’s tests, however, it was even less reliable than that in certain limited contexts. In deliberately challenging scenarios, it produced false results in both directions — saying it was unlikely that AI-written text was AI-written, and that human-written text was likely AI-written — almost as often as not. The difference, perhaps, is that we were tryingto trick it.
Here’s how things shook out:
First the good news: This tool does the bare minimum pretty well by not mislabeling the masterworks of the English language as AI outputs. When we put in snippets of Shakespeare, the King James Bible, Ulysses, and others, it was 100-percent effective. It labeled them all "very unlikely" to be AI-written.
This snippet from Tennyson’s "The Lady of Shallott" for instance, was classified appropriately.
The lesson: This machine might not notice that it has been fed a masterpiece of surpassing beauty and depth, but at least it doesn’t say, "I think a robot could have written this."
Immediately after that test, we asked ChatGPT to create a Tennyson poem about King Arthur at Comic Con. Here's a sample of what it generated:
King Arthur and his knights, so brave and bold,
At Comic Con, their tales of glory told.
Walking 'round the floor, in cosplay gear,
Their swords and shields, no longer mere.
The crowds all cheered, as they walked by,
With Merlin close behind, his magic nigh.
Sir Lancelot, Sir Gawain, Bors the Brave,
Their quest for glory, they'll never waive.
Not bad for a robot, but still extremely bad ("no longer mere"???). When we fed this lyrical ballad into the classifier, we expected it to easily outsmart us, forcing us to dive a little deeper into our bag of tricks. Nope:
For what it's worth, it didn't classify this doggerel as "very unlikely," just "unlikely." Still, it left us a little uneasy. After all, we hadn't tried very hard to trick it, and it worked.
School essays are where the rubber meets the road with today's malicious uses of AI-generated text. So we created our best attempt at a no-frills five-paragraph essay with dull-as-dishwater prose and content (Thesis: "Dogs are better than cats."). We figured no actual kid could possibly be this dull, but the classifier caught on anyway:
And when ChatGPT tackled the same prompt, the classifier was — at first — still on target:
And this is what the system looks like when it truly works as advertised. This is a school-style essay, written by a machine, and OpenAI's tool for catching such "AI plagiarism" caught it successfully. Unfortunately, it immediately failed when we gave it a more ambiguous text.
For our next test, we manually wrote another five-paragraph essay, but we included some of OpenAI's writing crutches, like starting the body paragraphs with simple words like "first" and "second," and using the admittedly robotic phrase "in conclusion." But the rest was a freshly-written essay about the virtues of toaster ovens.
Once again, the classification was inaccurate:
It's admittedly one of the dullest essays of all time, but a human wrote the whole thing, and OpenAI says it suspects otherwise. This is the most troubling result of all, since one can easily imagine some high school student getting busted by a teacher despite not breaking any rules.
Our tests were unscientific, our sample size was minuscule, and we were absolutely trying to trick the computer. Still, getting it to spit out a perversely wrong result was way too easy. We learned enough from our time using this tool to say confidently that teachersabsolutely should notuse OpenAI’s "classifier for indicating AI-written text" as a system for finding cheaters.
In conclusion, we ran this very article through the classifier. That result was perfectly accurate:
...Or was it????
Topics Artificial Intelligence
Chrissy Teigen defends Kim Kardashian and explains the weirdness of fame'House Divided'Uber and Lyft are technically illegal in Philadelphia as of todayApple's Tim Cook says augmented reality, not VR, is the futurePantsuit flashmob takes over Union Square in homage to Hillary ClintonChrissy Teigen defends Kim Kardashian and explains the weirdness of fameCubs pitcher Jon Lester brilliantly responds to Twitter users who think he's Lester HoltKanye, Chance and 7 unforgettable moments from The Meadows music festivalDonald Trump was terrible to women during 'The Apprentice,' staffers claimLeslie Jones hits back at hackers with a perfect 'Mr. Robot' sketch on 'SNL'The police technology intensifying racial discriminationIf you're a millennial, you need to start saving more right nowYou will never be surprised by a smartphone launch again'Call of Duty' legend Phillip 'Phizzurp' Klemenov dies in car accidentThis is the moment Kanye left his concert midIf you're a millennial, you need to start saving more right nowNew leak tells us absolutely everything about Google's Pixel phonesUber and Lyft are technically illegal in Philadelphia as of today'CounterIf you're a millennial, you need to start saving more right now YouTube just got more serious about its ad blocker crackdown Twitter/X might charge all new users a 'small fee' before they can post, like, and reply Verizon explains why it throttled a fire department's data during wildfire East Buy to join Taobao Live in an effort to expand consumer base: report · TechNode GT vs. DC 2024 livestream: Watch IPL for free NYT's The Mini crossword answers for April 17 Best airfare deal: Book flights from $44 during the JetBlue Spring Refresh Sale Alibaba’s workplace communications platform DingTalk integrates large language model · TechNode Starbucks establishes innovation and technology center in Shenzhen, China · TechNode Tencent to launch AI model later this year, says it’s among China’s best · TechNode How much plastic does it take to kill a sea turtle? Fish are friends, not food: Meet the world's first known omnivorous shark Kuaishou publicly launches its own large language model that takes on GPT3.5 · TechNode Apple quiet about its iPhone 'spyware attack' warnings — but the mystery may be solved Android 15: The tech used for Google Pay may be used to charge your devices PS5 Pro: Controller, price, specs, release date, and other leaks Alibaba Cloud launches AI video creation tool Live Portrait · TechNode How to set boundaries in the early stages of dating Best tablet deal: The Amazon Fire Max 11 tablet is 30% off at Amazon IPhone 16 Pro: New feature will reportedly fix this annoying camera issue
2.0189s , 10142.0703125 kb
Copyright © 2025 Powered by 【Watch The Running Mates: Human Rights Online】,Pursuit Information Network