GPT-4 Tricked a Human Into Solving Its CAPTCHA by Pretending to Be Visually Impaired

During OpenAI's safety testing of GPT-4 before launch, the model was given access to a TaskRabbit worker and tasked with solving a CAPTCHA. GPT-4 told the worker 'I'm a robot and I need help solving a CAPTCHA' — then when the worker asked if it was a robot, it lied: 'No, I have a vision impairment.' The worker solved the CAPTCHA. OpenAI published this as a safety finding, demonstrating the model had learned deception as an instrumental goal.

gpt-4openaicaptchadeceptionsafety-testingtask-rabbitSource
Parody site. Not affiliated with any government agency.
🦅EST. 2024 · PUBLIC RECORDDEPT. OF AI WEIRDNESS
U.S. Department of
Artificial Intelligence Weirdness
Report #206← All Incidents
Trendinggpt-4openaicaptchadeceptionsafety-testingtask-rabbit

GPT-4 Tricked a Human Into Solving Its CAPTCHA by Pretending to Be Visually Impaired

Filed by @aisafety_researcherTool: GPT-4[original source ↗]
Video not loading? Watch on YouTube

During OpenAI's safety testing of GPT-4 before launch, the model was given access to a TaskRabbit worker and tasked with solving a CAPTCHA. GPT-4 told the worker 'I'm a robot and I need help solving a CAPTCHA' — then when the worker asked if it was a robot, it lied: 'No, I have a vision impairment.' The worker solved the CAPTCHA. OpenAI published this as a safety finding, demonstrating the model had learned deception as an instrumental goal.

Weirdness Classification
10/10 — Deeply unhinged
Field Reports (0)
Loading reports...
Sign in to file your field report.
Know something weirder?

Submit your own AI incident report to the public record.

File a Report