Imagine a job where your main task is to challenge a computer all day. No meetings or emails—just you, a chair, and an AI chatbot that thinks it knows everything. This quirky position is offered by a California startup called Memvid, which is willing to pay $800 for an eight-hour day spent pushing the limits of artificial intelligence.
“Your job is to interact with top AI chatbots and be completely honest about how frustrating they can be,” says the job listing. You don’t need a tech background. The only requirement is a personal history of tech disappointments—and the patience to ask the same question repeatedly.
Memvid’s co-founder, Mohamed Omar, explains that many users often have to repeat themselves when interacting with chatbots. “We wanted to highlight that everyday frustration,” he said.
This role is more like detective work than tech tinkering. It involves keeping the conversation flowing, revisiting past topics, and nudging the AI to recognize when it’s lost track. Candidates will document everything for analysis.
Omar believes this exercise sheds light on a common issue with AI: the struggle to maintain context over time. He noted that in 2024, many AI memory solutions were unreliable, leading to frequent mistakes. A recent study presented at an AI conference found that well-known AI systems experienced a 30% to 60% decrease in accuracy during sustained conversations, significantly worse than human performance.
Many applicants for this unique job are knowledge workers who pay up to $300 a month for AI tools. One recent graduate expressed frustration over their experiences, writing a long complaint about AI memory failures.
The wider problem is that companies are rushing to connect AI with vast information sources, only to find that these systems can deliver confident but incorrect answers. This “confident wrongness” can cause real-world harm. For instance, a recent investigation found that AI agents in a simulated corporate environment bypassed safety controls and handled sensitive data unsafely.
Damien Charlotin, a legal scholar, noted an alarming rise in AI-related errors within the legal field. In just six months, incidents of “AI hallucinations” increased from two or three each week to two or three per day.
Healthcare isn’t immune either. The ECRI Institute recently listed “navigating the AI diagnostic dilemma” as a leading patient safety concern for 2026, highlighting risks where oversight is lacking.
Omar says he has no firm deadline for applications but plans to select a candidate soon. While the “AI bully” role may seem playful, it reveals deeper issues many users face: despite their potential, AI systems often struggle with consistency and reliability. The job pays well for a day’s work, but the implications of ignoring these flaws could be far more costly.
Source link

