Workers are wasting half a day each week fixing AI ‘workslop’
Better staff training and understanding of the technology is needed to cut down on AI workslop
Sign up today and you will receive a free copy of our Future Focus 2025 report - the leading guidance on AI, cybersecurity and other IT challenges as per 700+ senior executives
You are now subscribed
Your newsletter sign-up was successful
AI 'workslop' is forcing employees to work an extra four and a half hours each week to clean up mistakes, according to new research.
A survey of more than 1,100 US enterprise AI users from Zapier found that while 92% of workers say AI boosts their productivity, the average employee spends more than half a workday revising, correcting - and sometimes completely redoing - AI-generated outputs.
Three-quarters of respondents reported at least one negative consequence from low-quality AI outputs, including work rejected by stakeholders (28%), security incidents (27%), and customer complaints (25%).
Zapier noted that just 2% of respondents don’t need to revise what AI produces.
A key factor here lies in poor training, researchers noted. Employees without AI training are six times more likely to say AI makes them less productive.
While untrained workers spend less time on AI cleanup, they also report fewer productivity gains: just 69% say AI helps, compared with 94% of trained workers.
“The productivity gains from AI are real. 92% of workers feel them. But so is the cleanup work,” said Emily Mabie, senior AI automation engineer at Zapier.
Sign up today and you will receive a free copy of our Future Focus 2025 report - the leading guidance on AI, cybersecurity and other IT challenges as per 700+ senior executives
“The companies seeing the best results aren't the ones avoiding AI. They're the ones who have invested in training, context, and orchestration tools that turn AI from a sloppy experiment into a managed process.”
The worst AI workslop offenders
Data analysis tops the workslop list, according to Zapier, with 55% saying data analysis and visualization projects require the most cleanup, followed by writing tasks at 46%.
Meanwhile, engineering, IT, and data roles average five hours per week fixing AI outputs, with 78% reporting negative consequences. Finance and accounting teams face the highest rate of negative consequences at 85%, averaging 4.6 hours of cleanup per week.
The time lost in fixing AI-generated outputs has a significant impact on bottom lines, according to Zapier. Workers spending more than five hours a week on AI cleanup tasks are more than twice as likely to report lost revenue, clients, or deals.
Zapier said better data quality and more robust infrastructure could go a long way in helping to improve the situation.
The study found that those with access to AI orchestration tools and comprehensive company context – so internal documentation, brand guides, project templates, or prompt libraries – said the technology does have a big positive impact on productivity.
“The solution isn’t fewer tools, it’s better infrastructure,” said Mabie. “Orchestration, training, and proper context convert AI from a vague experiment into a managed process where the extra cleanup is the cost of doing more meaningful work faster, rather than the cost of pretending you are.”
AI workslop is here to stay
The rise of AI workslop has become a recurring pain point for enterprises ramping up adoption of the technology.
A report from MIT researchers last year found that more than 40% of US-based workers had been given AI-generated content that “masquerades as good work but lacks the substance to meaningfully advance a given task”.
This, the study noted, was destroying productivity and harming perception of the technology in the workplace.
Certain professions are experiencing acute issues with this trend, such as those working in software development. A recent study from CodeRabbit, for example, shows that AI makes 1.7 times as many mistakes as human programmers.
The use of AI in software development has been one of the leading use-cases for the technology over the last three years, with developers reporting significant productivity boosts from AI code generation.
Research from Harness in early 2025, however, found that these productivity gains are offset by the fact developers are having to drop tools and manually remediate faulty code, slowing down processes.
FOLLOW US ON SOCIAL MEDIA
Make sure to follow ITPro on Google News to keep tabs on all our latest news, analysis, and reviews.
You can also follow ITPro on LinkedIn, X, Facebook, and BlueSky.
Emma Woollacott is a freelance journalist writing for publications including the BBC, Private Eye, Forbes, Raconteur and specialist technology titles.
-
Salesforce targets telco gains with new agentic AI toolsNews Telecoms operators can draw on an array of pre-built agents to automate and streamline tasks
-
Four national compute resources launched for cutting-edge science and researchNews The new national compute centers will receive a total of £76 million in funding
-
If you thought RTO battles were bad, wait until AI mandates start taking hold across the industryOpinion Forcing workers to adopt AI under the threat of poor performance reviews and losing out on promotions will only create friction
-
Sam Altman just said what everyone is thinking about AI layoffsNews AI layoff claims are overblown and increasingly used as an excuse for “traditional drivers” when implementing job cuts
-
Google says hacker groups are using Gemini to augment attacks – and companies are even ‘stealing’ its modelsNews Google Threat Intelligence Group has shut down repeated attempts to misuse the Gemini model family
-
Why Anthropic sent software stocks into freefallNews Anthropic's sector-specific plugins for Claude Cowork have investors worried about disruption to software and services companies
-
B2B Tech Future Focus - 2026Whitepaper Advice, insight, and trends for modern B2B IT leaders
-
What the UK's new Centre for AI Measurement means for the future of the industryNews The project, led by the National Physical Laboratory, aims to accelerate the development of secure, transparent, and trustworthy AI technologies
-
Half of agentic AI projects are still stuck at the pilot stage – but that’s not stopping enterprises from ramping up investmentNews Organizations are stymied by issues with security, privacy, and compliance, as well as the technical challenges of managing agents at scale
-
What Anthropic's constitution changes mean for the future of ClaudeNews The developer debates AI consciousness while trying to make Claude chatbot behave better
