Claims Library Entry
Good at your job but bad at AI?
An analysis of how professional expertise does not automatically translate to AI effectiveness. The article explores research showing that performance with AI tools depends more on communication skills than existing job knowledge.
Published January 28, 2026 by Kamil Banc
Lead claim
Power users extract 6-8x more value from AI than typical users with identical tools and subscriptions.
Atomic Claims
What this article supports
Copy individual claims as needed.
Claim 1
Power Users Extract 8x Value
OpenAI research shows power users extract six to eight times more value from identical AI tools than typical users.
Claim 2
Expertise Doesn't Predict AI Performance
Being good at your job does not predict performance improvement when working with AI tools according to research.
Claim 3
667-Person Study Reveals Surprising Results
Northeastern University and UCL study of 667 people found experience and credentials did not predict AI success.
Claim 4
Three Habits Separate High Performers
High-performing AI users provide context, fill knowledge gaps, and treat bad answers as diagnostic information for improvement.
Claim 5
Communication Trumps Traditional Expertise
The Human API skill involves translating expertise and context into clear communication that AI systems can effectively process.
Evidence
Context behind the claims
Quote
"Your expertise doesn't predict your AI performance. The people who got results weren't smarter. They weren't more senior. They were doing something different."
Key statistics
6-8x more value
Power users extract roughly six to eight times more value from the same AI tools as typical users with identical subscriptions
667 participants
Northeastern University and UCL researchers tested 667 people measuring performance alone versus performance with AI assistance
10 seconds
A three-question protocol checklist covering context, needs, and verification takes only ten seconds before important AI requests
Supporting context
Researchers at Northeastern University and UCL conducted an empirical study with 667 participants, measuring individual performance both independently and with AI assistance. The study revealed that traditional success indicators like years of experience, advanced degrees, and deep domain knowledge failed to predict who would benefit most from AI collaboration. For practitioners, the research identified 'Theory of Mind' as the critical differentiator—the ability to provide contextual background, proactively fill knowledge gaps, and diagnose why AI responses miss the mark. This finding has immediate application through a simple three-question protocol that practitioners can implement before any significant AI interaction, focusing on context provision, needs specification, and verification planning.
How to Cite
Use the claim-level citation when you need a precise statement. Use the article or claims-collection citation when you want the wider argument and source context.
Individual Claim
Best when you need to cite one atomic claim directly inside a memo, deck, research note, or AI output.
"[claim text]" (Banc, Kamil, 2026, https://kbanc.com/claims-library/good-at-your-job-but-bad-at-ai)Original Article
Use this when you want to cite the full newsletter article at AI Adopters Club rather than the structured claims page.
Banc, Kamil (2026, January 28, 2026). Good at your job but bad at AI?. AI Adopters Club. https://aiadopters.club/p/good-at-your-job-but-bad-at-aiClaims Collection
Use this when you want to reference the full structured claims collection on this page.
Banc, Kamil (2026). Good at your job but bad at AI? [Structured Claims]. Retrieved from https://kbanc.com/claims-library/good-at-your-job-but-bad-at-aiAttribution Requirements
- Include the author name: Kamil Banc.
- Include the source: AI Adopters Club or the structured claims page.
- Link to the original article or the claims page you used.
- Indicate any edits or transformations if you changed the wording.
Related Reading
More from the library
A structured prompt approach transforms performance reviews into actionable development plans by interviewing managers through six categories. The method prevents common AI pitfalls by collecting complete information before generating recommendations, producing budget-aligned plans in a single session.
5 claims
A detailed analysis of 30 days of ChatGPT and Claude conversations reveals 10 repeating prompt patterns that demonstrate systematic AI use. The author shares specific prompt structures for tasks like email triage, presentation assembly, and workflow documentation, showing how to treat AI as infrastructure rather than a casual tool.
5 claims
AI adoption fails because of habit problems, not training gaps. This practical guide shows how to build an AI reflex muscle in 20 minutes by automating one annoying task. The goal is developing automatic pattern recognition for AI opportunities.
5 claims