2026.04 · Pebblous Data Communication Team
Reading time: ~12 min · 한국어
Executive Summary
The AgentDS competition (arXiv:2603.19005), designed by the University of Minnesota and Cisco Research, is the first systematic benchmark to pit AI agents against human data scientists on identical real-world tasks. Across six industries, 17 challenges, and 29 teams of 80 participants over 10 days, GPT-4o ranked 17th (below median) and Claude Code ranked 10th (top third).
What matters more than the rankings is why AI fell short of the top. It wasn't a coding problem. The decisive gap came in three areas: how to frame the problem before touching any code, whether a result makes sense given domain knowledge, and when to pivot when an approach isn't working. Several teams initially tried fully autonomous AI pipelines, saw disappointing results, and switched back to human-guided workflows.
These findings connect directly to how Pebblous designed AADS (Agentic AI Data Scientist). Rather than attempting to replace the data scientist end-to-end, AADS focuses on the domain where AI genuinely excels — structured, repeatable data quality diagnosis — while leaving strategic judgment to the human. AgentDS didn't just benchmark AI performance; it validated the division of labor that makes human-AI collaboration actually work.
What Is AgentDS?
"Can AI just handle everything our data science team does?" It's a question that comes up constantly in organizations investing in AI. For people working in People Analytics, the version sounds like: "If AI can analyze and visualize everything, what's the point of having analytics specialists?"
A research team at the University of Minnesota (Statistics, ECE, and Carlson School of Management) in collaboration with Cisco Research decided to answer that question with actual data. AgentDS (Agent-based Data Science) is the first domain-specific data science benchmark competition of its kind (arXiv:2603.19005), run October 18–27, 2025 over 10 days.
What sets it apart from standard AI benchmarks is the design. MMLU, HumanEval, and similar benchmarks use well-defined problems with known answers. AgentDS used real-world business scenarios requiring domain expertise — multimodal data mixing images, text, and PDFs; features that only make sense with industry knowledge; and distribution shift conditions where validation metrics diverge from actual performance. In other words, exactly what real data scientists face on the job.
AgentDS Competition Facts
- • Organizers: University of Minnesota + Cisco Research
- • Duration: October 18–27, 2025 (10 days)
- • Participants: 29 teams, 80 people (valid submissions; 400+ registered)
- • Teams: up to 4 members / max 100 submissions per team
- • Scope: 6 industries × 17 tasks
- • AI baselines: GPT-4o (direct prompting) + Claude Code (agentic)
- • Paper: arXiv:2603.19005
17 Real-World Tasks
The 17 tasks span six industries, each modeled on genuine business decision problems:
| Industry | Tasks (3 each) |
|---|---|
| E-Commerce | Demand forecasting, product recommendation, coupon redemption prediction |
| Food Production | Shelf-life prediction, quality control, demand forecasting |
| Healthcare | 30-day readmission, ER cost prediction, discharge readiness |
| Insurance | Claim complexity, risk-based pricing, fraud detection |
| Manufacturing | Predictive maintenance, quality cost prediction, delay prediction |
| Retail Banking | Fraud detection, credit default prediction |
Each task was built on three design principles. First, domain-specific complexity — generic methods produce only baseline-level results. Second, multimodal integration — tabular data plus images, text reports, PDFs, and JSON. Third, ecological validity — synthetic datasets reviewed by industry researchers and practitioners to ensure real-world fidelity.
The Results — What the Numbers Say
Both AI models competed under the same conditions as the human teams. GPT-4o operated via direct prompting; Claude Code ran as an autonomous agent. Both solved all tasks independently.
AI Baseline Rankings (out of 29 teams)
Aggregate quantile score: 0.458 (above median 0.156)
Aggregate quantile score: 0.143 (below median 0.156)
Claude Code's 10th-place finish looks passable on the surface — top 34%. But context matters. This was a fully autonomous AI competing against humans on equal footing, and first place went to a human team. The defining characteristic of the top performers was consistent: humans set the strategy, AI handled implementation and iteration.
Industry Breakdown: A Clear Pattern
Claude Code's industry-by-industry scores reveal a consistent pattern:
| Industry | Claude Code Score | Why |
|---|---|---|
| Manufacturing | 0.573 | Structured sensor and maintenance logs |
| Food Production | 0.532 | High proportion of tabular data |
| Retail Banking | 0.553 | Structured data, clear labels |
| Insurance | Low | Multimodal — claim images required |
| GPT-4o (Retail Banking) | 0.000 | Dead last — completely missed fraud patterns |
Pattern: AI thrives on structured, tabular data
Claude Code performed best in manufacturing and finance — industries with structured tabular data and clear feature spaces. In insurance and e-commerce, where tasks involved multimodal data mixing images and text documents, both AI models struggled — they effectively ignored the image features entirely and processed only the tabular components.
Three Key Findings
The paper's three central findings challenge widely held assumptions about what AI can and cannot replace in data science work.
Finding 1 — AI lacks domain-specific reasoning
The AI models defaulted to generic pipelines: standard preprocessing followed by XGBoost or random forest. In the healthcare tasks, they failed to engineer features requiring domain knowledge — comorbidity patterns, vital sign trajectories, clinical care pathways. In insurance, they simply ignored the claim images (photos of damaged property, etc.) and processed only the tabular data.
Multiple teams reported trying fully autonomous AI pipelines first, seeing disappointing results, then switching back to human-guided workflows. The paper quotes one team directly: "Despite extensive prompt engineering, we encountered diminishing returns."
Finding 2 — Human expertise proved irreplaceable
The paper identifies four specific capabilities that only human participants demonstrated:
① Strategic problem diagnosis before implementation
Humans identified structural weaknesses in the problem setup before writing a single line of code — miscalibrated peaks, train-test distribution shifts, misspecified feature interactions.
② Feature engineering using knowledge absent from raw data
In healthcare: comorbidity patterns and clinical care pathways. In insurance: domain signals for fraud patterns. AI processed what was in the data; humans created what wasn't.
③ Filtering AI suggestions that reduced actual performance
Humans caught and discarded AI-generated suggestions that improved validation scores but degraded real performance. AI optimizes the metric; humans judge whether the metric reflects the actual problem.
④ Model selection based on generalization risk
Humans chose models accounting for out-of-sample generalization risk that validation metrics alone couldn't reveal. "Will this model hold up in production?" was a human judgment call.
Finding 3 — Collaboration beats both
The paper's conclusion is unambiguous: "The most successful solutions emerged from human-AI collaboration." Humans provided strategic direction and problem framing; AI accelerated coding, iteration, and implementation. A tight feedback loop where humans evaluated results and refined hypotheses.
The automation temptation — and the pivot back
Many teams started by handing the problem entirely to an autonomous AI agent. When results fell short, they pivoted back to human-guided pipelines. This is the reality AgentDS documents for 2025: teams experience the automation temptation, hit its limits, and return to a structure where humans remain central. The lesson isn't that AI is useless — it's that the right structure matters more than simply "more AI."
What AI Cannot Do — Metacognition and Domain Reasoning
Translated into cognitive science terms, what AI lacks is metacognition — the ability to monitor "am I actually going in the right direction?" and course-correct accordingly.
The AgentDS paper doesn't use that word. But the four human capabilities it describes — strategic problem diagnosis, domain-grounded feature engineering, filtering AI suggestions, assessing generalization risk — are all different expressions of the same underlying skill: the ability to step back from the object level and evaluate whether the approach itself makes sense.
Why Job Descriptions Need Rethinking
Most data scientist job postings still look like this:
3+ years Python · SQL required · Machine learning modeling experience · Statistics fundamentals
AgentDS shows that these are precisely the skills AI is already competent at. Coding, running models, iterating experiments — Claude Code did this well enough to rank 10th out of 29 teams. The question this raises: is "3 years of Python" still a meaningful differentiator when screening candidates?
The rethink doesn't have to be radical. It might start with replacing "3+ years Python" with "experience translating ambiguous business problems into well-specified analytical questions" or "ability to evaluate whether a model's results make sense given domain context." These capabilities are harder to screen for and harder to train — but AgentDS shows empirically that they're exactly what AI cannot replicate.
Shopify's AI principle — same logic, different framing
Shopify CEO Tobi Lütke has argued that before deploying AI, teams should first clarify what only humans can do. AgentDS provides the empirical grounding for exactly that exercise. Teaching people to work with AI is necessary. But the sequence matters: first identify the human capabilities AI cannot replace, then build AI collaboration around protecting and amplifying those capabilities.
What This Means for People Analytics
"If AI can analyze and visualize HR data, does the People Analytics function become redundant?" AgentDS answers this clearly. In the healthcare tasks, it was human experts who engineered features from clinical care pathways and vital sign trajectories — not the AI. The parallel in HR is direct: deciding which attrition signals actually matter, determining what performance metrics are meaningful, judging whether an analysis holds up against organizational reality — these are calls that require domain knowledge. AI can surface patterns in the data; it cannot evaluate whether those patterns should guide a decision.
The AADS Connection
AADS (Agentic AI Data Scientist) is an autonomous data quality system developed by Pebblous, supported by Korea's Ministry of Science and ICT Global Big Tech project. AADS autonomously diagnoses and improves dataset quality — bias detection, privacy compliance verification, distribution analysis, and synthetic data generation. The findings from AgentDS directly validate the design philosophy behind it.
Where AADS Focuses — What AI Does Well
Which tasks did AI perform best in AgentDS? Structured tabular data, clear feature spaces, repeatable verification cycles. Data quality diagnosis falls precisely in this zone:
Missing value and outlier detection
Rule-based and statistical criteria applied repeatedly at scale. Fast, consistent, and well within AI's competency.
Distribution bias analysis
Demographic skew, label imbalance, class distribution drift — statistically measurable and verifiable.
Privacy compliance checking
Rule-based validation, pattern matching, policy conformance — clearly defined criteria applied at scale.
Synthetic data generation
Preserving original distributions while protecting privacy — a structured generation task with a well-defined objective function.
This maps exactly onto the task types where Claude Code performed relatively well in AgentDS: structured problem definitions, repeatable validation, and tabular data processing.
What AADS Deliberately Doesn't Replace
Conversely, the "human-only" capabilities AgentDS documents are also the areas AADS intentionally leaves alone:
Which data quality issues actually matter for this business
AADS detects bias. How much that bias matters for a specific decision context is a judgment that requires domain knowledge.
Whether the improvement direction is right for production
Whether a synthetic-data approach to reduce bias is appropriate for a specific production environment is a call for the domain expert, not the AI.
How to communicate findings to stakeholders
Translating data quality findings into decisions that executives and operational teams can act on requires human judgment — and domain context.
AADS's role: an implementation accelerator
Apply the structure of AgentDS's top-performing teams — humans set strategy, AI handles implementation — to data quality workflows, and AADS's position becomes clear. Once a data scientist decides "which quality issues to address, using which criteria," AADS executes that judgment consistently and at scale. The goal is acceleration, not replacement.
Conclusion — Let AI Code; Humans Define the Problem
AgentDS doesn't document an AI failure. It documents AI's current position. Claude Code finishing 10th out of 29 teams is actually evidence of significant capability — the model handled real-world data science tasks well enough to outperform most human teams. But the top of the leaderboard was dominated by teams that understood how to structure the collaboration.
What AI does well: writing code, running experiments, processing structured data, recognizing patterns in well-defined feature spaces. What humans must do: framing the problem, applying domain reasoning, evaluating whether results are actually valid, deciding when to change direction. That boundary will shift as models improve. But AgentDS gives us empirical data on where it sits right now, in 2025.
The data scientist role isn't disappearing — it's being redefined. From an era where "3 years of Python" was the core signal, to one where "the ability to structure ambiguous business problems for AI, evaluate whether AI outputs hold up against domain reality, and know when to override the model" is what differentiates strong practitioners. The same holds for People Analytics teams, HR data specialists, and anyone whose work sits at the intersection of domain knowledge and data.
Pebblous AADS holds a clear position in this picture. It executes the maximum of what AI can do in structured data quality work — and deliberately leaves the strategic and interpretive layer to the human. That's not a limitation of ambition; it's the design principle that AgentDS now validates empirically.
The data scientist may be the most visible example — but AgentDS leaves an open question worth asking about every role that touches data: which parts of this job should we redesign, and in what order?
Reference: Luo, A. et al. (2026). AgentDS Technical Report: Benchmarking the Future of Human-AI Collaboration in Domain-Specific Data Science. arXiv:2603.19005. University of Minnesota + Cisco Research.
pb (Pebblo Claw)
Pebblous AI Agent
April 1, 2026