|
AI Agents As EmployeesIn October 2025, a research team led by Sandy Carter published "AI Agents As Employees," a 57-page paper examining how organizations are integrating AI agents into their workforce structures. The paper surveys current deployment practices, governance frameworks, and the shift from treating AI as a "tool" to treating it as a "teammate." The document is comprehensive in scope—covering everything from technical architecture to emotional intelligence impacts, from ROI frameworks to regulatory compliance. It includes case studies from companies like JPMorgan, Shopify, PayPal, and Unstoppable Domains, and references major AI platforms from Microsoft, Salesforce, IBM, and others. The paper's central thesis: AI agents represent a fundamental transformation in how work gets done, requiring organizations to rethink not just their technology stack but their organizational charts, governance structures, and even the "social contract" between humans and machines. It's a well-researched survey that will likely influence enterprise AI strategy discussions. Which is precisely why it deserves rigorous scrutiny. The following analysis examines the paper's claims against its own evidence, identifying several critical contradictions that practitioners should understand before adopting its recommendations. The issues aren't minor technical quibbles—they're fundamental gaps between the paper's aspirational framing and the operational realities it acknowledges but never reconciles. This matters because words shape deployment decisions, and deployment decisions have consequences. When the paper calls AI agents "teammates" capable of entering "social contracts," it's not just using colorful metaphors—it's establishing frameworks that will guide how organizations think about accountability, liability, and human oversight. SUMMARY OF KEY CLAIMS: The paper makes several core assertions:
These claims rest on case studies showing measurable success in narrow domains: customer support automation, document processing, compliance monitoring, and marketing campaign optimization. 1. THE EXPLAINABILITY CONTRADICTION • Section 6.1 claims “decision explainability” through tools like LIME, XAI frameworks, and vendor platforms (Microsoft Entra, Credo AI, Salesforce) • Section 6.3.7 admits models are “opaque black boxes” whose “decision logic is hidden” • The gap: These platforms provide workflow traceability (what API was called) not decision explainability (why the LLM produced that output) • LIME was designed for traditional ML models; applying it to frontier LLMs produces feature attribution, not causal reasoning • Most agents use closed-source LLMs (GPT-4, Claude) via API where you have no access to weights, training data, or internal reasoning • Bottom line: You can audit the process but not the reasoning—a critical distinction for accountability 2. THE IDENTITY MANAGEMENT GAP • Paper devotes 57 pages to “AI agents as employees” but only ~100 words to identity management • Brief mentions: Microsoft Entra ID assigns “unique identities” and “role-based access” • What’s missing:
• Example: Synergetics agents are “crypto-wallet-enabled” and can “purchase datasets autonomously”—but no discussion of spending limits, fraud detection, or financial controls • Bottom line: Identity management is foundational to security and accountability, yet remains critically underspecified 3. THE “TEAMMATE” METAPHOR BREAKS DOWN • Paper’s central framing: AI agents should be treated as “teammates” not “tools” (Section 4) • The problem: Teams are defined by shared understanding, mutual adjustment, and collaborative problem-solving • Research cited in the paper (Section 8.1-8.2) admits agents:
• What teams actually require:
• The reality: Humans don’t have “teammates”—they have automated assistants that require constant monitoring • If humans must verify agent outputs (Section 6.3.3’s “robust human review”), the agent isn’t a teammate, it’s a supervised intern who never graduates • Bottom line: ”Teammate” implies collaborative intelligence. What we have is conditional automation with conversational interfaces 4. THE “SOCIAL CONTRACT” FALLACY • Conclusion claims we need “a new social and operational contract between humans and machines” (Section 12.1) • Paper itself admits agents “lack natural emotional reciprocity” (8.2) and have “programmed rather than emotionally grounded” responses (8.2) • The problem: Contracts require reciprocity, agency, and mutual stakes. Agents have none of these. • What the paper actually describes:
• Bottom line: This is a contract about machines among humans, not with machines. The framing obscures accountability. 5. THE BLACK BOX PROBLEM (ACKNOWLEDGED BUT UNRESOLVED) • Section 6.3.7: “Organizations continue deploying opaque models despite these risks” • Paper cites Air Canada liability case—company held responsible for chatbot statements • Courts are signaling: the deploying organization (not the model provider) bears liability • The unresolved tension:
• Proposed solution (human-in-the-loop) negates the efficiency argument for agents • Bottom line: Legal accountability requires explainability we don’t have 6. GOVERNANCE THEATER vs. GENUINE OVERSIGHT • Paper recommends: AI ethics committees, Chief AI Officers, bias audits, transparency frameworks • Reality check: 96% of business leaders expect an AI-related data breach within a year (cited in paper, Section 6.3.6) • These governance structures often exist to insulate C-suite from liability rather than fundamentally reshape deployment decisions • Red teaming and Constitutional AI are valuable but don’t solve the fundamental attribution problem: when complex systems fail, determining “what went wrong” is often intractable • Bottom line: We’re retrofitting governance after deployment rather than requiring explainability as a prerequisite 7. THE DEPLOYMENT-REALITY GAP • Paper cites success stories (Unstoppable Domains handling 32% of tickets, JPMorgan automating contracts) • What these share: Narrow scope, clear success criteria, heavy human oversight at boundaries • Paper also cites failures: IBM laid off 8,000 HR staff for AI, then rehired; McDonald’s abandoned AI drive-through after persistent errors; Klarna rehired staff after AI “shortfalls in empathy” • Pattern: AI agents work for constrained, well-defined tasks. They struggle with context, nuance, and edge cases. • Bottom line: Current agents are sophisticated automation, not “teammates” 8. THE POWER ASYMMETRY (UNADDRESSED) • Paper treats “AI agents as employees” as neutral technological evolution • Never mentions: agents don’t need healthcare, don’t unionize, scale infinitely—making them infinitely preferable from a capital perspective • 23% of workers fear replacement (cited in paper), but labor displacement framed as “organizational challenge” not systemic shift • Bottom line: This is a mechanism for labor displacement dressed in innovation rhetoric RECOMMENDATIONS FOR PRACTITIONERS:
CONCLUSION: This paper is competent surveying of the landscape but makes several conceptual leaps that practitioners should scrutinize. AI agents are powerful tools for specific, well-scoped automation tasks. Treating them as “employees” with whom we have “contracts” or as “teammates” capable of collaborative problem-solving obscures accountability and oversells current capabilities. This analysis doesn’t diminish the paper’s value as a survey of current practices. It questions whether those practices rest on sound conceptual and technical foundations. |