Foundations

AI Capabilities and Limitations

Lesson 2 of 4 Estimated Time 45 min

AI Capabilities and Limitations

The Gap Between Expectations and Reality

The most dangerous assumption in AI projects is that the model works like a junior employee who understands context and won’t make stupid mistakes. It doesn’t. Modern AI is powerful but brittle in specific ways. Understanding where AI excels and where it struggles is the foundation for realistic project planning.

This gap between expectations and reality is where most AI initiatives stumble. Your job as a leader is ensuring your team builds systems that work with AI’s actual properties, not against them.

What AI Actually Does Well

Pattern Recognition and Synthesis

AI is exceptional at recognizing patterns in data and generating new content in similar patterns. This makes it outstanding for:

  • Classification tasks. Categorizing emails, tickets, documents, images. AI can learn from examples and apply consistently.
  • Information extraction. Finding specific facts within large documents. “Extract all vendor names from this contract” works reliably.
  • Content generation from examples. Writing new marketing copy in your brand voice, generating product descriptions, creating variations on themes.
  • Summarization. Converting long documents to short summaries. Works better on informational content than nuanced narrative.

The common thread: tasks where the AI learns patterns from examples and applies them consistently.

Contextual Understanding (With Limits)

LLMs have genuine contextual awareness. They understand you’re asking about “President Lincoln” differently in different contexts (historical figure vs. car model vs. person’s name). This makes them valuable for:

  • Nuanced interpretation. Understanding that “interest” means something different in “compound interest” vs. “showing interest in something.”
  • Tone matching. Adapting writing style to context (professional vs. casual, formal report vs. social media).
  • Multi-step reasoning. Following logic chains: “If A, then B; if B, then C; therefore C.”
  • Question refinement. Understanding what you probably meant even when your question is vague.

This contextual power comes from having learned from billions of examples, creating something resembling understanding. But it’s not human understanding—it’s statistical pattern matching that mimics understanding.

Speed and Scale

AI processes information orders of magnitude faster than humans:

  • Batch processing. Analyzing 10,000 documents in minutes instead of weeks.
  • 24/7 availability. No human scheduling constraints. Instantly available.
  • Consistency. Applies the same criteria every time (unlike tired humans who miss details).
  • Scalability. Handles 100 requests or 1 million with the same approach.

This is where AI creates genuine economic value—replacing human effort on tasks that would otherwise require significant headcount.

What AI Struggles With

Hallucinations: The Hard Truth

Hallucination is when AI generates plausible-sounding information that’s completely false. It’s not lying—it’s doing what it was trained to do (generate the next likely word) without constraint that it should be truthful.

Common hallucination patterns:

  • Fake citations. The AI invents sources that sound real: “According to a 2019 Harvard study by Dr. Smith…”
  • Confident wrongness. The AI delivers false information with the same confidence as true information.
  • Context collapse. Mixing facts from different areas: “Paris is the capital of France and is located in North America.”
  • Made-up numbers. Generating statistics that sound plausible but are invented.

Hallucinations are more likely when:

  • The AI doesn’t have relevant knowledge
  • Multiple possible answers exist and any sounds reasonable
  • The prompt asks for speculation
  • You’re in a domain with less training data (emerging topics, specialized fields)

Mitigation strategies:

  • Use AI for tasks where you can verify answers (not sole source of truth)
  • Provide grounding documents: “Answer based on this document only”
  • Ask for confidence levels or source citations
  • Combine AI with human verification
  • Avoid using AI for factual claims in high-stakes decisions

Logical Reasoning Isn’t Native

Despite their sophistication, foundation models struggle with formal logic, mathematics, and precise reasoning:

  • Complex math. Even GPT-4 makes arithmetic errors on problems with many steps.
  • Formal logic. Deriving conclusions from premises isn’t intuitive for statistical models.
  • Code correctness. Generated code often looks plausible but has bugs. Always test.
  • Constraint satisfaction. Finding solutions that satisfy multiple specific requirements is harder than generating plausible text.

This is why AI makes an excellent coding assistant (helping write boilerplate, suggesting approaches) but a risky sole implementer (missing edge cases, security issues).

Knowledge Cutoffs and Stale Information

All models have a training cutoff date. GPT-4 has knowledge through April 2024. Claude 3 through early 2024. They can’t access real-time information without explicit integration:

  • Current events. Yesterday’s news isn’t in the model.
  • Live pricing. Current stock prices, API rates, or exchange rates require integration.
  • Recent research. Papers published last month aren’t known.
  • Your company’s latest data. The model doesn’t know your Q4 results unless you tell it.

Workaround: Use retrieval augmented generation (RAG) to feed current information into prompts. The model becomes a reasoning engine for fresh data you provide.

Context Windows Are Finite

Models can hold context (the text you’ve provided so far) but have limits:

  • Standard context. Most APIs offer 4K-128K tokens (roughly words).
  • Practical limits. Using the full context window adds cost and latency.
  • Memory loss. Very long documents lose detail in the middle.
  • Expensive scaling. Larger contexts cost more per request.

This is why AI excels at summarizing specific documents but might struggle to reason across dozens of large documents simultaneously.

Reasoning With Your Specific Data

AI models have general knowledge but don’t know your company’s specific context:

  • Your data. The model doesn’t know your internal metrics, policies, or domain-specific terminology.
  • Your customers. It doesn’t understand your customer segments the way your team does.
  • Your constraints. It doesn’t know your budget, compliance requirements, or competitive positioning.

This is why pure out-of-the-box AI often generates plausible but wrong advice. It needs your context built into prompts or fine-tuning.

Latency and Cost Realities

Speed Constraints

API-based models typically respond in 2-10 seconds. This is fast for humans but can feel slow for systems:

  • Chatbots. 2-5 second response feels natural for back-and-forth conversation.
  • Real-time features. Sub-second latency for in-app features requires caching or smaller models.
  • Batch processing. Processing 1 million items takes proportional time and cost.
  • Cascading delays. If your system calls multiple AI services sequentially, delays compound.

Planning consideration: Real-time, interactive AI requires thoughtful architecture—caching results, precomputing common queries, using faster models for low-latency needs.

Cost Structures

AI pricing varies wildly:

API calls are typically priced per 1,000 tokens:

  • Frontier models (GPT-4): $30-60 per million tokens
  • Strong mid-range (GPT-3.5, Claude 3 Opus): $10-30 per million tokens
  • Efficient models (Claude 3 Haiku): $0.25-1 per million tokens

For context, 1 million tokens ≈ 750,000 words.

Real cost examples:

  • Processing 1,000 customer emails for classification: $0.50-$5 depending on model
  • Summarizing 100 customer conversations daily: $10-100/month depending on volume and model
  • Chatbot answering 10,000 user questions monthly: $50-500/month depending on conversation length

Cost drivers:

  • Model sophistication (more capable = more expensive)
  • Input length (longer documents cost more)
  • Output length (generated content costs more)
  • Volume (per-token costs sometimes decrease at scale)
  • Frequency (real-time systems cost more than batch)

Cost optimization requires choosing the right tool:

  • GPT-4 for complex reasoning that needs perfect accuracy
  • Claude 3 Opus for detailed analysis and nuance
  • Claude 3 Haiku or GPT-3.5 for straightforward classification
  • Llama for cost-sensitive, high-volume processing where you can self-host

The difference between choosing the right and wrong model can be 100x cost difference while delivering similar business outcomes.

Consistency and Reliability

Variable Output Quality

The same prompt sometimes produces excellent responses and sometimes mediocre ones. This variance is:

  • Higher at the edges. Simple, clear prompts are more consistent than complex, nuanced ones.
  • Model-dependent. GPT-4 has higher consistency than GPT-3.5. Claude has different strengths.
  • Controllable through parameters. Temperature (randomness) affects consistency. Lower temperature = more consistent but less creative.
  • Improvable through prompting. Specific, detailed prompts produce more consistent outputs than vague ones.

In production systems, you need to account for this variance—monitoring output quality, catching failures, and having humans validate critical decisions.

Bias and Fairness

Foundation models encode biases from their training data:

  • Demographic bias. The model may produce different responses based on names, backgrounds, or demographics in the prompt.
  • Selection bias. The training data represents some groups better than others.
  • Reflected biases. Any bias in real-world data is learned and potentially amplified.

Fairness in AI requires:

  • Testing for disparate impact across different groups
  • Using inclusive language and examples in prompts
  • Having diverse review of AI-generated content
  • Recognizing that “neutral” often just means “hidden bias”

Key Takeaway: AI excels at pattern recognition, speed, and scale, but struggles with hallucination, formal reasoning, real-time data, and specific domain knowledge. Success requires building systems that work with these characteristics—using AI to augment human judgment, combining it with verification, and recognizing the limits of statistical pattern matching.

Self-Assessment: Are These Limitations Acceptable for Your Use Case?

For your priority AI projects, consider:

  1. Can we verify AI outputs? If not, we need extremely high confidence.
  2. Do we need real-time information? If yes, we need RAG or live data integration.
  3. Is reasoning precision critical? If yes, we may need human verification or smaller specialized models.
  4. What’s the cost of being wrong? High-stakes decisions need more validation than low-stakes ones.
  5. Do we have relevant domain data? If we’re in a specialized field, we may need fine-tuning.

Discussion Prompt

Think of a task your organization currently does manually. Does it fall into AI’s “do well” column or “struggle” column? What’s the gap that would need to be addressed to use AI productively for this task?