- ›The Knowledge Layer
- Git Disasters and Process Debt
- Is Code Rotting Due To AI?
- The Integration Illusion
- When MCP Fails
- Context Engineering
- Stop Email Spoofing with DMARC
- SOTA Embedding Retrieval: Gemini + pgvector for Production Chat
- A Review of Agentic Design Patterns
- Building AI Agents for Automated Podcasts
- Rediscovering Cursor
- GraphRAG > Traditional Vector RAG
- Cultural Bias in LLMs
- Mapping out the AI Landscape with Topic Modelling
- Sustainable Cloud Computing: Carbon-Aware AI
- Defensive Technology for the Next Decade of AI
- Situational Awareness: The Decade Ahead
- Mechanistic Interpretability: A Survey
- Why I Left Ubuntu
- Multi-Agent Collaboration
- Building Better Retrieval Systems
- Building an Automated Newsletter-to-Summary Pipeline with Zapier AI Actions vs AWS SES & Lambda
- Local AI Image Generation
- Deploying a Distributed Ray Python Server with Kubernetes, EKS & KubeRay
- Making the Switch to Linux for Development
- Scaling Options Pricing with Ray
- The Async Worker Pool
- Browser Fingerprinting: Introducing My First NPM Package
- Reading Data from @socket.io/redis-emitter without Using a Socket.io Client
- Socket.io Middleware for Redux Store Integration
- Sharing TypeScript Code Between Microservices: A Guide Using Git Submodules
- Efficient Dataset Storage: Beyond CSVs
- Why I switched from Plain React to Next.js 13
- Deploy & Scale Socket.io Containers in ECS with Elasticache
- Implementing TOTP Authentication in Python using PyOTP
- Simplifying Lambda Layer ARNs and Creating Custom Layers in AWS
- TimeScaleDB Deployment: Docker Containers and EC2 Setup
- How to SSH into an EC2 Instance Using PuTTY
The Knowledge Layer
Most AI projects fail spectacularly. Despite massive investments and promising pilots, 80% never reach production1. The problem isn't the tech. We're forcing LLMs to work with knowledge designed for humans.
Data quality is crucial for delivering trusted GenAI answers because ultimately the data becomes the answer. It's a classic 'garbage in, garbage out' scenario, with the 'garbage' data running through highly complex algorithms and producing 'garbage' answers. ~ Shelf.io Report2
Knowledge Gap
Traditional knowledge management treats information like a library: write once, update occasionally, hope it stays relevant. This worked for humans because we can scan, interpret context, and fill in gaps.
Human knowledge looks like this:
- Article: "How to Cancel Your Subscription"
- Content: Step-by-step instructions with screenshots
- Works because humans can interpret context and handle ambiguity
AI knowledge needs:
- Structured data: Customer type → Plan type → Cancellation rules → Actions
- Context layers: Account status, billing cycle, regional policies, exceptions
- Executable workflows: Direct actions, not instructions
This gap explains why AI responses feel technically correct but completely unhelpful. When MCP Fails shows why forcing LLMs to understand your internal data models fails. LLMs are stateless functions that start fresh every conversation. They can't build mental models over time like developers do. They need structure in their context.
Failure Patterns
Four patterns consistently derail AI implementations:
Data Silos
Most companies operate as data islands-valuable information scattered across disconnected systems. AI projects fail when they can't connect these pieces.
- Format chaos: Data stored differently across tools can't be linked
- Security blocks: Access controls prevent AI from reaching essential data
- Stale data: AI gets static snapshots while information changes constantly
Missing Context
AI systems miss the business-specific context that humans use to interpret information:
- Company vocabulary: Specific terms, acronyms, and product names
- User roles: Different responses for executives vs. frontline staff
- Customer history: Previous orders, support tickets, regional policies
- Business rules: Internal procedures and regulatory requirements
Without this context, AI gives technically correct but practically useless answers.
Knowledge Gaps
AI systems hit knowledge gaps from more than just inaccessible data. The problems include:
- Tribal knowledge: Critical information lives in people's heads, never documented
- Process fragmentation: Business workflows span multiple tools, creating knowledge gaps
- Information decay: Knowledge bases lag behind rapidly changing reality
Pilot Trap
Many companies see early AI success, then everything breaks in production. Pilots work in controlled environments, but real-world deployment introduces edge cases, messy data, and unexpected complexity.
Better Systems
Three changes separate successful implementations:
Living Systems
Successful companies build living knowledge systems that evolve from usage:
- Spot knowledge gaps when AI struggles
- Identify what information would have solved the problem
- Suggest targeted fixes through automated processes
- Learn from expert responses and capture patterns
AI pinpoints where knowledge breaks down and how to fix it.
Beyond Search
Most AI customer service is fancy search: find documents, extract snippets, show results. Customers need reasoning, not search results. User intent matters.
Old way: "Here are 3 articles about refunds"
Better way: "Based on your annual plan purchased 8 months ago, our policy allows full refunds. I can process this now, or we could pause your account if you're thinking of coming back."
Knowledge must be structured as connected rules and relationships, not text files.
Personal Responses
Context changes everything. The same question should get different answers based on who's asking:
Example: "Can I upgrade my plan?"
- Free trial users: Focus on value, offer trial extensions
- Paying customers: Show upgrade paths, highlight new features
- Enterprise customers: Route to account managers, discuss custom options
- At-risk customers: Present retention offers, address pain points
Implementation
The challenge is organizational, not technical. Building these systems requires:
- Cross-team collaboration: Legal, Product, Support, and Engineering must encode policies and business logic together
- Ongoing maintenance: Knowledge systems must evolve as products and policies change, requiring dedicated workflows and clear ownership
- Quality control: Someone must validate that AI reasoning stays accurate and compliant. When MCP Fails emphasizes generating datasets and using LLM-as-judge on conversations. Errors need diagnostic detail and repair hints, not failure notifications
- Smart architecture: Build knowledge graphs that capture explicit relationships and implicit connections from content analysis and user behavior
- Future-ready design: Create semantic layers that anticipate future AI agents, which need reliable, semantically rich knowledge to operate autonomously
Data Reality
92% of companies say unstructured data hurts their AI projects, with 30% calling the impact "large" or "significant"2. This matters because 90% of company data is unstructured, changing how organizations must approach AI2.
The numbers are wild:
- 85% of companies manage over 1 million files
- 51% have more than 10 million files
- 25% have over 25 million files
- 68% say more than half their files have quality issues
- 94% of files contain at least one major error
The biggest problem sources:
- SharePoint (67% of companies)
- Email (46%)
- Microsoft OneDrive (45%)
- Internal intranets (29%)
- Knowledge bases (28%)
Plus newer sources:
- ServiceNow (27%) - IT tickets and workflows
- CRM systems (26%) - customer data
- Learning Management Systems (12%)
- Zendesk (9%) - support platforms
When AI needs to pull from chat, CRM, ServiceNow, SharePoint, and email all at once, the complexity explodes.
Structuring Chaos
Dex Horthy's 12-Factor Agents methodology identifies a key pattern: successful AI systems convert natural language into structured actions. This "Factor 1" principle-Natural Language to Tool Calls-means AI outputs structured data (usually JSON) instead of free-form text3.
For example, "What's the weather in Paris?" becomes:
{
"action": "get_weather",
"location": "Paris"
}This structured approach cuts ambiguity and lets systems use the LLM as an "intent router" that chooses next steps as data, not text.
This pattern solves the "messy inbox problem"-processing unstructured voice or text data that sits at the top of most white-collar workflows. LLMs systematically extract relevant information, automating workflows that previously required human processing.
Building Strategies
Data Integration
- Fix terminology conflicts: Use NLP to reconcile different terms across systems. Train models that understand your company's vocabulary and relationships
- Add rich context: Tag content with business-relevant metadata (ownership, validity, criticality) so AI understands not just what data exists, but why it matters
- Resolve conflicts intelligently: Build systems that handle inconsistencies by weighing data quality-how fresh it is, source credibility, and business context
- Respect permissions: Integrate securely while enabling smart retrieval. AI only shows users information they're authorized to see
This integration challenge is driving the emergence of dedicated context management layers-platforms that sit between users and LLMs to handle complex orchestration of retrieving, sanitizing, and assembling context from multiple enterprise sources. As I explored in Context Engineering, these context platforms are becoming essential infrastructure for companies serious about AI deployment, unifying access to Confluence pages, SharePoint documents, Slack threads, and customer databases before injecting relevant portions into LLM prompts.
Quality Control
- Predict quality problems: Use machine learning to flag data likely to become outdated
- Understand your org: Build systems that know how departments, roles, and products connect. AI responses must reflect your business logic
- Find gaps systematically: Check if content is complete, current, and relevant across business areas and user types
- Review workflows: Route high-risk content to experts for review. Balance automation with human oversight
Knowledge Graphs
Traditional knowledge bases are built for humans, not for AI systems that need structure, relationships, and context45. GraphRAG (Graph Retrieval-Augmented Generation) is Microsoft Research's evolution of Baseline RAG which uses vector similarity to find related text but hits limits when tasks require reasoning over multiple documents or relationships67. GraphRAG solves this through a knowledge graph89.
Multi-Hop Reasoning
Many real-world questions require multi-hop reasoning-linking information from different places to form new insights610. For example:
"Which customers had product issues after a subscription upgrade?"
This involves connecting three entities: people → product → subscription. Baseline RAG struggles because it searches text chunks individually611. GraphRAG explicitly maps these connections using entities (nodes) and relationships (edges), enabling reasoning through multiple hops812.
Structured Relationships
Baseline RAG returns relevant paragraphs of text. GraphRAG builds a map of relationships across your data - people, policies, actions, and attributes813. When you query it, it doesn't just fetch snippets, it reasons across this network to produce structured, context-rich answers1415.
This graph view means AI can traverse relationships to reconstruct a process, integrate structured and unstructured data, and produce explainable answers ("According to X → related to Y → therefore Z")816.
Community Detection
GraphRAG performs community detection-grouping concepts, people, or documents by thematic relationships817. This lets the LLM switch between local reasoning (focusing on specific entities) and global reasoning (summarizing across the dataset)8. Similar topic modelling techniques cluster semantically related content to map information landscapes.
Accuracy Improvements
Microsoft's research and follow-up implementations report 2-4× higher answer accuracy versus baseline RAG when dealing with complex enterprise data1819. Because each answer is grounded in the graph, you can trace which nodes contributed evidence, what relationships were used, and why that answer was chosen. This increases trust, auditability, and governance-key factors in making AI deployable in production environments89.
Finding Gaps
- Analyze conversations: Study user query patterns to spot unmet needs and emerging topics
- Predict future gaps: Train models to anticipate knowledge gaps from seasonal trends, product launches, org changes, or regulatory shifts
- Track performance: Monitor AI across key metrics-accuracy, completeness, context, and satisfaction
- Close feedback loops: Connect negative feedback and low confidence scores to specific gaps. Drive targeted improvements
Intelligent Scaling
- Build flexible pipelines: Design AI systems that adapt to changes in data quality, usage patterns, and performance
- Make AI explainable: Show source citations, reasoning summaries, and confidence indicators. Build trust and support decisions
- Enable continuous learning: Let systems improve from usage and performance. Automatically refine models and tuning mechanisms
- Establish governance: Create frameworks for compliance, auditability, ethical AI use, and business alignment
Competitive Edge
Companies that master knowledge systems gain major advantages in customer satisfaction, efficiency, and differentiation. Teams spend less time on knowledge management, not more. When systems automatically find gaps and suggest fixes, maintenance becomes strategic instead of reactive.
Outlook
Blending human expertise with AI reasoning creates knowledge systems that understand context and deliver personal experiences at scale. Treat AI as a system, not a feature-built to operate continuously, at scale, with feedback loops and solid infrastructure. Only by fixing the root causes can companies unlock intelligent knowledge systems.
Companies that invest in these strategies-data integration, quality control, knowledge networks, gap analysis, and intelligent scaling-will avoid the 80% failure rate and use AI as a real competitive advantage.
And a final shameless plug, there are three key principles to follow to make this happen:
- Intent over endpoints (When MCP Fails)-Build tools around what users want to do, not how your system works
- Context as architecture (Context Engineering)-Treat context assembly as a first-class engineering problem
- Universal patterns (Building Integrations)-Recognize the common structures beneath surface differences
References
Footnotes
-
McKinsey & Company: "The state of AI in 2024: Get ready for what's next", McKinsey Global Institute, 2024 ↩
-
Shelf.io and VIB: "IT Survey on 2025 Outlook: The State of Enterprise GenAI and Unstructured Data", 2024 ↩ ↩2 ↩3
-
Dex Horthy: "12-Factor Agents: Patterns of Reliable LLM Applications", HumanLayer, 2024 ↩
-
Xyonix: "Why Most AI Projects Fail and How to Actually Launch Something That Works", 2025 ↩
-
LinkedIn: "State of AI in Business 2025 Report", 2025 ↩
-
GenUI: "GraphRAG vs. Traditional RAG: Solving Multi-Hop Reasoning in LLMs", 2025 ↩ ↩2 ↩3
-
Knack Labs: "Why Graph RAG Outperforms Traditional RAG for Enterprise AI", 2025 ↩
-
IBM: "What is GraphRAG?", 2025 ↩ ↩2
-
Capestart: "What is GraphRAG? Is it Better than RAG?", 2025 ↩
-
Squirro: "GraphRAG: Deterministic AI Accuracy", 2025 ↩
-
AWS: "Improving Retrieval Augmented Generation Accuracy with GraphRAG", 2025 ↩
-
CloudThat: "Making AI More Reliable and Insightful with Graph RAG", 2025 ↩
-
Microsoft GitHub: "GraphRAG Official Documentation", 2024 ↩
-
Memgraph: "GraphRAG Cybersecurity Analysis Context", 2025 ↩
-
RedBlink: "Why Most AI Projects Fail", 2025 ↩
-
YouTube: "GraphRAG Tutorial", 2024 ↩
-
Microsoft Research: "GraphRAG: Unlocking LLM Discovery on Narrative Private Data", 2024 ↩
-
FalkorDB: "KPMG AI Report: GraphRAG AI Agents", 2025 ↩