Skip to main content

Enterprise AI Analysis: Navigating Knowledge with LLMs vs. Search

This analysis, from the enterprise AI solutions experts at OwnYourAI.com, deconstructs the academic paper "To Google or To ChatGPT? A Comparison of CS2 Students' Information Gathering Approaches and Outcomes" by Aayush Kumar, Daniel Prol, Amin Alipour, and Sruti Srinivasa Ragavan. We translate its critical findings into actionable strategies for corporate learning, knowledge management, and AI system design, revealing the hidden costs of "quick answers" and charting a course for building AI that fosters genuine expertise.

Executive Summary for Business Leaders

The study reveals a critical paradox for the enterprise: while Large Language Models (LLMs) like ChatGPT provide rapid answers, they can hinder deep, conceptual understanding, especially for complex topics. Employees may complete tasks faster but fail to grasp the underlying principles, leading to a "brittle" knowledge base within your organization. This phenomenon, which we term the "Productivity-Proficiency Gap," has significant long-term implications for innovation, problem-solving, and quality.

  • Finding 1: Different Tools, Different Thinking. Employees using search engines (like Google) tend to use broad, keyword-based queries, exposing them to diverse, context-rich resources. In contrast, LLM users ask specific, narrow questions, which can lead them down "rabbit holes" and prevent them from seeing the bigger picture.
  • Finding 2: The Illusion of Understanding. For difficult subjects, learners using LLMs performed significantly worse on conceptual tests than those using traditional web search. They could often find a solution but couldn't explain *why* it worked, a major risk for enterprise codebases and complex systems.
  • The Enterprise Imperative: Relying on off-the-shelf LLMs for internal training and knowledge transfer is a high-risk strategy. The optimal approach is a custom-built, hybrid AI knowledge system that combines the speed of conversational AI with the depth of curated, multimodal content (documentation, videos, best-practice guides). This ensures employees not only get answers but truly build sustainable skills.

The following analysis breaks down the research data and provides a roadmap for implementing such a system to close the Productivity-Proficiency Gap in your organization.

Book a Meeting to Future-Proof Your Team's Expertise

Section 1: The Core Research - A Tale of Two Learning Strategies

The study conducted a controlled experiment with 32 intermediate computer science students. Each student was tasked with learning two new programming concepts of varying difficulty: a moderately difficult concept (Immediately Invoked Function Expressions or IIFE) and a more complex one (Currying). For one concept, they could only use ChatGPT (the "AI" group); for the other, they used traditional web resources like Google search, tutorials, and videos (the "NoAI" group). Their performance was then measured through a conceptual quiz and a practical debugging task.

This setup provides a powerful analogue for the modern enterprise environment, where employees must constantly learn new technologies, processes, and internal systems. The choice between consulting a general-purpose AI and searching an internal knowledge base (like Confluence or SharePoint) mirrors the study's core comparison.

How Learners Allocated Their Time

The study tracked how students spent their 15-minute learning sessions. In both scenarios, learning from examples was dominant, but the NoAI group also dedicated significant time to watching videos, highlighting the demand for richer media. Interestingly, users spent slightly more time "locating" information with ChatGPT, which contradicts the assumption that direct answers save time. This is likely due to the cognitive load of formulating precise prompts and parsing conversational history.

AI (ChatGPT)
NoAI (Web Search)

Section 2: Decoding User Behavior - Keyword Search vs. Conversational Inquiry

A crucial finding lies in *how* users sought information. The study analyzed the structure of 148 web queries and 237 LLM prompts, revealing fundamentally different cognitive approaches.

  • Web Search (NoAI): Dominated by keyword-based queries (42%). Users cast a wide net, relying on the search engine to surface a variety of ranked, context-rich documents. This method encourages exploration and synthesis of information from multiple sources.
  • LLM (AI): Characterized by explicit questions and requests for explanation (39%). Users engaged in a more conversational, iterative dialogue, asking significantly more follow-up questions. This creates a focused but potentially narrow learning path, heavily dependent on the AI's initial response.

Query Phrasing: Search Engine vs. LLM

This chart illustrates the stark difference in how users frame their information needs. Search engine users "search," while LLM users "ask." This has profound implications for designing internal knowledge systems. A system that only supports one mode of inquiry will fail to meet the needs of all users.

Enterprise Takeaway: Your internal AI knowledge system must be a hybrid. It needs a powerful semantic search to handle keyword-based exploration *and* a sophisticated conversational interface for targeted, explanatory dialogues. Relying on one or the other creates an incomplete solution.

Section 3: The Performance Paradox - When Quick Fixes Erode Deep Knowledge

This is the most critical insight for any organization leveraging AI. While both groups performed similarly on the simpler task, the results for the complex "Currying" task were alarming.

Conceptual Quiz Scores on the Difficult Task (Currying)

On a quiz designed to test deep understanding, students who used traditional web search significantly outperformed those who used ChatGPT. The LLM users struggled to connect concepts, suggesting they acquired fragmented facts rather than a cohesive mental model. The average score for the NoAI group was 2.0, while the AI group averaged -0.69 (on a scale where incorrect answers gave negative points).

The practical debugging task further exposed this gap. While the raw success rate was similar, the researchers interviewed students about their solutions. The results revealed a category of "Correct (LoU)" - solutions that worked, but where the student had a Lack of Understanding.

Debugging Task Outcomes: The Hidden Cost of "AI Magic"

The chart below shows the breakdown of outcomes. A significant portion of the AI group fell into the "Correct (LoU)" category. In an enterprise setting, this translates to an employee copy-pasting code that fixes a bug but introduces technical debt or violates best practices because they don't understand *how* it works. This is a ticking time bomb for quality and maintainability.

Correct (with Understanding)
Correct (Lack of Understanding)
Incorrect

Section 4: Enterprise Applications & A Custom AI Roadmap

The research is a clear mandate: enterprises cannot simply deploy a generic LLM and expect effective knowledge transfer. A strategic, custom approach is required. At OwnYourAI.com, we build systems based on a "Guided Expertise" framework, which mitigates the risks identified in the study.

Section 5: Interactive ROI Calculator - The Value of Deep Knowledge

The "Productivity-Proficiency Gap" has real financial costs: time wasted on rework, senior developer time spent re-explaining concepts, and bugs shipped to production. Use our calculator to estimate the potential value of implementing a custom AI knowledge system that fosters deep understanding instead of shallow fixes.

Section 6: Test Your Understanding

Take this short quiz to see if you've grasped the key takeaways from this analysis.

Conclusion: Build AI That Builds Expertise, Not Dependency

The study "To Google or To ChatGPT?" is a landmark paper for the enterprise AI era. It proves that the path to information dramatically shapes the quality of learning. While generic LLMs offer impressive speed, they can inadvertently create a workforce that is proficient at getting answers but deficient in true understanding.

The future of competitive advantage lies in building a knowledgeable, adaptable team. This requires moving beyond off-the-shelf AI and investing in custom knowledge systems that are designed for learning, not just answering. These systems blend conversational ease with curated, context-rich resources, guide users toward holistic understanding, and provide analytics to identify knowledge gaps before they become business problems.

Ready to build an AI that empowers your team with genuine expertise?

Schedule Your Custom AI Strategy Session

Ready to Get Started?

Book Your Free Consultation.

Let's Discuss Your AI Strategy!

Lets Discuss Your Needs


AI Consultation Booking