News
Jun 2, 2025
Now that you understand what LLMs are and their limitations, let's tackle the practical question that matters most: how do you actually harness this powerful technology to perform useful tasks?
When vendors talk about "AI solutions," they're often using different techniques to make LLMs do specific work. Understanding these core approaches will help you cut through the marketing jargon and evaluate what's actually happening behind the scenes. Here are the main techniques that developers and companies use to transform raw LLM capabilities into practical applications:
Think of prompt engineering as learning to speak "AI language." Just as you might phrase requests differently when speaking to different staff members based on their expertise, how you "ask" an AI system to perform a task dramatically affects the results.
Behind the scenes, vendors are:
For example, a vendor's document processing system isn't just "asking" the AI to summarize content. They've likely spent months refining instructions like: "Create a concise summary of this patient encounter note, focusing on diagnosis, treatment plan, and follow-up instructions. Format the summary with clear headings for each section." These precise instructions, invisible to the end user, are what make the difference between an unreliable demo and a production-ready solution.
RAG is a powerful approach that addresses one of LLMs' biggest weaknesses: their tendency to hallucinate when they don't know something. Here's how it works:
The key insight with RAG is that it's essentially an extension of prompt engineering—the system is automatically injecting relevant facts into the prompt and instructing the LLM to "answer based only on the information provided." This creates a powerful combination: the LLM's language capabilities with the accuracy of verified information.
Behind the scenes, RAG systems work through a process called "vectorization"—converting your documents into a special mathematical format (think of it as creating a digital "fingerprint" of each piece of information). When someone asks a question, the system converts that question into the same mathematical format and finds the most similar "fingerprints" in your knowledge base, pulling that information into the prompt.
This is particularly valuable in healthcare, where accuracy matters. A RAG-powered system answering questions about insurance coverage would first retrieve the actual policy details before generating an explanation, dramatically reducing hallucination risks.
For more sophisticated use cases, AI agents can coordinate multi-step processes by breaking complex tasks into manageable pieces:
This multi-agent approach is one of the most powerful ways to reduce errors. For example, one agent might generate content while another specialized "critic" agent reviews that content for accuracy against trusted sources. This creates a system of checks and balances that significantly improves reliability.
For example, an AI agent processing a new patient referral might extract information from the referral document, check insurance eligibility, identify appropriate providers based on specialties needed, find available appointments, and draft the initial communication—all as a coordinated workflow rather than isolated tasks. A verification agent could then double-check that the extracted information matches the original referral document before the workflow proceeds.
While prompt engineering, RAG, and AI agents can work effectively with general-purpose LLMs, some vendors offer "fine-tuned" models—custom versions specifically trained on healthcare or other domain-specific data. But what does this really mean, and what are the tradeoffs?
When a vendor mentions their "fine-tuned model," they're typically referring to additional training they've conducted on a specific dataset to adapt a general-purpose LLM to a particular domain or task. This process involves exposing the model to thousands of examples to adjust its internal parameters. Here's what you should understand about fine-tuning:
However, recent research reveals significant downsides to fine-tuning that many vendors won't mention:
When evaluating vendor solutions, it's important to understand their approach to model development. Some vendors heavily market their "fine-tuned" models as a competitive advantage, while others focus on sophisticated prompting and RAG techniques with general-purpose models. Neither approach is inherently superior—what matters is the actual performance and reliability of the solution with your specific use cases and data. A well-designed system using prompt engineering and RAG often delivers better results than a poorly fine-tuned model, and at significantly lower risk. When vendors emphasize their "custom fine-tuned models," ask critical questions about how they maintain safety, prevent capability degradation, and handle ongoing updates.
When considering AI solutions for your practice, remember this golden rule: Never implement an AI system without thoroughly testing it on your own data first. Here's why this matters:
Be extremely skeptical of any vendor that resists testing their AI solution with your actual data before full implementation. A reputable AI partner will insist on running pilots to demonstrate performance and identify potential issues before full deployment.
The good news is that hallucinations can be managed through thoughtful implementation strategies. The appropriate approach depends on your risk assessment:
For high-risk contexts (where incorrect information could impact patient care or compliance):
For medium-risk contexts (where errors would cause inefficiency but not harm):
For low-risk contexts (where occasional errors are tolerable):
The right approach depends on your specific needs, but the key principle remains: AI implementations should be matched with appropriate oversight mechanisms based on a realistic assessment of both the benefits and risks.
Curious about Basata? Learn how you can give your admin team superpowers.
Get in touch today and learn how Basata can give your admin team superpowers.