• | 8:00 am

The most innovative AI applications today have this in common

The most forward-thinking developers are blending domain intelligence and trust to make LLMs work for corporate America.

The most innovative AI applications today have this in common
[Source photo: FC]

Generative AI models don’t know what they don’t know. Ask a question on a subject they haven’t encountered in their training, and they might just make something up. For an individual playing around with the tech, that’s an annoyance. For a company, it’s a nightmare.

That’s why the best AI application developers are building their own AI around the foundation model, leveraging unique technology or data to get the good out of LLMs while managing the bad. When done well, these apps or services can do far more—and create far more real-world value—than an LLM-powered chatbot can by itself.

“In an LLM-enabled world, it is quite tempting for a large model to feel like a hammer, and everything else a nail,” says Chris Kauffman, a partner at the VC firm General Catalyst. “In production, the truth is more nuanced,” he says, adding that the most forward-thinking builders are creating what he calls “integrated circuits of intelligence” with this blended approach.

Some of the most effective AI applications add domain expertise that doesn’t exist in the vast world of internet content, which trains most LLMs. Casetext adds value to its legal assistant AI, CoCounsel, via specialized data. CoCounsel gets its basic text summarization and writing capabilities from OpenAI’s GPT-4 language model, but the LLM also has access to “ground truth” information from proprietary databases of verified legal data. Interplay Learning, which develops training resources for electricians and others in the skilled trades, married its extensive multimedia knowledge base with an LLM to launch its training assistant SAM (Skill Adviser and Mentor), which allows workers to ask specific questions when they’re in the middle of a job. And Seekr, which makes a search engine that scores news content on its reliability, fine-tunes its LLM with media expertise collected from a panel of independent journalists. It also trains the AI on data from a large repository of well-reported, well-written news articles that it’s painstakingly collected. The goal is an AI tool that functions something like a human editor with a distaste for spin, exaggeration, partisan bias, and clickbait headlines.

Building that kind of trust into generative AI is the other essential value that applied AI purveyors are seeking to add. For law firms, Casetext built guardrails around Open AI’s GPT-4 to prevent it from making up facts—case details or key findings, for example—which could be disastrous in a legal context. It also added data privacy and security measures so that the model’s output meets professional legal standards. When Salesforce launched its Einstein AI framework in March 2023, it didn’t just let customers of its popular cloud customer relationship management (CRM) software bring generative AI to their sales, marketing, IT, and customer support groups by choosing a third-party LLM such as Anthropic or OpenAI—or using a Salesforce one. Einstein connects whichever LLM customers want to use with that company’s proprietary business data so that the output of the generative AI tools is informed by the company’s business intelligence and communication style. Because many CEOs and CIOs fear the “leakage” of that valuable currency, Salesforce built the Einstein Trust Layer, which masks the proprietary data before it passes through an LLM.

Although OpenAI and other LLM makers continue to add new capabilities to their offerings in the effort to create “one model to do it all,” AI app developers will continue to use LLMs as just one layer of a larger system. “We think it is imperative to take a long-term view and build with malleability and modularity in mind,” says General Catalyst’s Kauffman, citing the critical need to be able to swap new models in and out, integrate new data, and “incorporate new computational paradigms as they emerge.”

  Be in the Know. Subscribe to our Newsletters.

ABOUT THE AUTHOR

Mark Sullivan is a senior writer at Fast Company, covering emerging tech, AI, and tech policy. Before coming to Fast Company in January 2016, Sullivan wrote for VentureBeat, Light Reading, CNET, Wired, and PCWorld More

More Top Stories:

FROM OUR PARTNERS