- | 8:00 am
The most innovative AI applications today have this in common
The most forward-thinking developers are blending domain intelligence and trust to make LLMs work for corporate America.
Generative AI models don’t know what they don’t know. Ask a question on a subject they haven’t encountered in their training, and they might just make something up. For an individual playing around with the tech, that’s an annoyance. For a company, it’s a nightmare.
That’s why the best AI application developers are building their own AI around the foundation model, leveraging unique technology or data to get the good out of LLMs while managing the bad. When done well, these apps or services can do far more—and create far more real-world value—than an LLM-powered chatbot can by itself.
“In an LLM-enabled world, it is quite tempting for a large model to feel like a hammer, and everything else a nail,” says Chris Kauffman, a partner at the VC firm General Catalyst. “In production, the truth is more nuanced,” he says, adding that the most forward-thinking builders are creating what he calls “integrated circuits of intelligence” with this blended approach.
Some of the most effective AI applications add domain expertise that doesn’t exist in the vast world of internet content, which trains most LLMs. Casetext adds value to its legal assistant AI, CoCounsel, via specialized data. CoCounsel gets its basic text summarization and writing capabilities from OpenAI’s GPT-4 language model, but the LLM also has access to “ground truth” information from proprietary databases of verified legal data. Interplay Learning, which develops training resources for electricians and others in the skilled trades, married its extensive multimedia knowledge base with an LLM to launch its training assistant SAM (Skill Adviser and Mentor), which allows workers to ask specific questions when they’re in the middle of a job. And Seekr, which makes a search engine that scores news content on its reliability, fine-tunes its LLM with media expertise collected from a panel of independent journalists. It also trains the AI on data from a large repository of well-reported, well-written news articles that it’s painstakingly collected. The goal is an AI tool that functions something like a human editor with a distaste for spin, exaggeration, partisan bias, and clickbait headlines.
Although OpenAI and other LLM makers continue to add new capabilities to their offerings in the effort to create “one model to do it all,” AI app developers will continue to use LLMs as just one layer of a larger system. “We think it is imperative to take a long-term view and build with malleability and modularity in mind,” says General Catalyst’s Kauffman, citing the critical need to be able to swap new models in and out, integrate new data, and “incorporate new computational paradigms as they emerge.”