The furor has calmed since last November when Generative AI first rose to peak hype. The breathless excitement about the size of OpenAI’s GPT-4, Google’s Bard, and Meta’s LLaMA and the broad range of tasks they can fulfill has given way to more measured consideration of the value they bring to enterprise buyers. Are the number of GenAI training parameters relevant for an enterprise with a specific business goal? Might a smaller model trained only on relevant language data be the future of Generative AI?
This is an age-old question in the tech space. Many technologies enter mainstream awareness with a general-purpose application that must evolve to serve enterprise needs. Take speech recognition as an example. Applying the technology to drive customer service prompts required a different model from applying the technology to identify a customer for security purposes, or disambiguating the street names “Market” from “Marquette” for navigation guidance.
A similar evolution is already underway with Generative AI. Commentators have been arguing that LLMs will get smaller, faster, and less expensive to build and use as enterprises identify specific use cases and develop specific models for them. We share our own perspective specifically for marketing use cases in our CMO’s Guide to Achieving Impact With Generative AI.
In this post, I discuss why smaller may be better for enterprise technology buyers looking to understand the future of Generative AI.
LLMs like GPT-4, Bard, and LLaMa represent a massive step forward in the discipline of natural language processing (NLP). Trained on a vast corpus of general-purpose text from the internet, they work by predicting the next likely word or set of words in a sentence, based on what they’ve seen in their training data.
When they first achieved mainstream attention, there seemed to be endless potential for them to evolve toward artificial general intelligence, a theoretical model for an AI that can reason through any task.
The current reality looks a bit different. Based on the nature of their training data and how people are using them, there appear to be three immediate use cases for Generative AI.
The first and most straightforward GenAI use case focuses on tactical writing tasks. No surprise there!
Depending on the user’s purpose, that text might be used as a first draft for some written output, such as an email, a resume, or a report. Gartner predicts that by 2025, 30% of outbound marketing messages from large organizations will be synthetically generated, up from less than 2% in 2022.
The second immediate use case is to replace search with a more fluent and specific alternative. Over the past year or so, millions of people have used ChatGPT to answer a question that in the past they would have posed to a search engine. Instead of search results, snippets, or FAQs responses, they get full sentences. They can even get source materials if they ask.
Third, GenAI is emerging as a powerful tool for mining language data for ideas, opportunities, or sources of risk. For example, financial institutions and retailers are using it to mine transaction data to spot behavior associated with risk or fraud. Health researchers are hoping GenAI can help identify novel pharmaceuticals.
Beyond text, there are other use cases are emerging for the text-to-image models, which allow users to quickly translate a verbal request into a corresponding visual (i.e. “Create a photo-realistic painting of a rottweiler running after a ball on the beach”).
These use cases validate the importance of LLMs for solving broad and common problems related to completing basic writing or information-gathering tasks. They cannot do more than that on their own, however. To the extent that they do, it is because users are “hacking” the models with more sophisticated prompts.
What does that mean for enterprises?
Employees write tactical texts and conduct both basic research and fact checks for their jobs all the time. For many of these general-purpose tasks, a core LLM large language model will do just fine.
Ask a general-purpose LLM to perform a specialized task for an enterprise, however, and that LLM will almost always underperform. It simply wasn’t trained to execute on specific language-driven tasks aligned to a concrete performance goal.
Look at that picture of a rottweiler as a low-stakes example. It is recognizably a dog in that breed. The fact that I could create it within less than a minute is a testament to the power of the underlying model. One reason why that speed is possible is because LLMs that are general in nature use a “limit approaches” construct to fulfill the purpose they were designed for. Put simply, they are designed to get close enough, not make an exact match. Fuzzy logic is both the power of a general-purpose LLM and its weakness.
Look again at that dog picture and fuzzy is what you get. For example, the foreleg and shoulder are clearly human with black fur on them.
Does it matter?
Not if all I am doing is using a picture to illustrate a point in a blog post. It would matter a great deal, however, if I was using a text-to-visuals model to design a product that has utility and safety requirements. Or reporting on public company sales data to the SEC. Or mapping precise exit routes for emergency evacuation procedures.
For important tasks aligned with business goals, enterprise buyers can’t tolerate imprecision. Yet that is inevitable when too many of a model’s training parameters are irrelevant to the task at hand. It is like asking a Boeing 747 to do the job of a General Dynamics F16. Both are aircraft, but the Boeing wouldn’t just be the wrong choice for an active combat zone. It would be a liability.
For that reason, the future of Generative AI for enterprises applications must meet standards in the following areas:
The good news for enterprise technology buyers is that the options for enterprise Generative AI are already diverse and growing well beyond the familiar, consumer-oriented tools. The five standards of reliability, efficiency, effectiveness, security, and innovation are deeply embedded in the DNA of organizations creating Generative AI models and solutions from the outset to solve a defined business problem.
Nvidia, for example, has an enterprise Generative AI model it licenses for third-party developer adoption. So does Amazon for its AWS customer base. Large technology companies have also integrated Generative AI capabilities into their platforms. Examples include Microsoft 365, SAP, and Adobe Firefly. Finally, the start up community is bringing an array of specific point solutions with Generative AI models trained to solve specific problems.
A key difference between enterprise Generative AI and general purpose LLMs lies in the following functional capabilities:
Given the use cases for Generative AI and the specific enterprise needs standards GenAI must address, consider the following four options for the future of Generative AI in your enterprise:
The launch of GPT-3 and others provided solutions to a few of the technical problems that made model development challenging for private enterprises. Specifically, the transformer architecture gave data scientists a new approach that sped up language model training. Any enterprise can theoretically use these techniques to create their own language models for in-house purposes, if they have the staff and the budget to do it.
Pros:
Cons:
Enterprises can license or download LLMs and refine their functionality. There are two primary methods for doing that. One is by fine-tuning the model using specialized data sets to narrow the parameters that the LLM uses. Another is through prompt engineering, which uses specific instructions to focus the LLM to deliver a specific output, sometimes by focusing only on certain parameters. Think of fine turning or prompt engineering as making a large language model intentionally smaller for a specific purpose.
Pros:
Cons:
As noted above, there have been a string of announcements from enterprise technology service providers about the Generative AI functionality they have developed for their solutions. Microsoft has integrated GPT-4 into Office 365; Salesforce has language generation capabilities in its marketing cloud, etc. The language model underpinning the Generative AI functionality may be large and general purpose, or it may be specialized, depending on how the vendor developed it.
Pros:
Cons:
A range of point solutions have emerged to address specific Generative AI use cases. Persado Motivation AI, for example, is an enterprise Generative AI solution that optimizes the digital marketing language used to motivate customers to engage and act. Other marketing-focused solutions include Descript.ai for video editing and podcasting, and Peppertype.ai for content and video creation.
Some point solutions start with a licensed LLM and build their solution on top of it. Others, like Persado, build language models using specific data sets focused on the use case. In our case, we focus on brand campaign performance.
Pros:
Cons:
The major Generative AI models brought the technology into the mainstream. For an enterprise to adopt a new technology widely, however, it must have a more precise business value proposition (efficiency, effectiveness, innovation) and clear governance parameters (reliability and security). While there will be a place for general purpose LLMs in that eco-system, there are also huge opportunities today for enterprises to adopt point solutions founded on smaller and more refined language models designed to solve a problem, or unlock an opportunity.
To learn more about how Persado Motivation AI helps brands optimize customer engagement, reach out for a free trial today.
The post The future of Generative AI language models is “small” appeared first on Persado.
Podcasts are a fun and easy way to learn about machine learning.
TL;DR We asked o1 to share its thoughts on our recent LNM/LMM post. https://www.artificial-intelligence.show/the-ai-podcast/o1s-thoughts-on-lnms-and-lmms What…
Palantir and Grafana Labs’ Strategic PartnershipIntroductionIn today’s rapidly evolving technological landscape, government agencies face the…
Amazon SageMaker Pipelines includes features that allow you to streamline and automate machine learning (ML)…
When it comes to AI, large language models (LLMs) and machine learning (ML) are taking…
Cohere's Command R7B uses RAG, features a context length of 128K, supports 23 languages and…