It is a myth that a single defining large language model (LLM) or other type of generative AI model will define all use cases.
Many corners of the technology market have come to be defined by a handful of companies. The nature of generative AI, especially for enterprises, suggests we will be looking at thousands of models or more.
The reasons vary, but it is already clear that some models are good at summarization, others are good at bulleted lists, others are good at reasoning, and so forth. Industries, lines of business, and companies have very different editorial tones for expression of knowledge. All these should be considered when choosing your models.
Generative AI models consume large amounts of computing resources. The large funding rounds required for companies creating foundation models is just one testament to these costs.
Potentially high compute costs are one reason why using the right model for the job is so important. The larger the model, the more it costs to query.
Your enterprise model doesn’t need to know the words to every Taylor Swift song to generate a summary report on next quarter’s sales goals by region. Context is king and you need to be selective in just how much IQ a model requires for your use case.
Just as past “bring your own device” and “bring your own app” movements raised “shadow IT” concerns, some financial institutions I work with have shut down access to publicly available generative AI for fear that models could leak proprietary information.
Some public generative AI services may leverage user data for future training sets, potentially exposing proprietary data. Let’s say a bank is exploring a merger for a large industrial client, and someone in the M&A department queries a public model, asking, “What are some good takeover targets for XYZ Industries?” If that information contributes to the model’s training data, the service could become trained to answer this question for anyone. By default, Google Cloud AI services do not use private data in this way.
Most customers I speak to are worried about the security of the questions they ask models, the content they train on, and their models’ outputs. And they probably should be.
The accuracy and reliability of generative AI has been one of the biggest topics around the new technology. An algorithm is designed to give an answer no matter what, and in some cases, generative AI models can produce answers that aren’t true.
Every company I know has deeply invested in creating verifiable facts and data. It’s essential for enterprise customers to use models and a technology architecture that is grounded in the factuality of their data.
Most generative AI models punt on this enterprise data requirement. It is essential, especially in regulated industries, to not punt.
Enterprise customers have many information sources: pricing, HR, legal, finance, etc. But I dont know of any company that allows open access to all of this information.
Some business leaders are increasingly interested in building all their information into a large language model, so it can conceivably answer all questions, whether that’s at the organizational level or the global level.
After a company thinks through how they can keep their information private and factual, they quickly realize the next step: How do I manage who can ask questions of my models, and at what level?