The Don’ts: Taking on More Work Than Necessary
Avoid models that try to do everything. Generative AI technologies such as OpenAI can have as many as 1 trillion parameters. That’s not consumable for most entities; it uses a lot of computing resources, and it’s a lot more than many predictive models will need, Winterich says.
Fortunately, a groundswell of open-source models with fewer but more specific variants has lowered the barrier to entry for organizations looking to put generative AI into practice.
“That makes it easy to say, this model is good for finance, translation, search and so on,” Winterich says. “That’s where the rubber hits the road.”
Approach infrastructure with caution
Along the same lines, the evolution of purpose-built predictive models means agencies don’t need to max out their infrastructure deployments. For example, large language models such as Mistral 7 can run on a single graphics processing unit.
“You don’t need to rent a collocated data center or a room full of computers to do knowledge-based search,” Winterich says.
With agencies increasingly interested in training models on their own data sets, on-premises infrastructure can provide privacy and peace of mind, he adds. The scalability of exascale supercomputers such as HPE Cray can further help agencies leverage the right resources at the right time.
Click the banner to learn why platform engineering improves efficiency in your organization.