Exploring Complexity Ep. 2
Why the world feels more complex—and why that feels hard. Why more problems are complex problems. Why organizations struggle with complexity.
Enterprises face a critical choice in their generative AI adoption strategy: fine-tuning or Retrieval-Augmented Generation (RAG)? While fine-tuning has been the go-to approach for early adopters, a new study suggests that RAG may be the more powerful and sustainable path forward.
Key points:
Enterprises face a critical choice in their generative AI adoption strategy: fine-tuning or Retrieval-Augmented Generation (RAG)? While fine-tuning has been the go-to approach for early adopters seeking to quickly adapt genAI to their needs, a new study suggests that RAG may be the more powerful and sustainable path forward.
To date, fine-tuning has been the favored approach for enterprises looking to harness genAI. By training foundational language models (LLMs) on domain-specific data, businesses can rapidly develop custom applications tailored to their needs. This plug-and-play simplicity has made fine-tuning the entry point for many—a16z's research shows that 72% of enterprises rely on fine-tuning while only 22% rely on RAG.
However, the popularity of fine-tuning may owe more to timing than true technical superiority. As the first widely accessible adaptation technique, fine-tuning naturally attracted early adopters eager to experiment with genAI. The publicity around prominent fine-tuned models like BloombergGPT further fueled this trend.
However as enterprises move beyond initial pilots into large-scale deployment, the limitations of fine-tuning are coming into sharper focus. Training LLMs from scratch is staggeringly resource-intensive, requiring vast computational power and bespoke technical talent. Fine-tuned models also struggle to keep pace with the rapid evolution of genAI, leaving enterprises at risk of being leapfrogged by nimbler competitors.
The Artificiality Weekend Briefing: About AI, Not Written by AI