Instead of fine-tuning an LLM as a first approach, try prompt architecting instead

Date:

Share post:


Amid the generative AI eruption, innovation directors are bolstering their business’ IT department in pursuit of customized chatbots or LLMs. They want ChatGPT but with domain-specific information underpinning vast functionality, data security and compliance, and improved accuracy and relevance.

The question often arises: Should they build an LLM from scratch, or fine-tune an existing one with their own data? For the majority of companies, both options are impractical. Here’s why.

TL;DR: Given the right sequence of prompts, LLMs are remarkably smart at bending to your will. The LLM itself or its training data need not be modified in order to tailor it to specific data or domain information.

Exhausting efforts in constructing a comprehensive “prompt architecture” is advised before considering more costly alternatives. This approach is designed to maximize the value extracted from a variety of prompts, enhancing API-powered tools.

TL;DR: Given the right sequence of prompts, LLMs are remarkably smart at bending to your will.

If this proves inadequate (a minority of cases), then a fine-tuning process (which is often more costly due to the data prep involved) might be considered. Building one from scratch is almost always out of the question.

The sought-after outcome is finding a way to leverage your existing documents to create tailored solutions that accurately, swiftly, and securely automate the execution of frequent tasks or the answering of frequent queries. Prompt architecture stands out as the most efficient and cost-effective path to achieve this.

What’s the difference between prompt architecting and fine-tuning?

If you are considering prompt architecting, you have likely already explored the concept of fine-tuning. Here is the key distinction between the two:

While fine-tuning involves modifying the underlying foundational LLM, prompt architecting does not.

Fine-tuning is a substantial endeavor that entails retraining a segment of an LLM with a large new dataset — ideally your proprietary dataset. This process imbues the LLM with domain-specific knowledge, attempting to tailor it to your industry and business context.

In contrast, prompt architecting involves leveraging existing LLMs without modifying the model itself or its training data. Instead, it combines a complex and cleverly engineered series of prompts to deliver consistent output.

Fine-tuning is appropriate for companies with the most stringent data privacy requirements (e.g., banks)



Source link

Lisa Holden
Lisa Holden
Lisa Holden is a news writer for LinkDaddy News. She writes health, sport, tech, and more. Some of her favorite topics include the latest trends in fitness and wellness, the best ways to use technology to improve your life, and the latest developments in medical research.

Recent posts

Related articles

Artifact takes on X and Threads with new Posts feature

Artifact, a platform built by Instagram’s co-founders, is launching the ability for users to make posts. Up...

Can Bird’s Spin acquisition give it the lift it needs?

After being delisted from the NYSE, regaining investor confidence won't be easy It’s been over a year since...

Mark Zuckerberg went 33 minutes without saying ‘metaverse’ at his keynote

When Meta rebranded from Facebook two years ago, the word “metaverse” went mainstream. Even people outside the...

Generative AI stickers are coming to Meta’s apps

At its Meta Connect event today, CEO Mark Zuckerberg announced that generative AI stickers would be coming...

Ask Sophie: How would a government shutdown affect the H-1B visa process?

Sophie Alcorn Contributor Sophie Alcorn is the founder of Alcorn Immigration Law in Silicon Valley and 2019 Global Law...

Sam Altman backs Slope’s $30M round to digitize, scale B2B payments

Slope, a business-to-business payments platform for enterprise companies, closed on a venture round of $30 million to...

Mythos Ventures grabs $14M for inaugural fund to invest in AI

Mythos Ventures, an early-stage venture capital firm started by Vishal Maini, closed $14 million in capital commitments...

Gem Security raises $23M for its cloud security platform

Cloud detection and response company Gem Security today announced that it has raised a $23 million Series...