In the world of LLMs, there are four primary levels of training and fine-tuning. Each level serves a specific purpose and involves different amounts of data and computational resources. Let’s explore these levels in detail.
Pretraining is the foundation of creating a new base model. This level is used when you want to develop a model from scratch or create a specialized base model for specific domains like mathematics, medicine, or other specialized topics.
Instruction tuning is the process of fine-tuning a pretrained model to follow specific instructions or prompts.
This is the most common fine tuning, since base model has compressed most the world knowledge, you
only have to train them on your task.
Text-SQL, TEXT-API, Question Answering, Text to action, Retrievals, Generate HTML design from prompt
Enhancing the model’s understanding of task-specific prompts.
Fine tune small model on synthetic dataset from a bigger model for cost reduction.
Fine tune small model on to prevent long repetitive prompt and speed.
Fine tune big model on hard task that father modal cant solve.
dataset.jsonl
Copy
{"instruction": "Some important task with context ?", "output": "Boo" }{"instruction": "Some important task with context ?", "output": "Boo" }{"instruction": "Some important task with context ?", "output": "Boo" }
Chat tuning is a specialized form of instruction tuning that focuses on improving the model’s conversational abilities.
So if build an interactive chatbot this is what you need to train
Each level of LLM training serves a specific purpose in the development and refinement of these powerful AI models. While pretraining creates the foundation, the subsequent levels of fine-tuning allow for more specialized and targeted improvements. Understanding these levels can help in choosing the right approach for developing or customizing LLMs for specific applications.