Fine-tuning Open Source LLMs with Mistral
Andrea, a Computing Engineer at CERN, and Josep, a Data Scientist at the Catalan Tourist Board, will walk you through the steps needed to customize the open-source Mistral LLM.
Jul 16, 2024
RelatedSee MoreSee More
tutorial
Mistral 7B Tutorial: A Step-by-Step Guide to Using and Fine-Tuning Mistral 7B
The tutorial covers accessing, quantizing, fine-tuning, merging, and saving this powerful 7.3 billion parameter open-source language model.
Abid Ali Awan
12 min
tutorial
Fine-Tuning LLaMA 2: A Step-by-Step Guide to Customizing the Large Language Model
Learn how to fine-tune Llama-2 on Colab using new techniques to overcome memory and computing limitations to make open-source large language models more accessible.
Abid Ali Awan
12 min
tutorial
Codestral API Tutorial: Getting Started With Mistral’s API
To connect to the Codestral API, obtain your API key from Mistral AI and send authorized HTTP requests to the appropriate endpoint (either codestral.mistral.ai or api.mistral.ai).
Ryan Ong
9 min
tutorial
A Comprehensive Guide to Working with the Mistral Large Model
A detailed tutorial on the functionalities, comparisons, and practical applications of the Mistral Large Model.
Josep Ferrer
12 min
tutorial
LlaMA-Factory WebUI Beginner's Guide: Fine-Tuning LLMs
Learn how to fine-tune LLMs on custom datasets, evaluate performance, and seamlessly export and serve models using the LLaMA-Factory's low/no-code framework.
Abid Ali Awan
12 min
tutorial
Fine-Tuning Llama 3 and Using It Locally: A Step-by-Step Guide
We'll fine-tune Llama 3 on a dataset of patient-doctor conversations, creating a model tailored for medical dialogue. After merging, converting, and quantizing the model, it will be ready for private local use via the Jan application.
Abid Ali Awan
19 min