Track
Llama 4 With vLLM: A Guide With Demo Project
Learn how to deploy and use Meta's LLaMA 4 Scout with vLLM on RunPod for both text completion and multimodal inference.
Apr 10, 2025 · 12 min read
Learn AI with these courses!
5 hours hr
Course
Fine-Tuning with Llama 3
2 hr
1.4K
Course
Retrieval Augmented Generation (RAG) with LangChain
3 hr
5.2K
See More
RelatedSee MoreSee More
blog
Meta's Llama 4: Features, Access, How It Works, and More
Learn about the Llama 4 suite of large language models, including Llama 4 Scout, Llama 4 Maverick, and the in-training Llama 4 Behemoth.
Alex Olteanu
8 min
blog
Llama 3.2 Guide: How It Works, Use Cases & More
Meta releases Llama 3.2, which features small and medium-sized vision LLMs (11B and 90B) alongside lightweight text-only models (1B and 3B). It also introduces the Llama Stack Distribution.
Alex Olteanu
8 min
Tutorial
Llama Stack: A Guide With Practical Examples
Llama Stack is a set of standardized tools and APIs developed by Meta that simplifies the process of building and deploying large language model applications.
Hesam Sheikh Hassani
8 min
Tutorial
Llama 3.3: Step-by-Step Tutorial With Demo Project
Learn how to build a multilingual code explanation app using Llama 3.3, Hugging Face, and Streamlit.
Dr Ana Rojo-Echeburúa
12 min
Tutorial
vLLM: Setting Up vLLM Locally and on Google Cloud for CPU
Learn how to set up and run vLLM (Virtual Large Language Model) locally using Docker and in the cloud using Google Cloud.
François Aubry
12 min
Tutorial
Llama 3.2 and Gradio Tutorial: Build a Multimodal Web App
Learn how to use the Llama 3.2 11B vision model with Gradio to create a multimodal web app that functions as a customer support assistant.
Aashi Dutt
10 min