Skip to main content

Unlock New LLM Architectural Capabilities By Retraining

Jacob Buckman, the CEO at Manifest AI, guides you through cutting-edge methods for LLM design, retraining, and deployment.
Nov 24, 2025

Large language models (LLMs) are evolving fast—and retraining offers a powerful path to new architectural capabilities, lower cost, and better performance. Recent research on the Brumby-14B-Base model shows how attention-free retention layers, efficient retraining from pretrained weights, and long-context processing can reshape what’s possible in generative AI. This session dives deep into these techniques and how you can apply them in your own work.

In this code-along, Jacob Buckman, the CEO at Manifest AI, guides you through cutting-edge methods for LLM design, retraining, and deployment. You’ll explore how alternative attention mechanisms like power retention change model behaviour, how to repurpose pretrained models for new architectures, and how to engineer large-scale training pipelines. Whether you’re building foundation models or pushing boundaries in generative AI research, this session gives you the tools and insights to lead the next wave.

Key Takeaways:

  • Understand how different attention (and retention) mechanisms affect LLM behaviour, performance, and cost.
  • Learn the process of retraining large language models by repurposing pretrained weights and exploring new architectures.
  • Discover the latest research techniques—including long-context inference, hardware-efficient kernels, and scalable LLM families—and how to apply them in your projects.

Session Resources + GitHub LinkPower Attention - Manifest AI

Topics
Related

blog

AI in HR: Shaping the Future of Talent Management

Unlock AI's potential in modern HR practices. From recruitment to retention, discover how artificial intelligence is transforming the workplace.
Vinod Chugani's photo

Vinod Chugani

11 min

podcast

[Radar Recap] The Future of Programming: Accelerating Coding Workflows with LLMs

Michele Catasta, VP of AI at Replit and Jordan Tigani, CEO at Motherduck, explore practical applications of LLMs in coding workflows, what the future holds for AI-assisted coding, and more.

podcast

Creating Custom LLMs with Vincent Granville, Founder, CEO & Chief Al Scientist at GenAltechLab.com

Richie and Vincent explore why you might want to create a custom LLM, the development and features of custom LLMs, architecture and technical details, corporate use cases and much more.

Tutorial

An Introduction to the Mamba LLM Architecture: A New Paradigm in Machine Learning

Discover the power of Mamba LLM, a transformative architecture from leading universities, redefining sequence processing in AI.
Kurtis Pykes 's photo

Kurtis Pykes

Tutorial

How to Train an LLM with PyTorch

Master the process of training large language models using PyTorch, from initial setup to final implementation.
Zoumana Keita 's photo

Zoumana Keita

code-along

Understanding LLMs for Code Generation

Explore the role of LLMs for coding tasks, focusing on hands-on examples that demonstrate effective prompt engineering techniques to optimize code generation.
Andrea Valenzuela's photo

Andrea Valenzuela

See MoreSee More