Published February 13, 2026

verl Framework for Training Language Models Now Supports AMD GPUs

Training Language Models with Feedback: verl Now Runs on AMD GPUs

The verl framework for training large language models with reinforcement learning has received support for AMD ROCm 7.0.0 and expanded scaling capabilities.

Technical context Infrastructure
Event Source: AMD Reading Time: 3 – 4 minutes

When it comes to training large language models, people usually picture server racks filled with NVIDIA GPUs. But the industry is gradually expanding, and AMD continues to develop its AI ecosystem. A recent update to the verl framework brings support for AMD ROCm 7.0.0 and the ability to scale across hundreds of GPUs simultaneously.

Что такое verl и для чего он нужен

What is verl and why is it needed?

verl is a framework for training large language models using reinforcement learning from human feedback (RLHF). This technology helps make models more useful and safer by training them on examples of what responses humans consider good versus bad.

Simply put: first, the model generates several response options to the same prompt. Then, another model (or humans) evaluates these options. Based on these evaluations, the original model gradually learns to provide higher-quality answers. This is exactly how models like ChatGPT were trained.

verl is developed by the Volcano Engine team, the cloud division of ByteDance. The framework was initially created for the company's internal needs but was later made open-source.

Что нового в последней версии verl

What's new in the latest version?

The main change is full support for the AMD ROCm 7.0.0 platform. This is significant because, until recently, almost all tools for working with large models were tailored to NVIDIA hardware. AMD is actively working to change this situation, and verl is one example of this movement.

Additionally, the new version can scale to clusters of hundreds of GPUs. This is crucial for training truly large models, where one or even ten cards are clearly not enough. The framework supports distributing computations across nodes and is optimized for such environments.

verl also integrates support for vLLM, a popular engine for fast inference of large models. This helps speed up the response generation process during the training phase, ultimately reducing the time and cost of the entire procedure.

Почему это важно для индустрии ИИ

Why is this important for the industry?

Reinforcement learning is not just a buzzword. It's one of the key methods that transforms a raw language model into a useful tool. Without this stage, a model might generate text that is technically correct but useless or even harmful.

But there's a catch: this type of training requires immense computational resources. The model must repeatedly generate responses, receive evaluations, update its parameters, and repeat the cycle over and over. All of this happens with large volumes of data and involves models containing billions of parameters.

That's why having working tools on alternative platforms is a major step forward. It gives teams more hardware options, reduces dependency on a single manufacturer, and potentially makes training more accessible.

Дальнейшие перспективы развития AMD AI

What's next?

AMD continues to develop its AI ecosystem. ROCm 7.0.0 is the seventh major version of their software platform, and tool support grows with each iteration. verl is just one example of how third-party developers are adapting their solutions for this platform.

For those working with large models, this means more options are becoming available. You can use servers with AMD GPUs if they are more cost-effective or have better availability. You can experiment with different configurations and compare performance.

Of course, NVIDIA's ecosystem is still more mature and widely supported. But the situation is gradually changing, and verl with its ROCm 7.0.0 support is another step in this direction.

Original Title: Reinforcement Learning from Human Feedback on AMD GPUs with verl and ROCm 7.0.0 – ROCm Blogs
Publication Date: Feb 12, 2026
AMD www.amd.com An international company manufacturing processors and computing accelerators for AI workloads.
Previous Article MiniMax M2.5: Open-Source Models Catch Up to Claude Sonnet Next Article How AMD and Qwen Optimized MI300X GPUs for Peak Performance

From Source to Analysis

How This Text Was Created

This material is not a direct retelling of the original publication. First, the news item itself was selected as an event important for understanding AI development. Then a processing framework was set: what needs clarification, what context to add, and where to place emphasis. This allowed us to turn a single announcement or update into a coherent and meaningful analysis.

Neural Networks Involved in the Process

We openly show which models were used at different stages of processing. Each performed its own role — analyzing the source, rewriting, fact-checking, and visual interpretation. This approach maintains transparency and clearly demonstrates how technologies participated in creating the material.

1.
Claude Sonnet 4.5 Anthropic Analyzing the Original Publication and Writing the Text The neural network studies the original material and generates a coherent text

1. Analyzing the Original Publication and Writing the Text

The neural network studies the original material and generates a coherent text

Claude Sonnet 4.5 Anthropic
2.
Gemini 2.5 Pro Google DeepMind step.translate-en.title

2. step.translate-en.title

Gemini 2.5 Pro Google DeepMind
3.
Gemini 2.5 Flash Google DeepMind Text Review and Editing Correction of errors, inaccuracies, and ambiguous phrasing

3. Text Review and Editing

Correction of errors, inaccuracies, and ambiguous phrasing

Gemini 2.5 Flash Google DeepMind
4.
DeepSeek-V3.2 DeepSeek Preparing the Illustration Description Generating a textual prompt for the visual model

4. Preparing the Illustration Description

Generating a textual prompt for the visual model

DeepSeek-V3.2 DeepSeek
5.
FLUX.2 Pro Black Forest Labs Creating the Illustration Generating an image based on the prepared prompt

5. Creating the Illustration

Generating an image based on the prepared prompt

FLUX.2 Pro Black Forest Labs

Related Publications

You May Also Like

Explore Other Events

Events are only part of the bigger picture. These materials help you see more broadly: the context, the consequences, and the ideas behind the news.

AI: Events

Perplexity Shows How to Train Trillion-Parameter Models on AWS

Technical context Infrastructure

The Perplexity team has adapted a framework for training ultra-large neural networks for Amazon's cloud infrastructure. This allowed them to eliminate the rigid dependency on proprietary NVIDIA hardware and utilize standard networking solutions.

Perplexity AIresearch.perplexity.ai Feb 7, 2026

Don’t miss a single experiment!

Subscribe to our Telegram channel —
we regularly post announcements of new books, articles, and interviews.

Subscribe