Running large language models (LLMs) locally on AMD systems has become more accessible, thanks to Ollama. This guide will focus on the latest Llama 3.2 model, published by Meta on Sep 25th 2024, Meta's Llama 3.2 goes small and multimodal with 1B, 3B, 11B and 90B models. Here’s how you can run these models on various AMD hardware configurations and a step-by-step installation guide for Ollama on both Linux and Windows Operating Systems on Radeon GPUs.
Supported AMD GPUs
Ollama supports a range of AMD GPUs, enabling their product on both newer and older models. Here you can find the list of supported GPUs by Ollama:
Installation and Setup Guide for Ollama
Linux
- System Requirements:
- Ubuntu 22.04.4
- AMD GPUs with latest AMD ROCm ™ software installed
- Install ROCm 6.1.3 from below instructions
- Install Ollama via single command
- Download and run llama-3.2 model:
Windows
- System Requirements:
- Windows 10 or Higher
- Supported AMD GPUs with driver installed
- For Windows installation you can simply download and install Ollama from here:
https://ollama.com/download
Once installed, simply open PowerShell and run:
As simple as that, you are ready to chat with your local LLM.
You can find the list of all available models from Ollama here https://ollama.com/library
Conclusion
The extensive support for AMD GPUs by Ollama demonstrates the growing accessibility of running LLMs locally. From consumer-grade AMD Radeon™ RX graphics cards to high-end AMD Instinct ™ accelerators, users have a wide range of options to run models like Llama 3.2 on their own hardware. This flexible approach to enable innovative LLMs across the broad AI portfolio, allows for greater experimentation, privacy, and customization in AI applications across various sectors.