AMD is excited to release one-step diffusion models that demonstrate the readiness of AMD Instinctâ„¢ MI250 accelerators for model training and further research. The models are designed to provide performance comparable to traditional full-step diffusion models while maintaining the efficiency needed for training on data center systems or deployment on edge devices, such as AI-enabled PCs and laptops.
Full technical blog AMD Nitro Diffusion: One-Step Text-to-Image Generation Models
ROCm 6.3 marks a significant milestone for the AMD open-source platform, introducing advanced tools and optimizations to elevate AI, ML, and HPC workloads on AMD Instinct GPU accelerators. ROCm 6.3 is engineered to empower a wide range of customers—from innovative AI startups to HPC-driven industries—by enhancing developer productivity
This blog delves into the standout features of this release, including seamless SGLang integration for accelerated AI inferencing, a re-engineered FlashAttention-2 for optimized AI training and inference, the introduction of multi-node Fast Fourier Transform (FFT) to revolutionize HPC workflows and more. Explore these exciting updates and more as ROCm 6.3 continues to drive innovation across industries.
more
In recent years, the rapid development of artificial intelligence technology, especially the progress in large language models (LLMs), has garnered significant attention and discussion. From the emergence of ChatGPT to subsequent models like GPT-4 and Llama, these language models have demonstrated remarkable capabilities in natural language processing, generation, understanding and reasoning. Continuing AMD tradition of open-sourcing models and code to help the community advance together, we are excited to release our first series of fully open 1 billion parameter language models, AMD OLMo.
To dive deeper into the three stages of training, and AMD OLMo model results, please reference the full article here: Introducing the First AMD 1B Language Models: AMD OLMo
In the ever-evolving landscape of artificial intelligence, large language models (LLMs) like GPT-4 and Llama have garnered significant attention for their impressive capabilities in natural language processing and generation. However, small language models (SLMs) are emerging as an essential counterpart in the AI model community offering a unique advantage for specific use cases. AMD is excited to release its very first small language model, AMD-135M with Speculative Decoding. This work demonstrates the commitment to an open approach to AI which will lead to more inclusive, ethical, and innovative technological progress, helping ensure that its benefits are more widely shared, and its challenges more collaboratively addressed.
moreRunning large language models (LLMs) locally on AMD systems has become more accessible, thanks to Ollama. This guide will focus on the latest Llama 3.2 model, published by Meta on Sep 25th 2024, Meta's Llama 3.2 goes small and multimodal with 1B, 3B, 11B and 90B models. Here’s how you can run these models on various AMD hardware configurations and a step-by-step installation guide for Ollama on both Linux and Windows Operating Systems on Radeon GPUs.
more
Calling all AI enthusiasts, data scientists, developers, and high-performance computing professionals! The latest release of AMD ROCm 6.2 is here, packed with innovative enhancements designed to elevate your computational tasks.
more
AMD Instinct MI300X accelerator is a strong choice for deploying large language models due to its ability to address cost, performance, and availability challenges. By providing a reliable ,efficient alternative, and a strong ROCm ecosystem AMD helps businesses maintain robust AI operations and achieve optimal performance.
moreThe PyTorch deep-learning framework is a foundational technology for machine learning and artificial intelligence. Discover our perspective on the evolution of this enormously successful open-source project and learn about some of the factors behind the unprecedented success of PyTorch.
moreDiscover AMD AI solutions with day-0 support for Meta's Llama 3.1, delivering efficiency from cloud to edge
more