#rocm

2026-01-25

Towards Automated Kernel Generation in the Era of LLMs

#CUDA #Triton #ROCm #LLM

hgpu.org/?p=30511

Valentin Churavy :julia:vchuravy@fosstodon.org
2026-01-25

#JuliaCon 2026 will have a minisymposium on "Julia, GPUs, and Accelerators" hosted by Julian Samaroo, Tim Besard, and @vchuravy find out more on pretalx.com/juliacon-2026/talk and submit your talks through the Call for Proposals juliacon.org/2026/cfp/ until Feb 28th!

#julialang #cuda #rocm #opencl #gpu

The JuliaGPU community has been a strong presence at JuliaCon for many years, and continues to be a strong foundation of Julia's overall computing ecosystem. In 2026, we propose to organize a minisymposium specifically focused on the usage and programming of GPUs (and other accelerators) in Julia. There is some overlap with an HPC minisymposium, however we intend to focus our track on very GPU-specific content or low-level details that make JuliaGPU tick. Additionally, material relating to non-GPU devices (such as TPUs, APUs, IPUs, etc.) are very welcome!
2026-01-24

AMD dropped ROCm support for Radeon VII (gfx906) at v6.2.4. I wanted to run local LLMs on it anyway.

Found a community-maintained image on r/LocalLLaMA that packages ROCm 7.1 with llama.cpp for gfx906. One docker pull later, I had llama.cpp + Ollama + Qdrant + Open WebUI running on "unsupported" hardware.

Docker Hub is a library of solved problems.

Full story: bit.ly/4pTk3zf

#Docker #DockerCaptain #LocalLLM #AMD #ROCm #OpenSource #SelfHosted #MachineLearning

金のニワトリ (@gosrum)

ROCm 7.2가 공개되었습니다. (링크 포함) AMD의 Radeon Open Compute 스택 업데이트로, Radeon/Ryzen 기반의 네이티브 리눅스 호환성 관련 문서가 안내되어 있어 GPU 컴퓨팅 및 머신러닝 개발 환경에 중요한 드라이버/플랫폼 업데이트입니다.

x.com/gosrum/status/2014315158

#rocm #amd #gpu #ml

2026-01-21
If you can run FLUX and SDXL locally, you really don’t need cloud AI services anymore.
No subscriptions, no prompt logging, no rate limits, just full control.

And no, you don’t need an RTX 5090 or a Radeon AI PRO R9700 (I’d love to have one though xD).
Every image here was rendered on an AMD RX 6700 XT, a mid-range GPU from 2021, not even officially supported by ROCm.

Local models. Local compute. Local freedom.

#SDXL #StableDiffusion #LocalAI #GenerativeArt #OpenSource #FOSS #PrivacyFirst #NoCloud #AMD #RX6700XT #ROCm #Linux #AIArt #DiffusionModels #CreativeCoding #MidrangeGPU #SelfHosted #Decentralized
Alexey Skobkinskobkin@gts.skobk.in
2026-01-21

✅ Заставить полноценный FLUX1.dev работать без OOM на картинках 1024x1024
✅ Также успешно запустить квантизированный FLUX под ROCm несмотря на то, что у Invoke в образе с ROCm поломана bitsandbytes
✅ Смириться, что генерировать в нормальном разрешении и с нормальными параметрами - это меееедленно.

#log #GenAI #Invoke #FLUX #server #AMD #ROCm #troubleshooting #memory #OOM

2026-01-20

vLLM 0.13 chạy thành công trên Windows 11 với RX 7900 XT qua ROCm (TheRock), đạt ~3.4 tok/s dùng TRITON_ATTN. ROCM_ATTN crash, TRITON_ATTN ổn định hơn. Cài đặt còn thủ công, nhiều phần thiếu biên dịch → dùng fallback Python. Chi tiết trong log. #vLLM #ROCm #AI #Triton #LLM #WindowsML #vLLM0.13 #AIInference #TínhToánLocal #MôHìnhNgônNgữ

reddit.com/r/LocalLLaMA/commen

2026-01-19

Neuer Blogartikel: Ubuntu-Chroot auf Debian für volle ROCm-GPU-Beschleunigung in KI-Workflows. Inklusive Automatic1111, ComfyUI und Text-Generierung:

404lifenotfound.freeddns.org/p

#Linux #ROCm #AMD #GPU #AI #KI #StableDiffusion #ComfyUI #PyTorch #Virtualization #Chroot #Debian #Ubuntu #Developer #Tech

2026-01-18

🚀 Cập nhật ổn định ROCm + Linux cho card đồ họa Strix Halo (tháng 1/2026) đã ra mắt! Phiên bản mới cải thiện hiệu năng và tính ổn định trên môi trường Linux, hỗ trợ tốt hơn cho các tác vụ AI và compute. Đánh dấu bước tiến đáng chú ý cho người dùng Linux yêu thích GPU AMD. #ROCm #Linux #StrixHalo #CậpNhật #Tech #Gaming #AI #Ubuntu #AMD 🚀

reddit.com/r/LocalLLaMA/commen

2026-01-18

🛠️ ROCm+Linux trên AMD Strix Halo: Cấu hình ổn định tháng 1/2026 đã được công bố! 🎥 Video mới mô tả chi tiết các thiết lập hoạt động, nguyên nhân gây lỗi ban đầu và bảng tham khảo đầy đủ. 🚀 Tham khảo công cụ trên GitHub để tối ưu hiệu năng GPU. #ROCm #Linux #AMD #StrixHalo #GPU #TinCôngNghệ #LinuxVN

reddit.com/r/LocalLLaMA/commen

2026-01-14

🖥️ Hai card ASRock Radeon AI Pro R9700 trên CachyOS đang “nấu” LLM: chạy Vulkan vì ROCm không ổn, mỗi card một agent (Qwen‑3‑vl‑30b‑a3b, Devstral2‑24b, Qwen‑3‑coder). Mô hình 52 GB Qwen‑Next‑32B chạy trên 2 GPU cho ~28 token/s, nhưng thiếu P2P gây bottleneck. Đã chuyển sang các agent quantized riêng để tăng năng suất, kết hợp qdrant, Neo4j, Postgres cho bộ nhớ dài. #ASRock #RadeonAI #CachyOS #GPU #LLM #Vulkan #ROCm #AI #CôngNghệ #TinTứcViet

reddit.com/r/LocalLLaMA/commen

ℒӱḏɩę 💾☮∞🎶♲☀🔋Lydie@tech.lgbt
2026-01-12

Local "AI" is a fun hobby for me. I made my best "AI slop" yet with my solar powered 7900XTX. 4kWh of sun power used. No LMI communities near data centers were harmed.

#AI #localai #AMD #rocm #COMFYUI

With the cluster and storage down for maintenance this week, I thought about doing some #pytorch development locally an my laptop. No dedicated GPU, but a recent #intel Core Ultra 2 CPU. So first I went for the CPU-only packages.

Only yesterday I realized that the Windows task manager shows something called #npu in its charts. Fair enough, let's see whether pytorch can use it ...

I have to say, the #xpu pytorch integration has gone a long way. In my case, I could use it as a drop-in replacement for #cuda / #rocm pytorch, yet with the obvious performance differences between HPC GPUs and a laptop iGPU.

docs.pytorch.org/docs/stable/n

Always nice to have alternatives. Especially for basic testing I don't need big (expensive) GPUs.

Benjamin Carr, Ph.D. 👨🏻‍💻🧬BenjaminHCCarr@hachyderm.io
2026-01-07

#AMD Just Showed Off Its Own DGX Spark Rival, The Ryzen #AIHalo: Full #ROCm Support, Day-0 Support For Leading #AI Models, Available In Q2
AMD Ryzen AI MAX family, codenamed #StrixHalo, has seen adoption, from laptops to handhelds and Mini PC; every consumer PC segment. High-performance and premium SoCs offer amazing performance thanks to #Zen5 #CPU, #RDNA 3.5 #GPU, & #XDNA 2 #NPU architectures. AMD developed its own Mini PC featuring these chips, called the #RyzenAIHalo.
wccftech.com/amd-showed-off-it

2025-12-27

If you do #LLM on #AMD #Rocm platforms, this really speeds up the processing - tested this on Whisper

export TORCH_ROCM_AOTRITON_ENABLE_EXPERIMENTAL=1

https://rocm.docs.amd.com/projects/radeon-ryzen/en/latest/docs/install/installryz/native_linux/install-pytorch.html

2025-12-20

AMD Radeon AI PRO R9700: Kết quả benchmarks ROCm & Vulkan với llama.cpp. Sử dụng CPU Ryzen 7 5800X, RAM 64GB, ROCm 7.1.1. ROCm nhanh hơn trong xử lý đầu vào (prompt), Vulkan hiệu quả hơn ở token generation. Một số mô hình như gpt-oss 20B MXFP4 MoE đạt ~18K output words (Vulkan 32 batch). Ghi chú: Hiệu suất giảm khi tăng context dài, ROCm ổn định hơn. #AMD #AI #ROCm #Vulkan #llamaCPP #Benchmark #CôngNghệViSửLiệu #ThửNghiệmMáyTính #AMD_R9700 #GPU_Amsterdam #TríTuệNhânTạo

reddit.com/r/

Benjamin Carr, Ph.D. 👨🏻‍💻🧬BenjaminHCCarr@hachyderm.io
2025-12-17

#ZLUDA For #CUDA On Non-NVIDIA GPUs Enables #AMD #ROCm 7 Support
ZLUDA is working on bringing CUDA to non-NVIDIA GPUs. While there were prior versions focused originally on Intel GPUs and then for a while AMD-financed work on Radeon/ROCm support, the current take is on being a multi-vendor CUDA implementation and with a special focus on getting CUDA #AI workloads up and running.
phoronix.com/news/ZLUDA-ROCm-7

2025-12-16

Cần hỗ trợ chạy LLAMA.cpp trên hệ thống Arch với GPU AMD. Không có binary sẵn và cài ROCm gặp khó khăn. Ai có kinh nghiệm hỗ trợ! #LLAMAcpp #ArchLinux #AMD #ROCM #CầnHỗTrợ #GPU

reddit.com/r/LocalLLaMA/commen

Client Info

Server: https://mastodon.social
Version: 2025.07
Repository: https://github.com/cyevgeniy/lmst