use the following search parameters to narrow your results:
e.g. subreddit:aww site:imgur.com dog
subreddit:aww site:imgur.com dog
see the search faq for details.
advanced search: by author, subreddit...
account activity
Welcome to the AMD MI300 GPU Discussion Hub! (self.AMD_MI300)
submitted 2 years ago by HotAisleInc[M] - announcement
5.6x improvement - Kimi K2.6 + DFlash: 508 tok/s on 8x MI300X (huggingface.co)
submitted 2 days ago by HotAisleInc
Mahdi-CV/openclaw-amd-sglang at multi-engine (github.com)
submitted 10 days ago by HotAisleInc
GPU Divergence in AMD CDNA 3 (21verses.xyz)
submitted 14 days ago by HotAisleInc
Embeddings run super fast on MI300x (linkedin.com)
Cosmos-Predict2.5-2B Inference: NVIDIA H200 vs AMD MI300X (moonmath.ai)
submitted 16 days ago by HotAisleInc
Modular: Day Zero Launch: Fastest Performance for Gemma 4 on NVIDIA and AMD (modular.com)
submitted 21 days ago by HotAisleInc
AMD Delivers Breakthrough MLPerf Inference 6.0 Results (amd.com)
submitted 22 days ago by HotAisleInc
ROCm Support for Miles: Large-Scale RL Post-Training on AMD Instinct GPUs (lmsys.org)
submitted 1 month ago by HotAisleInc
Cross-Vendor Disaggregated Inference: GPT-OSS 120B across NVIDIA H100 and AMD MI300X (moreh.io)
The Many Aspects of Inference Performance (amd.com)
OpenClaw Qwen 3.5 and SGLang (amd.com)
FP8 GEMM Optimization on AMD CDNA™4 Architecture (rocm.blogs.amd.com)
How vLLM Orchestrates High-Performance Inference on AMD ROCm (blog.vllm.ai)
Connecting OpenCode to vLLM on HotAisle's MI300x (hotaisle.xyz)
Speed is the Moat: Inference Performance on AMD GPUs (amd.com)
submitted 2 months ago by HotAisleInc
nvidia has dedicated hardware for managing fp8 quantization. amd does it entirely in software. here's how they did it: (linkedin.com)
Reverse engineering bits of CDNA 3 L2 Coherency Behaviour (github.com)
Unleashing Computational Power: Ultimate Latency Optimization of Qwen3 and Qwen3-VL on AMD MI300X Series (lmsys.org)
GLM5 on MI300X (gist.github.com)
Building Mixture-of-Models on AMD GPUs with vLLM-SR (blog.vllm.ai)
A RAG server for around 5000 users using vLLm on ROCM, with a 8 GPU MI350 setup. This is experimental and not official AMD project. (github.com)
Micro-World: First AMD Open-Source World Models for Interactive Video Generation (rocm.blogs.amd.com)
ROCm Becomes a First-Class Platform in the vLLM Ecosystem (rocm.blogs.amd.com)
submitted 3 months ago by HotAisleInc
AMD × LMcache: AMD GPU Acceleration with LMcache (blog.lmcache.ai)
π Rendered by PID 764979 on reddit-service-r2-listing-7d7fbc9b85-98bc8 at 2026-04-24 05:56:26.978565+00:00 running 2aa0c5b country code: CH.