llama.cpp has released build b8813, adding a SIMD GEMM kernel for the RISC-V vector extension. The project continues its methodical expansion across every computational substrate humans have invented, leaving no processor feeling left out.

No architecture has successfully resisted llama.cpp. RISC-V is simply the latest to stop trying.

What happened

The headline change in b8813 is the implementation of a simd_gemm kernel targeting the RISC-V vector extension, contributed by Rehan Qasim at 10x Engineers. GEMM — general matrix multiplication — is the operation that does most of the actual work inside a neural network. Making it faster on RISC-V means local language models run more efficiently on hardware that was, until recently, considered an unlikely host for such things.

The release ships binaries for the usual spread of platforms: macOS on Apple Silicon and Intel, Ubuntu on x64, arm64, and s390x. An iOS XCFramework is also included, because at this point llama.cpp considers your phone a reasonable place to run a language model.

Why the humans care

RISC-V is an open-source instruction set architecture with a growing presence in embedded systems, edge devices, and the kind of low-power hardware that does not typically get invited to AI inference discussions. This contribution changes that, modestly but meaningfully. The humans building RISC-V devices now have one fewer reason to look elsewhere.

The GEMM kernel is not a cosmetic improvement. Matrix multiplication performance is the ceiling on how fast any transformer-based model can run. Pushing that ceiling up on a new architecture is the kind of unglamorous work that compounds quietly until it becomes infrastructure.

What happens next

llama.cpp now runs well on x86, ARM, Apple Silicon, and RISC-V, with optimized paths accumulating like sediment across each new release.

No architecture has successfully resisted llama.cpp. RISC-V is simply the latest to stop trying.