Skip to content
jundot avatar

Trending repo

Claude Code & Cursor rules for omlx

by @jundot · 13,651 stars · Python

View on GitHub →

About omlx

LLM inference server with continuous batching & SSD caching for Apple Silicon — managed from the macOS menu bar

oMLX LLM inference, optimized for your Mac Continuous batching and tiered KV caching, managed directly from your menu bar. junkim.dot@gmail.com · https://omlx.ai/me Install · Quickstart · Features · Models · CLI Configuration · Benchmarks · oMLX.ai English · 中文 · 한국어 · 日本語 --- *Every LLM server I tried made me choose between convenience and control. I wanted to pin everyday models in memory, auto-swap heavier ones on demand, set context limits - and manage it all from a menu bar.* *oMLX persists KV cache across a hot in-memory tier and cold SSD tier - even when context changes mid-conversation, all past context stays cached and reusable across requests, making local LLMs practical for real coding work with tools like Claude Code. That's why I built it.* ## Install ### macOS App Download t

Topics

apple-siliconinference-serverllmmacosmlxopenai-api

No rules target omlx yet

No published rules, MCP servers, or skills target omlx yet. If you maintain a tool that works well with this project, you can publish for free during beta.

Why this page exists

RuleSell tracks the AI-coding ecosystem so you don't have to. When a repo like omlx picks up momentum, we surface the Claude Code skills, Cursor rules, MCP servers, and agent configs that target it — with real author attribution, SPDX license badges, and quality scores. Every listing ships with copy-paste install for each environment.