mlx-omni-server

MLX Omni Server is a local inference server powered by Apple's MLX framework, specifically designed for Apple Silicon (M-series) chips. It implements OpenAI-compatible API endpoints, enabling seamless integration with existing OpenAI SDK clients while leveraging the power of local ML inference.

Python v0.5.2 📦 2.0K/mo
674 stars  ·  83 forks  ·  Updated  ·  MIT

Install

pip install mlx-omni-server
plain text: /sdks/mlx-omni-server/install.txt
View on GitHub
📄 Documentation
function-callinggenaimlxopenaiopenai-apistructured-outputstttoolstts

Version history All releases ↗

## What's Changed * Logger now obeys `--log-level` command line arg. Added CORS support by @aperaham in https://github.com/madroidmaq/mlx-omni-server…

**Full Changelog**: https://github.com/madroidmaq/mlx-omni-server/compare/v0.5.0...v0.5.1

## What's Changed * Support Anthropic API by @madroidmaq in https://github.com/madroidmaq/mlx-omni-server/pull/66 **Full Changelog**: https://gi…

**Full Changelog**: https://github.com/madroidmaq/mlx-omni-server/compare/v0.4.8...v0.4.9

**Full Changelog**: https://github.com/madroidmaq/mlx-omni-server/compare/v0.4.6...v0.4.8

**Full Changelog**: https://github.com/madroidmaq/mlx-omni-server/compare/v0.4.5...v0.4.6

**Full Changelog**: https://github.com/madroidmaq/mlx-omni-server/compare/v0.4.4...v0.4.5

## What's Changed * Support speculative-decoding/draft-model by @madroidmaq in https://github.com/madroidmaq/mlx-omni-server/pull/52 * Refine model …

Quality signals

Score
Stars674
Forks83
Last updated
LicenseMIT

Supported APIs (2)

apple.comapple.com

Language

Python
Browse all Python SDKs →
📦 Package statsPyPI ↗
Monthly downloads2.0K