Tech News
← Back to articles

Show HN: We made our own inference engine for Apple Silicon

read original related products more articles

uzu

A high-performance inference engine for AI models on Apple Silicon. Key features:

Simple, high-level API

Hybrid architecture, where layers can be computed as GPU kernels or via MPSGraph (a low-level API beneath CoreML with ANE access)

Unified model configurations, making it easy to add support for new models

Traceable computations to ensure correctness against the source-of-truth implementation

Utilizes unified memory on Apple devices

Quick Start

First, add the uzu dependency to your Cargo.toml :

[ dependencies ] uzu = { git = " https://github.com/trymirai/uzu " , branch = " main " , package = " uzu " }

... continue reading