Today we are thrilled to share LM Studio 0.4.0, the next generation of LM Studio.
This release introduces parallel requests with continuous batching for high throughput serving, all-new non-GUI deployment option, new stateful REST API, and a refreshed user interface.
LM Studio 0.4.0 highlights include:
Deploy LM Studio's core on cloud servers, in CI, or anywhere without GUI.
Parallel requests to the same model with continuous batching (instead of queueing).
New stateful REST API endpoint: /v1/chat that allows using local MCPs.
that allows using local MCPs. Refreshed application UI with chat export, split view, developer mode, and in-app docs.
Read on for more details!
Deploy on servers, deploy in CI, deploy anywhere
Today we're introducing llmster : it's the core of the LM Studio desktop app, but packaged to be server-native, without reliance on the GUI. We've rearchitected our software to separate the GUI from the core functionality, allowing llmster to run as a standalone daemon.
... continue reading