Documentation
Running LLMs Locally
User Interface
Advanced
Command Line Interface - lms
API
REST Endpoints
LM Studio SDK (TypeScript)
Run LM Studio as a service (headless)
Advanced
Starting in v0.3.5, LM Studio can be run as a service without the GUI. This is useful for running LM Studio on a server or in the background on your local machine.
Running LM Studio as a service consists of several new features intended to make it more efficient to use LM Studio as a developer tool.
To enable this, head to app settings (Cmd
/ Ctrl
+ ,
) and check the box to run the LLM server on login.
Enable the LLM server to start on machine login
When this setting is enabled, exiting the app will minimize it to the system tray, and the LLM server will continue to run in the background.
Useful when utilizing LM Studio as an LLM service with other frontends or applications.
Load models on demand
/v1/models
will return all downloaded models, not only the ones loaded into memory/v1/models
will return only the models loaded into memoryThe counterpart to loading models on-demand is unloading them from memory at the right time.
LM Studio 0.3.9 introduces TTL (unload after some time) and Auto-Evict (limit how many JIT models are loaded at one time).
Read about Idle TTL and Auto-Evict.
Your last server state will be saved and restored on app or service launch.
To achieve this programmatically, you can use the following command:
lms server start
If you haven't already, bootstrap lms
on your machine by following the instructions here.
Chat with other LM Studio developers, discuss LLMs, hardware, and more on the LM Studio Discord server.
Please report bugs and issues in the lmstudio-bug-tracker GitHub repository.