On-Device ML Engineer, Inferencing
LM Studio is the world's most popular desktop application for running AI locally. The performance and flexibility of our inferencing engines is core to what people love about LM Studio and what makes us unique. In this role you will bring up new model support in LM Studio's inferencing libraries as well as in open source engines (e.g. llama.cpp, MLX, and more).
As an On-Device ML Engineer working on LM Studio, you are expected to be able to implement new model code across inferencing libraries, port implementations from (e.g. PyTorch / HF transformers) to other libraries (e.g. llama.cpp), or implement models based on academic papers.
Bring up new AI model architectures across LM Studio's integrated inferencing libraries and ship them to millions of users worldwide. Collaborate with experts in the open-source community to make new model support available as soon as possible. Leverage software engineering skills and excellent intuition to architect and implement extensible on-device ML software. Work with LM Studio systems and app engineers to make the latest AI technologies useful and pliable through UIs and APIs.
The expected salary range for this position is:
$150,000 - $350,000 USD
Element Labs is an AI software company based out of New York, USA. We build LM Studio, a desktop application and developer platform for building on-device AI applications. Our mission is to make local AI accessible, useful, and ubiquitous. We are looking for teammates who want to take a huge swing at this mission with us.