One stop solution to your remote job hunt!

By signing up you get access to highly customizable remote jobs newsletter, An app which helps you in your job hunt by providing you all the necessary tools.

Try Worqstrap Remote Jobs commitment free for 7 days, no credit card necessary.
OR
Subscribe to our highly customizable newsletter to get remote jobs from top remote job boards delivered to your inbox.
Zep AI 24 days ago
cafulltimemachine learningsan franciscous / remote (us; ca)
Apply Now

"

Zep is building the long-term memory layer for the LLM application stack. We have a large and active open-source community and recently launched our cloud service. We are seeking an experienced ML Engineer to join our startup. As a critical member of our small, high-performance team, you will be responsible for model selection, evaluation, and performance, and the processes and tools around supporting activities.

We are a remote-first organization. Zep is funded by YC, Engineering Capital, and angels such as Guillermo Rauch (Vercel).

Responsibilities

- Full lifecycle of LLM development: model selection, finetuning, and model and prompt evaluation.

- Implement model and prompt evaluation infrastructure, process, and metrics to support rapid deployment of new product

- Supporting and optimizing low-latency inference solutions using Triton Inference Server, TensorRT-LLM, and vllm

- Collaborate with the team to develop and deploy LLM applications

- Contribute to the entire development lifecycle, from ideation to deployment and maintenance

Requirements

- 5+ years of experience in ML engineering, with 12 months or more focus on LLM applications

- Strong understanding of LLM models and LLM inference

- Expertise in model evaluation approaches and metrics

- Hands-on experience with low-latency inference technologies (e.g., Triton Inference Server, TensorRT-LLM, vllm)

- Highly proficient in Python

- Strong familiarity with Kubernetes

- Strong familiarity with AWS and Azure

- Familiarity with guided LLM output techniques

Nice to Have

- Experience with Rust or Go

- Contributions to LLM-related open-source projects

- Knowledge of advanced NLP techniques and architectures

Benefits

- Directly impact the development of the future LLM application stack

- Competitive salary and equity compensation

- Flexible work hours and remote work options

- Health, dental, and vision insurance

",