Machine Learning Engineer
Journee
Job Summary
Journee is building Ojin, a real-time generative AI platform for developers. As the first dedicated Machine Learning Engineer, you will be crucial in shaping this platform. Your role involves deploying, optimizing, and maintaining real-time generative AI models for text, audio, and video, ensuring performance and cost-effectiveness. You will also research, test, and integrate new state-of-the-art open-source models, fine-tune them for customer needs, and collaborate on building MLOps infrastructure. This position requires close collaboration with cross-functional teams to define the product roadmap.
Must Have
- 3+ years of hands-on experience in machine learning engineering
- Proficiency in Python, modern asynchronous Python (asyncio)
- Deep experience with ML frameworks like PyTorch, TensorFlow, or JAX
- Solid understanding of MLOps principles
- Experience with tools for model serving, monitoring, and automation
- Experience with model optimization techniques (quantization, distillation, pruning)
- Professional fluency in English
Good to Have
- Experience with real-time generative models for audio or video
- Experience with AWS
- Experience with containerization (Docker) and orchestration technologies
- Experience working in a fast-paced startup environment
- Contributions to open-source ML projects
Perks & Benefits
- Flexibility to work in Berlin or remotely (CET +/- 2 hours)
- Pivotal role as the first dedicated ML expert
- High degree of autonomy and direct impact on product and technology
- Opportunity to work on cutting-edge projects with top global brands
- Strong company values: high performance, transparency, continuous learning, humbleness, and pride
Job Description
The Role
This is a full-time position, either in our Berlin office or remote (CET +/- 2h)
As our first dedicated Machine Learning Engineer, you will play a pivotal role in shaping the future of our real-time generative AI platform. This is a unique opportunity for a mid-to-senior level engineer to take ownership of our ML-powered features, from experimentation with new models to their deployment and optimization in a production environment.
You will be at the heart of our mission to make real-time generative AI accessible and affordable. You will work closely with a cross-functional team of product managers, web, and systems engineers to bring our vision to life.
What You'll Do
- Deploy and optimize models: Deploy, optimize, and maintain real-time generative AI models for text, audio and video on our inference stack. You'll be responsible for ensuring they are performant, reliable, and highly cost-effective.
- Experiment and integrate: Research, test, and integrate new state-of-the-art open-source models into our platform to expand our offerings.
- Fine-tuning: Fine-tune models to meet specific customer needs and improve user experience.
- Build and maintain ML infrastructure: Collaborate with the engineering team to build and maintain our MLOps infrastructure, including our inference serving stack and model management systems.
- Cross-functional collaboration: Work closely with product, design, and engineering to define the roadmap for our ML-powered products and features.
What We're Looking For
- Experience: You have 3+ years of hands-on experience in machine learning engineering, with a proven track record of deploying and optimizing models in production environments.
- Technical skills: You are proficient in Python and have deep experience with modern, asynchronous Python (asyncio) and ML frameworks like PyTorch, TensorFlow, or JAX.
- MLOps: You have a solid understanding of MLOps principles and experience with tools for model serving, monitoring, and automation.
- Model optimization: You have experience with model optimization techniques (e.g., quantization, distillation, pruning) for real-time performance.
- Problem solver: You are a pragmatic problem-solver, able to balance technical trade-offs with business needs.
- Collaborative spirit: You have excellent verbal and written communication skills, with the ability to convey complex technical concepts clearly and effectively. You enjoy working in a collaborative, cross-functional team environment.
- Adaptability: You are highly adaptable and thrive in dynamic, fast-paced startup environments.
- Language: Professional fluency in English, both spoken and written.
Nice-to-Haves
- Experience with real-time generative models for audio or video.
- Experience with AWS.
- Experience with containerization (Docker) and orchestration technologies.
- Experience working in a fast-paced startup environment.
- Contributions to open-source ML projects.
Why work with us
- Flexibility: Work in Berlin or remotely in a similar time zone (CET +/- 2 hours).
- Pivotal role: As the first dedicated ML expert on the team, you will have the opportunity to grow with the company.
- Autonomy & impact: High degree of autonomy and direct impact on our product and technology, working on cutting-edge projects with top global brands.
- Values: We have strong values in place: high performance, transparency, continuous learning, humbleness, and pride.