You will lead the development of end-to-end multimodal agentic AI pipelines for a novel physical device. This role involves architecting the full interaction loop—from camera-based scene understanding to cloud reasoning and action. You will bridge the gap between computer vision and large language models to create a seamless user experience on hardware.
AI Engineer at Alcor Labs
Are you tired of building chatbots? Join a high-growth Bay Area startup where you’ll build the 'brain' for a physical device. We are looking for an AI Engineer to own the end-to-end multimodal loop, from camera-based scene understanding to agentic reasoning and action. If you have experience shipping AI products in a startup environment and want to define how humans interact with hardware through vision and language, this is the role for you.
Overview
Role overview
Company
Alcor Labs
Bay Area AI hardware startup building multimodal intelligence for physical devices
Responsibilities
What you will do
- Design and implement end-to-end AI pipelines that process camera data to perform complex scene understanding and reasoning.
- Build the full interaction loop connecting hardware sensors to cloud-based LLMs and translating outputs into device actions.
- Optimize multimodal model performance for latency and accuracy to ensure real-time responsiveness on an edge-tethered device.
Candidate profile
Who this is a fit for
- Proven experience building and shipping multimodal AI features or agentic loops within a fast-moving startup environment.
- Strong background in camera-based scene understanding, computer vision, and integrating large language models for reasoning.
- Deep technical expertise in Python and modern AI frameworks, with a track record of deploying models to production devices.
What makes it remarkable
Why this role is remarkable
- Work at the cutting edge of AI hardware, moving beyond pure software to build intelligence that interacts with the physical world.
- Join a fast-paced, high-growth startup environment where your code directly dictates the capabilities of a shipped physical product.
- Drive the architectural decisions for multimodal reasoning, integrating vision and language models into a cohesive, agentic system.
Jack & Jill
How Jack & Jill work together
Meet Jack
Jack gets to know what you're great at and what you want next, then searches 15 million jobs daily and helps you discover roles at companies like this.
How does this work?
Jack's an AI agent for job searching and career coaching. He works for you.
Jill is the AI recruiter working for the company. She recruits from Jack's network.
If it's a match and the company wants to meet you, they'll make the intro. In the meantime, if you'd like, Jack will send you excellent alternatives.