About the role
- We are looking for a Senior Camera & Vision Engineer to lead the development of camera-based perception in our robotics systems. This is a hands-on role for someone who deeply understands camera hardware, ISP tuning, and modern computer vision and ML techniques—and enjoys turning raw sensor data into reliable, real-time perception outputs.
You will own camera selection and configuration, tune imaging pipelines, develop and integrate vision algorithms, and work closely with perception, ML, and software teams to build a robust, scalable perception stack. This role sits at the intersection of hardware, algorithms, and software, with a strong focus on modern ML-based vision systems.
What you'll do
- Own camera-based perception pipelines from sensor selection to production deployment
- Evaluate and select camera sensors, modules, and optics based on system and application requirements
- Tune and optimize ISP pipelines (image quality, color, HDR, noise, latency, synchronization)
- Develop and integrate computer vision and ML-based algorithms into the perception stack
- Work with perception and software teams to integrate cameras with other sensors (LiDAR, IMU, radar, etc.)
- Design and deploy real-time vision systems on embedded and edge compute platforms
- Implement and optimize models for tasks such as object detection, tracking, and segmentation; depth estimation from cameras (mono, stereo, multi-view); optical flow and motion estimation
- Stay current with state-of-the-art vision models, tools, and frameworks, and apply them pragmatically to products
- Debug perception issues across the full stack—from sensor and ISP to algorithms and ML inference
Required Qualifications
- Bachelor's degree or higher in Computer Science, Electrical Engineering, Robotics, or a related field
- 7+ years of experience in camera, computer vision, or perception engineering
- Strong experience with camera sensors, optics, and ISP tuning
- Deep understanding of computer vision and ML fundamentals
- Hands-on experience developing and deploying ML-based vision models
- Strong software engineering skills in C++ and/or Python
- Experience integrating perception algorithms into production systems
- Ability to collaborate effectively with hardware, systems, and software teams
Preferred Qualifications
- Experience in robotics, autonomous systems, AR/VR, or advanced driver assistance systems (ADAS)
- Familiarity with modern vision and ML frameworks (PyTorch, TensorFlow, ONNX, OpenCV, etc.)
- Experience with embedded or edge AI deployment (NVIDIA Jetson, CUDA, TensorRT, etc.)
- Background in depth estimation, stereo vision, SLAM, or visual-inertial systems
- Experience with optical flow and motion estimation techniques
- Track record of shipping camera-based perception systems to production
Benefits (subject to location and local regulations)
- Competitive salary and equity
- Comprehensive medical, dental, and vision coverage
- 401(k) retirement plan
- Flexible vacation and time-off policy
- Collaborative, fast-paced, and inclusive work environment
- Opportunity to work on cutting-edge technologies with a highly cross-functional team
About Lyte
Lyte builds perception systems for Physical AI operating in complex environments. The company combines custom silicon, integrated sensors, and software into a unified platform for robotics, mobility, and next-generation automation. Lyte is headquartered in Sunnyvale, California, with a global presence. For more information, visit www.lyte.ai
If you’re excited about building impactful technology in a dynamic, hands-on environment, we’d love to hear from you!
The pay range for this role is:
150,000 - 300,000 USD per year (Bay Area, CA)