Posted 1mo ago

Senior Perception Engineer

@ Origin
Bengaluru, Karnataka, India
OnsiteFull Time
Responsibilities:Develop pipelines, Calibrate sensors, Benchmark models
Requirements Summary:3+ years in computer vision and 3D perception; strong CV fundamentals; C++ (Python optional); PyTorch/TensorRT; familiarity with localization/SLAM; able to work with real-world data.
Technical Tools Mentioned:C++, Python, PyTorch, TensorRT, ROS/ROS2, OpenCV, PCL, Nvidia Deepstream, GStreamer, Holoscan
Save
Mark Applied
Hide Job
Report & Hide
Job Description

Description
About Origin
Origin (previously 10xConstruction) is building general-purpose autonomous robots for US construction to tackle rising costs, safety risks, and labour shortages. Our modular, multi-trade platform combines purpose-built hardware with real-time site intelligence to navigate complex environments and execute tasks with precision. Trained in high-fidelity simulation and already deployed on live sites, our robots deliver 5x faster execution, 250%+ margin expansion, and significant cost savings. Join India’s most talent-dense robotics team consisting of individuals from IITs, Stanford, UCLA, etc.

About the Role
You will work on building and optimizing 3D perception systems that enable robots to understand and interact with complex real-world environments. The goal is to develop robust perception pipelines that work reliably across both simulation and real-world construction sites, ensuring accurate scene understanding, localization, and decision-making.

Key Responsibilities
3D Perception Development & Scene Understanding

  • Design, implement, and deploy realtime 3D perception pipelines leveraging LiDAR, IMU, Stereo, and RGB cameras.
  • Develop algorithms for spatial and temporal data interpretation to enable high-fidelity semantic scene understanding.
  • Optimize ego-motion estimation and localization modules to ensure seamless integration with downstream planning and control tasks.



Deep Learning

  • Train and integrate deep learning models required for semantic world understanding and surface finish classification for quality control
  • Collect and curate high-quality datasets (real and synthetic) and automate training pipelines and experiment tracking.
  • Benchmark and optimize perception models for edge devices to ensure real-time performance in resource-constrained environments



Sensor Fusion & Localization

  • Design and implement sensor fusion strategies combining visual, inertial, and spatial data
  • Architect and implement sophisticated sensor fusion strategies that integrate classical methods with deep learning approaches to convert noisy, asynchronous data from heterogeneous sensors into a unified environment representation.



Calibration

  • Develop and automate robust extrinsic and intrinsic calibration procedures (Camera, LiDAR, IMU) using target-based and/or targetless methods.
  • Design and implement algorithms for online calibration drift detection and self-healing to maintain system integrity during long-term deployments.
  • Establish rigorous quantitative metrics to objectively evaluate and certify calibration quality.



Collaboration

  • Partner with Navigation, Manipulation, and Cloud teams to ensure perception outputs are optimized for downstream path planning, grasping, and remote fleet monitoring.
  • Define, track system-level and own system level KPIs and perception metrics to identify regressions across software iterations.



Requirements

  • 3+ YoE
  • Strong fundamentals in computer vision and 3D perception
  • Proficiency in C++ (Python is a plus)
  • Familiarity with PyTorch, TensorRT
  • Basic understanding of Localisation and SLAM
  • Ability to work with real-world data and debug complex systems



Nice to Have

  • Experience with Nvidia Deepstream, GStreamer, Holoscan
  • Experience with ROS/ROS2
  • Hands-on experience with LiDAR, RGB-D cameras, or IMUs
  • Familiarity with OpenCV, PCL, or similar libraries
  • Experience working on robotics or vision-based projects