About us:
Aeva’s mission is to bring the next wave of perception to a broad range of applications from automated driving to industrial robotics, consumer electronics, consumer health, security, and beyond. Aeva is transforming autonomy with its groundbreaking sensing and perception technology that integrates all key LiDAR components onto a silicon photonics chip in a compact module. Aeva 4D LiDAR sensors uniquely detect instant velocity in addition to 3D position, allowing autonomous devices like vehicles and robots to make more intelligent and safe decisions.
Role Overview:
We're looking for an engineer who can own data collection, scalable data systems, and sensor data workflows. You'll work cross-functionally with perception, operations, and labeling teams to build fast, reliable, and production-grade data pipelines.
What you will be doing
Develop Python and C++ infrastructure to ingest, validate, and organize camera and LiDAR data from our collection vehicles — file handling, data verification, metadata extraction, and database writes.Run camera, lidar, gps calibration workflows on the data collection vehicles. Debug hardware/software integration issues on data collection rigs including sensor connectivity, logging failures, storage limits, and network dropoutsBuild and maintain ETL pipelines that move high-volume sensor data from vehicles to internal storage and annotation vendorsCollaborate with external annotation and labeling vendors: package data for delivery, validate returned labels, and flag quality issuesCollaborate with the operations team to plan, support, and execute data collection runs
What You Have
Bachelor's degree in Computer Science, Electrical Engineering, or a related field; or equivalent practical experience3+ years of industry experience in software development, with production Python and C++ (2+ years with a Master's)Strong fundamentals in software engineering — data structures, algorithms, system design, and writing code that is readable, testable, and maintainableHands-on experience designing and operating data pipelines at scale, including ownership through incidents and iterationsFamiliarity with robotics or autonomous systems data — sensor logs, binary formats, or perception datasets is a plusExperience with databases as part of a production data workflow, including schema design and data integrity considerationsAbility to work independently and cross-functionally in a fast-moving environment with ambiguous requirements