About the Company:
Chorus transforms the way organizations make, move and manage goods with advanced sensor technology providing granular, parcel-level visibility. Leveraging proprietary sensor technology, AI-powered algorithms, and existing infrastructure, Chorus enables asset orchestration at an unprecedented scale. Chorus began as a project at X, The Moonshot Factory, an innovation lab that is part of Alphabet, Google's parent company. Visit the Chorus website for more details.
About the Role:
We're seeking a Senior Data Engineer with proven experience designing and implementing scalable data platforms that support analytics, reporting, and machine learning workloads. You'll build cloud-based infrastructure that powers our core products, requiring both technical depth and architectural vision.
You'll establish systems and processes that ensure data quality and reliability while making forward-thinking decisions that accommodate rapid growth. Though we build on the cloud, you should understand the principles underneath the layers of abstraction to generate performant designs.
Working within our Data/ML team, you'll collaborate closely with ML/Data Engineers while serving as a bridge between our Platform Engineering team and Reporting & Analytics team.
As part of our growing startup, you'll thrive in a fast-paced environment, balance ideal solutions with practical implementation, and contribute to our collaborative culture.
What You'll Do:
- Design, implement, and maintain ETL pipelines that efficiently transfer data from upstream sources to our analytics data warehouse
- Oversee our analytics data warehouse to deliver high-performance, reliable data products while establishing standards that ensure quality and usability
- Partner with ML Engineers to optimize data flows for models, including improvements to Feature Stores and training pipelines
- Develop production services that transform ML predictions into fully functional products
Required Skills:
- Keen Ability to work cross functionally and a true desire to work at a startup.
- Demonstrated ability to own complex data-intensive systems end-to-end
- Comfortable programming in both Java and Python
- Experience building production ETL pipelines
- Familiarity with a variety of databases (eg, SQL/NoSQL/Columnar)
- Familiarity with a variety of data processing platforms (eg Kafka, Spark)
- Ability to work Hybrid in either Mountain View, CA or Boulder CO.
It Would Be Nice if you also have some of:
- Experience building on public clouds (we use GCP)
- Experience with underlying infrastructure (k8s, HDFS, etc.)
- Familiarity with streaming and complex data pipeline architectures.
- Familiarity with ML-based products and infrastructure
- Management skills/experience