About the Project
Weβre building a next-generation environmental intelligence platform for the global food system. Our mission is to accelerate the planetβs progress toward Net Zero carbon emissions.
Responsibilities
-
Design and implement scalable, maintainable data architectures aligned with business objectives
-
Build robust, reusable data pipelines and systems with a long-term view on performance, reliability, and support
-
Assemble and manage large, complex, structured, and unstructured datasets to meet functional and non-functional requirements
-
Develop and maintain ETL/ELT pipelines using Apache Spark, Databricks, Azure Data Factory, and related technologies
-
Optimize data pipelines, storage, and query performance (including Spark and Delta Tables optimizations)
-
Collaborate with Product Owners and cross-functional teams in an agile environment
-
Partner with Analytics, Data Science, and Operations teams to enable advanced analytics and data-driven decision-making
-
Define and promote best practices, reusable patterns, and reference architectures across the data platform
-
Ensure data security, privacy, and segregation across regions, data centers, and Azure environments
-
Identify and implement process improvements, including automation and infrastructure redesign for scale and performance
-
Mentor and support other engineers, contributing to technical growth across the team
Requirements
-
6+ years of experience in data engineering, analytics engineering, or a closely related role
-
6+ years of hands-on experience with big data systems, relational databases, or data mining
-
4+ years of experience with data engineering languages such as Python, PySpark, SQL, Scala, or R
-
2+ years of experience with cloud platforms and/or DevOps practices
-
Strong experience with Apache Spark and Databricks
-
Solid understanding of relational and NoSQL databases (e.g., MySQL, SQL Server)
-
Proven experience with data modeling (Star/Snowflake schemas, SCD types)
-
Experience building and orchestrating data pipelines and workflows
-
English - Upper-Intermediate or higher
Nice to Have
-
Experience with Azure cloud services
-
Familiarity with hierarchical data models or graph-based data structures
-
Hands-on experience optimizing Spark workloads and Delta Tables
-
Background in sustainability, climate tech, or supply chain data
-
Bachelorβs degree in Computer Science, Engineering, or a related technical field