A
Contract
On-site
Cambridge, England, United Kingdom
Data Engineer
Description

Position: Data Engineer

Location: Cambridge / Luton, UK (Hybrid 2-3 days onsite in a week)

Duration: Long Term B2B Contract

Job Description:

The ideal candidate with a minimum of 5 +years of experience having strong experience working with Snowflake, DBT, Python, and AWS to deliver ETL/ELT Pipelines using different resources.

• Proficiency in Snowflake data warehouse architecture Design, build, and optimize ETL/ELT pipelines using DBT (Data Build Tool) and Snowflake.

• Experience with DBT (Data Build Tool) for data transformation and modelling. Implement data transformation workflows using DBT (core/cloud).

• Strong Python programming skills for automation and data processing. Leverage Python to create automation scripts and optimize data processing tasks.

• Proficiency in SQL performance tuning, and query optimization techniques using snowflake.

• Troubleshoot and optimize DBT models, and Snowflake performance.

• Knowledge of CI/CD, version control (Git) tools. Experience with orchestration tools such as Airflow,

• Strong analytical and problem-solving skills with an ability to work in agile development environment independently.

• Ensure data quality, reliability, and consistency across different environments.

• Collaborate with other data engineers, data analysts, and business stakeholders to understand data needs and translate them into engineering solutions.

• Certification in AWS, Snowflake, or DBT is a plus.



Requirements

Position: Data Engineer

Location: Cambridge / Luton, UK (Hybrid 2-3 days onsite in a week)

Duration: Long Term B2B Contract

Job Description:

The ideal candidate with a minimum of 5 +years of experience having strong experience working with Snowflake, DBT, Python, and AWS to deliver ETL/ELT Pipelines using different resources.

• Proficiency in Snowflake data warehouse architecture Design, build, and optimize ETL/ELT pipelines using DBT (Data Build Tool) and Snowflake.

• Experience with DBT (Data Build Tool) for data transformation and modelling. Implement data transformation workflows using DBT (core/cloud).

• Strong Python programming skills for automation and data processing. Leverage Python to create automation scripts and optimize data processing tasks.

• Proficiency in SQL performance tuning, and query optimization techniques using snowflake.

• Troubleshoot and optimize DBT models, and Snowflake performance.

• Knowledge of CI/CD, version control (Git) tools. Experience with orchestration tools such as Airflow,

• Strong analytical and problem-solving skills with an ability to work in agile development environment independently.

• Ensure data quality, reliability, and consistency across different environments.

• Collaborate with other data engineers, data analysts, and business stakeholders to understand data needs and translate them into engineering solutions.

• Certification in AWS, Snowflake, or DBT is a plus.