Cloud Data Engineer-AWS (Remote, Canada)
We are looking for a Senior Data Engineer to join our team. We are seeking an experienced Senior Data Engineer with a strong background in AWS DevOps and data engineering to join our team. In this role, you will manage and optimize our data infrastructure, focusing on both data engineering and DevOps responsibilities. A key aspect of this role involves deploying machine learning models to AWS using SageMaker, so expertise with AWS and SageMaker is essential. Experience with Snowflake is highly desirable, as our data environment is built around Snowflake for analytics and data warehousing.
Responsibilities:
- Design, develop, and maintain ETL pipelines to ensure reliable data flow and high-quality data for analytics and reporting.
- Build and optimize data models, implementing best practices to handle large volumes of data efficiently in Snowflake.
- Create and maintain complex SQL queries and transformations for data processing and analytics.
- Conduct orchestration and scheduling through Apache Airflow.
- Document data pipelines, architecture, and processes, maintaining clear and updated technical documentation.
- Design, develop, and maintain ETL pipelines to ensure reliable data flow and high-quality data for analytics and reporting.
- Build and optimize data models, implementing best practices to handle large volumes of data efficiently in Snowflake.
- Create and maintain complex SQL queries and transformations for data processing and analytics.
- Conduct orchestration and scheduling through Apache Airflow.
- Document data pipelines, architecture, and processes, maintaining clear and updated technical documentation.
- Architect, build, and maintain data science data and models infrastructure on AWS, focusing on scalability, performance, and cost-efficiency.
- Collaborate with Data Scientists to deploy machine learning models on AWS SageMaker, optimizing model performance and ensuring secure deployments.
- Automate deployment and monitoring of ML models using CI/CD pipelines and infrastructure-as-code (IaC) tools such as Terraform or AWS CloudFormation.
- AWS specific tasks (EC2, S3, RDS, VPC, CloudFormation, AutoScaling, CodePipeline, CodeBuild, CodeDeploy, ECS/EKS, cost management, etc.)
- Set up and manage monitoring solutions (e.g., CloudWatch) to ensure data pipelines and deployed models are operating effectively.
Qualifications:
- Bachelor’s or Master’s degree in Computer Science, Data Engineering, or a related field.
- 5+ years of experience in Data Engineering with at least 3+ years working in AWS environments.
- Strong knowledge of AWS services, specifically SageMaker, Lambda, Glue, and Redshift.
- Hands-on experience deploying machine learning models in AWS SageMaker.
- Proficiency in DevOps practices, including CI/CD pipelines, containerization (Docker, ECS, EKS), and infrastructure-as-code (IaC) tools like Terraform or CloudFormation.
- Advanced SQL skills and experience in building and maintaining complex ETL workflows.
- Proficiency in Python, with additional skills in Java or Scala
- Practical experience with Airflow for DAG management and data orchestration.
- Proficient in version control (GIT) and containerized deployment with Docker and managed services such as AWS Fargate, ECS, or EKS.
- Effective communication, Result oriented approach.
Compensation range for candidates in Canada, the specific range should be determined prior to posting the role.
Pay range for role if based in Canada.
$120,000 - $170,000 CAD
Apply for this job
*
indicates a required field