Data Engineering Analyst at Sanofi
Toronto, ON, Canada -
Full Time


Start Date

Immediate

Expiry Date

05 Feb, 25

Salary

0.0

Posted On

26 Jan, 25

Experience

4 year(s) or above

Remote Job

No

Telecommute

No

Sponsor Visa

No

Skills

Data Warehouse, Orchestration, Job Scheduling, Logging, Infrastructure, Devops, Snowflake, Containerization, Airflow, Computer Science, Kubernetes, Data Structures, Github, Query Optimization, Database Optimization, Code, Automation, Algorithms, Postgresql, Python, Etl

Industry

Information Technology/IT

Description

Reference No. R2782648
Position Title: Data Engineering Analyst
Department: Data Engineering - RWD
Location: Toronto Ontario, Hybrid 60% at-home, 40% in-office / week

TECHNICAL SKILLS:

  • 4+ years of relevant experience developing backend, integration, data pipelining, and infrastructure with relevant technologies and tools (Snowflake, AWS, Spark, Informatica/IICS or equivalent).
  • Bachelor’s degree in computer science, engineering, or similar quantitative field of study
  • Expertise in database optimization and performance improvement
  • Expertise in Python, PySpark, and Snowpark
  • Experience data warehouse and object-relational database (Snowflake and PostgreSQL) and writing efficient SQL queries
  • Experience in cloud-based data platforms (Snowflake, AWS)
  • Proficiency in developing robust, reliable APIs using Python and FastAPI Framework
  • Experience with job scheduling and orchestration (Airflow is a plus)
  • Expert in ELT and ETL & Experience working with large data sets and performance and query optimization.
  • Understanding of data structures and algorithms
  • Experience in modern testing framework (SonarQube, K6 is a plus)
  • Strong collaboration skills, willingness to work with others to ensure seamless integration of the server-side and client-side
  • Knowledge of DevOps best practices and associated tools, a plus, especially in the setup, configuration, maintenance, and troubleshooting of associated tools:
  • Containers and containerization technologies (Kubernetes, Argo, Red Hat OpenShift)
  • Infrastructure as code (Terraform)
  • Monitoring and Logging (CloudWatch, Grafana)
  • CI/CD Pipelines (JFrog Artifactory)
  • Scripting and automation (Python, GitHub, Github actions)
Responsibilities
  • Establish technical designs to meet Sanofi requirements aligned with the architectural and Data standards
  • Ownership of the entire back end of the application, including the design, implementation, testing, and troubleshooting of the core application logic, databases, data ingestion and transformation, data processing and orchestration of pipelines, APIs, CI/CD integration and other processes
  • Fine-tune and optimize queries using Snowflake platform and database techniques
  • Optimize ETL/data pipelines to balance performance, functionality, and other operational requirements
  • Assess and resolve data pipeline issues to ensure performance and timeliness of execution
  • Assist with technical solution discovery to ensure technical feasibility
  • Assist in setting up and managing CI/CD pipelines and development of automated tests
  • Developing and managing microservices using python
  • Conduct peer reviews for quality, consistency, and rigor for production-level solution
  • Design application architecture for efficient concurrent user handling, ensuring optimal performance during high usage periods
  • Promote best practices and standards for code management, automated testing, and deployments
  • Own all areas of the product lifecycle: design, development, test, deployment, operation, and support
  • Create detail documentation on Confluence to be able to support and maintain codebase and its functionality
Loading...