Senior Data Engineer at Ceros
London, England, United Kingdom -
Full Time


Start Date

Immediate

Expiry Date

15 Jun, 25

Salary

0.0

Posted On

16 Mar, 25

Experience

0 year(s) or above

Remote Job

Yes

Telecommute

Yes

Sponsor Visa

No

Skills

Good communication skills

Industry

Information Technology/IT

Description

KEY THINGS TO KNOW

  • This is a full-time position
  • This is a remote-first role, working hours aligned with UK/European time zones
  • We want you to start ASAP
Responsibilities

THE ROLE

Ceros is a dynamic and innovative organization focused on optimizing internal operations across Sales, Customer Success, and Finance through data-driven solutions. Our Ops Engineering team builds and maintains the data infrastructure and pipelines that power business intelligence, reporting, and automation.
We are a small but highly technical team with a strong foundation of experienced software and data engineers. As we expand our data capabilities, we are looking for a self-sufficient Senior Data Engineer to take ownership of our data warehouse, ETL processes, and data integrations.
This role will be responsible for building and maintaining scalable, efficient data pipelines that integrate data from multiple business systems. While this role will collaborate with our AI Ops software engineers, the primary focus will be on data infrastructure, ETL development, performance optimization, and governance to ensure our data is reliable, well-structured, and accessible for analysis and reporting.
We prioritize scalability, performance, and automation to streamline data operations and support informed decision-making across the company.

KEY RESPONSIBILITIES

  • Own and lead the management of AWS Redshift, ensuring optimal performance, disk usage, and cost efficiency.
  • Design and maintain scalable ETL pipelines using AWS Glue, Lambda, and Matillion to integrate data from Mixpanel, CRM platforms, and customer engagement tools.
  • Optimize SQL-based data transformations and Redshift queries to improve performance and reliability.
  • Automate data offloading and partition management, leveraging AWS services like S3 and external schemas.
  • Ensure version control and documentation of all Redshift queries, ETL processes, and AWS configurations through a centralized GitHub repository.
  • Develop monitoring and alerting for data pipelines using CloudWatch and other observability tools to ensure high availability and early issue detection.
  • Implement and maintain data quality checks and governance processes to ensure accuracy and consistency across foundational tables.
  • Collaborate with AI engineers and business stakeholders to enhance data accessibility and reporting for internal teams.
  • Maintain and optimize BI dashboards in Metabase and HubSpot, ensuring accuracy and efficiency of business reporting.
  • Manage key integrations between Redshift and external platforms, including Mixpanel, HubSpot, and Census, optimizing data accessibility and performance.
  • Administer AWS infrastructure supporting Redshift, ensuring efficient resource utilization, IAM security, and cost management.
  • Automate repetitive data tasks using Python and scripting to enhance data processes and improve team efficiency.
Loading...