Data Platform Architect at Baubap
, , Mexico -
Full Time


Start Date

Immediate

Expiry Date

26 Feb, 26

Salary

0.0

Posted On

28 Nov, 25

Experience

10 year(s) or above

Remote Job

Yes

Telecommute

Yes

Sponsor Visa

No

Skills

Data Engineering, Data Architecture, AWS, SQL, Python, Pipeline Orchestration, Data Warehousing, Data Lakes, Distributed Systems, Performance Tuning, CI/CD, Mentoring, Schema Design, Optimization, Monitoring, Alerting

Industry

Financial Services

Description
The mission: Design and implement the foundational data architecture that will power our next stage of growth. This role is hybrid: highly strategic in defining long-term standards and architecture, while also hands-on in building the first pipelines, storage layers, and frameworks that make data reliable, scalable, and accessible. Unlike roles limited to pipeline maintenance, this position is responsible for the entire backbone of Baubap’s data ecosystem: migrating analytical workloads away from the monolithic architectures, standardizing schema practices, enabling distributed storage, and ensuring financial and behavioral data can be trusted at terabyte scale. This person will also be responsible for mentoring and growing the Data Engineering team, establishing a culture of quality and ownership The expected outcome: Define and implement the transition from monolithic RDS analytical workloads to a scalable data warehouse/lake. Standardize schema and database practices across the company. Build and maintain pipelines (batch and real-time) that handle terabyte-scale datasets from multiple sources (mobile, financial, behavioral). Implement monitoring, validation, and alerting to ensure reliable data delivery. Ensure financial datasets (payments, collections, reports) are reliable, reconcilable, and auditable, reducing inconsistencies. Own the AWS-based data stack (Aurora, Redshift, S3, Airflow, DMS, Glue), balancing reliability, cost, and scalability. Evaluate and introduce tools for orchestration, observability, and optimization. Mentor backend and data engineers, creating a culture of excellence in data engineering. Define architectural standards, coding practices, and documentation templates. The day to day tasks: Build and maintain ETL/ELT pipelines (batch + streaming). Optimize large-scale queries and data models for performance. Manage and partition storage solutions (S3, Redshift) for efficiency. Audit and fix bottlenecks in financial data flows. Collaborate with data scientists, analysts, and backend engineers to translate needs into infrastructure. Document architecture, schemas, and pipelines for long-term clarity. Propose and implement improvements in cost efficiency, resilience, and reliability. Why YOU should apply: 7+ years in data engineering/architecture, with proven experience at TB+ scale. Experience designing and implementing data platforms Strong expertise in AWS (Aurora RDS, Redshift, S3, Glue, DMS). Mastery of SQL and Python for data workflows. Experience with pipeline orchestration (Airflow, Step Functions, etc.). Solid knowledge of data warehousing and data lakes (partitioning, schema design, optimization). Familiarity with distributed systems and performance tuning at scale. Experience balancing strategic design and hands-on execution. Knowledge of CI/CD for data pipelines, access control, monitoring. Ability to communicate trade-offs and designs to both technical and non-technical stakeholders. Experience mentoring engineers or growing data teams. What we can offer: Being part of a multicultural, highly driven team of professionals 20 vacation days / year + 75% holiday bonus (Prima Vacacional) 1 month (proportional) of Christmas bonus (Aguinaldo) Food vouchers Health & Life insurance Competitive salary
Responsibilities
Design and implement the foundational data architecture for the company, transitioning from monolithic RDS analytical workloads to a scalable data warehouse/lake. Build and maintain pipelines that handle terabyte-scale datasets while ensuring data reliability and accessibility.
Loading...