Data engineer at Master-Works
Riyadh, Riyadh Region, Saudi Arabia -
Full Time


Start Date

Immediate

Expiry Date

02 Jun, 26

Salary

0.0

Posted On

04 Mar, 26

Experience

5 year(s) or above

Remote Job

Yes

Telecommute

Yes

Sponsor Visa

No

Skills

Kafka, Spark, Hadoop, Teradata CLDM, Data Modeling, ETL, SQL, Stored Procedures, Spark Streaming, Dataflow, HDFS, Hive, Sqoop, FSLDM, Data Vault Modeling, PySpark

Industry

IT Services and IT Consulting

Description
Master Works is hiring an experienced Data Engineer (5+ years) in Riyadh to design and optimize large-scale real-time and batch data pipelines within the Telecom domain. Design, develop, and maintain real-time and batch data pipelines leveraging Kafka, Spark, and Hadoop components. Must have understanding of Teradata CLDM, should know how to create new Data Model or modify existing data model based on business requirement Collaborate with business analysts and data architects to translate business requirements into robust data models and ETL frameworks. Apply Relational and Dimensional modeling techniques to design databases and ensure data is organized effectively for both operational and analytical purposes. Write, debug, and optimize SQL and Stored Procedures to ensure efficient data processing. Work closely with BI, Data Science, and Campaign teams to ensure seamless data availability for analytics Work closely with Data Architects, Analysts, and Business Stakeholders to translate business requirements into database solutions. Ensure all database design and code is well-documented and follows best practices for performance and maintainability. Involves designing fact and dimension tables for reporting and analytics purposes, often in a star or snowflake schema. Develop and maintain technical documentation (data flow diagrams, source-to-target mappings, architecture documents). Ensure that the Data Dictionary is always up-to-date, capturing all changes to the database schema, including newly created or modified tables, columns, views, Perform data quality checks, validation, and ensure end-to-end data accuracy and lineage. Support and troubleshoot real-time streaming jobs and ensure high availability of data pipelines Have-Must Strong expertise in real-time data integration using Kafka, Spark Streaming, or Dataflow. Hands-on experience with Hadoop ecosystem components (HDFS, Hive, Sqoop, Spark etc.). Strong Data Modeling concepts including FSLDM, CLDM, and Dimensional / Data Vault modeling. Deep understanding of Telecom domain (BSS/OSS, CDR, usage, revenue, and campaign data). Experience building and optimizing ETL pipelines and data ingestion frameworks for structured and unstructured data. Proficiency in SQL and distributed data processing using Hive, Spark SQL, or PySpark. Good understanding of data governance, data quality, and lineage frameworks. Strong analytical and problem-solving skills. Excellent communication and collaboration skills with cross-functional teams. Experience in working on data vartulization tools like (Tibco. Trino etc) Good-to-Have Familiarity with Data Catalogs, Metadata Management, and NDMO data governance standards. Experience with Data Catalogue tools. Familiarity with CI/CD pipelines, Git. Knowledge of ETL orchestration tools like Airflow, NiFi
Responsibilities
The Data Engineer will be responsible for designing and optimizing large-scale real-time and batch data pipelines within the Telecom domain, leveraging technologies like Kafka, Spark, and Hadoop components. This includes collaborating with stakeholders to translate business needs into robust data models, developing ETL frameworks, and ensuring data availability for analytics teams.
Loading...