DATA ENGINEER at Inetum
Madrid, Community of Madrid, Spain -
Full Time


Start Date

Immediate

Expiry Date

03 Jun, 26

Salary

0.0

Posted On

05 Mar, 26

Experience

2 year(s) or above

Remote Job

Yes

Telecommute

Yes

Sponsor Visa

No

Skills

Spark, Scala, Kafka, APIs, HDFS, SQL, CI/CD, Gitlab, Jenkins, Airflow, S3 Storage, Shell Script, Kubernetes, Elasticsearch, Kibana, Dataiku

Industry

IT Services and IT Consulting

Description
Company Description 🚀 Join Inetum – We're Hiring a DATA ENGINEER! 🚀 At Inetum, a leading international digital consultancy, we empower 27,000 professionals across 27 countries to shape their careers, foster innovation, and achieve work-life balance. Proudly certified as a Top Employer Europe 2024, we’re passionate about creating positive and impactful digital solutions. Job Description Hybrid work mode in Madrid. This service could imply a willingness to travel occasionally within Italy. The main responsibilities are: Data modeling and pipelines development with Spark on Scala in order to ingest and transform data from several sources (Kafka topics, APIs, HDFS, structured databases…). Data transformation and quality to ensure data consistency and accuracy. Set up CI/CD pipelines to automate deployments, unit testing and development management. The implementation of different orchestrators and scheduling processes to automate the data pipeline execution (Airflow as a service). Modifying the existing code as per business requirements and continuously improvement to achieve a better performance and maintainability. Ensuring the performance and security of the data infrastructure and following the best practice of Data engineering. Contributing to production support, incident and anomaly corrections, and to implement functional and technical evolutions to ensure the stability of production processes Writing technical documentation to ensure knowledge capitalization. Qualifications Good knowledge of Spark on Scala CI/CD tools (Gitlab, Jenkins…) HDFS and structured databases (SQL) Full understanding of Apache Airflow Streaming process (Kafka, event steam…) S3 storage Shell script Some knowledge of [*] Kubernetes Optionally/ as a plus [*] Elasticsearch and Kibana HVault Dataiku Dremio as tool to virtualize data Additional Information BUSINESS AND TRANSVERSAL SKILLS Knowledge of banking industry and processes Business / IT relationship Ability to understand, explain and support change Ability to Deliver / Results driven Ability to collaborate / Teamwork LANGUAGES English level B2 or higher Optionally, French level B1
Responsibilities
The main responsibilities involve developing data models and pipelines using Spark on Scala to ingest and transform data from various sources, ensuring data quality and consistency. This also includes setting up CI/CD pipelines, implementing orchestrators like Airflow, and continuously improving code for performance and maintainability.
Loading...