Data Engineer (PySpark) at GSSTech Group
Bengaluru, karnataka, India -
Full Time


Start Date

Immediate

Expiry Date

29 Jun, 26

Salary

0.0

Posted On

31 Mar, 26

Experience

5 year(s) or above

Remote Job

Yes

Telecommute

Yes

Sponsor Visa

No

Skills

PySpark, Informatica BDM, ETL, ELT, Data Mapping, Data Profiling, Impact Analysis, SQL, Data Warehousing, Data Quality, Debugging, Deployment, AI Technologies, Claude

Industry

IT Services and IT Consulting

Description
We are looking for a skilled Data Engineer to join the Data Engineering Chapter supporting the Group Operations team at ENBD. The ideal candidate will be responsible for building scalable data pipelines, performing data analysis, and delivering high-quality data solutions aligned with enterprise data models. Key Responsibilities Collaborate with the Group Operations Team to understand business and data requirements on a daily basis Perform Impact Assessment for new and existing data changes Conduct Technical Data Mapping and Data Profiling activities Design, develop, and maintain ETL pipelines for data extraction, transformation, and loading Build data solutions to feed into AECB application as per prescribed data models Develop and optimize data pipelines using PySpark on modern data platforms Ensure data quality, consistency, and integrity across systems Perform unit testing, debugging, and deployment of data solutions Leverage modern tools and AI technologies (e.g., Claude) to enhance development efficiency and reduce operational errors Work closely with cross-functional teams including business analysts, architects, and QA Required Skills & Qualifications Strong hands-on experience with PySpark and distributed data processing Experience in Informatica BDM Development (Big Data Management) Solid understanding of ETL/ELT concepts and data pipeline architecture Expertise in data mapping, data profiling, and impact analysis Experience working with large-scale data systems and cloud/data platforms Strong SQL skills and understanding of data warehousing concepts Familiarity with banking/financial domain (preferred but not mandatory) Knowledge of AI-assisted development tools (e.g., Claude) is a plus Good problem-solving and analytical skills Nice to Have Experience with AECB data/reporting systems Exposure to big data ecosystems (Hadoop/Spark clusters) Understanding of data governance and compliance standards
Responsibilities
The role involves building scalable data pipelines, performing data analysis, and delivering high-quality data solutions aligned with enterprise data models for the Group Operations team. Key tasks include designing, developing, and maintaining ETL pipelines using PySpark and ensuring data quality across systems.
Loading...