Big Data Developer
at Sii
15-402 Białystok, województwo podlaskie, Poland -
Start Date | Expiry Date | Salary | Posted On | Experience | Skills | Telecommute | Sponsor Visa |
---|---|---|---|---|---|---|---|
Immediate | 14 Aug, 2024 | Not Specified | 15 May, 2024 | N/A | Yarn,Git,Apache Spark,Elasticsearch,Scala,Hive,Github,Jenkins,Python,English,Ansible,Version Control Tools,Sql | No | No |
Required Visa Status:
Citizen | GC |
US Citizen | Student Visa |
H1B | CPT |
OPT | H4 Spouse of H1B |
GC Green Card |
Employment Type:
Full Time | Part Time |
Permanent | Independent - 1099 |
Contract – W2 | C2H Independent |
C2H W2 | Contract – Corp 2 Corp |
Contract to Hire – Corp 2 Corp |
Description:
Possible locations
- Bialystok
- Bydgoszcz
- Czestochowa
- Gdansk
- Gliwice
- Katowice
- Kielce
- Cracow
- Lublin
- Lodz
- Pila
- Poznan
- Rzeszow
- Szczecin
- Torun
- Warsaw
- Wroclaw
- Remote work
Technologies & tools
Must have:
- Python
- Spark
- Hadoop
Nice-to-have:
- Scala
- BigQuery
Join us as a Big Data Developer and work within the team of one of the largest financial institutions in the world. The Data team is primarily responsible for cloud migration and sourcing data from core banking systems, processed on the Hadoop platform. You will develop the platform for 52 clients across various countries.
Your role
- Understanding all aspects of the Big Data ecosystem to make informed decisions about network and hardware, operating system, configuration, and end-user tooling and provisioning
- Implementing and maintaining infrastructure based on designs provided by Big Data Architects, ensuring compliance with group operational and security standards and policies
- Managing core ecosystem components such as Hive, Unix, Shell Scripting, and Spark
- Understanding the interaction between the Hadoop ecosystem and other data sources to ensure seamless integration and data flow
- Ensuring standard operational requirements are met by implementing infrastructure for monitoring, contingency planning, and user provisioning
Your skills
- Hands-on experience in using Scala, Python, or Java and Unix/Linux environment on-premises
- Experience with developing RESTful APIs using Java Springboot
- Work experience in building data pipelines using Hadoop components (Apache Hadoop, Scala, Apache Spark, YARN, Hive, SQL)
- Solid experience in optimizing Spark jobs
- Previous work with time-series/analytics DB’s such as Elasticsearch
- Excellent knowledge of industry-standard version control tools (Git, GitHub) and automated deployment tools (Ansible & Jenkins)
- Basic shell-scripting knowledge
- Understanding of big data modeling techniques using relational and non-relational techniques
- Familiarity with SDLC (Software Development Lifecycle)
- Very good knowledge of English
Job no. 240514-FBMG
Responsibilities:
- Understanding all aspects of the Big Data ecosystem to make informed decisions about network and hardware, operating system, configuration, and end-user tooling and provisioning
- Implementing and maintaining infrastructure based on designs provided by Big Data Architects, ensuring compliance with group operational and security standards and policies
- Managing core ecosystem components such as Hive, Unix, Shell Scripting, and Spark
- Understanding the interaction between the Hadoop ecosystem and other data sources to ensure seamless integration and data flow
- Ensuring standard operational requirements are met by implementing infrastructure for monitoring, contingency planning, and user provisionin
REQUIREMENT SUMMARY
Min:N/AMax:5.0 year(s)
Information Technology/IT
IT Software - Application Programming / Maintenance
Software Engineering
Graduate
Proficient
1
15-402 Białystok, Poland