Big Data Developer

at  Sii

15-402 Białystok, województwo podlaskie, Poland -

Start DateExpiry DateSalaryPosted OnExperienceSkillsTelecommuteSponsor Visa
Immediate14 Aug, 2024Not Specified15 May, 2024N/AYarn,Git,Apache Spark,Elasticsearch,Scala,Hive,Github,Jenkins,Python,English,Ansible,Version Control Tools,SqlNoNo
Add to Wishlist Apply All Jobs
Required Visa Status:
CitizenGC
US CitizenStudent Visa
H1BCPT
OPTH4 Spouse of H1B
GC Green Card
Employment Type:
Full TimePart Time
PermanentIndependent - 1099
Contract – W2C2H Independent
C2H W2Contract – Corp 2 Corp
Contract to Hire – Corp 2 Corp

Description:

Possible locations

  • Bialystok
  • Bydgoszcz
  • Czestochowa
  • Gdansk
  • Gliwice
  • Katowice
  • Kielce
  • Cracow
  • Lublin
  • Lodz
  • Pila
  • Poznan
  • Rzeszow
  • Szczecin
  • Torun
  • Warsaw
  • Wroclaw
  • Remote work

Technologies & tools

Must have:

  • Python
  • Spark
  • Hadoop

Nice-to-have:

  • Scala
  • BigQuery

Join us as a Big Data Developer and work within the team of one of the largest financial institutions in the world. The Data team is primarily responsible for cloud migration and sourcing data from core banking systems, processed on the Hadoop platform. You will develop the platform for 52 clients across various countries.

Your role

  • Understanding all aspects of the Big Data ecosystem to make informed decisions about network and hardware, operating system, configuration, and end-user tooling and provisioning
  • Implementing and maintaining infrastructure based on designs provided by Big Data Architects, ensuring compliance with group operational and security standards and policies
  • Managing core ecosystem components such as Hive, Unix, Shell Scripting, and Spark
  • Understanding the interaction between the Hadoop ecosystem and other data sources to ensure seamless integration and data flow
  • Ensuring standard operational requirements are met by implementing infrastructure for monitoring, contingency planning, and user provisioning

Your skills

  • Hands-on experience in using Scala, Python, or Java and Unix/Linux environment on-premises
  • Experience with developing RESTful APIs using Java Springboot
  • Work experience in building data pipelines using Hadoop components (Apache Hadoop, Scala, Apache Spark, YARN, Hive, SQL)
  • Solid experience in optimizing Spark jobs
  • Previous work with time-series/analytics DB’s such as Elasticsearch
  • Excellent knowledge of industry-standard version control tools (Git, GitHub) and automated deployment tools (Ansible & Jenkins)
  • Basic shell-scripting knowledge
  • Understanding of big data modeling techniques using relational and non-relational techniques
  • Familiarity with SDLC (Software Development Lifecycle)
  • Very good knowledge of English

Job no. 240514-FBMG

Responsibilities:

  • Understanding all aspects of the Big Data ecosystem to make informed decisions about network and hardware, operating system, configuration, and end-user tooling and provisioning
  • Implementing and maintaining infrastructure based on designs provided by Big Data Architects, ensuring compliance with group operational and security standards and policies
  • Managing core ecosystem components such as Hive, Unix, Shell Scripting, and Spark
  • Understanding the interaction between the Hadoop ecosystem and other data sources to ensure seamless integration and data flow
  • Ensuring standard operational requirements are met by implementing infrastructure for monitoring, contingency planning, and user provisionin


REQUIREMENT SUMMARY

Min:N/AMax:5.0 year(s)

Information Technology/IT

IT Software - Application Programming / Maintenance

Software Engineering

Graduate

Proficient

1

15-402 Białystok, Poland