Big Data Engineer – AWS (EMR, S3, Glue, Spark, Hadoop) at Atlantis IT group
Toronto, ON, Canada -
Full Time


Start Date

Immediate

Expiry Date

23 Nov, 25

Salary

0.0

Posted On

23 Aug, 25

Experience

0 year(s) or above

Remote Job

Yes

Telecommute

Yes

Sponsor Visa

No

Skills

Good communication skills

Industry

Information Technology/IT

Description

Big Data Engineer AWS (EMR, S3, Glue, Spark, Hadoop)
Toronto
Role and ResponsibilitiesUnderstand requirement from product owners and translate into requirement and scope documentsDecide on the best fitment of technologiesservices that are available in scopeCreate Solution for data ingestion and data transformation using Hadoop services like Spark, Spark streaming, Hive, etc.Create technical design documents to communicate solutions to the team and mentor the team to develop the solutionBuild the solution with AWS and Hadoop Services as per design specificationsAssist team teams to build test cases and support with testing the solutionCoordinate with Upstream, Downstream and other supporting teams for production implementationProvide post-production support for Solutions implemented Please list 5 mandatory MUST HAVE skills and experience for this requirement. Please include skills related to technical as well as domain and non-technical skills and experience as applicable to the position.Candidate should have strong working experience with data services (EMR, S3, Glue) on AWS cloud.Strong hands-on experience on AWS EMR Glue (Spark with Scala) and AWS S3.Strong hands-on experience in Hadoop services like SparkIn-depth knowledge and extensive experience in building batch workloads on AWS using AWS EMR.Adept in analyzing and refining requirements, consumption query patterns and choosing the right technology fit like RDBMS, data lake and data warehouse.Should have knowledge of analytical data modelling on any of the RDBMSMPP platform Knowledge of PythonProven practical experience in migrating Hadoop based data lake from on-prem to AWS EMR on EMRFSPlease list 5 NICE TO HAVE but not mandatory skills and experience for this requirement Hands-on experience working with Terra bytes peta bytes scale data and millions of transactions per day.Skills to develop ETL pipeline using AirflowKnowledge of Spark streaming or any other streaming jobsAbility to deploy code using AWS Code Pipeline and Bit bucket is an added plusExpert in any of the following programming language Scala, Java and comfortable with working on Linux platform.Knowledge of Cloud based MPP Platfor

Responsibilities

Please refer the Job description for details

Loading...