Data Engineer at EVO OUTSOURCING SOLUTIONS PTE LTD
Singapore, Southeast, Singapore -
Full Time


Start Date

Immediate

Expiry Date

22 Apr, 25

Salary

0.0

Posted On

23 Jan, 25

Experience

6 year(s) or above

Remote Job

No

Telecommute

No

Sponsor Visa

No

Skills

Spark, Hive, Query Optimization, Database Design, Mapreduce, Kafka, Data Engineering, Analytical Skills, Yarn

Industry

Information Technology/IT

Description

Responsibilities

  • Design, build, and optimize scalable data pipelines using Scala and Hadoop frameworks.
  • Develop and implement techniques and analytics applications to transform raw data into meaningful information using data-oriented programming languages and visualisation software.
  • Write and optimize complex SQL queries for efficient data retrieval and transformation.
  • Troubleshoot and resolve performance issues in queries and data workflows.
  • Manage Hadoop-based data infrastructure, including HDFS, Hive, and related components.
  • Monitor system performance and optimize resource utilization in a distributed environment.

Skills/Requirement

  • Proficiency in Scala programming and functional programming principles.
  • Strong SQL skills, including query optimization and database design.
  • Hands-on experience with Hadoop ecosystem tools (HDFS, Hive, YARN, MapReduce).
  • Familiarity with other big data tools like Spark, Kafka, or Flink is a plus.
  • Strong analytical skills with the ability to troubleshoot and resolve complex data issues.
  • 6+ years of experience in data engineering, big data technologies, or a similar role.
Responsibilities
  • Design, build, and optimize scalable data pipelines using Scala and Hadoop frameworks.
  • Develop and implement techniques and analytics applications to transform raw data into meaningful information using data-oriented programming languages and visualisation software.
  • Write and optimize complex SQL queries for efficient data retrieval and transformation.
  • Troubleshoot and resolve performance issues in queries and data workflows.
  • Manage Hadoop-based data infrastructure, including HDFS, Hive, and related components.
  • Monitor system performance and optimize resource utilization in a distributed environment
Loading...