Data Engineer
at MiC3 International
Johannesburg, Gauteng, South Africa -
Start Date | Expiry Date | Salary | Posted On | Experience | Skills | Telecommute | Sponsor Visa |
---|---|---|---|---|---|---|---|
Immediate | 17 Feb, 2025 | Not Specified | 18 Nov, 2024 | N/A | Good communication skills | No | No |
Required Visa Status:
Citizen | GC |
US Citizen | Student Visa |
H1B | CPT |
OPT | H4 Spouse of H1B |
GC Green Card |
Employment Type:
Full Time | Part Time |
Permanent | Independent - 1099 |
Contract – W2 | C2H Independent |
C2H W2 | Contract – Corp 2 Corp |
Contract to Hire – Corp 2 Corp |
Description:
DATA INTEGRATIONS ENGINEER (NIFI)
As a Data Integration Engineer, you will be responsible for designing, implementing, and maintaining data integration solutions to handle real-time streaming data from various sources like IoT/IIoT protocols, third-party APIs, or even raw files. Your main objective will be to process data in real-time and provide valuable insights for our organization. You will work with a diverse range of Big Data tools and technologies. The successful candidate will have experience in embedded systems bringup requirements engineering management, systems integration, developing programs to drive HW and SW planning and articulating the big picture.
Additionally, you will be involved in the development of a Data Streaming platform using Nifi
Responsibilities:
- Data Integration Design: Collaborate with cross-functional teams to understand data requirements, source systems, and data formats.
- Design efficient data integration pipelines for real-time data streaming from multiple sources.
- Programming Languages: Develop custom data processing components and applications using Java and Python to meet specific business requirements.
- ETL Development: Implement Extract, Transform, Load (ETL) processes to ingest and transform data from various streaming sources into a format suitable for analysis and storage.
- Real-time Data Processing: Develop and optimize data processing workflows to ensure timely handling of streaming data, maintaining low-latency and high-throughput capabilities.
- Big Data Tools: Utilize and maintain various Big Data tools such as Apache NiFi, Spark, Kafka etc. to build scalable and robust data integration solutions.
- Message Broker Configuration: Set up and configure message brokers like RabbitMQ, AMQP, and Kafka to enable efficient data exchange between different systems and applications.
- IoT/IIoT Protocols Integration: Integrate and work with IoT/IIoT protocols such as MQTT, SNMP, CoAP, TCP, and WebSockets to capture data from edge devices and industrial systems.
- Data Quality and Validation: Implement data validation checks and data quality measures to ensure the accuracy and reliability of the integrated data.
- Performance Monitoring: Monitor the performance and health of data integration pipelines, making necessary adjustments to optimize data flow and resource utilization.
- Troubleshooting and Issue Resolution: Diagnose and resolve issues related to data integration, ensuring smooth and uninterrupted data streaming.
- Technical Requirements:
- Strong experience in designing and implementing data integration solutions for real-time streaming data.
- Proficiency in using Big Data technologies such as Apache NiFi, Apache Spark and Kafka
- Familiarity with message brokers like RabbitMQ, AMQP, and Kafka for data exchange and event-driven architectures.
- Hands-on experience with IoT/IIoT protocols such as MQTT, SNMP, CoAP, TCP, and WebSockets.
- Proficiency in programming languages such as Java and Python for developing custom data processing components.
- Knowledge of data quality assurance and validation techniques to ensure reliable data.
- Ability to troubleshoot and resolve issues related to data integration and streaming processes.
- Strong analytical and problem-solving skills, with a keen eye for detail.
- Excellent communication and teamwork skills to collaborate effectively with cross-functional teams.
- Experience with cloud-based platforms and distributed systems is advantageous.
- Have a never ending curious mindset to learn and work with new tools and technologies
Responsibilities:
- Data Integration Design: Collaborate with cross-functional teams to understand data requirements, source systems, and data formats.
- Design efficient data integration pipelines for real-time data streaming from multiple sources.
- Programming Languages: Develop custom data processing components and applications using Java and Python to meet specific business requirements.
- ETL Development: Implement Extract, Transform, Load (ETL) processes to ingest and transform data from various streaming sources into a format suitable for analysis and storage.
- Real-time Data Processing: Develop and optimize data processing workflows to ensure timely handling of streaming data, maintaining low-latency and high-throughput capabilities.
- Big Data Tools: Utilize and maintain various Big Data tools such as Apache NiFi, Spark, Kafka etc. to build scalable and robust data integration solutions.
- Message Broker Configuration: Set up and configure message brokers like RabbitMQ, AMQP, and Kafka to enable efficient data exchange between different systems and applications.
- IoT/IIoT Protocols Integration: Integrate and work with IoT/IIoT protocols such as MQTT, SNMP, CoAP, TCP, and WebSockets to capture data from edge devices and industrial systems.
- Data Quality and Validation: Implement data validation checks and data quality measures to ensure the accuracy and reliability of the integrated data.
- Performance Monitoring: Monitor the performance and health of data integration pipelines, making necessary adjustments to optimize data flow and resource utilization.
- Troubleshooting and Issue Resolution: Diagnose and resolve issues related to data integration, ensuring smooth and uninterrupted data streaming.
- Technical Requirements:
- Strong experience in designing and implementing data integration solutions for real-time streaming data.
- Proficiency in using Big Data technologies such as Apache NiFi, Apache Spark and Kafka
- Familiarity with message brokers like RabbitMQ, AMQP, and Kafka for data exchange and event-driven architectures.
- Hands-on experience with IoT/IIoT protocols such as MQTT, SNMP, CoAP, TCP, and WebSockets.
- Proficiency in programming languages such as Java and Python for developing custom data processing components.
- Knowledge of data quality assurance and validation techniques to ensure reliable data.
- Ability to troubleshoot and resolve issues related to data integration and streaming processes.
- Strong analytical and problem-solving skills, with a keen eye for detail.
- Excellent communication and teamwork skills to collaborate effectively with cross-functional teams.
- Experience with cloud-based platforms and distributed systems is advantageous.
- Have a never ending curious mindset to learn and work with new tools and technologie
REQUIREMENT SUMMARY
Min:N/AMax:5.0 year(s)
Information Technology/IT
IT Software - DBA / Datawarehousing
Software Engineering
Graduate
Proficient
1
Johannesburg, Gauteng, South Africa