Software Engineer - Snowflake at White Cloak Technologies
Manila, Metro Manila, Philippines -
Full Time


Start Date

Immediate

Expiry Date

07 Apr, 26

Salary

0.0

Posted On

07 Jan, 26

Experience

2 year(s) or above

Remote Job

Yes

Telecommute

Yes

Sponsor Visa

No

Skills

Confluent Kafka, Snowflake, Data Pipeline Development, Streaming Data Processing, Data Integration, Database Management, Security & Compliance, Optimization, Testing, Troubleshooting, Documentation, Collaboration, SQL, Python, Scala, ETL/ELT

Industry

Software Development

Description
We are looking for a talented and motivated Software Engineer with experience in Confluent, Snowflake and QLIK to join our team. This role involves designing, developing, and maintaining real-time data streaming and cloud-based data warehousing solutions using Confluent Kafka, Snowflake. The ideal candidate will work closely with cross-functional teams to build scalable, efficient, and secure data pipelines that drive business intelligence and analytics. Responsibilities: Data Pipeline Development: Design, develop, and optimize real-time and batch data pipelines using Confluent Kafka and Snowflake. Streaming Data Processing: Implement event-driven architectures leveraging Kafka Streams, ksqlDB, and Kafka Connect. Data Integration: Integrate structured and semi-structured data from multiple sources into Snowflake for analytics and reporting. Database Management: Develop, optimize, and maintain Snowflake-based data warehouses, ensuring high availability and performance. Security & Compliance: Implement data security best practices, including encryption, role-based access control (RBAC), and compliance with GDPR, HIPAA, or other relevant regulations. Optimization: Optimize query performance and storage efficiency in Snowflake using clustering, partitioning, and caching techniques. Testing: Conduct unit testing and data validation to ensure accuracy and reliability of data pipelines. Troubleshooting: Diagnose and resolve issues related to Kafka messaging, Snowflake storage, and data transformations. Documentation: Prepare and maintain technical documentation, including architecture diagrams, data flow processes, and best practices. Collaboration: Work with Data Scientists, BI Teams, and Cloud Engineers to provide real-time and batch data solutions. Maintenance: Monitor and maintain the stability of streaming pipelines and data warehouse operations. Qualifications: Strong understanding of real-time data streaming and event-driven architectures. Hands-on experience with Confluent Kafka (Kafka Streams, Kafka Connect, ksqlDB). Experience in Snowflake data warehouse development and administration. Proficiency in SQL, Python, or Scala for data processing and transformation. Familiarity with cloud platforms (AWS, Azure, GCP) for data pipeline orchestration. Experience with ETL/ELT processes and data pipeline automation. Knowledge of CI/CD for data pipelines, version control (Git), and DevOps tools. Strong problem-solving and analytical skills. Nice to have: Confluent Kafka Certification or Snowflake Certification. Experience with Apache Airflow, dbt, or other data orchestration tools. Knowledge of Big Data Technologies (Spark, Flink, or Druid). Understanding of GraphQL, REST APIs, and WebSockets for data integration. Experience working in finance, healthcare, or other regulated industries.
Responsibilities
The Software Engineer will design, develop, and maintain real-time data streaming and cloud-based data warehousing solutions. They will work closely with cross-functional teams to build scalable and efficient data pipelines for business intelligence and analytics.
Loading...