AI Data Engineer at Zonestra Technology LLc
Bettendorf, IA 52722, USA -
Full Time


Start Date

Immediate

Expiry Date

18 Jul, 25

Salary

0.0

Posted On

18 Apr, 25

Experience

3 year(s) or above

Remote Job

Yes

Telecommute

Yes

Sponsor Visa

No

Skills

Python, Computer Science, Relational Databases, Data Preparation, Analytical Skills, Communication Skills, Ec2, Natural Language Processing, Database Design, Machine Learning, Apache Spark, Data Modeling, Docker, Query Optimization, Distributed Systems, Containerization

Industry

Information Technology/IT

Description

DETAILS

Aquent Studios is seeking a highly skilled and motivated AI Data Engineer with experience in Langchain, large data sets, AWS, and expertise in SQL queries and databases. This role will involve working on the design, development, and maintenance of robust data pipelines, facilitating the integration of AI models, and processing large-scale datasets in cloud-based environments.
Our client is at the forefront of cutting-edge AI and data solutions. They empower organizations to leverage the power of data and artificial intelligence for growth, innovation, and efficiency. This role will join their talented team and contribute to the development of AI-driven systems using large-scale data solutions.

REQUIREMENTS

Experience:

  • 3+ years of experience as a Data Engineer, with a focus on AI or machine learning data pipelines.
  • Hands-on experience with Langchain for building and optimizing AI workflows and automation.
  • Strong experience with large-scale data management, including working with distributed systems and processing massive datasets.
  • Proficient in SQL and experience with relational databases (e.g., PostgreSQL, MySQL, SQL Server) and NoSQL databases.
  • Extensive experience with AWS, including services like S3, EC2, Lambda, Redshift, and RDS.

TECHNICAL SKILLS:

  • Proficient in Python
  • Strong understanding of database design, data modeling, and query optimization.
  • Familiarity with data warehousing concepts and tools.
  • Exerience working with data pipelines and workflow orchestration tools like Apache Airflow, or similar.
  • Knowledge of AI and machine learning concepts, particularly around data preprocessing and feature engineering.

Education:

  • Bachelor’s or Master’s degree in Computer Science, Data Engineering, Artificial Intelligence, or a related field.

Soft Skills:

  • Strong problem-solving and analytical skills.
  • Excellent communication skills, with the ability to explain technical concepts to non-technical stakeholders.
  • Ability to work in a fast-paced environment and manage multiple tasks effectively.
  • Team-oriented with a collaborative attitude.

PREFERRED QUALIFICATIONS:

  • Experience with Apache Spark, Kafka, or other big data technologies.
  • Familiarity with containerization (Docker, Kubernetes) for deploying scalable data solutions.
  • Knowledge of AI-specific data workflows, including data preparation for natural language processing, computer vision, or other AI applications.
Responsibilities
  • Data Engineering: Design, develop, and manage data pipelines to handle and process large datasets with a focus on scalability and efficiency.
  • Langchain Integration: Utilize Langchain to build and optimize AI workflows, automate processes, and enhance AI model capabilities with data-driven features.
  • Cloud Data Infrastructure: Manage and scale data infrastructure on AWS, using services such as S3, EC2, Lambda, Redshift, and others.
  • SQL & Database Expertise: Write complex SQL queries for data extraction, transformation, and analysis. Optimize database performance and ensure data integrity.
  • Collaboration: Work closely with Data Scientists, AI Engineers, and other stakeholders to ensure data is properly structured for machine learning models and other analytics use cases.
  • Data Quality & Governance: Monitor data quality, create processes for error handling, and ensure compliance with data privacy regulations and best practices.
  • Performance Optimization: Continuously optimize and fine-tune data processing pipelines and database queries for better performance and scalability.
  • Documentation: Maintain detailed documentation on the architecture, data processes, and code to ensure smooth collaboration and knowledge transfer.
Loading...