Senior Data Engineer - Azure Databricks at Blend360
Hyderabad, Telangana, India -
Full Time


Start Date

Immediate

Expiry Date

19 May, 26

Salary

0.0

Posted On

18 Feb, 26

Experience

2 year(s) or above

Remote Job

Yes

Telecommute

Yes

Sponsor Visa

No

Skills

Azure Databricks, Python, SQL, Apache Spark, PySpark, Spark SQL, Azure Data Factory, Delta Lake, ADLS Gen2, Data Modeling, ETL/ELT, Data Ingestion, Git, Data Quality, Security, Governance

Industry

Professional Services

Description
Company Description Blend is a premier AI services provider, committed to co-creating meaningful impact for its clients through the power of data science, AI, technology, and people. With a mission to fuel bold visions, Blend tackles significant challenges by seamlessly aligning human expertise with artificial intelligence. The company is dedicated to unlocking value and fostering innovation for its clients by harnessing world-class people and data-driven strategy. We believe that the power of people and AI can have a meaningful impact on your world, creating more fulfilling work and projects for our people and clients. For more information, visit www.blend360.com Job Description We are looking for an experienced Senior Azure Databricks Engineer with strong hands-on expertise in Python, SQL, and Apache Spark to design, build, and optimize scalable data pipelines and analytics solutions on the Azure cloud platform. The ideal candidate should have experience working with large datasets, distributed data processing, analytics use cases, and modern data engineering practices. Responsibilities Design, develop, and maintain scalable data pipelines using Azure Databricks Implement ETL/ELT workflows using PySpark, Spark SQL, and Python Implement pipelines for data ingestion using Azure Data Factory Optimize Spark jobs for performance, cost, and scalability Work with structured and semi-structured data (Parquet, Delta, JSON, CSV) Build and manage Delta Lake tables (ACID, time travel, schema evolution) Integrate Databricks with Azure Data Lake Storage (ADLS Gen2) Develop complex queries and transformations using SQL Collaborate with Data Science teams to prepare data for modelling use cases, ensuring appropriate transformations, feature generation, and storage. Follow best practices for security, access control, and governance in Azure Ensure data quality, validation, and monitoring using testing tools Deployment of solutions to Production environments Qualifications 4+ years of experience in Data Engineering, ideally supporting POS and SKU datasets. Handling high volume transactional datasets Strong hands-on experience with Azure Databricks Understanding of the Medallion Architecture and implementing it within Databricks Good understanding of data modelling techniques Proficiency in Python for data processing Strong knowledge of SQL (joins, window functions, performance tuning) Hands-on experience with Apache Spark / PySpark Experience working with Delta Lake Knowledge of Azure Data Lake Storage (ADLS Gen2) Understanding of distributed computing concepts Experience with Git version control Understanding of ML use cases and data considerations for model development
Responsibilities
The role involves designing, developing, and maintaining scalable data pipelines and ETL/ELT workflows primarily using Azure Databricks, PySpark, Spark SQL, and Python. Responsibilities also include optimizing Spark jobs, managing Delta Lake tables, and collaborating with Data Science teams to prepare data for modeling use cases.
Loading...