CONTINGENT - Data Engineer UAPI at RESOURCE METRIX
, , -
Full Time


Start Date

Immediate

Expiry Date

04 Apr, 26

Salary

0.0

Posted On

04 Jan, 26

Experience

2 year(s) or above

Remote Job

Yes

Telecommute

Yes

Sponsor Visa

No

Skills

Data Engineering, Python, SQL, Spark, PySpark, Airflow, dbt, Databricks, Kafka, Flink, Delta Lake, Snowflake, Redshift, Data Governance, Lineage, Audit Logging

Industry

IT Services and IT Consulting

Description
Benefits: Bonus based on performance Competitive salary Paid time off Client Summary: Internal Revenue Service (IRS) — Unified Application Programming Interface (UAPI) Program. UAPI modernizes IRS data access with secure APIs integrating legacy systems, modern databases, and AI/ML workflows. Work is primarily remote with limited on-site needs in Lanham, MD; Washington, DC; Ogden, UT; and Farmers Branch, TX. Eligibility & Suitability: U.S. citizen or Lawful Permanent Resident (position risk level dependent). Must pass IRS contractor staff‑like access screening: RAC, fingerprints (USAccess), eAPP (as needed), tax compliance check, Selective Service (males born after 12/31/1959), Security Awareness Training/UNAX, and role‑based cybersecurity training. Work involves SBU/FTI and requires adherence to IRM 10.5/10.8, Publication 4812, encryption, incident reporting (CSIRC/TIGTA), and Section 508 documentation/support. Position Summary: (OPENINGS: Multiple) Design/operate ETL/ELT pipelines for IRS UAPI; standardize ingestion and mapping from legacy sources; enable governed lakehouse (Delta Lake) with Spark/Databricks; orchestrate with Airflow/dbt; support streaming (Kafka/Flink), lineage/audit, validation logs, and API-ready datasets. Responsibilities: · Build batch and streaming pipelines (Spark/PySpark, Airflow, dbt) to support API data flows · Implement data validation, reconciliation, and quality checks; publish validation logs · Manage lakehouse storage (Delta Lake on S3/Azure) and warehouse models (Snowflake/Redshift) · Operate Kafka/Flink for real-time events; design schema evolution and governance · Deliver lineage/audit artifacts; contribute to monthly spec/mapping approvals · Optimize cost/performance; automate CI/CD for data workflows Qualifications: · 3+ years data engineering with Python/SQL and Spark/PySpark · Hands-on with Airflow, dbt, Databricks; Kafka/Flink streaming · Experience with Delta Lake, Snowflake/Redshift, and lakehouse patterns · Data governance, lineage, audit logging, and schema evolution · Preferred: IRS legacy data familiarity (MeF/CFOL/IDRS) and federal environments This is a remote position.
Responsibilities
Design and operate ETL/ELT pipelines for IRS UAPI, standardizing ingestion and mapping from legacy sources. Support streaming, lineage/audit, validation logs, and API-ready datasets.
Loading...