Senior Data Engineer

at  Lextegrity

Remote, Oregon, USA -

Start DateExpiry DateSalaryPosted OnExperienceSkillsTelecommuteSponsor Visa
Immediate10 Nov, 2024Not Specified11 Aug, 20245 year(s) or aboveData Transformation,Spark,Snowflake,Pycharm,Sql,Data Engineering,Pandas,Sqlalchemy,Cloud Services,Aws,Scalability,Explain Plan,Data Loading,Data Manipulation,Load,Data Sharing,Cli,Python,Use Case,Operators,Data Warehouse,Cost EfficiencyNoNo
Add to Wishlist Apply All Jobs
Required Visa Status:
CitizenGC
US CitizenStudent Visa
H1BCPT
OPTH4 Spouse of H1B
GC Green Card
Employment Type:
Full TimePart Time
PermanentIndependent - 1099
Contract – W2C2H Independent
C2H W2Contract – Corp 2 Corp
Contract to Hire – Corp 2 Corp

Description:

ABOUT LEXTEGRITY

Lextegrity is a global leader in the digital transformation of corporate risk management. Our integrated platform combines continuous spend monitoring of data from systems like SAP, Oracle, and Concur with workflow automation tools to help global organizations prevent and detect fraud, corruption, conflicts of interest, and other economic crimes.
Lextegrity was founded in 2017 by veteran legal and audit professionals with decades of experience in addressing corporate fraud. Organizations lose 5% of their annual revenues to fraud, yet most companies struggle with embedding compliance policies into real-time business practices. Our software solutions address this key problem.

POSITION SUMMARY

We are seeking a talented Senior Data Engineer to join the data platform group for our Compliance Monitoring product, focused on enabling efficient data to support our risk analytics engine that works to identify fraud, bribery, and corruption. The ideal candidate will have a deep understanding of the python and SQL programming languages and experience with data analytics. The candidate will work closely with the implementation teams, data scientists, and product stakeholders in evolving this unique product.

QUALIFICATIONS

  • 5+ years of experience working heavily (IE daily) with:
  • Python

Responsibilities:

  • Strong Python skillset
  • Advanced proficiency in Python for data engineering, with over 5 years of experience in developing optimized ETL pipelines for data transformation and analysis. Expert in utilizing Pandas and NumPy for efficient data manipulation, alongside SQLAlchemy for database interactions.
  • Demonstrated capability in leveraging Python to design and execute complex data models and ETL processes in Snowflake, ensuring high performance and scalability. Experience with Python-based data processing frameworks to support analytics and machine learning implementations.
  • Knowledge of integrating Python applications with cloud services and data orchestration tools, including AWS and Apache Airflow, to enhance data pipeline efficiency.
  • Built multiple batch data processing pipelines (or pipeline frameworks) to extract, transform / load, load / transform data.
  • Designed/Written data models and ETL using Snowflake as the data warehouse
  • Demonstrated ability to design and implement robust data models in Snowflake, optimizing for performance, scalability, and cost-efficiency.
  • Advanced understanding of Snowflake’s architecture and features, including data sharing and warehousing capabilities to support efficient data analysis.
  • Proven track record of deploying and managing Snowflake’s security and access control features, ensuring data integrity and compliance with privacy regulations.
  • Experience with Snowpark and Snowpipe is a plus, enhancing capabilities in building scalable data applications and automating data loading.
  • Worked extensively with AWS services (RDS, EMR, S3)
  • Spent time provisioning, using, and terminating them (either via Boto3, CDK, CLI, or Apache Airflow operators).
  • Experience with Spark (with PySpark) is a plus
  • Git / Github with PyCharm (or your IDE of choice).
  • Worked with Apache Airflow building DAG’s for ETL / ELT.
  • Built single-use case DAG’s.
  • Built frameworks / reusable components.
  • SQL Proficiency
  • Can utilize an explain plan to optimize query performance.
  • Proficient in SQL from the basics of join optimization / indexing to window functions and beyond.


REQUIREMENT SUMMARY

Min:5.0Max:10.0 year(s)

Information Technology/IT

IT Software - DBA / Datawarehousing

Software Engineering

Graduate

Proficient

1

Remote, USA