Senior Data Engineer
at Lextegrity
Remote, Oregon, USA -
Start Date | Expiry Date | Salary | Posted On | Experience | Skills | Telecommute | Sponsor Visa |
---|---|---|---|---|---|---|---|
Immediate | 10 Nov, 2024 | Not Specified | 11 Aug, 2024 | 5 year(s) or above | Data Transformation,Spark,Snowflake,Pycharm,Sql,Data Engineering,Pandas,Sqlalchemy,Cloud Services,Aws,Scalability,Explain Plan,Data Loading,Data Manipulation,Load,Data Sharing,Cli,Python,Use Case,Operators,Data Warehouse,Cost Efficiency | No | No |
Required Visa Status:
Citizen | GC |
US Citizen | Student Visa |
H1B | CPT |
OPT | H4 Spouse of H1B |
GC Green Card |
Employment Type:
Full Time | Part Time |
Permanent | Independent - 1099 |
Contract – W2 | C2H Independent |
C2H W2 | Contract – Corp 2 Corp |
Contract to Hire – Corp 2 Corp |
Description:
ABOUT LEXTEGRITY
Lextegrity is a global leader in the digital transformation of corporate risk management. Our integrated platform combines continuous spend monitoring of data from systems like SAP, Oracle, and Concur with workflow automation tools to help global organizations prevent and detect fraud, corruption, conflicts of interest, and other economic crimes.
Lextegrity was founded in 2017 by veteran legal and audit professionals with decades of experience in addressing corporate fraud. Organizations lose 5% of their annual revenues to fraud, yet most companies struggle with embedding compliance policies into real-time business practices. Our software solutions address this key problem.
POSITION SUMMARY
We are seeking a talented Senior Data Engineer to join the data platform group for our Compliance Monitoring product, focused on enabling efficient data to support our risk analytics engine that works to identify fraud, bribery, and corruption. The ideal candidate will have a deep understanding of the python and SQL programming languages and experience with data analytics. The candidate will work closely with the implementation teams, data scientists, and product stakeholders in evolving this unique product.
QUALIFICATIONS
- 5+ years of experience working heavily (IE daily) with:
- Python
Responsibilities:
- Strong Python skillset
- Advanced proficiency in Python for data engineering, with over 5 years of experience in developing optimized ETL pipelines for data transformation and analysis. Expert in utilizing Pandas and NumPy for efficient data manipulation, alongside SQLAlchemy for database interactions.
- Demonstrated capability in leveraging Python to design and execute complex data models and ETL processes in Snowflake, ensuring high performance and scalability. Experience with Python-based data processing frameworks to support analytics and machine learning implementations.
- Knowledge of integrating Python applications with cloud services and data orchestration tools, including AWS and Apache Airflow, to enhance data pipeline efficiency.
- Built multiple batch data processing pipelines (or pipeline frameworks) to extract, transform / load, load / transform data.
- Designed/Written data models and ETL using Snowflake as the data warehouse
- Demonstrated ability to design and implement robust data models in Snowflake, optimizing for performance, scalability, and cost-efficiency.
- Advanced understanding of Snowflake’s architecture and features, including data sharing and warehousing capabilities to support efficient data analysis.
- Proven track record of deploying and managing Snowflake’s security and access control features, ensuring data integrity and compliance with privacy regulations.
- Experience with Snowpark and Snowpipe is a plus, enhancing capabilities in building scalable data applications and automating data loading.
- Worked extensively with AWS services (RDS, EMR, S3)
- Spent time provisioning, using, and terminating them (either via Boto3, CDK, CLI, or Apache Airflow operators).
- Experience with Spark (with PySpark) is a plus
- Git / Github with PyCharm (or your IDE of choice).
- Worked with Apache Airflow building DAG’s for ETL / ELT.
- Built single-use case DAG’s.
- Built frameworks / reusable components.
- SQL Proficiency
- Can utilize an explain plan to optimize query performance.
- Proficient in SQL from the basics of join optimization / indexing to window functions and beyond.
REQUIREMENT SUMMARY
Min:5.0Max:10.0 year(s)
Information Technology/IT
IT Software - DBA / Datawarehousing
Software Engineering
Graduate
Proficient
1
Remote, USA