Negotiable
Outside
Remote
USA
Summary: The Data Engineer role requires a highly experienced professional with over 10 years in software engineering, specifically using SPARK with Python and SQL. The position emphasizes expertise in AWS data engineering services and Oracle, along with a strong understanding of ETL processes. The role can be performed remotely or based in New York, New York, and is classified as outside IR35.
Key Responsibilities:
- Utilize SPARK with Python and SQL for software engineering tasks.
- Work with RDBMS databases, modeling, ETL, and SQL concepts.
- Read and interpret logic, schema, and stored procedures in Oracle.
- Implement AWS data engineering services including batch processing, EMR, S3, Glue, and Lambda.
- Utilize Informatica for ETL processes.
- Script using Unix/Python.
- Implement Storage Area Networks (SAN) and related technologies.
- Create infrastructure programmatically in AWS using various services.
Key Skills:
- 10+ years of experience in software engineering using SPARK with Python and SQL.
- Knowledge of RDBMS database, modeling, ETL, and SQL concepts.
- Expertise in Oracle.
- Experience with AWS data engineering services.
- Familiarity with Snowflake (nice to have).
- Proficiency in Informatica.
- Unix/Python scripting skills.
- Understanding of Storage Area Networks (SAN) and related technologies.
- Ability to programmatically create infrastructure in AWS.
Salary (Rate): undetermined
City: New York
Country: USA
Working Arrangements: remote
IR35 Status: outside IR35
Seniority Level: undetermined
Industry: IT
Role: Data Engineer Contract: W2 Only
Required Skills:
- 10+ years of experience software engineering using SPARK with Python and SQL
- Need to know RDBMS database, modeling, ETL and SQL concepts
- Expertise in Oracle is a must to read logic, schema, stored procedures
- AWS data engineering services, is a must (batch, EMR, S3, glue, lambda, etc)
- Snowflake is a nice to have but will be first considered
- Informatica is a must (need ETL concepts)
- Unix/Python for scripting is a must
- Spark is a nice to have
- Implement Storage Area Networks (SAN) and related technologies such as iSCSI and NDMP networking protocols
- Programmatically create infrastructure in AWS, leveraging Autoscaling Groups, Security Groups, Route53, S3, and IAM.