Data Engineer

Data Engineer

Posted 1 day ago by 1751529049

Negotiable
Outside
Remote
USA

Summary: The Data Engineer role requires a highly experienced professional with over 10 years in software engineering, specifically using SPARK with Python and SQL. The position emphasizes expertise in AWS data engineering services and Oracle, along with a strong understanding of ETL processes. The role can be performed remotely or based in New York, New York, and is classified as outside IR35.

Key Responsibilities:

  • Utilize SPARK with Python and SQL for software engineering tasks.
  • Work with RDBMS databases, modeling, ETL, and SQL concepts.
  • Read and interpret logic, schema, and stored procedures in Oracle.
  • Implement AWS data engineering services including batch processing, EMR, S3, Glue, and Lambda.
  • Utilize Informatica for ETL processes.
  • Script using Unix/Python.
  • Implement Storage Area Networks (SAN) and related technologies.
  • Create infrastructure programmatically in AWS using various services.

Key Skills:

  • 10+ years of experience in software engineering using SPARK with Python and SQL.
  • Knowledge of RDBMS database, modeling, ETL, and SQL concepts.
  • Expertise in Oracle.
  • Experience with AWS data engineering services.
  • Familiarity with Snowflake (nice to have).
  • Proficiency in Informatica.
  • Unix/Python scripting skills.
  • Understanding of Storage Area Networks (SAN) and related technologies.
  • Ability to programmatically create infrastructure in AWS.

Salary (Rate): undetermined

City: New York

Country: USA

Working Arrangements: remote

IR35 Status: outside IR35

Seniority Level: undetermined

Industry: IT

Detailed Description From Employer:

Role: Data Engineer Contract: W2 Only

Required Skills:

  • 10+ years of experience software engineering using SPARK with Python and SQL
  • Need to know RDBMS database, modeling, ETL and SQL concepts
  • Expertise in Oracle is a must to read logic, schema, stored procedures
  • AWS data engineering services, is a must (batch, EMR, S3, glue, lambda, etc)
  • Snowflake is a nice to have but will be first considered
  • Informatica is a must (need ETL concepts)
  • Unix/Python for scripting is a must
  • Spark is a nice to have
  • Implement Storage Area Networks (SAN) and related technologies such as iSCSI and NDMP networking protocols
  • Programmatically create infrastructure in AWS, leveraging Autoscaling Groups, Security Groups, Route53, S3, and IAM.