Data Engineer (Spark, AML) - W2 Only

Data Engineer (Spark, AML) - W2 Only

Posted 1 week ago by 1750938609

Negotiable
Outside
Remote
USA

Summary: The Data Engineer role focuses on developing applications using various programming languages and cloud technologies. Candidates should have extensive experience in data engineering, particularly with real-time data processing and distributed computing tools. The position is contract-based and requires a W2 employment arrangement. This role is fully remote, allowing for flexibility in work location.

Key Responsibilities:

  • Develop applications using Python, SQL, Scala, or Java.
  • Utilize public cloud services such as AWS, Microsoft Azure, or Google Cloud.
  • Work with distributed data/computing tools like MapReduce, Hadoop, Hive, EMR, Kafka, Spark, Gurobi, or MySQL.
  • Implement real-time data and streaming applications.
  • Manage NoSQL databases such as Mongo or Cassandra.
  • Oversee data warehousing solutions using Redshift or Snowflake.
  • Utilize UNIX/Linux for basic commands and shell scripting.
  • Apply Agile engineering practices in development processes.

Key Skills:

  • 7+ years of application development experience.
  • 4+ years of experience with public cloud platforms.
  • 4+ years of experience with distributed data/computing tools.
  • 4+ years of experience in real-time data processing.
  • 4+ years of experience with NoSQL databases.
  • 4+ years of data warehousing experience.
  • 4+ years of UNIX/Linux experience.
  • 2+ years of experience with Agile practices.

Salary (Rate): undetermined

City: undetermined

Country: USA

Working Arrangements: remote

IR35 Status: outside IR35

Seniority Level: undetermined

Industry: IT

Detailed Description From Employer:

Role: Data Engineer
Location: Remote

Type: Contract

Note: W2 only

Data Engineers Requirement

7+ years of experience in application development including Python, SQL, Scala, or Java

4+ years of experience with a public cloud (AWS, Microsoft Azure, Google Cloud)

4+ years experience with Distributed data/computing tools (MapReduce, Hadoop, Hive, EMR, Kafka, Spark, Gurobi, or MySQL)

4+ year experience working on real-time data and streaming applications

4+ years of experience with NoSQL implementation (Mongo, Cassandra)

4+ years of data warehousing experience (Redshift or Snowflake)

4+ years of experience with UNIX/Linux including basic commands and shell scripting

2+ years of experience with Agile engineering practices-