Data Engineer (AWS Lakehouse / CDK / CI-CD)

Data Engineer (AWS Lakehouse / CDK / CI-CD)

Posted 6 days ago by NetStrive Consulting

Negotiable
Undetermined
Remote
Remote

Summary: The Data Engineer role focuses on building and deploying production data pipelines on AWS, utilizing technologies such as PySpark, Glue, and Lambda. The position requires hands-on experience with infrastructure management through CDK and CI/CD pipelines. This is a contract-to-hire opportunity lasting over 12 months, with a remote working arrangement. Candidates should possess relevant AWS certifications and experience in data pipeline development.

Key Responsibilities:

  • Build and deploy production data pipelines on AWS using PySpark/Glue, Lambda, and dbt.
  • Manage infrastructure through CDK and CI/CD pipelines.
  • Develop event-driven data pipelines using AWS-managed services.
  • Implement infrastructure as code with CDK, CloudFormation, or Terraform.
  • Deploy applications via CI/CD pipelines.
  • Utilize strong SQL and dbt Core for data transformations and testing.
  • Work with Lambda/API ingestion patterns and S3-based lakehouse architecture.

Key Skills:

  • AWS Certified Data Analytics -- Specialty.
  • AWS Certified Developer -- Associate.
  • Databricks Certified Data Engineer Associate.
  • Pipeline development with AWS Glue (PySpark) or Spark on EMR.
  • Experience with AWS-managed services (Lambda, EventBridge, SQS, Kinesis).
  • Infrastructure as code with CDK, CloudFormation, or Terraform.
  • Deployment via CI/CD pipelines (GitHub Actions, CodePipeline, etc.).
  • Strong SQL + dbt Core for transformations and testing.
  • Lambda / API ingestion patterns and S3-based lakehouse experience.

Salary (Rate): undetermined

City: undetermined

Country: undetermined

Working Arrangements: remote

IR35 Status: undetermined

Seniority Level: undetermined

Industry: IT

Detailed Description From Employer:

Role: Data Engineer (AWS Lakehouse / CDK / CI-CD)

Duration: 12+ Month Contract-to-Hire

Location: Remote

Hands-on data engineer who builds and deploys production data pipelines on AWS using PySpark/Glue, Lambda, and dbt, with infrastructure managed through CDK and CI/CD pipelines.

Must Have / Target Certifications

  • AWS Certified Data Analytics -- Specialty
  • AWS Certified Developer -- Associate
  • Databricks Certified Data Engineer Associate

Must-Have Signals

  • Pipeline development with AWS Glue (PySpark) or Spark on EMR
  • Experience building event-driven data pipelines using AWS-managed services (Lambda, EventBridge, SQS, Kinesis)
  • Infrastructure as code with CDK, CloudFormation, or Terraform
  • Deployment via CI/CD pipelines (GitHub Actions, CodePipeline, etc.)
  • Strong SQL + dbt Core for transformations and testing
  • Lambda / API ingestion patterns and S3-based lakehouse experience


Thanks

John

Netstrive Consulting, LLC

450, E. 22nd St, Suite 178,

Lombard, IL 60148

Direct:

Desk No: Ext-404

Fax:

Web:

Email:

Note: This email is not intended to be a solicitation. Please accept our apologies and reply in the subject heading with REMOVE to be removed from our mailing list.