Negotiable
Undetermined
Remote
Remote
Summary: The Data Engineer role focuses on building and deploying production data pipelines on AWS, utilizing technologies such as PySpark, Glue, and Lambda. The position requires hands-on experience with infrastructure management through CDK and CI/CD pipelines. This is a contract-to-hire opportunity lasting over 12 months, with a remote working arrangement. Candidates should possess relevant AWS certifications and experience in data pipeline development.
Key Responsibilities:
- Build and deploy production data pipelines on AWS using PySpark/Glue, Lambda, and dbt.
- Manage infrastructure through CDK and CI/CD pipelines.
- Develop event-driven data pipelines using AWS-managed services.
- Implement infrastructure as code with CDK, CloudFormation, or Terraform.
- Deploy applications via CI/CD pipelines.
- Utilize strong SQL and dbt Core for data transformations and testing.
- Work with Lambda/API ingestion patterns and S3-based lakehouse architecture.
Key Skills:
- AWS Certified Data Analytics -- Specialty.
- AWS Certified Developer -- Associate.
- Databricks Certified Data Engineer Associate.
- Pipeline development with AWS Glue (PySpark) or Spark on EMR.
- Experience with AWS-managed services (Lambda, EventBridge, SQS, Kinesis).
- Infrastructure as code with CDK, CloudFormation, or Terraform.
- Deployment via CI/CD pipelines (GitHub Actions, CodePipeline, etc.).
- Strong SQL + dbt Core for transformations and testing.
- Lambda / API ingestion patterns and S3-based lakehouse experience.
Salary (Rate): undetermined
City: undetermined
Country: undetermined
Working Arrangements: remote
IR35 Status: undetermined
Seniority Level: undetermined
Industry: IT
Detailed Description From Employer:
Role: Data Engineer (AWS Lakehouse / CDK / CI-CD)
Duration: 12+ Month Contract-to-Hire
Location: Remote
Hands-on data engineer who builds and deploys production data pipelines on AWS using PySpark/Glue, Lambda, and dbt, with infrastructure managed through CDK and CI/CD pipelines.
Must Have / Target Certifications
- AWS Certified Data Analytics -- Specialty
- AWS Certified Developer -- Associate
- Databricks Certified Data Engineer Associate
Must-Have Signals
- Pipeline development with AWS Glue (PySpark) or Spark on EMR
- Experience building event-driven data pipelines using AWS-managed services (Lambda, EventBridge, SQS, Kinesis)
- Infrastructure as code with CDK, CloudFormation, or Terraform
- Deployment via CI/CD pipelines (GitHub Actions, CodePipeline, etc.)
- Strong SQL + dbt Core for transformations and testing
- Lambda / API ingestion patterns and S3-based lakehouse experience
Thanks
John
Netstrive Consulting, LLC
450, E. 22nd St, Suite 178,
Lombard, IL 60148
Direct:
Desk No: Ext-404
Fax:
Web:
Email:
Note: This email is not intended to be a solicitation. Please accept our apologies and reply in the subject heading with REMOVE to be removed from our mailing list.