Negotiable
Outside
Remote
USA
Summary: The ETL Data Engineer role focuses on developing and maintaining ETL pipelines to support decision-making processes within the organization. The position requires strong expertise in data engineering, SQL, and Python, with a preference for candidates located in the Pacific Time Zone. The role is primarily remote, with a preference for candidates from the San Francisco area, and is structured as a W2 contract for six months. The engineer will collaborate with data scientists and analysts to optimize data structures and ensure data quality.
Key Responsibilities:
- Build core business data pipelines.
- Design data models and schemas to meet business and engineering requirements.
- Define and implement data quality checks to ensure ongoing data consistency.
- Perform SQL tuning to optimize data processing performance.
- Write clean, well-tested, and maintainable code, prioritizing scalability and cost efficiency.
- Conduct code reviews to uphold code quality standards.
- Produce high quality documentation to facilitate ownership transfer and ongoing support.
- Collaborate with internal and external partners to remove blockers, provide support, and achieve results.
- Develop, implement, and maintain robust Airflow ETL pipelines to support business and product decisions.
- Productionalize SQL and business logic, and perform data modeling to optimize data structures.
- Optimize and troubleshoot unhealthy pipelines to ensure data reliability and performance.
- Implement data quality validation frameworks to maintain data integrity.
- Build essential data feeds, such as a lifetime value (LTV) pipeline for writers, and create reliable data sources for dashboards.
- Collaborate closely with data scientists and engineers on the writer team to support their data needs.
- Work with the central data platform team for technical guidance and code reviews.
Key Skills:
- 5+ years of professional experience in data engineering or a related field.
- Strong expertise in SQL and experience with Spark and/or Trino.
- Proficiency in Python.
- Strong data modelling skills and a deep understanding of ETL processes.
- Experience building and optimizing complex data models and pipelines.
- Extensive experience with Apache Airflow for building and managing ETL pipelines.
- Familiarity with the modern data stack, including Hive-based or partitioned data lake structures.
- Strong sense of ownership and accountability for your work.
- Excellent critical thinking and problem-solving skills.
Salary (Rate): undetermined
City: San Francisco
Country: USA
Working Arrangements: remote
IR35 Status: outside IR35
Seniority Level: undetermined
Industry: IT
------------ONLY W2------------------ONLY W2-----------------ONLY W2----------------ONLY W2----------------
*********Candidates must be from Pacific Time Zone.***********
Title: ETL Data Engineer
Location: San Francisco, CA 94107 (Remote)
Duration: 6 Months Contract
Job Description:
Terms of Employment
- W2 Contract, 6 Months
- Remote Opportunity (Candidates from San Francisco area highly preferred)
- Shift Schedule: 08:00 AM-05:00 PM
Overview:
- Data powers all the decisions we make. It is the core of our business, helping us create a great transportation experience for our customers, and providing insights into the effectiveness of our services and products. You will be part of core Rider team, focusing on developing ETL pipelines that support decision-making processes for demand, finance and competitive data. Your contributions will help data scientists, analysts, and business leaders make informed decisions that drive success.
- You will need to evaluate multiple approaches, and implement solutions based on fundamental principles, best practices, and supporting data. By architecting, building, and launching robust data pipelines, you will enable seamless access to insights that fuel critical functions such as Analytics, Data Science, and Engineering.
Responsibilities
- Build core business data pipelines.
- Design data models and schemas to meet business and engineering requirements.
- Define and implement data quality checks to ensure ongoing data consistency.
- Perform SQL tuning to optimize data processing performance.
- Write clean, well-tested, and maintainable code, prioritizing scalability and cost efficiency.
- Conduct code reviews to uphold code quality standards.
- Produce high quality documentation to facilitate ownership transfer and ongoing support.
- Collaborate with internal and external partners to remove blockers, provide support, and achieve results.
- Develop, implement, and maintain robust Airflow ETL pipelines to support business and product decisions.
- Productionalize SQL and business logic, and perform data modeling to optimize data structures.
- Optimize and troubleshoot unhealthy pipelines to ensure data reliability and performance.
- Implement data quality validation frameworks to maintain data integrity.
- Build essential data feeds, such as a lifetime value (LTV) pipeline for writers, and create reliable data sources for dashboards.
- Collaborate closely with data scientists and engineers on the writer team to support their data needs.
- Work with the central data platform team for technical guidance and code reviews.
Required Skills & Experience
- 5+ years of professional experience in data engineering or a related field.
- Strong expertise in SQL and experience with Spark and/or Trino.
- Proficiency in Python.
- Strong data modelling skills and a deep understanding of ETL processes.
- Experience building and optimizing complex data models and pipelines.
- Extensive experience with Apache Airflow for building and managing ETL pipelines.
- Familiarity with the modern data stack, including Hive-based or partitioned data lake structures.
- Strong sense of ownership and accountability for your work.
- Excellent critical thinking and problem-solving skills.
Preferred Skills & Experience
- Experience working directly with cross-functional teams (data analytics, data science, engineering) to align data engineering solutions with business goals.
- Experience working at a large, well-known tech company (e.g., Meta, Google, Netflix), especially with consumer-facing products.
- Bachelor of Science in Computer Science or a related field (e.g., Statistics, Mathematics, Electrical Engineering).
- Experience working on diverse, multi-project backlogs and adapting to changing priorities.
- Familiarity with data quality frameworks and data validation best practices.
Sincerely,
Preetam Raj
Lead Technical Recruiter
nTech Workforce Inc.
D: EXT: 726
E: