Data Engineer

Data Engineer

Posted 1 week ago by Appiness Inc.

Negotiable
Undetermined
Remote
Remote

Summary: We are seeking a detail-oriented Data Integration Engineer responsible for designing, developing, and managing data integration solutions. The ideal candidate will have experience in integrating data across various systems and building ETL/ELT pipelines to ensure data accuracy and quality. This role is crucial for enabling seamless data flow to support business intelligence and operational needs.

Key Responsibilities:

  • Design and implement data integration workflows between internal and external systems, including APIs, databases, SaaS applications, and cloud platforms.
  • Develop and maintain scalable ETL/ELT pipelines for structured and unstructured data using tools like Informatica, Talend, SSIS, Apache NiFi, or custom Python/SQL scripts.
  • Build and manage real-time and batch data pipelines leveraging technologies like Kafka, Spark Streaming.
  • Ensure high data quality, accuracy, and consistency during data ingestion and transformation.
  • Implement data validation, cleansing, deduplication, and monitoring mechanisms.
  • Contribute to metadata management, data lineage, and data catalog initiatives.
  • Collaborate with data engineers, business analysts, data scientists, and application teams to understand integration needs and deliver effective solutions.
  • Troubleshoot and resolve data integration and pipeline issues in a timely manner.
  • Provide documentation and knowledge transfer for developed solutions.
  • Support data movement across hybrid environments (on-prem, cloud, third-party systems).
  • Work with DevOps or platform teams to ensure scalability, security, and performance of data integration infrastructure.

Key Skills:

  • Bachelor’s degree in Computer Science, Information Systems, Engineering, or a related field.
  • 4–8 years of experience in data integration, data engineering, with strong ETL and SQL.
  • Strong experience with integration tools such as Informatica, Talend, MuleSoft, SSIS, or Boomi.
  • Proficient in SQL, Python, and scripting for data manipulation and automation.
  • Experience with cloud data platforms (Google Cloud Platform) and services such as Google Cloud Dataflow.
  • Familiarity with REST/SOAP APIs, JSON, XML, and flat file integrations.
  • Experience with message queues or data streaming platforms (Kafka, RabbitMQ, Kinesis).
  • Understanding of data warehousing concepts and tools (e.g., Snowflake, Redshift, BigQuery).
  • Knowledge of data security, privacy, and compliance best practices (HIPAA, GDPR, etc.).
  • Prior experience in industries like healthcare, fintech, or e-commerce is a plus.
  • Strong problem-solving and debugging skills.
  • Excellent communication and collaboration abilities.
  • Ability to manage multiple priorities and deliver in a fast-paced environment.
  • Attention to detail and a commitment to delivering high-quality work.

Salary (Rate): £60,000 yearly

City: undetermined

Country: undetermined

Working Arrangements: remote

IR35 Status: undetermined

Seniority Level: undetermined

Industry: IT

Job Summary:

We are looking for a detail-oriented and technically skilled Data Integration Engineer to design, develop, and manage robust data integration solutions. The ideal candidate will have hands-on experience in integrating data across disparate systems, building ETL/ELT pipelines, and ensuring the accuracy, quality, and consistency of enterprise data. You will play a key role in enabling seamless data flow between systems to support business intelligence, analytics, and operational needs.

Key Responsibilities:

Integration Design & Development

  • Design and implement data integration workflows between internal and external systems, including APIs, databases, SaaS applications, and cloud platforms.
  • Develop and maintain scalable ETL/ELT pipelines for structured and unstructured data using tools like Informatica, Talend, SSIS, Apache NiFi, or custom Python/SQL scripts.
  • Build and manage real-time and batch data pipelines leveraging technologies like Kafka, Spark Streaming.

Data Quality & Governance

  • Ensure high data quality, accuracy, and consistency during data ingestion and transformation.
  • Implement data validation, cleansing, deduplication, and monitoring mechanisms.
  • Contribute to metadata management, data lineage, and data catalog initiatives.

Collaboration & Troubleshooting

  • Collaborate with data engineers, business analysts, data scientists, and application teams to understand integration needs and deliver effective solutions.
  • Troubleshoot and resolve data integration and pipeline issues in a timely manner.
  • Provide documentation and knowledge transfer for developed solutions.

Platform & Infrastructure Support

  • Support data movement across hybrid environments (on-prem, cloud, third-party systems).
  • Work with DevOps or platform teams to ensure scalability, security, and performance of data integration infrastructure.

Qualifications:

  • Bachelor’s degree in Computer Science, Information Systems, Engineering, or a related field.
  • 4–8 years of experience in data integration, data engineering, with strong ETL and SQL.
  • Strong experience with integration tools such as Informatica, Talend, MuleSoft, SSIS, or Boomi.
  • Proficient in SQL, Python, and scripting for data manipulation and automation.
  • Experience with cloud data platforms (Google Cloud Platform) and services such as Google Cloud Dataflow.
  • Familiarity with REST/SOAP APIs, JSON, XML, and flat file integrations.

Preferred Skills:

  • Experience with message queues or data streaming platforms (Kafka, RabbitMQ, Kinesis).
  • Understanding of data warehousing concepts and tools (e.g., Snowflake, Redshift, BigQuery).
  • Knowledge of data security, privacy, and compliance best practices (HIPAA, GDPR, etc.).
  • Prior experience in industries like healthcare, fintech, or e-commerce is a plus.

Soft Skills:

  • Strong problem-solving and debugging skills.
  • Excellent communication and collaboration abilities.
  • Ability to manage multiple priorities and deliver in a fast-paced environment.
  • Attention to detail and a commitment to delivering high-quality work.