Dataflux Architect

Dataflux Architect

Posted Today by Edge Global

Negotiable
Undetermined
Remote
Remote or Florida

Summary: The Dataflux Architect role involves leading the architectural strategy and migration of enterprise data quality, MDM, and data integration workloads from SAS DataFlux to the Databricks Lakehouse Platform. This hands-on, client-facing position requires expertise in DataFlux and modern data architecture, focusing on translating legacy systems into scalable solutions. The architect will also provide technical leadership and mentorship to engineering teams while ensuring alignment with enterprise architecture and governance standards.

Key Responsibilities:

  • Lead end-to-end solution architecture for the DataFlux to Databricks migration, including current-state assessment, gap analysis, target-state design, and migration roadmap.
  • Reverse-engineer and document existing DataFlux jobs, data services, business rules, QKBs, and MDM hub configurations to produce a complete logical inventory.
  • Design the target Databricks Lakehouse architecture with Delta Lake, Unity Catalog governance, and Delta Live Tables pipelines.
  • Define the strategy for migrating standardization, parsing, matching, clustering, and survivorship logic from DataFlux into Databricks-native patterns.
  • Architect MDM target-state for party, product, location, and reference data domains; define golden record logic, hierarchy management, and stewardship workflows.
  • Establish data quality frameworks using Delta Live Tables expectations, Great Expectations, or Databricks Lakehouse Monitoring.
  • Partner with enterprise architecture, data governance, and security teams to align on Unity Catalog design, lineage, RBAC, and PII handling.
  • Provide technical leadership and mentorship to a blended onshore/offshore engineering team; conduct design reviews and enforce engineering standards.
  • Serve as the senior client-facing technical advisor - present architecture decisions, trade-offs, and migration progress to stakeholders.
  • Own technical risk identification and mitigation across the migration lifecycle, including cutover strategy and decommissioning of DataFlux infrastructure.

Key Skills:

  • 10+ years of enterprise data architecture experience, with a minimum of 5 years of hands-on experience designing and deploying solutions on SAS DataFlux.
  • Deep working knowledge of DataFlux Data Management Server, Architect jobs, Profile jobs, data services, and the QKB.
  • Demonstrated experience with DataFlux match codes, clustering, entity resolution, and survivorship rule design.
  • Strong architectural experience across MDM domains - Customer/Party, Product, Location, Vendor, Employee, and Reference Data.
  • Hands-on experience with at least one enterprise MDM platform in addition to DataFlux.
  • Production experience architecting solutions on Databricks, including Delta Lake, Unity Catalog, and Delta Live Tables.
  • Strong PySpark and Spark SQL skills; able to design performant patterns for large-scale workloads.
  • Broad fluency across data quality, data governance, data modeling, and metadata management.
  • Experience implementing data governance tooling and familiarity with regulatory and privacy frameworks.
  • Prior experience leading DataFlux modernization or sunset programs is preferred.

Salary (Rate): undetermined

City: undetermined

Country: undetermined

Working Arrangements: remote

IR35 Status: undetermined

Seniority Level: undetermined

Industry: IT

Detailed Description From Employer:

Title: Dataflux Architect
Location: NYC, NY(Remote)

Job Type: Contract

Job Description :

Architect (ATC)Engagement OverviewOur client is undertaking a strategic modernization initiative to migrate their enterprise data quality, MDM, anddata integration workloads from SAS DataFlux (dfPower Studio, Data Management Studio, and the DataFlux DataManagement Server) to the Databricks Lakehouse Platform.

We are seeking a Senior Onshore DataFlux SolutionArchitect to lead the architectural strategy, target-state design, and migration blueprint for this multi-phaseprogram.

This role is a hands-on, client-facing leadership position responsible for translating legacy DataFluxlogic, business rules, and MDM constructs into a modern, scalable Databricks-native architecture leveraging DeltaLake, Unity Catalog, and Delta Live Tables.

Key Responsibilities

Lead end-to-end solution architecture for the DataFlux to Databricks migration, including current-stateassessment, gap analysis, target-state design, and migration roadmap.

Reverse-engineer and document existing DataFlux jobs, data services, business rules, QKBs (QualityKnowledge Bases), and MDM hub configurations to produce a complete logical inventory.

Design the target Databricks Lakehouse architecture (medallion: bronze/silver/gold) with Delta Lake, UnityCatalog governance, and Delta Live Tables pipelines that replicate or improve upon DataFlux DQ and MDMfunctionality.

Define the strategy for migrating standardization, parsing, matching, clustering, and survivorship logic from DataFlux into Databricks-native patterns

(PySpark, SQL, and partner tools such as Reltio, Informatica CDQ,or Zingg where appropriate).

Architect MDM target-state for party, product, location, and reference data domains; define golden recordlogic, hierarchy management, and stewardship workflows on the Lakehouse.

Establish data quality frameworks (DQ rules, scorecards, exception handling) using Delta Live Tablesexpectations, Great Expectations, or Databricks Lakehouse Monitoring as DataFlux replacements.

Partner with the client's enterprise architecture, data governance, and security teams to align on UnityCatalog design, lineage, RBAC, and PII handling.

Provide technical leadership and mentorship to a blended onshore/offshore engineering team; conduct designreviews and enforce engineering standards.

Serve as the senior client-facing technical advisor - present architecture decisions, trade-offs, and migrationprogress to Director and VP-level stakeholders.

Own technical risk identification and mitigation across the migration lifecycle, including cutover strategy,parallel run validation, and decommissioning of DataFlux infrastructure.

Required Qualifications

DataFlux Expertise (Non-Negotiable)

10+ years of enterprise data architecture experience, with a minimum of 5 years of hands-on experiencedesigning and deploying solutions on SAS DataFlux (dfPower Studio and/or Data Management Studio).

Deep working knowledge of DataFlux Data Management Server, Architect jobs, Profile jobs, data services,and the QKB (Quality Knowledge Base) - including authoring custom definitions, regex libraries, phonetics,and locale-specific rules.

Demonstrated experience with DataFlux match codes, clustering, entity resolution, and survivorship rule design at enterprise scale.

Proven ability to reverse-engineer complex, undocumented DataFlux job flows and translate them intomodern equivalents.Master Data Management (MDM)

Strong architectural experience across MDM domains - Customer/Party, Product, Location, Vendor, Employee, and Reference Data.

Hands-on experience with at least one enterprise MDM platform in addition to DataFlux: Informatica MDM,Reltio, Profisee, IBM InfoSphere MDM, or Stibo STEP.

Expertise in match/merge logic, golden record creation, hierarchy management, cross-reference (XREF)design, and data stewardship workflows.Databricks & Modern Data Stack

Production experience architecting solutions on Databricks, including Delta Lake, Unity Catalog, Delta LiveTables, Workflows, and the medallion architecture pattern.

Strong PySpark and Spark SQL skills; able to design performant patterns for large-scale matching, deduplication, and DQ workloads.

Working knowledge of cloud platforms (Azure, AWS, or Google Cloud Platform) and modern ingestion tools (Fivetran, ADF,Airflow, dbt).Data Domains & Governance

Broad fluency across data quality, data governance, data modeling (3NF, dimensional, Data Vault), and metadata management.

Experience implementing data governance tooling (Collibra, Alation, Atlan, or Unity Catalog-native governance).

Familiarity with regulatory and privacy frameworks (HIPAA, GDPR, CCPA, SOX) and their impact on MDM and DQ design.

Preferred Qualifications

Prior experience leading at least one DataFlux modernization or sunset program.

Databricks certifications (Data Engineer Professional, Solutions Architect Professional).

Experience in healthcare payer, financial services, or insurance verticals.

Background in consulting or professional services - comfortable with SOW-driven delivery and billableutilization expectations.