Data Architect (SME Level)

Active TS/SCI clearance required.

We’re hiring a Data Engineer to assess, modernize, and operate production data pipelines that pull from enterprise operational systems and turn them into reliable, integrated datasets analysts and decision-makers can trust. A lot of this work is detective work: source systems often have limited documentation, non-standard schemas, and quirky APIs. We’re looking for someone with strong fundamentals and the curiosity to reverse-engineer a messy system until it makes sense — not someone who needs a clean spec to get started. You’ll work hands-on with Python, SQL, and orchestration frameworks like Airflow or Prefect, design dimensional models, and build the integrations that make the data usable.

Highlights

  • Strong Python and advanced SQL — complex queries, optimization, performance tuning
  • ETL/ELT from diverse sources: SaaS platforms, databases, files, streams
  • Comfortable working with semi-structured data (JSON, XML) and unfamiliar schemas
  • Cloud data platforms — AWS, Azure, or GCP
  • Dimensional modeling and data warehouse fundamentals
  • Data quality, validation, and pipeline observability built in from the start
  • Production pipelines with Airflow, Prefect, or similar orchestration
  • API integrations — OAuth/SSO, rate limiting, pagination, retry logic, error handling
  • Git-based workflows and solid software engineering habits

Bonus

  • Experience extracting data from IT operations or service management systems (ServiceNow, network management platforms, monitoring tools)
  • PySpark for distributed processing; dbt for transformations
  • Terraform or CloudFormation; CI/CD for data pipelines
  • Streaming (Kafka, Kinesis); data quality tooling (Great Expectations, Soda, Monte Carlo)
  • Docker/Kubernetes for data workloads

Location: Washington, DC metro area

To apply for this job email your details to jobs@cleverdba.com