The Data Engineer will support the design, development, and optimization of modern data pipelines and cloud-based data platforms. This role involves building scalable ETL/ELT processes, integrating structured and unstructured data sources, and enabling secure, reliable data delivery for analytics, machine learning, and operational systems.
Requirements
- Design and implement robust ETL/ELT data pipelines using tools such as Apache Airflow, Spark, dbt, Kafka, or Databricks.
- Develop and optimize workflows that ingest, transform, and publish data from diverse sources into data lakes and data warehouses.
- Support batch and real-time data integration using scalable streaming frameworks and distributed processing engines.
- Manage infrastructure-as-code deployments and collaborate with DevOps teams to ensure data pipeline reliability, scalability, and security.
- Work closely with data scientists, analysts, and product teams to deliver data that supports BI dashboards, ML pipelines, and reporting.
- Apply best practices in data governance, schema versioning, and metadata management.
- Ensure systems are compliant with applicable data privacy regulations and security frameworks, including FISMA and Section 508.
- Document pipeline workflows, technical specifications, and data lineage to support auditability and transparency.
Benefits
- Generous medical, dental, and vision plans
- Remote working opportunities
- Flexibility to balance quality work and personal life