AirflowApacheDockerJavaLinuxPythonSparkTypeScriptBashData EngineeringAnalyticsApache SparkApache AirflowPodmanGitVersion ControlSource Control
About this role
Role Overview
Design, develop, and maintain data workflows using Apache Airflow
Build and manage complex DAGs, operators, sensors, and scheduling logic
Develop scalable automation pipelines to support data engineering and analytics initiatives
Integrate workflows with distributed data processing engines such as Apache Spark
Collaborate with cross-functional teams to ensure workflows are secure, efficient, and production-ready
Troubleshoot and optimize pipeline performance and reliability
Automate manual processes using Bash and Python
Support containerized deployments using Docker, containerd, or Podman
Maintain version control and code quality using Git
Requirements
Active TS/SCI w/ Polygraph
Master’s degree in Computer Science or related field + 3 years of relevant experience or Bachelor’s degree in Computer Science or related field + 5 years of relevant experience
7 years of relevant Software Engineering experience
Strong experience using Linux CLI and Linux tools
Experience writing Bash scripts to automate processes
Recent development experience in Python and Java
Hands-on experience with Apache Airflow (DAG design, scheduling, operators, sensors)
Experience with distributed big data processing engines (e.g., Apache Spark)
Experience with container technologies (Docker, containerd, Podman)