Administer and maintain the Databricks platform, including workspace management and cluster optimization
Manage and optimize Unity Catalog for data governance and access control
Build and support ETL and ELT data pipelines using Spark and cloud native tools
Optimize Delta Lake architecture and overall lakehouse performance
Support AI and machine learning workflows within the platform
Implement and maintain CI/CD pipelines for data and ML operations
Manage IAM and RBAC across AWS, Azure, and GCP environments
Ensure platform security and compliance with DoD data policies
Monitor system performance and troubleshoot data platform issues
Collaborate with engineering and analytics teams to support data driven solutions
Requirements
Master’s degree in Computer Science, Information Technology, Data Science, or related field OR 11 years of equivalent experience in data platforms and cloud engineering
Experience administering Databricks in an enterprise environment
Strong experience with ETL or ELT pipeline development
Experience working with Delta Lake and lakehouse architectures
Hands-on experience with Python and SQL or NoSQL databases
Experience with Apache Spark and distributed data processing
Experience with cloud platforms such as AWS, Azure, or GCP
Experience implementing CI/CD pipelines and DevSecOps practices
Working knowledge of IAM, RBAC, and cloud security principles