Job Description:
• Administer and maintain the Databricks platform, including workspace management and cluster optimization
• Manage and optimize Unity Catalog for data governance and access control
• Build and support ETL and ELT data pipelines using Spark and cloud native tools
• Optimize Delta Lake architecture and overall lakehouse performance
• Support AI and machine learning workflows within the platform
• Implement and maintain CI/CD pipelines for data and ML operations
• Manage IAM and RBAC across AWS, Azure, and GCP environments
• Ensure platform security and compliance with DoD data policies
• Monitor system performance and troubleshoot data platform issues
• Collaborate with engineering and analytics teams to support data driven solutions
Requirements:
• Master’s degree in Computer Science, Information Technology, Data Science, or related field OR 11 years of equivalent experience in data platforms and cloud engineering
• Experience administering Databricks in an enterprise environment
• Strong experience with ETL or ELT pipeline development
• Experience working with Delta Lake and lakehouse architectures
• Hands-on experience with Python and SQL or NoSQL databases
• Experience with Apache Spark and distributed data processing
• Experience with cloud platforms such as AWS, Azure, or GCP
• Experience implementing CI/CD pipelines and DevSecOps practices
• Working knowledge of IAM, RBAC, and cloud security principles
Benefits:
• None specified