Zachary Piper Solutions is a leading federal health IT solutions provider, and they are seeking a Data Engineer II to join their team. The Data Engineer II will focus on building systems, pipelines, and infrastructure to make data usable for analysts and data scientists, while contributing to challenging projects and growing their career.
Responsibilities:
- Build and maintain PySpark data pipelines in the Databricks environment, Optimize Spark jobs performance and resource usage, identifying and addressing bottlenecks and inefficiencies in backend systems
- Design, develop, and maintain high-quality backend software components and services, ensuring functionality, performance, and scalability
- Research and build proof of concepts in the data space
- Write clean, well-structured, and maintainable code, adhering to established coding standards and best practices
- Perform thorough code reviews, providing constructive feedback to peers and identifying potential risks or areas for improvement
- Debug and resolve defects, proactively identifying and addressing potential issues before they impact users
- Create and maintain comprehensive technical documentation
Requirements:
- 5+ years of related experience plus a bachelors degree in computer science, computer engineering, or a related field
- Strong technical skills including Python, Apache Spark, R, SAS, and solid understanding of software design patterns, data structures, algorithms, and distributed computing
- Experience building data pipelines (ETL/ELT) and managing databases and SQL performance
- Experience working with cloud platforms (AWS/AZURE/GCP)
- Technical Skills: SQL (heavy), Python/Scala, ETL Pipelines, Cloud Platforms (AWS/GCP/Azure), Databricks/Spark, dbt/airflow, Data Modeling (star schemas, warehouses)
- Experience working in Agile development environments
- Databricks/Apache Spark certifications