Blueprint is a technology solutions firm headquartered in Bellevue, Washington, focused on solving complex problems for organizations. They are seeking a Senior Data Engineer to design, develop, and maintain scalable data infrastructure solutions while ensuring efficient data flow across systems.
Responsibilities:
- Architect, implement, and maintain scalable data architectures to meet client data processing and analytics needs
- Collaborate with cross-functional teams to understand data requirements and translate them into effective data pipeline solutions
- Develop, optimize, and maintain ETL processes to ensure the timely and accurate movement of data across systems
- Implement best practices for data pipeline orchestration and automation using tools like Azure Data Factory or Apache Airflow
- Leverage cloud platforms (Azure and/or AWS) to build and optimize data solutions, including services like Azure Synapse Analytics, Azure Blob Storage, AWS Redshift, S3, or Glue
- Utilize Databricks for big data processing, analytics, and machine learning workflows
- Lead Unity Catalog migration efforts, ensuring seamless transition and optimal data organization for governance and access control
- Establish and enforce data governance policies and procedures, ensuring data quality, integrity, and accuracy
- Optimize data processing and query performance for large-scale datasets within Databricks and cloud environments
- Document data engineering processes, architecture, and configurations for future scalability
- Collaborate with data scientists, analysts, and other stakeholders to provide the necessary infrastructure for their needs
Requirements:
- Bachelor's or Master's degree in Computer Science, Information Technology, or a related field
- Minimum of 5 years of experience in data engineering roles
- Proven expertise in Databricks, including Unity Catalog migration for data governance and organization
- Proficiency in Azure (e.g., Azure Blob Storage, Azure Synapse Analytics, Azure Functions) and/or AWS (e.g., S3, Redshift, Glue, EMR, Lambda)
- Strong programming skills in languages such as Python, Scala, or Java
- Experience with data modeling, schema design, and database optimization
- Proficiency in using orchestration tools like Azure Data Factory or Apache Airflow
- Familiarity with version control systems and collaboration tools
- Strong troubleshooting skills and the ability to implement effective solutions for complex data issues
- Strong communication and interpersonal skills, with the ability to work collaboratively in a team-oriented environment
- Proactive in staying updated with industry trends and emerging technologies in data engineering