Thermo Fisher Scientific is a world leader in serving science, dedicated to impactful work and innovative thinking. The Senior Databricks Engineer will architect and optimize data solutions to support the company's digital transformation strategy, focusing on building workflows within cloud-based systems.
Responsibilities:
- Architect, build, and optimize data solutions that support Thermo Fisher Scientific’s digital transformation strategy
- Build, develop, and deploy scalable data pipelines and ETL/ELT processes using Databricks
- Engineer robust data solutions to integrate enterprise data sources, including ERP, CRM, laboratory, and manufacturing systems
- Develop reusable frameworks and templates to accelerate data delivery and ensure consistency across domains
- Implement and maintain high-performance data connections across Databricks, Snowflake, and Iceberg environments
- Author and optimize complex SQL queries, transformations, and data models for analytics and reporting use cases
- Support data Lakehouse and data mesh initiatives to enable seamless access to trusted data across the organization
- Apply data governance, lineage, and security controls using Unity Catalog, Delta Live Tables, and related technologies
- Partner with compliance and cybersecurity teams to uphold data privacy, GxP, and regulatory standards
- Establish monitoring, auditing, and optimization processes for ongoing data quality assurance
- Collaborate with data scientists, architects, and business partners to build and implement end-to-end data solutions
- Serve as a technical mentor and leader with vision within the CRG data engineering community
- Contribute to critical initiatives for digital platform modernization and advanced analytics enablement
Requirements:
- Bachelors degree or equivalent
- Minimum of 8 years professional experience in data engineering or data platform development is required
- Minimum of 5 years of hands-on experience with Databricks and Apache Spark in production environments is required
- Demonstrated expertise with Snowflake and Apache Iceberg is required
- Strong proficiency in SQL and experience optimizing queries on large, distributed datasets is required
- Proven experience with cloud-based data platforms (Azure preferred; AWS or GCP acceptable) is required
- Strong understanding of data modeling, ETL/ELT pipelines, and data governance practices is required
- Experience implementing Unity Catalog or CI/CD pipelines for data workflows
- Experience in life sciences, biotech, or manufacturing environment
- Strong interpersonal and communication skills with ability to collaborate across global and multi-disciplinary teams