Bayer is a company driven to solve the world’s toughest challenges and striving for a world where 'Health for all Hunger for none’ is a reality. They are seeking a Senior Data Engineer to implement data flows, support data streaming systems, and develop business intelligence reports, among other responsibilities.
Responsibilities:
- Delivers the designs set by more senior members of the data engineering community
- Implement data flows to connect operational systems, data for analytics and business intelligence (BI) systems
- Document source-to-target mappings
- Re-engineer manual data flows to enable scaling and repeatable use
- Support the build of data streaming systems
- Write ETL (extract, transform, load) scripts and code to ensure the ETL process performs optimally
- Develop business intelligence reports that can be reused
- Build accessible data for analysis
- Show an awareness of the need to translate technical concepts into non-technical language
- Understand what communication is required with internal and external stakeholders
- Undertake data profiling and source system analysis
- Present clear insights to colleagues to support the end use of the data
- Design, build and test data products based on feeds from multiple systems, using a range of different storage technologies, access methods or both
- Create repeatable and reusable products
- Show an awareness of opportunities for innovation with new tools and uses of data
- Deliver data solutions and integration pipelines in accordance with agreed organizational standards that ensure services are resilient, scalable and future-proof
- Explain the concepts and principles of data modeling
- Produce, maintain and update relevant data models for an organization's specific needs
- Reverse-engineer data models from a live system
- Work with metadata repositories to complete complex tasks such as data and systems integration impact analysis
- Maintain a repository to ensure information remains accurate and up to date
- Explain the types of problems in databases, data processes, data products and services
- Design, code, test, correct and document simple programs or scripts under the direction of others
- Understand the core technical concepts related to the role, and apply them with guidance
- Correctly execute test scripts under supervision
- Understand the role of testing and how it works
Requirements:
- Proficiency in programming language such as Python or Java
- Experience with Big Data technologies such as Hadoop, Spark, and Kafka
- Familiarity with ETL processes and tools
- Knowledge of SQL and NoSQL databases
- Strong understanding of relational databases
- Experience with data warehousing solutions
- Proficiency with cloud platforms
- Expertise in data modeling and design
- Experience in designing and building scalable data pipelines
- Experience with RESTful APIs and data integration
- Relevant certifications (e.g., GCP Certified, AWS Certified, Azure Certified)
- Bachelor's degree in Computer Science, Data Engineering, Information Technology, or a related field
- Strong analytical and communication skills
- Ability to work collaboratively in a team environment
- High level of accuracy and attention to detail