Assemble and manage large, complex datasets that meet both non-functional and functional business requirements.
Identify, design, and implement internal process improvements to enhance scalability, optimize data delivery, and automate manual processes.
Build and maintain optimal data pipeline architecture for efficient extraction, transformation, and loading of data from various sources.
Provide technical expertise and support for data-related issues.
Partner with data and analytics experts to enhance functionality and performance of data systems.
Design and deploy frameworks and microservices to serve data assets to data consumers.
Performs root cause analysis on data and processes to identify opportunities for improvement.
Act as a trusted advisor and partner to business leads, understanding their needs, interpreting data drivers, and aligning data strategies with business goals.
Evaluate customer requirements and contribute to the development of structural necessities within the data architecture.
Ensure database implementation procedures comply with relevant regulations.
Organize data within cloud storage and implement appropriate security measures.
Collaborate cross-functionally with multiple teams within the organization.
Works independently and collaboratively in a team environment.
Excellent analytical and problem-solving abilities.
Strong attention to detail to ensure accuracy and predictability of data pipelines.
Effectively communicates with engineers, product managers, and analysts to understand data needs.
Commitment to building secure, resilient, and fault-tolerant architectures with clear documentation and support procedures.
Requirements
Bachelor’s degree required in data engineering, data analytics, computer engineering, or a related field.
Minimum of 3 years’ experience in data engineering or database analytics.
Must have strong programming skills in Python, Scala, and SQL.
Advanced SQL knowledge and experience working with relational databases.
Prior work with various data management technologies such as RedShift, Postgres, or Snowflake is preferred.