Design, build, and scale robust ETL pipelines to support complex data workflows while ensuring high performance, reliability, and adaptability to evolving business needs.
Automate and manage data ingestion from diverse sources (databases, APIs, cloud platforms), ensuring system resilience, fault tolerance, and failover readiness.
Optimize data storage, processing, and retrieval layers to balance performance, scalability, and cost efficiency across the data platform.
Modernize and enhance legacy data systems by identifying gaps, implementing architectural improvements, and aligning solutions with future business requirements.
Lead technical excellence within the data engineering function through mentorship, code reviews, best-practice enforcement, and adoption of advanced tools and frameworks.
Ensure end-to-end data quality, integrity, and governance by implementing validation, monitoring, testing, and compliance-focused data controls.
Collaborate cross-functionally with analytics, product, engineering, DevOps, and business stakeholders to translate requirements into scalable data models and transformations.
Drive a data-driven culture and long-term data strategy by enabling self-service analytics, maintaining clear documentation, leading training initiatives, and contributing to architecture roadmaps and governance policies.
Requirements
10+ years of hands-on experience in Data Engineering
Bachelor’s degree in Data Engineering, Computer Science, Data Analytics, or a related field is required. Master’s degree preferred.
Advanced proficiency in Python and SQL, with proven experience in ETL pipeline development.
Experience with cloud data platforms such as AWS, GCP, or Azure, including cloud-native data engineering tools and services.
Strong understanding of modern data architecture patterns, including batch processing, streaming, and event-driven systems, along with industry best practices.
Demonstrated ability to optimize data workflows, troubleshoot complex data issues, and ensure high performance, scalability, and reliability of data systems.
Strong project management skills, with the ability to work independently, manage priorities, and deliver high-quality outcomes in a fast-paced environment.
Nice to Have Hands-on experience with modern data platforms and tools such as data warehouses/lakehouses (Snowflake, Redshift, BigQuery), big data or streaming frameworks (Spark, Kafka), and data orchestration tools (Airflow or equivalents).
Previous experience working with large-scale healthcare or insurance organizations.
Tech Stack
Airflow
Amazon Redshift
AWS
Azure
BigQuery
Cloud
ETL
Google Cloud Platform
Kafka
Python
Spark
SQL
Benefits
Culture of Relentless Performance: Join a technology-driven team with a 99% project success rate and over 30% year-over-year revenue growth.
Competitive Pay and Benefits : Comprehensive compensation package including health insurance and relocation support.
Work From Anywhere Culture : Flexibility to work remotely.
Growth Mindset: Access to certification programs, mentorship, internal mobility, and continuous learning opportunities.
Global Impact: Work on impactful projects with leading global clients.
Welcoming Multicultural Environment: Inclusive, collaborative, and supportive workplace with regular team-building activities.
Social Sustainability Values: Commitment to IT education, community empowerment, fair practices, environmental sustainability, and gender equality.