Clara IT Systems is seeking a Data Engineer to design, build, and optimize scalable ETL/ELT pipelines. The role involves developing data pipelines for batch and real-time processing while ensuring data quality and collaborating with cross-functional teams including data scientists and analysts.
Responsibilities:
- Design, build, and optimize scalable ETL/ELT pipelines
- Develop data pipelines for batch and real-time processing
- Work with large-scale structured & unstructured datasets
- Implement solutions across AWS / Azure / GCP environments
- Ensure data quality, governance, and performance optimization
- Collaborate with cross-functional teams including data scientists and analysts
Requirements:
- Strong programming in Python / Java / Scala
- Advanced SQL skills
- Experience with Apache Spark / PySpark
- Hands-on with at least one cloud platform (AWS / Azure / GCP)
- Experience with Airflow / Kafka or similar tools
- Snowflake / Redshift / BigQuery
- Docker / Kubernetes
- CI/CD pipelines
- Data modeling & warehousing concepts