
Job Title: Data Engineer
Job Description: • Design, build, and maintain scalable data pipelines using PySpark and Python • Develop and optimize complex SQL queries for large datasets • Implement and manage ETL/ELT processes ensuring data quality and reliability • Collaborate with business and product teams to translate data requirements into solutions • Build and maintain data warehouse solutions • Handle large-scale data processing using Hadoop/Big Data technologies • Perform performance tuning and optimization of data workflows Required Skills: • Strong hands-on experience with PySpark and Python • Advanced proficiency in SQL • Solid experience in ETL processes and data warehousing • Familiarity with Hadoop ecosystem and Big Data technologies • Experience working with large datasets in distributed environments • Good communication and business understanding Good to Have: • Experience with Apache Airflow • Exposure to cloud platforms (AWS, GCP, Azure) • Knowledge of data lakes and modern data architectures • Experience with streaming tools (Kafka, Spark Streaming)