CloudIngest is a company specializing in data solutions, and they are seeking an Azure Cosmos Data Engineer. The role involves building end-to-end data pipelines and working with Azure Synapse and Cosmos DB to support enterprise applications with large datasets.
Responsibilities:
- Strong experience building end-to-end data pipelines (ETL/ELT)
- Hands-on with cloud platforms (Azure preferred)
- Experience working on enterprise/global applications with large datasets
- Strong hands-on experience with Azure Cosmos DB (required)
- Min 2 years Experience with Azure Synapse Analytics (required) for analytics and data warehousing
- Experience with Azure Data Factory, Databricks (PySpark), and Data Lake
- Strong programming in Python, SQL, PySpark
- Good understanding of SQL + NoSQL databases
- Experience in international data mapping / data standardization across regions
- Exposure to distributed systems and multi-region data processing
- Understanding of data partitioning, replication, and performance optimization
- Experience supporting global applications / enterprise-wide analytics platforms
- Strong collaboration with cross-functional and global teams
- Ability to translate business requirements into data solutions
- Experience working in fast-paced, high-impact environments
- Ownership mindset (handling large-scale, business-critical pipelines)
Requirements:
- Expertise in Azure Synapse, Cosmos DB for potential role
- Experience in international data mapping, data conversion, set up
- Azure Synapse Analytics: Data integration, big data processing
- Databricks: Apache Spark, Delta Lake, notebooks
- SQL: Advanced querying, performance optimization, database management
- PySpark: ETL processes, data transformation, performance tuning
- Data Pipelines: Design, development, orchestration, monitoring
- Data Modeling & Database Design: Schema design, normalization/denormalization
- ETL/ELT Tools: Azure Data Factory
- NoSQL Database: Azure Cosmos DB
- Relational Database: SQL Server and Oracle
- Version Control Systems: Git proficiency
- Strong experience building end-to-end data pipelines (ETL/ELT)
- Hands-on with cloud platforms (Azure preferred)
- Experience working on enterprise/global applications with large datasets
- Strong hands-on experience with Azure Cosmos DB
- Min 2 years Experience with Azure Synapse Analytics for analytics and data warehousing
- Experience with Azure Data Factory, Databricks (PySpark), and Data Lake
- Strong programming in Python, SQL, PySpark
- Good understanding of SQL + NoSQL databases
- Experience in international data mapping / data standardization across regions
- Exposure to distributed systems and multi-region data processing
- Understanding of data partitioning, replication, and performance optimization
- Experience supporting global applications / enterprise-wide analytics platforms
- Strong collaboration with cross-functional and global teams
- Ability to translate business requirements into data solutions
- Experience working in fast-paced, high-impact environments
- Ownership mindset (handling large-scale, business-critical pipelines)