Cognizant is seeking a Data Engineer specializing in Talend and Databricks to design and deliver scalable data integration solutions. The role involves building large-scale data pipelines, developing ETL solutions, and processing various healthcare data formats while collaborating with cross-functional teams.
Responsibilities:
- Build and optimize large-scale data pipelines using Databricks to support data engineering and processing needs
- Develop ETL solutions using Talend Data Integration and Talend Data Mapper to handle complex transformations
- Process healthcare data formats—including X12 (834/837), JSON, XML, flat files, and Excel—to meet business and compliance requirements
- Integrate, validate, and reconcile data across MongoDB, relational databases (Oracle, SQL Server), and target systems
- Work closely with the Test Data Management team to perform ETL operations and ensure accurate data movement into Teradata
Requirements:
- Proven hands-on experience with Databricks for data engineering, ETL processing, and large-scale data workflows
- Strong expertise in Talend Data Integration and Talend Data Mapper, including complex mapping and transformation logic
- Experience working with healthcare data formats, including X12 (834/837), JSON, XML, Excel, and flat files
- Solid working knowledge of MongoDB and relational databases (Oracle, SQL Server), including data validation and query development
- Basic to intermediate skills in Java and/or Python to support custom transformations, utilities, and pipeline troubleshooting
- Experience processing EDI and MongoDB data and loading into Teradata environments
- Experience supporting test data management initiatives or large enterprise data programs
- Familiarity with data quality frameworks, metadata management, or automation of data validation processes
- Exposure to cloud data ecosystems or modern data engineering tools