Develop and maintain data pipelines based on the business requirements in the Hadoop ecosystem by following the coding standards and procedures
Perform Unit/SIT/UAT testing and work with team members to debug and evaluate the business results
Profile data to assist with defining the data elements, propose business term mappings, and define data quality rules
Provide data analysis for data ingestion, standardization and curation efforts ensuring all data is understood from a business context
Work closely with Product Owners, Product Managers, Program Manager, Scrum Masters and Data Engineers in a Scaled Agile framework
Work with the Data Office to ensure the lineage of all data assets are properly documented in the appropriate enterprise metadata repositories
Requirements
Bachelor's degree in a quantitative field (such as Engineering, Computer Science) with 1+ years’ experience or in lieu of degree 3+ years of Information Technology experience within financial services
Must have knowledge on building graphs, dml functions, psets ,plans and ability to write abstracted, reusable code components in Ab initio GDE
Must have knowledge in writing shell scripts, complex sql queries and familiar with CI-CD tools such as CloudBees and Bitbucket
Must have knowledge with any of the databases such as Hive, Oracle and MySQL
Must have knowledge on the any of the scheduling tools like StoneBranch ,Control Center or any other similar scheduling tools
Good to have experience working in Financial Industry or Credit processing experience
Good to have knowledge on Cloud such as AWS services
Hands on experience with Ab initio ETL tool and any of the databases such as Hive, Oracle and MySQL
Excellent written and oral communication skills & Experience working iteratively in a fast-paced agile environment.
Tech Stack
AWS
Cloud
ETL
Hadoop
MySQL
Oracle
SQL
Benefits
best-in-class employee benefits and programs that cater to work-life integration and overall well-being