Design and implement scalable, secure data architectures for analytics and operational use cases in fraud detection and risk management.
Build and maintain efficient, reliable data pipelines and ETL/ELT workflows, optimizing for performance and cost.
Manage large, high-scale datasets; establish and enforce data integrity, lineage, and quality controls.
Collaborate with data scientists to prepare training datasets and operationalize machine learning models for fraud detection.
Utilize technologies such as Cassandra, PostgreSQL, Python, PySpark, and Databricks to deliver robust solutions.
Implement cloud-based data solutions on Microsoft Azure to support scalability, resiliency, and security.
Establish and uphold data governance policies, standards, and best practices across the data lifecycle and coordinate releases and updates across complex infrastructure and cross-functional teams, ensuring seamless deployment and minimal disruption.
Take ownership of deployments and environments, manage tasks effectively, and demonstrate accountability for results.
Requirements
10+ years of experience in data architecture and analytics, with direct focus on fraud detection and risk management.
5+ years of hands-on programming experience with Java and Python, including SQL expertise and experience with high-scale relational databases; NoSQL experience (such as Cassandra, HBase, or DynamoDB) preferred.
3+ years of technical leadership experience designing and delivering data solutions at scale and managing large data volumes and applying data management best practices, including partitioning, indexing, caching, and performance tuning.
2+ years’ experience preparing and training machine learning models using fraud data, including feature engineering and model operationalization.
Proficiency with Cassandra, PostgreSQL, Python, Spark/PySpark, Databricks, and Azure data services.
Familiarity with pipeline automation, orchestration, and incremental data loading processes and knowledge of supervised and unsupervised learning approaches and techniques such as PCA and XGBoost.
Experience in data mapping, standardization, data quality rules, and creating derived attributes for analytics.
Bachelor’s degree in data science, Computer Science, Engineering, Mathematics, or a related field, or equivalent experience.
Must currently possess valid and unrestricted U.S. work authorization to be considered for this role.
Tech Stack
Azure
Cassandra
Cloud
DynamoDB
ETL
HBase
Java
NoSQL
Postgres
PySpark
Python
Spark
SQL
Benefits
Annual incentive opportunity which may be delivered as a mix of cash bonus and equity awards