Capnexus is a comprehensive services provider specializing in retail software solutions. They are seeking a highly skilled Senior AWS Data Engineer to lead data architecture, pipeline development, and ERP integration for an AI-powered modernization project in the construction industry.
Responsibilities:
- Lead data architecture design, API assessment, and ETL requirements gathering during the Discovery & Design phase
- Develop and configure CMIC ERP API integration to establish reliable data exchange between the ERP system and the AWS platform
- Design and implement data pipelines using AWS Glue for ETL processing of subcontractor documents and ERP data
- Integrate Amazon Textract to extract structured data from insurance certificates, bonding letters, and financial documents
- Build and maintain data models to support AI-powered validation, risk profiling, and executive reporting
- Configure Amazon S3 data lake architecture to store and manage raw, processed, and curated data assets
- Implement AWS Lambda and AWS Step Functions to orchestrate data workflows and automated processing pipelines
- Develop and expose data through Amazon API Gateway to support application and dashboard consumption
- Ensure data quality, validation, and integrity across all integration points and pipeline outputs
- Conduct data integration testing and support user acceptance testing (UAT) for data-dependent features
- Collaborate with Full Stack, AI/ML, and DevOps team members to ensure seamless end-to-end data flows
- Contribute to knowledge transfer documentation, data pipeline runbooks, and operations guides
Requirements:
- 5+ years of data engineering experience, with at least 2+ years working in AWS cloud environments
- Strong proficiency in SQL and experience with AWS database services including RDS, DynamoDB, and Aurora
- Hands-on experience with AWS Glue for ETL development and data pipeline orchestration
- Experience integrating with ERP systems or enterprise APIs, with ability to assess and document API capabilities
- Working knowledge of Amazon Textract for automated document data extraction
- Proficiency with Amazon S3 for data lake design, storage management, and lifecycle policies
- Experience with AWS Lambda and AWS Step Functions for serverless data workflow orchestration
- Solid understanding of data modeling, data quality validation, and integration testing practices
- Familiarity with Amazon API Gateway for exposing data services to downstream consumers
- Strong analytical, problem-solving, and communication skills with the ability to work in Agile/Scrum teams
- Experience with Amazon Bedrock or AI/ML-integrated data pipelines
- Familiarity with CMIC ERP or similar construction industry enterprise resource planning systems
- Experience with GraphQL APIs or AWS AppSync
- Knowledge of TypeScript or Python for pipeline scripting and Lambda function development
- AWS Certification (Data Analytics Specialty, Database Specialty, or Solutions Architect)
- Background in construction, insurance, or financial services industries