Data team is seeking a data engineering contractor to support a growing AI focus data engineering team. They are responsible for the design and development of database models, data/batch processing jobs, data manipulation, data mining, and data extraction/transformation/loading using Python. Immediate projects connect AI applications to newly built and evolving data sources. This is a hands-on, technical role suited for someone who is comfortable working with data infrastructure and application layer integration.
Responsibilities:
Write and optimize SQL queries to extract, transform, and load structured and unstructured data.
Provide scoping, estimating, planning, design, development, and support services to support project initiatives.
Design, configure, deploy and maintain Python Scripts to support AI project initiatives.
Collaborate on AWS infrastructure, including working with S3 for data storage and AWS Glue for ETL pipeline development.
Contribute to Python-based front-end applications, supporting current interfaces and preparing for a potential future migration to Angular.
Design, development, test, and implement application code.
Track progress and intervene as needed to eliminate barriers and ensure delivery.
Resolve or escalate problems and manage risk for both development and production support.
Maintain deep knowledge and awareness of technical & industry best practices and trends, especially in technology & methodologies.
Mandatory Skills:
At least 4+ years of Developer experience specifically focused on Data Engineering.
Strong Hands-on experience in Data Engineering development using Python and Pyspark as an ETL tool.
Hands-on experience in AWS services like Glue, RDS, S3, Step functions, Event Bridge, Lambda, MSK (Kafka), EKS etc.
Hands-on experience in Databases like Postgres, SQL Server, Oracle, Sybase.
Hands-on experience with SQL database programming, SQL performance tuning, relational model analysis, queries, stored procedures, views, functions and triggers.
Strong technical experience in Design (Mapping specifications, HLD, LLD), Development (Coding, Unit testing).
Good knowledge in CI/CD DevOps process and tools like Bitbucket, GitHub, Jenkins.
Strong foundation and experience with data modeling, data warehousing, data mining, data analysis and data profiling.
Strong experience with Agile/SCRUM methodology.
Good communication and inter-personal skills.
Nice to Have:
Familiarity with AI/ML data workflows or LLM-adjacent tooling.