International Business Consultants LLC (IBC) is a global strategic consulting firm that helps organizations achieve digital transformation. The Data Engineer will develop and implement machine learning models, analyze datasets, and collaborate with various teams to deliver scalable data solutions.
Responsibilities:
- Develop and implement machine learning models for prediction, ranking, and decision support (e.g., disposition estimation, cycle time forecasting, and precedent search)
- Analyze large-scale structured and unstructured datasets to extract features, identify patterns, and support model development
- Partner with data engineers, AI engineers, and business stakeholders to translate real-world problems into scalable modeling solutions
- Evaluate model performance using appropriate metrics, historical outcomes, and edge case analysis; iterate to improve accuracy and reliability
- Design and maintain feedback loops to incorporate user actions (e.g., overrides, corrections) into continuous model improvement
- Communicate model behavior and results through clear visualizations, summaries, and explainability techniques
- Document methodologies, assumptions, and model behavior to support transparency and reproducibility
- Design, build, and maintain scalable batch and streaming data pipelines for ingesting and processing large datasets
- Transform, model, and optimize data for analytics, reporting, and downstream applications
- Implement data validation, monitoring, and security controls to ensure data quality, reliability, and compliance
- Contribute to the design and evolution of the data platform with a focus on scalability, performance, and maintainability
- Collaborate with BI, analytics, AI, and product teams to deliver data solutions aligned with business needs
- Develop automated workflows and observability mechanisms to ensure pipeline reliability and system visibility
- Create and maintain documentation for pipelines, data models, and platform components
- Design and implement LLM-powered workflows to extract and transform data from unstructured and semi-structured sources (e.g., documents, images, spreadsheets)
- Build and orchestrate AI-driven processes for tasks such as document parsing, classification, and entity extraction
- Develop and optimize LLM-based solutions using prompting, retrieval, and (where appropriate) fine-tuning
- Integrate AI workflows into production systems, ensuring reliability, scalability, and maintainability
- Evaluate LLM outputs for accuracy, consistency, and edge cases; implement improvements through iteration and testing
- Collaborate with data engineers, software engineers, and business stakeholders to translate requirements into technical solutions
- Ensure adherence to security, governance, and responsible AI standards
Requirements:
- Expertise in Data Engineering, with experience in creating and optimizing data pipelines
- Proficiency in Data Modeling and foundational knowledge of database design principles
- Hands-on experience with Extract, Transform, Load (ETL) processes and Data Warehousing solutions
- Strong skills in Data Analytics and working knowledge of data visualization tools
- Proven problem-solving and debugging abilities in data processing and pipeline management
- Bachelor's degree in Computer Science, Data Science, Engineering, or related field
- Strong communication skills and ability to work collaboratively in cross-functional teams
- Familiarity with cloud data platforms (e.g., AWS, Azure, Google Cloud) is a plus
- Experience with programming languages such as Python, SQL, or Java is highly desirable