Midwest Loan Services is seeking a Data Engineer to build a modern, scalable data platform as part of their transformation into a cloud-first, API-first enterprise. The role involves developing secure data ingestion pipelines, curating data for analytics, and supporting the modernization of the data ecosystem.
Responsibilities:
- Design and implement scalable data pipelines in Azure using Databricks, Spark, Delta Lake, DBT, Dagster, Airflow, and Parquet
- Develop data ingestion workflows from multiple sources (e.g. SFTP, vendor APIs) into Azure Data Lake
- Build and maintain Bronze/Silver/Gold layer transformations within a medallion architecture
- Implement data quality, data deduplication, and validation logic across ingestion layers
- Develop parameterized and reusable notebooks for batch and streaming jobs
- Create robust merge/update logic into Delta Lake using optimized key-based partitioning
- Collaborate with business and application teams to identify data integration requirements
- Support integration with downstream APIs, Power BI models, and SQL-based reporting
- Implement monitoring, logging, and lineage for data pipelines using tools like Unity Catalog and Azure Monitor
- Participate in code reviews, technical design discussions, and backlog grooming
- SQL Server Development & Optimization – Develop, optimize, and maintain SQL Server stored procedures, functions, views, and indexing strategies to ensure high-performance data processing
- ETL & Data Integration – Design and manage ETL/ELT processes using SQL Server Integration Services (SSIS) and SQL batch jobs to extract, transform, and load data efficiently
Requirements:
- Bachelor's degree in computer science, Engineering, or a related field (or equivalent experience)
- 5+ years of hands-on experience building data pipelines and distributed data systems
- Strong experience in Databricks, Delta Lake, and Azure-based big data tools
- Familiarity with version control (Git), CI/CD practices, and agile delivery
- SQL Server & Database Management – Expertise in T-SQL, stored procedures, indexing, performance tuning, and query optimization
- Strong understanding of data ingestion patterns and partitioning strategies
- Proficient in PySpark/SQL with a focus on performance tuning
- Solid grasp of modern data lake architecture and structured streaming
- Excellent problem-solving and debugging skills
- Ability to collaborate with business, product, and technology teams
- Strong communication skills and documentation discipline
- Experience with financial or regulated data environments preferred
- Experience in the mortgage servicing or mortgage lending industry is a plus
- Streaming & Big Data (Future Roadmap): Delta Lake, Databricks, Kafka (preferred but not required)