Northrop Grumman is a leading technology company that offers revolutionary systems impacting lives worldwide. They are seeking a Principal Data Engineer to design, build, and maintain data pipelines that transform enterprise-wide systems into actionable information for finance stakeholders.
Responsibilities:
- Design & Deploy Pipelines – Develop production‑grade data pipelines using Python/PySpark and SQL in Databricks to ingest, cleanse, and transform finance‑related data
- Data Integration – Build and sustain data pipeline processes from SAP ERP and other financial reporting systems
- Data Modeling – Convert high‑level design specs into physical, dimensional data models optimized for finance analytics (e.g., cost visibility, financial forecasting, earned value management)
- Reliability & Monitoring – Proactively monitor pipeline health, troubleshoot production issues, and implement fixes to guarantee high data availability and accuracy for downstream reporting
- Business Partnership – Collaborate with senior engineers, finance analysts, and business partners to understand analytical needs and translate complex questions into technical solutions
- Quality, Security & Governance – Enforce data‑quality standards, security controls, and governance best practices; maintain thorough documentation and adhere to code‑quality guidelines
Requirements:
- Candidates must have a either a Master's degree or a Bachelor's degree in Engineering, Computer Science, Software Engineering, or a similar related technical discipline
- 4 years of hands‑on data‑engineering experience is required
- 1 year focused on finance‑domain data analytics solutions is needed
- Proven experience extracting and integrating data from SAP ERP
- Experience with SQL and Python for data manipulation and transformation is needed
- Experience with Databricks or equivalent (example: Snowflake) is a must
- The ideal candidate will have a Bachelor's degree in Data Science and 6 years of experience with finance‑domain data analytics solutions
- Strong working knowledge of dimensional modeling principles
- Familiarity with specific SAP data models and modules
- Understanding of fundamental data governance and Master Data Management (MDM) concepts
- Comfortable working with Git (or comparable version‑control system)