Greystar is a leading global real estate platform that offers expertise in property management and development. They are seeking a Senior Data Engineer/Architect to join their D2AI team, focusing on leveraging Azure SQL and Databricks to develop data capabilities for customer-facing applications in the real estate sector.
Responsibilities:
- 100% hands-on development – Azure SQL, Cosmos DB, Databricks, and Azure OpenAI: develop and unit test database and AI pipeline code, including T-SQL, stored procedures, functions, views, and LLM prompt orchestration layers
- Own and maintain the Databricks data ingestion & output pipelines for end programs such as greystar.com and Microsoft Customer Insights CDP platform, including Delta Lake table optimization, schema evolution, and medallion architecture (Bronze/Silver/Gold) design
- Architect and maintain AI-ready data structures — clean, well-typed, and optimized for feature engineering and model consumption
- Participate in the design of databases and feature stores, applying normalization or denormalization as appropriate to support both operational and ML workloads
- Create, deploy, and maintain ADF and Databricks Workflows pipelines, adhering to Greystar's standards and documented best practices
- Perform analysis of complex data and document findings, leveraging AI-assisted tooling and notebooks to surface insights faster
- Prepare and curate data for prescriptive and predictive modeling — including feature extraction, data labeling pipelines, and training/test dataset construction
- Combine raw data from disparate external sources; build and support complex ingestions including real-time streaming (Event Hub) and batch patterns
- Collaborate closely with data scientists, ML engineers, and application developers consuming the data — ensuring outputs are well-documented, versioned, and model-ready
- Integrate and support AI/ML model outputs back into data products — scoring pipelines, inference result storage, and feedback loops for model monitoring
- Play a direct role in the maintenance, technical support, documentation, and administration of databases and Databricks environments, including Unity Catalog governance
- Ensure standards are followed by participating in code reviews — including review of AI prompt logic, notebook structure, and pipeline configuration
- Act as a thought leader: bring forward new ideas in AI, data engineering, and LLM integration — and then build them
Requirements:
- 6+ years relevant and progressive data engineering experience
- Deep Technical knowledge and experience in Microsoft Azure architecture, including Azure PaaS databases, Synapse, ADF pipelines, Azure functions, Event Grids etc
- 3+ Years of Experience with Cosmos DB
- 3+ Years of Experience with Data Bricks
- Hands-on skills working with data pipelines using SQL and No-SQL databases
- Minimum of 1 year of relevant experience working with Azure Data Lakes Gen 2
- Experience with Power Platform / Power BI
- Experience in engineering practices such as code refactoring, design patterns, CI/CD, and highly scalable data applications
- Experience developing batch ETL pipelines; real-time pipelines are a plus
- Knowledge of advanced data engineering concepts such as dimensional modeling, ETL, data governance, data warehousing, structured and unstructured data
- Good Experience with Python, machine learning frameworks and statistics
- Knowledge of Agile software development process
- Excellent problem-solving skills and experience
- Strong communication and collaboration skills
- "Self-starter" attitude and the ability to make decisions with minimal guidance from others
- Innovative and passionate about your work and the work of your teammates
- Ability to comprehend and analyze operational systems and ask appropriate questions to determine how to improve, migrate or modify the solution to meet business needs
- Bachelor's Degree in computer science, information technology, business management, information systems, or equivalent experience
- Experience / Familiarity with D365 Customer Insights platform / Dataverse is a plus
- Databricks or other Certifications are nice to have but not required
- Advanced degree preferred