Veritis Group Inc is seeking a Data Engineer to work remotely. The role involves building data marts and mini data warehouses using GCP, BigQuery, and Python, with a focus on data aggregation and API development.
Responsibilities:
- Building data marts using the same GCP stack
- Involve dealing with numbers and large volumes of claims data
- Require aggregations and stuff
- Building a portal that shows holistic data for a provider
- Involves API work
- Building Python APIs to read analytics databases like BigQuery
- Candidates will heavily lean on the data side for building data marts where data is already aggregated
- Concern about the network and provider team giving the right requirements for consolidating data
- Need to reach out to multiple teams as the data is very different
- Difficulty in defining metrics like 'claims paid data'
- May need to supplement expertise from another domain
Requirements:
- Strong hands-on Python programming
- Spark/PySpark
- GCP (BigQuery, Dataproc, Google Cloud Functions, GKE, Cloud SQL - would not consider all must haves, just general awareness of the GCP ecosystem and data services)
- Experience working with various data types and structures
- Good BigQuery PySpark skills
- Experience in data warehousing
- Experience building data warehouses
- Background in working with transaction data (not just master data)
- Experience with facts and aggregations
- API building with Python
- Strong Python REST API developers
- AI experience - building AI systems, models, or building inference pipelines and processing data 'for AI'