AR5 Labs builds AI-powered automation tools for data analytics, including their flagship product, PlotStudio AI. They are seeking a Data Engineer to develop the data integration layer for PlotStudio AI, enabling users to connect their existing data sources seamlessly and enhancing the platform's capabilities.
Responsibilities:
- Design and build production-ready connectors to major data warehouses (Snowflake, BigQuery, Redshift, Databricks)
- Develop integrations with relational and NoSQL databases (PostgreSQL, MySQL, MongoDB)
- Build and maintain API integrations with common SaaS platforms (HubSpot, Salesforce, Stripe, Google Analytics, Shopify, and others)
- Develop ETL/ELT pipelines for ingesting, validating, and transforming data from external sources into formats our AI agents can analyze
- Define data schemas, validation rules, and error handling strategies for diverse and unpredictable data sources
- Implement secure authentication flows (OAuth 2.0, API keys, service accounts) for third-party data connections
- Collaborate with the AI/ML team to ensure ingested data integrates cleanly with PlotStudio's data profiling, cleaning, and analysis pipeline
- Optimize connector performance for large datasets, including incremental syncs and query pushdown where applicable
- Write technical documentation for connector APIs and integration patterns
Requirements:
- 2+ years of experience in data engineering, backend engineering, or a similar role focused on data systems
- Strong proficiency in Python and SQL
- Hands-on experience with at least one major cloud data warehouse (Snowflake, BigQuery, Redshift, or Databricks)
- Familiarity with ETL/ELT tools and frameworks (dbt, Airflow, Fivetran, or similar)
- Experience building and consuming REST APIs and working with third-party API documentation
- Solid understanding of data modeling, schema design, and data quality practices
- Comfortable working with cloud platforms (AWS, GCP, or Azure)
- Clear written and verbal communication skills — you can explain technical decisions to non-technical teammates
- Experience with Electron or desktop application architectures
- Familiarity with real-time data streaming tools (Kafka, Spark Streaming)
- Experience implementing OAuth 2.0 flows and managing third-party API authentication securely
- Exposure to LLMs, AI-powered applications, or agentic AI systems
- Contributions to open-source data tools or connector frameworks