Create and optimize end-to end data pipelines and perform and enhance real-time data processing.
Identify client needs and provide customized solutions across many major data platforms (Snowflake, Databricks, and more) atop all major cloud platforms.
Research and prototype new technologies using EMR, Kafka, Spark, Airflow, Flink, etc.
Perform coding using various programming and scripting languages to automate data ingestion and improve data management processes.
Support and develop data flows by developing processes that verify, standardize, and scale data input, transformation and storage.
Design and implement Data Architectures, customize storage and extraction, metadata, and information repositories.
Utilize programming languages like Kafka, Java, Scala, Python and Open Source RDBMS and NoSQL databases and Cloud based data warehousing services such as Snowflake and Redshift.
Work on a variety of projects involving data architecture, data models, data migration, data integration, data analysis, and visualization.
Work with clients from Latin America and communicate with them in Spanish regarding complex technological concepts.
Serve as subject matter expert on data engineering and management best practices.
Stay on top of tech trends, experimenting with and learning new technologies, participating in internal & external technology communities, and mentoring other members of the engineering community.
Work with Data Scientists and Data Owners to understand use cases and hypotheses to test with data.
Establish data governance models and contribute to strategies for enterprise data management.
Perform extensive data profiling and analysis based on the client's data and needs.
Work with software development teams and stakeholders to define reporting requirements.
Support project delivery on Data Warehouse/BI projects for external clients, including partnering with subject matter experts on project execution.
Support clients in developing and implementing data engineering process improvement plans.
Requirements
Education: Bachelor’s degree in Information Technology or a related field
Experience: 4 years, including 4 years of working in the position of Data Engineer, Software Developer or related experience
experience using traditional RDMS systems (MSSQL, Oracle, Postgres, MySQL)
experience building Continuous Integration/Continuous Deployment + infrastructure as code
Data Architecture experience
experience using Kafka, Java, Scala, Python
experience using NoSQL databases and Cloud based data warehousing services such as Redshift or Databricks
experience working within a team in an Agile development environment, managing stakeholders, defining requirements and participating in Agile ceremonies
experience translating technical solutions into business outcomes
experience developing pipelines in Apache Airflow
experience using Snowflake designing data pipelines and building analytics solutions for cloud data projects
Special Skills or Other Requirements: Fluent in Spanish (to communicate with clients from Latin America)
Tech Stack
Airflow
Amazon Redshift
Apache
Cloud
Java
Kafka
MySQL
NoSQL
Open Source
Oracle
Postgres
Python
RDBMS
Scala
Spark
Benefits
Competitive salary – Great starting salaries for well qualified candidates
Generous Paid Time Off policy
Ippon offers flexible time off to help you be your best
Health, dental, and vision insurance – We are dedicated to helping our employees live healthier lives through comprehensive health programs
401k with company match
Ippon offers an industry leading 401(k) matching plan
Family Leave
We support the importance of family and offer maternity, paternity and family medical leave plans
A fun, smart, and healthy work environment, focused on our values and teamwork