Work with stream-processing, batch processing projects/frameworks/tools to create data workflows
Develop and maintain data pipelines from databases and SaaS applications
Write complex SQL queries and work with NoSQL databases
Perform unit testing, system integration testing and assist with user acceptance testing
Create and maintain detailed documentation of the technical design, operational support and maintenance procedures for all data pipeline tasks
Ensure data quality and compliance with development, architecture, reporting, and regulatory standards throughout the entire data pipeline
Collaborate with the rest of the Engineering Team, subject matter experts and department leaders to understand, analyze, build and deliver the data they need to power the business
Must-Haves:
Bachelor’s Degree in Computer Science or equivalent experience required
2+ years of experience in designing and software development
Hands on experience with Java or Python
Experience with Data stream processing and a streaming platform such as Apache Kafka, AWS Kinesis, Spark or Flink
Knowledge and experience working with a variety of data stores and formats (Relational, Non-relational, Flat, CSV, Excel, external APIs and data stores)
Written and spoken communication skills in English (B2 category)
Nice-to-Haves:
Experience consuming data from web services, SOAP and REST technologies, HTML, XML and JSON
Experience with a batch data processing and workflow orchestrator, such as Airflow, Luigi, Pinball, Chronos, or similar
Familiarity with ETL and experience in creating pipelines for large data sets
Experience with different kinds of databases: SQL (MySQL, SQL Server), NoSQL (MongoDB, DynamoDB, HBase), columnar (Redshift, Parquet), memory (Redis)
Experience with Amazon Web Services
Home office and remote work
Performance bonus
Professional development program
3rd pillar
Budget for home office equipment
Flexible working hours
High-standard health care
Yoga lessons
E-scooters
Multisport/Golem card
Healthy snacks
Fully covered meal tickets
Nice-to-Haves:
Experience consuming data from web services, SOAP and REST technologies, HTML, XML and JSON
Experience with a batch data processing and workflow orchestrator, such as Airflow, Luigi, Pinball, Chronos, or similar
Familiarity with ETL and experience in creating pipelines for large data sets
Experience with different kinds of databases: SQL (MySQL, SQL Server), NoSQL (MongoDB, DynamoDB, HBase), columnar (Redshift, Parquet), memory (Redis)
Experience with Amazon Web Services
Pracujeme na vývoji technologickej platformy, vďaka ktorej si milióny Američanov dokážu vybrať zdravotné poistenie.
Pracujeme na vývoji technologickej platformy, vďaka ktorej si milióny Američanov dokážu vybrať zdravotné poistenie.