Data Engineer
About The Position
- Lead development projects of critical, high-availability, cloud-scale services and APIs
- Support millions of clients with large amounts of data and scalability in mind
- Take part in all development stages – from design to deployment
- Develop and deploy real time/batch data processing pipelines using the latest technologies
- Design and build high-availability, cloud-scale data pipelines (ETLs)
Requirements
- A team player, highly motivated and a fast learner
- Ability to assume ownership of goals and products
- Passion for designing scalable, distributable and robust platforms and analytic tools
- +3 years of experience in large scale, distributed server side, backend development
- +3 years of experience writing and maintaining data pipelines (ETLs)
- +3 years of experience using JVM languages such as scala, java
- Experience with Spark, Kafka, Airflow, Mongo, Elastic, HDFS or similar technologies
- Experience with system monitoring (Prometheus, Influx or similar)
- Experience with Docker, Linux, CI/CD, Kubrenetes, etc
- Experience in working with cloud providers (e.g., AWS, GCP)
Advantages
- Experience in leading software development projects from nothing to production
- Experience in microservices architecture and flexible system design
- Full-stack capabilities, JavaScript
- Experience with scripting languages such as Python, Groovy
- Experience working closely with ML teams
- Good presentation skills