
Senior Site Reliability Engineer - Data Pipeline
- Česko
- Trvalý pracovní poměr
- Plný úvazek
- We're taking autonomous search mainstream, making product discovery more intuitive and conversational for customers, and more profitable for businesses.
- We're making conversational shopping a reality, connecting every shopper with tailored guidance and product expertise - available on demand, at every touchpoint in their journey.
- We're designing the future of autonomous marketing, taking the work out of workflows, and reclaiming the creative, strategic, and customer-first work marketers were always meant to do.
- Your task is to build and maintain an ecosystem where engineers can safely and efficiently develop, debug and operate their services running in GCP, Kubernetes using DataFlow, DataProc and Python with Go
- You make sure the services have high level of observability, enabling us to provide quality service for our customers
- Further services can scale vertically and horizontally based on current load, operational and telemetric data (OTEL, Prometheus, Victoria Metrics)
- Team have enough insights about health of our services (Grafana, Alerting, PageDuty)
- You helps the team to fulfill security requirements given ISO and SOC2 audits, by enforce security principles like key distribution, key rotation, authorisation & authentication on service level, data encryption at transit, data isolation, resource limitations, quality of service, audit logs (mainly by Enovy proxies)
- You contribute to our tooling, so we have tools in place for debugging, troubleshoot and performance testing
- You automate manual/semi-manual steps deployment and instance setup
- You have hands on on L3 support and incident resolutions
- CI pipelines have linters, security scans, code smell detection enabling engineers to produce quality MRs
- You can articulate how your contributions have transformed the way engineers work and think by fostering a strong DevOps/SRE culture.
- You can demonstrate how impactful your work as an SRE or DevOps Engineer can be in connection to business success
- You understand the importance of you build - you run it principle and you love the feeling you own it
- You are mindful of the costs associated with running our service, which translates into effective vertical and horizontal pod autoscaling and detailed telemetry insights.
- You believe the infrastructure as a code is the only thing that can bring stability into chaos
- Terraform is your daily bread, and HELM deployments are your second-best friend
- You use telemetry data and metrics to provide feedback to engineers on how the application and services behave
- You can navigate yourself in complex service architecture by using distributed debugging
- You have experience with Python and a solid grasp of engineering practices
- A big advantage is, if you have an experience with Go, or with ETL pipelines
- Python, GO
- Apache Kafka, Kubernetes, GitLab
- Google Cloud Platform, BigQuery, BigLake Table
- Open formats IceBerg, Avro, Parquet
- DataFlow, Apache Beam, DataProc, Spark,
- Mongo, Redis
- … and much more 🙂
- During the first 30 days, you will get to know the team, the company, and the most important processes. You'll work on your first tasks. We will help you to get familiar with our codebase and our product.
- During the first 90 days, you will start contributing to team's L3 rotation, troubleshooting, debugging that should let you better understand what is what and perhaps also to come with fresh ideas on how to improve our services and monitoring.
- During the first 180 days, you'll become an integral part of the team by actively contributing to the team's projects as well as to onCall rotation.