We are seeking a Senior Data DevOps Engineer to design, build and operate robust cloud-based data platforms, supporting data engineering and machine learning teams with scalable, production-ready infrastructure.
Technologies we are interested in...
Cloud Platforms
· AWS: EC2, S3, IAM, VPC; Glue, Redshift, EMR, Kinesis; SageMaker
· Azure: VM, Storage, RBAC, Networking; Synapse, Data Factory, Event Hubs; Azure ML
· GCP (basic): BigQuery, Pub/Sub, Dataflow
Infrastructure, Containers & Automation
· Infrastructure as Code: Terraform (AWS & Azure), Terragrunt, ARM / Bicep
· Containers & orchestration: Docker, Kubernetes (EKS / AKS), Helm
· Configuration management: Ansible, Chef, Puppet or similar
· CI/CD: Jenkins, TeamCity, Bamboo, GoCD or similar
Data, Streaming & ML Platforms
· Apache Spark, Apache Kafka
· Delta Lake / Lakehouse architectures
· Databricks
· Workflow orchestration: Airflow or Prefect
· MLOps & ML platforms: MLflow, Feature Stores (Feast), Model Serving (KFServing, TorchServe)
Big Data, Monitoring & Security
· Hadoop: HDFS, YARN; Cloudera or Hortonworks; Ambari, Cloudera Manager; Hive, HUE, Spark; Hadoop Security (intermediate)
· Monitoring & observability: Prometheus, Grafana, Datadog, Zabbix, Nagios
· Search & analytics: Elasticsearch (intermediate)
· Security best practices and authentication/authorization: LDAP, Kerberos, SAML
Programming
· Python (minimum 2 years hands-on experience)
· One scripting language: Bash, Perl, or Groovy