Cloud Platform Engineer

📍Prague I Brno

🕰 Part-Time I Full-Time

Are you passionate about building robust, scalable cloud infrastructure that powers the next generation of data-driven applications? Do you thrive on automating everything through code and ensuring seamless delivery pipelines? We are looking for a Cloud Platform Engineer who bridges the gap between infrastructure and data engineering. If you have an ownership mindset, love solving complex platform puzzles, and want to work in a high-impact, cloud-native environment, we’d love to meet you.

What You’ll Do

As a Cloud Platform Engineer at DataSentics, you will be the backbone of our technical delivery, ensuring our data and AI solutions run on world-class infrastructure. You will work closely with data engineers and scientists to build stable, secure, and automated environments. Your core responsibilities will include:

  • Architecting the Foundation: Design and implement cloud infrastructure (primarily AWS/Azure) using Infrastructure as Code (Terraform).
  • Automating Delivery: Build and maintain robust CI/CD pipelines (GitLab/GitHub) to ensure frequent and reliable deployments.
  • Platform Enablement: Own and solve platform-related topics around Databricks, including Unity Catalog integration, permissions, and service principal management.
  • Containerization: Manage and orchestrate services using Docker and Kubernetes (or Azure Container Apps) to host internal data applications.
  • Data Ops & Security: Set up monitoring, logging, and alerting systems while ensuring top-tier security through managed identities, secrets management, and access control.

Why Join Us?

At DataSentics, we are a team of tech enthusiasts who believe that great data science is only possible with great engineering. We combine deep technical expertise with a practical, product-driven approach. You’ll be part of a culture that values technical excellence, continuous learning, and the freedom to suggest and implement the best tools for the job.

Our Projects & Technology Stack

We build enterprise-grade data platforms and internal applications across industries like finance, retail, and manufacturing.

  • Cloud Ecosystems: Deep dives into AWS and Azure.
  • Data Orchestration: Advanced environments involving Databricks job clusters, complex pipelines, and automated permissions.
  • Modern DevOps: Heavy use of Terraform, Docker, and Git-based automation.
  • How we work: Agile teams (2–5 people), remote-first culture, and a focus on "building it right" the first time.
  • Flexibility: Work from anywhere with optional offices in Prague and Brno; flexible roles starting at 30 hours/week.

What We’re Looking For

Must-haves:

  • Cloud Proficiency: Experience with cloud platforms, ideally AWS/Azure.
  • Automation Expert: Proven track record with CI/CD (GitLab/GitHub pipelines) and Terraform.
  • Container Skills: Hands-on experience with Docker (Kubernetes or Azure Container Apps is a plus).
  • Platform Troubleshooting: Ability to solve complex integrations involving Databricks and data applications.
  • Mindset: A strong sense of ownership, independence, and the "investigative" skills needed to fix broken systems.

Nice-to-haves:

  • Familiarity with Databricks administration (Unity Catalog, permissions, service principals).
  • Basic to intermediate Python skills for automation and scripting.
  • Understanding of Data Engineering lifecycles (pipelines, orchestration, storage).
  • Experience with cloud security (Secrets, IAM) and monitoring/alerting tools.

What You’ll Get

  • The opportunity to shape the infrastructure of cutting-edge AI and Data products.
  • A collaborative environment focused on innovation and technical excellence.
  • Flexibility, remote work options, and a supportive, curious team culture.

Curious to know more? If you're excited about infrastructure as code and want to empower data teams to build amazing things, let’s talk!