Skip to content
mimi

Lead Data Scientist for LLMs

Dynatrace

Linz · flexible Lead Today

About the role

About

Lead Data Scientist for LLMs at Dynatrace, driving the end-to-end LLM stack from data and ingestion to prompt design, tooling, safety, and optimization. You’ll own architecture, evaluation, and LLMOps governance to enable enterprise-grade AI capabilities. This role combines hands-on DS work with technical leadership to scale generative AI across complex, multicloud environments. You’ll mentor teams, shape RAG standards, and partner with cross-functional peers to deliver impactful AI solutions.

Responsibilities

  • Own the LLM system architecture including retrieval pipelines, prompt and tool design, routing, safety layers, and telemetry with a focus on quality, latency, and cost.
  • Define and standardize RAG content ingestion, chunking, hybrid retrieval, reranking, query understanding, and output contracts.
  • Develop and maintain evaluation strategies spanning accuracy, grounding, safety, privacy, determinism, latency, and cost.
  • Formalize LLMOps practices: versioning for prompts/datasets/models, experiment governance, registries, and promotion criteria (dev→staging→prod).
  • Lead tool/agent design: API schema for function calls, error handling, recovery, self-correction, and guardrails.
  • Assess build-vs-buy options (managed vs open-source/self-hosted) considering performance, IP, privacy, and compliance.
  • Provide mentorship on prompting, retrieval design, evaluations, and safe deployment; lead reviews of prompts, pipelines, and evaluation reports.
  • Hands-on end-to-end RAG implementation: ingestion, chunking, embeddings, hybrid search, rerank, prompt assembly, tool calls, post-processing.
  • Engineer robust prompts/tools with templates, multi-turn strategies, and structured outputs (JSON Schema/Pydantic).
  • Select and tune models and embeddings; apply LoRA/PEFT or distillation when warranted.
  • Create eval corpora and KPIs for accuracy, grounding, and tool success; implement guardrails for PII/PHI and safety scoring.
  • Ship resilient services with analytics, alerts, SLOs, and drift/cost monitoring; optimize for scale via caching, batching, and routing by intent.

Requirements

  • Advanced CS/AI/ML degree or equivalent
  • 7+ years DS/ML experience
  • 3+ years NLP/LLMs
  • Shipped production systems
  • 5+ years professional Python
  • 3+ years data engineering for unstructured data (text processing, embedding-friendly preprocessing)
  • 1+ years proven RAG expertise (embeddings, retrieval, reranking, chunking)
  • 1+ years evaluation depth (offline/online for accuracy, grounding, safety)
  • 1+ years safety/privacy (PII/PHI redaction, policy enforcement)
  • 1+ years LLMOps (prompt/version management, experiment tracking, monitoring)
  • Excellent communication and ability to explain trade-offs

Benefits

  • remote options
  • hybrid work
  • international team
  • career development program
  • visa and relocation support
  • flexible work arrangements

Skills

Python

Don't send a generic resume

Paste this job description into Mimi and get a resume tailored to exactly what the hiring team is looking for.

Get started free