Skip to content
mimi

Senior ai security engineer

IDecisions

India · On-site Full-time Senior Today

About the role

Company

We partner with enterprises to advise, build, secure, and operationalize AI systems at scale. Our focus is on developing Generative AI (Gen AI), Agentic AI, and Reinforcement Learning-driven systems, while embedding security, governance, and risk controls directly into AI workflows. We enable organizations to safely deploy LLMs, autonomous agents, and adaptive decisioning systems in regulated, mission‑critical environments.

Job Description

As a Senior AI Security Engineer (Gen AI, Agentic AI & Reinforcement Learning), you will lead the design and implementation of secure, scalable, and adaptive AI systems, including LLM‑based applications, agentic workflows, and RL‑driven decision engines. This role goes beyond traditional security—you will build intelligent, self‑improving security review systems using agentic frameworks (Lang Graph, Lang Chain, Lang Smith) and reinforcement learning techniques to continuously enhance AI risk evaluation, policy enforcement, and approval workflows. You will collaborate closely with AI/ML engineers, platform teams, and governance stakeholders to embed autonomous, learning‑based security mechanisms into enterprise AI ecosystems.

Key Responsibilities

Gen AI, Agentic AI & RL Security Architecture Design and secure LLM, RAG, multi‑agent, and RL‑driven systems

  • Implement security controls for:
    • Autonomous decision‑making agents
    • RL‑based adaptive systems
    • Tool‑using and API‑integrated agents
  • Ensure safe exploration and bounded behavior in RL environments

Agentic AI + Reinforcement Learning for Security Automation (Core Focus)

  • Build agentic AI pipelines using:
    • Lang Graph → multi‑step, stateful security workflows
    • Lang Chain → LLM orchestration and tool integration
    • Lang Smith → observability, tracing, and evaluation
  • Develop RL‑enhanced security agents that:
    • Learn from past approval decisions
    • Optimize risk scoring and classification over time
    • Continuously improve policy enforcement accuracy
  • Implement feedback loops (human‑in‑the‑loop + automated) to train:
    • Risk evaluation agents
    • Compliance validation agents
  • Automate end‑to‑end intake → evaluation → approval pipelines for Gen AI and Agentic AI use cases

Reinforcement Learning Implementation & Governance

  • Design and implement RL models for adaptive security decisioning:
    • Policy optimization
    • Risk‑based prioritization
    • Dynamic access control adjustments
  • Apply safe RL techniques:
    • Reward shaping aligned with compliance and security policies
    • Constraint‑based RL (safe exploration boundaries)
  • Monitor and mitigate risks such as:
    • Reward hacking
    • Unsafe policy learning
    • Drift in learned behaviors
  • Integrate RL models into AI governance workflows for continuous improvement

AI Risk, Governance & Compliance

  • Translate frameworks such as:
    • NIST AI RMF
    • EU AI Act
    • OWASP Top 10 for LLMs
      into automated, adaptive controls
  • Build dynamic risk scoring systems enhanced by RL:
    • Adversarial Risk Score
    • Model Drift Index
    • Policy Compliance Confidence Score
  • Generate real‑time AI risk heat maps and approval recommendations
  • Implement policy‑as‑code + policy‑learning systems

Security Assessment & Red Teaming

  • Conduct AI/LLM/RL system security assessments
  • Perform red teaming across:
    • Prompt injection scenarios
    • Agent tool misuse
    • RL policy exploitation
  • Evaluate vulnerabilities in:
    • RAG pipelines
    • Multi‑agent coordination
    • RL training environments

AI/ML Lifecycle & LLMOps/RLOps Security

  • Secure the full lifecycle:
    • Data ingestion, labeling, and validation
    • Model training (LLM + RL) with GPU isolation and sandboxing
    • Deployment, inference, and continuous learning loops
  • Implement RLOps + LLMOps security controls
  • Ensure:
    • Model lineage and provenance
    • Secure feedback loops
    • Version control for policies and learned behaviors

Monitoring, Incident Response & Observability

  • Build AI + RL‑aware monitoring systems
  • Detect anomalies in:
    • LLM outputs
    • Agent decisions
    • RL policy shifts
  • Develop incident response playbooks for autonomous systems
  • Create executive dashboards linking AI + RL risk to business KPIs

Data Security & Access Control

  • Implement fine‑grained and adaptive access controls
  • Secure:
    • RAG knowledge bases
    • Vector databases
    • RL training datasets
  • Ensure compliance with data privacy and residency requirements

Thought Leadership

  • Act as an SME in:
    • AI Security
    • Agentic AI systems
    • Reinforcement Learning security
  • Research emerging risks in:
    • Autonomous AI systems
    • Self‑improving models
    • Multi‑agent + RL ecosystems

Qualifications Required

  • Bachelor’s degree in Computer Science, Engineering, or related field
  • 3–5+ years of experience in cybersecurity (application, cloud, or data security)
  • Strong experience in automation, scripting, and security tool development
  • Hands‑on experience with:
    • Gen AI / LLM applications
    • AI threat modeling and risk assessment
    • Deep understanding of AI threat vectors:
      • Prompt injection
      • Data leakage
      • Adversarial attacks
  • Experience with Azure or AWS cloud security ecosystems

Preferred (Strong Differentiators)

Gen AI & Agentic AI

  • Hands‑on experience with:
    • Lang Chain
    • Lang Graph
    • Lang Smith
  • Experience building agentic workflows and multi‑agent systems
  • Experience securing RAG pipelines and LLM applications

Reinforcement Learning (Highly Valued)

  • Experience implementing Reinforcement Learning models:
    • Policy optimization
    • Reward function design
    • Decision‑making systems
  • Familiarity with:
    • RLHF (Reinforcement Learning from Human Feedback)
    • Safe RL and constrained optimization
  • Experience integrating RL into:
    • Automation workflows
    • Security decision systems
  • Understanding of RLOps pipelines and lifecycle management

Security & Governance

  • Familiarity with:
    • OWASP Top 10 for LLMs
    • NIST AI RMF, EU AI Act, ISO 42001
  • Experience with:
    • Microsoft Sentinel, Azure Monitor, Purview, Key Vault
    • Policy‑as‑code and automated compliance frameworks
  • Knowledge of data privacy regulations (GDPR, DORA, etc.)

Requirements

  • Bachelor’s degree in Computer Science, Engineering, or related field
  • 3–5+ years of experience in cybersecurity (application, cloud, or data security)
  • Strong experience in automation, scripting, and security tool development
  • Hands-on experience with Gen AI / LLM applications
  • AI threat modeling and risk assessment
  • Deep understanding of AI threat vectors: Prompt injection, Data leakage, Adversarial attacks
  • Experience with Azure or AWS cloud security ecosystems

Responsibilities

  • Design and secure LLM, RAG, multi-agent, and RL-driven systems
  • Implement security controls for autonomous decision‑making agents, RL‑based adaptive systems, and tool‑using/API‑integrated agents
  • Ensure safe exploration and bounded behavior in RL environments
  • Build agentic AI pipelines using LangGraph, LangChain, and LangSmith
  • Develop RL‑enhanced security agents that learn from past approval decisions, optimize risk scoring, and improve policy enforcement with feedback loops
  • Automate end‑to‑end intake, evaluation, and approval pipelines for Gen AI and Agentic AI use cases
  • Design and implement RL models for adaptive security decisioning, policy optimization, risk‑based prioritization, and dynamic access‑control adjustments
  • Apply safe RL techniques such as reward shaping and constraint‑based RL
  • Monitor and mitigate risks like reward hacking, unsafe policy learning, and drift in learned behaviors
  • Integrate RL models into AI governance workflows for continuous improvement
  • Translate frameworks (NIST AI RMF, EU AI Act, OWASP Top 10 for LLMs) into automated adaptive controls
  • Build dynamic risk scoring systems enhanced by RL (Adversarial Risk Score, Model Drift Index, Policy Compliance Confidence Score)
  • Generate real‑time AI risk heat maps and approval recommendations
  • Implement policy‑as‑code and policy‑learning systems
  • Conduct AI/LLM/RL system security assessments and red‑team exercises (prompt injection, agent tool misuse, RL policy exploitation)
  • Evaluate vulnerabilities in RAG pipelines, multi‑agent coordination, and RL training environments
  • Secure the full AI/ML lifecycle: data ingestion, labeling, validation, model training with GPU isolation, deployment, inference, and continuous learning loops
  • Implement RLOps and LLMOps security controls (model lineage, provenance, secure feedback loops, version control)
  • Build AI + RL‑aware monitoring systems to detect anomalies in LLM outputs, agent decisions, and RL policy shifts
  • Develop incident response playbooks for autonomous systems and executive dashboards linking AI + RL risk to business KPIs
  • Implement fine‑grained and adaptive access controls for RAG knowledge bases, vector databases, and RL training datasets, ensuring data‑privacy and residency compliance
  • Act as subject‑matter expert in AI security, agentic AI systems, and reinforcement‑learning security, researching emerging risks

Skills

Generative AI (Gen AI)Large Language Models (LLM)Retrieval Augmented Generation (RAG)Agentic AIReinforcement Learning (RL)RLHFSafe RLLangChainLangGraphLangSmithAzure cloud securityAWS cloud securitySecurity automationScriptingSecurity tool developmentAI threat modelingPrompt injection mitigationData leakage preventionAdversarial attack defensePolicy‑as‑codeNIST AI RMFEU AI ActOWASP Top 10 for LLMsMicrosoft SentinelAzure MonitorAzure PurviewAzure Key VaultData privacy regulations (GDPR, DORA)RLOpsLLMOpsCI/CD for security

Don't send a generic resume

Paste this job description into Mimi and get a resume tailored to exactly what the hiring team is looking for.

Get started free