Skip to content
mimi

Generative AI Engineer

HireTalent - Staffing & Recruiting Firm

Rockville · Hybrid Contract Today

About the role

Project Description

The Generative AI Engineer works with moderate supervision across two equally weighted domains: (1) large-scale data pipeline development processing market events in a cloud environment, and (2) design and development of agentic AI systems including LLM-powered regulatory data assistants, MCP servers, and agent harness architectures. This position contributes to overall product quality throughout the software development lifecycle.

Responsibilities

  • Build and maintain ETL/ELT pipelines using Apache Spark, Hive, and Trino across S3-based data lake environments
  • Develop and optimize SQL for large-scale surveillance datasets including window functions, multi-table joins, and complex aggregations
  • Build and engineer big data systems (EMR-on-EC2, EMR-on-EKS) and develop solutions on analytical platforms (SageMaker, Domino, Dataiku)
  • Participate in data quality monitoring, anomaly detection, and production incident investigation
  • Develop AI agent systems using AWS Bedrock and agent frameworks (Strands Agents SDK, LangChain/LangGraph, or equivalent)
  • Build agent harness architectures combining LLM reasoning with deterministic execution - skill/RAG-based SQL generation and structured output validation
  • Implement agent memory, context management, and tool integration (MCP servers, API connectors, data catalog lookups) across the data lake
  • Build evaluation frameworks for agent accuracy - paraphrase robustness, routing precision, and structural consistency
  • Stay informed of advances in LLM frameworks (LangGraph, Google ADK, AWS Strands) and emerging AI capabilities
  • Write clean, well-tested code; contribute to CI/CD Jenkins pipelines and infrastructure-as-code on AWS
  • Ensure secure handling of RCI and sensitive regulatory data across both data pipelines and agent outputs - auditable execution traces
  • Adhere to *** and team standards for secure development practices and technology policies
  • Partner across teams, communicate technical information at the appropriate level, and maintain documentation on Confluence/Wiki
  • Actively learn from senior team members; contribute to process improvement in line with ***'s values of collaboration, expertise, innovation, and responsibility

Essential Technical Skills

Data Engineering & Big Data Technologies

  • Experience building data pipelines using Apache Spark (PySpark preferred) and SQL
  • Experience with SQL query engines (Hive, Trino/Presto, or similar) and cloud data platforms (AWS S3, EMR, Lambda)
  • Understanding of common issues like data skew and strategies to mitigate it, working with large data volumes, and troubleshooting job failures due to resource limitations, bad data, and scalability challenges
  • Real-world experience with debugging and mitigation strategies

Generative AI & Agentic Systems

  • Practical experience building LLM-powered agent systems that use tools and produce structured outputs (not just chatbot interfaces)
  • Hands-on experience with at least one agent framework: LangChain, LangGraph, AWS Strands, or equivalent
  • Working knowledge of prompt engineering, RAG architectures, and context/memory management
  • Experience with foundation model APIs (Anthropic Claude, Amazon Nova, OpenAI, or similar)
  • Memory Architecture: Understanding of agent memory tiers - working memory, episodic memory, semantic memory - and strategies for context persistence, pruning, and retrieval across sessions
  • Agent Harness Design: Familiarity with harness patterns that wrap LLM reasoning with deterministic guardrails, tool routing, verification loops, and graceful degradation

AI Tool Proficiency

  • Hands-on experience with AI development tools (GitHub Copilot, Q Developer, ChatGPT, Claude, etc.)
  • Experience with spec-driven development - using structured specifications to guide AI code generation, review, and validation
  • Ability to leverage AI pair programming for code suggestions, debugging, refactoring, and automated test generation

Cloud Technologies

  • Experience with AWS services like S3, EMR, EMR on EKS, Lambda, Bedrock, Step Functions, etc.
  • Hands-on experience using S3 with Spark (e.g., dealing with file formats, consistency issues)
  • Familiarity with AWS Bedrock for foundation model invocation, knowledge bases, guardrails, and agent orchestration
  • Exposure to Google Cloud Vertex AI (model garden, grounding, agent builder) or equivalent managed AI platforms
  • Familiarity with AWS monitoring and logging tools (CloudWatch, CloudTrail) for production workloads

Programming – Python

  • Proficiency in Python for data engineering and automation
  • Ability to write clean, modular, and performant code
  • Experience with functional programming concepts (e.g., immutability, higher-order functions)
  • Strong understanding of collections, concurrency, and memory management

SQL Skills (Window Functions, Joins, Complex Queries)

  • Proficiency with SQL window functions, multi-table joins, and aggregations
  • Ability to write and optimize complex SQL queries
  • Experience handling edge cases like NULLs, duplicates, and ordering

Good to Have

  • AWS Bedrock AgentCore (memory, identity, tool gateway)
  • Model Context Protocol (MCP) server development and integration
  • Agent evaluation harnesses and agentic patterns (draft-verification, compile-style generation)
  • Fine-tuning foundation models for domain-specific tasks (LoRA, PEFT, or managed fine-tuning via Bedrock/Vertex AI)
  • Local model execution with Ollama, vLLM, or similar for development and experimentation
  • Vector databases (FAISS, Pinecone, OpenSearch)
  • Docker, Kubernetes, and Amazon EKS for containerized workloads
  • Infrastructure as Code (Terraform, CloudFormation)
  • Experience with CI/CD pipelines (Jenkins, GitLab CI, GitHub Actions, ArgoCD)
  • Experience with monitoring and observability tools (Prometheus, Grafana, ELK stack)
  • AWS certifications (AI Practitioner, Solutions Architect, or Kubernetes certifications like CKA/CKAD)

Skills

AWS BedrockAWS EMRAWS LambdaAWS S3Apache SparkCI/CDCloudFormationDockerEMR on EKSETLGoogle Cloud Vertex AIGrafanaHiveInfrastructure as CodeJenkinsKubernetesLangChainLangGraphLLMOpenAIPythonRAGSageMakerSQLTerraformTrino

Don't send a generic resume

Paste this job description into Mimi and get a resume tailored to exactly what the hiring team is looking for.

Get started free