Skip to content
mimi

AI Software Engineer

Novaedge

Remote · France Contract Mid Level Today

About the role

About

We are seeking a hands-on AI Software Engineer to design, build, and deploy intelligent backend systems that power conversational AI, automation, and data-driven decision engines. You’ll collaborate with data scientists, ML engineers, and product teams to integrate LLM-based models (OpenAI, Anthropic, Meta Llama, etc.) into scalable microservices and internal tools.

Key Responsibilities

  • Design and develop Python-based backend systems supporting AI/LLM workflows, APIs, and data pipelines.
  • Build scalable microservices and vector-database integrations (e.g., Milvus, Pinecone, FAISS) for retrieval-augmented generation (RAG) pipelines.
  • Integrate and orchestrate LLMs using APIs (OpenAI, Anthropic, Hugging Face, vLLM, Triton, or similar).
  • Work closely with data engineering to optimize data ingestion, preprocessing, and embeddings pipelines.
  • Implement asynchronous and distributed processing (Celery, Kafka, or Ray).
  • Deploy and monitor services on Docker/Kubernetes with CI/CD pipelines (GitHub Actions, Jenkins, or GitLab CI).
  • Maintain documentation, testing, and model performance metrics.
  • Collaborate with DevOps and security to ensure safe and reliable AI deployments.

Required Skills & Experience

  • 3+ years experience in backend or full-stack development with Python (FastAPI, Flask, or Django).
  • Proven experience integrating AI/ML or NLP systems (LLMs, embeddings, transformers, etc.).
  • Strong understanding of RESTful and async APIs, data serialization, and model inference optimization.
  • Familiarity with vector databases (Milvus, Pinecone, FAISS, Weaviate) and document chunking/embedding techniques.
  • Experience with SQL and NoSQL databases (PostgreSQL, MongoDB, Redis).
  • Hands-on with Docker, Kubernetes, and cloud environments (AWS / GCP / Azure).
  • Knowledge of MLOps workflows (model packaging, inference serving, versioning).
  • Experience with Git, CI/CD, and automated testing.

Nice to Have

  • Familiarity with AI voice technologies (Riva, ElevenLabs, VAPI SDK, or similar).
  • Experience with LangChain, LlamaIndex, or Haystack for RAG pipelines.
  • Exposure to NVIDIA Triton / TensorRT-LLM / vLLM for high-performance inference.
  • Understanding of prompt engineering, retrieval evaluation, and fine-tuning pipelines.
  • Experience contributing to open-source AI frameworks.

Why Join Us

  • Build real AI products — from voice agents to LLM-powered automation systems — not just prototypes.
  • Work with a high-performance engineering team using NVIDIA hardware and cutting-edge open-source tools.
  • 100% remote flexibility, cross-functional collaboration, and ownership of critical AI systems.

Skills

AWSCeleryDockerFastAPIFlaskGitGitHub ActionsGCPJenkinsKafkaKubernetesMongoDBNoSQLOpenAIPineconePostgreSQLPythonRayRedisRESTful APISQLTriton

Don't send a generic resume

Paste this job description into Mimi and get a resume tailored to exactly what the hiring team is looking for.

Get started free