Senior AI Platform Backend Engineer (LLM)

Epam
Municipality of Madrid, Spain
4 days ago

Role details

Contract type
Permanent contract
Employment type
Full-time (> 32 hours)
Working hours
Regular working hours
Languages
English

Job location

Municipality of Madrid, Spain

Tech stack

Artificial Intelligence
DevOps
Github
Python
Machine Learning
Message Broker
Natural Language Processing
NLTK
Recommender Systems
TensorFlow
Test Driven Development
Chatbots
PyTorch
Delivery Pipeline
Large Language Models
Prompt Engineering
Deep Learning
Backend
Keras
FastAPI
HuggingFace
Machine Learning Operations
Spacy
GPT
Jenkins
Microservices

Job description

Maintain and enhance CI/CD pipelines using tools such as GitHub Actions, AWS CodePipeline, Jenkins or ArgoCD

  • Design and develop backend architecture for AI Verification and ChatGPT services utilizing Python and FastAPI
  • Build, optimize and scale classifiers and tools leveraging machine learning, encoders and rule-based models
  • Architect and implement solutions following Domain-Driven Design (DDD) and Test-Driven Development (TDD) best practices
  • Design, develop, and maintain a production-grade LLM-as-a-judge service for verifying AI-generated content from source documents, leveraging frameworks such as HuggingFace Transformers, SpaCy, NLTK, and BM25
  • Build and maintain high-throughput Retrieval-Augmented Generation (RAG) services, including ingestion pipelines and message brokers
  • Perform prompt engineering, including techniques such as Chain-of-Thought and Few-Shot Learning, across various LLMs (OpenAI, Anthropic, Google, etc.)
  • Provide hands-on expertise and support for one or more leading AI frameworks (TensorFlow, Keras, PyTorch, BERT, etc.)
  • Demonstrate technical leadership in at least one AI specialization, such as graph recommendation systems, deep learning or natural language processing

Requirements

We are seeking a talented and proactive Senior AI Platform Backend Engineer (LLM) to lead the design, optimization and deployment of machine learning pipelines using MLOps in cloud environments. In this role, you will implement LLM-based solutions for chatbots and Retrieval-Augmented Generation (RAG) systems and develop robust DevOps/MLOps pipelines for production., Strong proficiency in Python and experience developing backend services with FastAPI

  • Experience building and maintaining CI/CD pipelines using tools such as GitHub Actions, AWS CodePipeline, Jenkins or ArgoCD
  • Hands-on experience designing and optimizing scalable machine learning models, including classifiers, encoders and rule-based systems
  • Experience working with large language models (LLMs) and frameworks such as HuggingFace Transformers, Spacy, NLTK and BM25
  • Proven ability to design, develop and maintain high-load Retrieval-Augmented Generation (RAG) services, including ingestion pipelines and message brokers

Benefits & conditions

Private health insurance

  • EPAM Employees Stock Purchase Plan
  • 100% paid sick leave
  • Referral Program
  • Professional certification
  • Language courses, Why Join EPAM
  • WORK AND LIFE BALANCE. Enjoy more of your personal time with flexible work options, 24 working days of annual leave and paid time off for numerous public holidays.
  • CONTINUOUS LEARNING CULTURE. Craft your personal Career Development Plan to align with your learning objectives. Take advantage of internal training, mentorship, sponsored certifications and LinkedIn courses.
  • CLEAR AND DIFFERENT CAREER PATHS. Grow in engineering or managerial direction to become a People Manager, in-depth technical specialist, Solution Architect, or Project/Delivery Manager.
  • STRONG PROFESSIONAL COMMUNITY. Join a global EPAM community of highly skilled experts and connect with them to solve challenges, exchange ideas, share expertise and make friends.

Apply for this position