Назад
Company hidden
5 часов назад

AI Engineer (LLM)

Формат работы
hybrid
Тип работы
fulltime
Грейд
middle/senior
Английский
b2
Страна
India
Вакансия из списка Hirify.GlobalВакансия из Hirify Global, списка международных tech-компаний
Для мэтча и отклика нужен Plus

Мэтч & Сопровод

Для мэтча с этой вакансией нужен Plus

Описание вакансии

Текст:
/

TL;DR

AI Engineer (LLM): Building, fine-tuning, deploying, and scaling large language model–based systems with an accent on LLM optimization, backend API development, and MLOps. Focus on taking LLMs from experimentation to production-ready, scalable AI solutions, implementing RAG pipelines, and ensuring efficient model serving and automated evaluation.

Location: Noida, Uttar Pradesh (Hybrid)

Company

hirify.global is a growing data product company that was founded in early 2020 and works primarily with Fortune 500 companies, delivering digital solutions to accelerate business growth through innovation.

What you will do

  • Design and implement traditional ML and LLM-based systems and applications.
  • Optimize model inference performance and cost efficiency.
  • Fine-tune foundation models for specific use cases and domains.
  • Build robust backend infrastructure for AI-powered applications.
  • Implement and maintain MLOps pipelines for AI lifecycle management.
  • Design and implement comprehensive traditional ML and LLM monitoring and evaluation systems.

Requirements

  • 4–8 years of relevant experience in LLMs, Backend Engineering, and MLOps.
  • Experience with LLM expertise including parameter-efficient fine-tuning (LoRA, QLoRA), inference optimization (quantization, pruning), prompt engineering (RAG), and model evaluation.
  • Proficiency in Python with FastAPI, Flask, or similar frameworks for backend engineering.
  • Experience with RESTful APIs, real-time systems, and vector/traditional databases.
  • Hands-on experience with cloud platforms like AWS, GCP, or Azure, focusing on ML services.
  • Expertise in MLOps, including model serving frameworks (vLLM, SGLang, TensorRT), Docker, Kubernetes, CI/CD, and workflow orchestration tools like Airflow.

Nice to have

  • Hands-on experience with LLM Frameworks like Transformers, LangChain, or LlamaIndex.
  • Knowledge of LLM-specific monitoring tools and general ML monitoring.
  • Experience with distributed training and inference setups.
  • Knowledge of model compression techniques (distillation, quantization).
  • Experience deploying models handling high-throughput, low-latency requirements.
  • Familiarity with recent LLM research and ability to implement novel techniques.

Culture & Benefits

  • Competitive salary and strong insurance package.
  • Extensive learning and development resources for employee growth.

Будьте осторожны: если работодатель просит войти в их систему, используя iCloud/Google, прислать код/пароль, запустить код/ПО, не делайте этого - это мошенники. Обязательно жмите "Пожаловаться" или пишите в поддержку. Подробнее в гайде →

Текст вакансии взят без изменений

Источник - загрузка...