Назад
Company hidden
3 дня назад

Founding Abuse Engineer (AI)

Формат работы
onsite
Тип работы
fulltime
Грейд
senior
Английский
c1
Страна
US
Вакансия из списка Hirify.GlobalВакансия из Hirify Global, списка международных tech-компаний
Для мэтча и отклика нужен Plus

Мэтч & Сопровод

Для мэтча с этой вакансией нужен Plus

Описание вакансии

Текст:
/

TL;DR

Founding Abuse Engineer (AI): Designing and implementing end-to-end platform misuse detection and enforcement systems for AI model evaluations with an accent on bot detection, leaderboard integrity, and adversarial defense. Focus on building scalable anti-abuse primitives, mitigating LLM-specific harms, and creating internal investigation tooling.

Location: Must be based in the Bay Area

Company

hirify.global is an open platform created by UC Berkeley researchers for evaluating the real-world performance and reliability of AI models.

What you will do

  • Own the overall abuse vision and strategy for the Arena platform, defining detection and enforcement goals.
  • Design and operate detection systems for bots, sybils, and coordinated rating-system manipulation.
  • Build reversible and auditable enforcement primitives, including rate limits, shadowbans, and model-side refusals.
  • Mitigate inference abuse and cost exploitation at the platform layer.
  • Implement detection for jailbreaks and multi-provider misuse across served models.
  • Develop internal investigator tooling to allow policy and T&S analysts to triage incidents efficiently.

Requirements

  • 6+ years of production software engineering experience, specifically building systems under adversarial conditions.
  • Direct experience in trust & safety, anti-abuse, anti-fraud, integrity, or risk engineering.
  • Strong SQL and data-analysis skills for pattern-finding and investigation.
  • Proficiency in a modern backend language such as Node.js, TypeScript, Python, or Go.
  • Must be based in the Bay Area.

Nice to have

  • Experience with LLM-specific adversarial inputs, including jailbreaks and prompt injection.
  • Background in securing voting, rating, or marketplace platforms against manipulation.
  • ML systems experience, including feature engineering and label acquisition.
  • Contributions to open-source trust & safety or adversarial-ML work.

Culture & Benefits

  • Competitive compensation and equity aligned with market standards.
  • Comprehensive health and wellness benefits, including medical, dental, and vision.
  • Opportunity to work on cutting-edge AI with a small, mission-driven team.
  • A culture that values transparency, trust, craftsmanship, and community impact.

Будьте осторожны: если работодатель просит войти в их систему, используя iCloud/Google, прислать код/пароль, запустить код/ПО, не делайте этого - это мошенники. Обязательно жмите "Пожаловаться" или пишите в поддержку. Подробнее в гайде →