Research Scientist, GenAI Safety Evaluations - Computer Vision - Meta
New York, NY
About the Job
The GenAI Evaluations Foundations team is looking to hire a Research Scientist with Vision experience for the Safety Evaluations Dev team. Safety Evaluations are key to help us both internally align as well as externally share on how the LLM is being safe in its responses to adversarial or unsafe prompts. When our models are safe, we can feel confident and comfortable open-sourcing our models for other developers and our internal product teams to use.Evaluations & benchmarks are what steer AI progress, as we can inject them in all stages of the model training. The sooner we can catch issues, the faster we can fix things resulting in us saving millions of dollars and compute cycles while avoiding harming Meta’s reputation in the world. As Llama further builds on its image understanding and image generation modalities, we are looking to hire a Research Scientist who is passionate about Safety and has experience designing evaluations and dataset for Computer vision models.
RESPONSIBILITIES
Research Scientist, GenAI Safety Evaluations - Computer Vision Responsibilities:
MINIMUM QUALIFICATIONS
Minimum Qualifications:
PREFERRED QUALIFICATIONS
Preferred Qualifications:
RESPONSIBILITIES
Research Scientist, GenAI Safety Evaluations - Computer Vision Responsibilities:
- Design and implement datasets to evaluate our LLMs on safety, with a focus on Vision.
- Adapt standard machine learning methods to best exploit modern parallel environments (e.g. distributed clusters, multicore SMP, and GPU).
- Work with a large and globally distributed team across multiple functions to understand the needs and align on goals and outcomes.
- Play a significant role in healthy cross-functional collaboration.
MINIMUM QUALIFICATIONS
Minimum Qualifications:
- Bachelor's degree in Computer Science, Computer Engineering, relevant technical field, or equivalent practical experience.
- PhD in Computer Science, Computer Engineering, relevant technical field
- 3+ year(s) of work experience in a university, industry, or government lab with emphasis on AI Research in machine learning, deep learning, and computer vision.
- Programming experience in Python and experience with frameworks such as PyTorch.
- Exposure to architectural patterns of large scale software applications.
- Domain-Relevant Research Publications accepted at peer-reviewed AI conferences (e.g., NeurIPS, CVPR, ICML, ICLR, ICCV, and ACL)
PREFERRED QUALIFICATIONS
Preferred Qualifications:
- Experience working with Safety or related areas.
- Direct experience in building evals for generative AI and LLM research
- First author publications at peer-reviewed AI conferences (e.g., NeurIPS, CVPR, ICML, ICLR, ICCV, and ACL).
Source : Meta