cover image
Appen

Appen

appen.com

2 Jobs

19,339 Employees

About the Company

Appen has been a leader in AI training data for over 25 years, providing high-quality, diverse datasets that power the world's leading AI models. Our end-to-end platform, deep expertise, and scalable human-in-the-loop services enable AI innovators to build and optimize cutting-edge models. We specialize in creating bespoke, human-generated data to train, fine-tune, and evaluate AI models across multiple domains, including generative AI, large language models (LLMs), computer vision, speech recognition, and more. Our solutions support critical AI functions such as supervised fine-tuning, reinforcement learning with human feedback (RLHF), model evaluation, and bias mitigation. Our advanced AI-assisted data annotation platform, combined with a global crowd of more than 1M contributors in over 200 countries, ensures the delivery of accurate and diverse datasets. Our commitment to quality, scalability, and ethical AI practices makes Appen a trusted partner for enterprises aiming to develop and deploy effective AI solutions. At Appen, we foster a culture of innovation, collaboration, and excellence. We value curiosity, accountability, and a commitment to delivering the highest-quality AI solutions. We support work-life balance with flexible work arrangements and a dynamic, results-driven environment. Employees have access to competitive pay, comprehensive benefits, and opportunities for continuous learning and career growth. Our team works closely with the world’s top technology companies and enterprises, tackling exciting challenges and shaping the future of artificial intelligence.

Listed Jobs

Company background Company brand
Company Name
Appen
Job Title
English Speakers in India, Bangladesh & Pakistan— Annotate AI Q&A for Health Domain!
Job Description
Job Title: AI Q&A Annotation Specialist – Health Domain Role Summary: Create natural‑language question‑answer pairs that align with SQL query outputs for health‑related datasets, ensuring linguistic clarity, contextual accuracy, and logical consistency between user intent and data. Expectations: Deliver high‑quality, realistic Q&A annotations, verify alignment with query logic and dataset results, and maintain consistency across all assigned datasets. Key Responsibilities: - Interpret SQL queries and corresponding result tables. - Draft clear, natural user‑style questions reflecting real‑world scenarios. - Provide precise answers fully aligned with SQL outputs. - Verify consistency between question intent, query logic, and data. - Review and refine annotations to meet quality standards. Required Skills: - Strong understanding of SQL query logic (filters, joins, groupings, aggregations, subqueries). - Familiarity with relational database concepts (schemas, tables, columns, keys). - Excellent written English and ability to craft clear, natural questions. - Ability to reason over data and map user intent to database output. - Preferable domain familiarity in health care. Required Education & Certifications: - Bachelor’s degree in Computer Science, Information Systems, or related field (preferred but not mandatory).
Any-martin-rieux, France
Remote
26-11-2025
Company background Company brand
Company Name
Appen
Job Title
Engineer Intern, GenAI Research (Summer Internship)
Job Description
**Job Title** Engineer Intern, GenAI Research (Summer Internship) **Role Summary** Support the GenAI research team by designing and implementing training, evaluation, and benchmarking pipelines for large language models (LLMs). Contribute to model improvement initiatives, develop Python tooling, and document experimental results to enhance production model quality and reliability. **Expectations** - Independently develop and iterate on supervised fine‑tuning pipelines using open‑source LLMs. - Create and apply scientific benchmarks and structured evaluation methods. - Identify performance gaps in production models and propose retraining or hyperparameter solutions. - Communicate findings clearly through technical documentation and reports. **Key Responsibilities** - Build a lightweight supervised fine‑tuning pipeline for open‑source LLMs. - Design and implement new benchmarks assessing scientific and performance criteria. - Analyze production model outputs to pinpoint measurable improvement areas. - Conduct targeted retraining and hyperparameter searches to boost model performance. - Deploy updated models while ensuring core characteristics remain stable. - Develop Python tools to automate training, evaluation, benchmarking, and experimentation workflows. - Implement rubric‑based scoring and “LLM as a judge” evaluation workflows. - Document experimental design, benchmark methodology, and results with precision. - Collaborate remotely with research team members and maintain clear written communication. **Required Skills** - Proficiency in Python; experience with ML frameworks (e.g., PyTorch, TensorFlow) and open‑source model ecosystems. - Hands‑on experience with large language models, including supervised fine‑tuning, prompt engineering, or model evaluation. - Ability to build and maintain machine‑learning pipelines or research infrastructure. - Experience with model performance improvement via retraining or hyperparameter tuning. - Familiarity with cloud platforms (AWS, Azure) for compute resources. - Strong problem‑solving skills; ability to use LLMs as development aids. - Excellent written communication for technical documentation. - Capability to work independently in a remote, research‑driven environment. **Required Education & Certifications** - Currently enrolled in or recently completed a Master’s or PhD program in Computer Science, Artificial Intelligence, Machine Learning, Computer Engineering, or a closely related technical field.
United states
Remote
Fresher
19-02-2026