Swift Engineer (5+ YOE) – AI / LLM Code Evaluation (Remote, Contract) at Mercor | Torre
Swift Engineer (5+ YOE) – AI / LLM Code Evaluation (Remote, Contract)
Report

Swift Engineer (5+ YOE) – AI / LLM Code Evaluation (Remote, Contract)

You'll refine AI's reasoning for code and systems design, directly influencing global developer assistance.
Emma highlights
This highlight was written by Emma’s AI. Ask Emma to edit it.
Full-time

Legal agreement: Contractor

Currency exchange and taxes to be paid by:

Company

Compensation
USD30 - 90/hour
Non-negotiable

Engagement length: Open ended

location_on
Remote (anywhere)
skeleton-gauges
You have opted out of job matches in .
To undo this, go to the 'Skills and Interests' section of your preferences.
Review preferences
Posted 6 days ago

Requirements and responsibilities


Company: Mercor. Type: Contract (Full-time or Part-time). Location: Remote (Worldwide). Language: Professional English required. Compensation: - USD $30 – $90/hour (depending on experience & evaluation performance). - Weekly payments via Stripe or Wise. - Flexible workload (project-based, scalable hours). Mission: - Work directly with leading AI teams to improve how large language models reason about code, systems design, and technical problem-solving. - You will evaluate and refine AI-generated responses, making them more accurate, reliable, and aligned with real-world engineering standards. Responsibilities: - Evaluate AI-generated answers to coding and system design problems. - Execute and validate code outputs. - Identify bugs, inefficiencies, and incorrect reasoning. - Assess code quality & readability. - Assess algorithmic correctness. - Assess system design logic. - Annotate responses with structured, actionable feedback. - Follow defined evaluation frameworks and quality benchmarks. Required Skills: Core: - Swift (expert level). - Software Engineering (5+ years). - Data Structures & Algorithms. - Systems Design. - Debugging & Code Review. - Problem Solving (Medium–Hard level). Technical: - Code Execution & Testing. - API Design & Backend Logic. - Performance Optimization. - Version Control (Git). AI / Evaluation Context: - Experience using LLMs in development workflows. - Ability to evaluate reasoning, not just outputs. Nice-to-Have Skills: - RLHF / AI Model Evaluation. - Competitive Programming. - Open-source contributions (merged PRs). - Multi-language experience (Python, JS, etc.). - Technical writing / explaining complex concepts. Ideal Candidate: - Degree in Computer Science or related field (BS/MS/PhD). - Strong real-world engineering background. - Detail-oriented and highly analytical. - Comfortable identifying subtle logic flaws and edge cases. - Able to work independently in async environments. What You Will Achieve: - Improve the quality and reasoning of AI-generated code. - Influence how AI systems assist developers globally. - Deliver high-quality evaluation outputs that directly impact model performance.

Indefinitely open

Optionally, you can add more information later (benefits, pre-screening questions, etc.)
check_circle

Payment confirmed

A member of the Torre team will contact you shortly

In the meantime, continue adding information to your job opening.