QA Software Test Engineer Code Review - Remote
YO IT Consulting
See how well this job matches your profile
Sign up to get an AI match score and generate a tailored application in seconds.
Get your match scoreTags
About the role
Role Overview
QA Software Test Engineer (Code Review) – Remote (Independent Contractor)
You will support the creation of code evaluation datasets used to assess AI behavior in coding-related interactions. The dataset focuses on natural-language reasoning, explanation quality, and technical judgment (not whether code executes correctly).
Responsibilities
- Craft realistic developer prompts across categories such as:
- code review
- debugging
- error diagnosis
- configuration and related scenarios
- Source and adapt content from real PRs to create authentic evaluation cases.
- Write clear, technically accurate responses that demonstrate strong reasoning and explanation quality.
- Ensure scenarios are chat-pasteable and include all necessary context inline (code snippets, error messages, logs, requirements).
Qualifications
- 2+ years of experience in software engineering, technical research, or educational content development.
- Degree in Software Engineering, Computer Science, or a related field (Bachelor’s minimum; advanced degree preferred).
- Strong proficiency in Python, JavaScript, Java, or C++.
- Experience with debugging, testing, and validating code.
- Comfortable with technical writing and high attention to detail.
Contract Details
- Fully remote, completed on your own schedule.
- Start: Immediate
- Duration: 1–2 months (may extend/shorten based on needs and performance)
- Hours: Part-time 15–25 hrs/week, flexible up to 40 hrs/week.
Hiring Process
- Upload resume
- 15-minute AI interview (conversational)
- Follow-up communication with next steps and onboarding details
About YO IT Consulting
YO IT Consulting is a consulting firm that supports high-impact research collaborations, including AI-focused projects with leading labs. In this role, it engages independent contractors to build evaluation assets such as code-related prompts and datasets used to assess AI reasoning and explanation quality.
Scraped 4/7/2026