Don’t be “AI-shy” (part 2)


Reader,

In Part 1, you saw how a bleeding-edge tech assessment for a QA Engineer looks in 2026.

Now you're going to learn why this is something you can expect to see more of.

"The share of new code relying on AI rose from 5% in 2022 to 29% in early 2025"

And the trend is expected to continue growing.

Sure, there's skepticism about how much AI can speed up high-quality code generation, but AI's ability to do so has only increased.

I'm not a believer that we'll achieve AGI (or that we need to for disastrous impacts on the workforce and environment to come about), but I do think you should expect to see more companies assessing an Engineering candidate's ability to use AI to generate code.

It's quickly becoming table stakes for meeting new productivity expectations.

And it's becoming a normal part of the job (for better or worse).

How do you assess "competence with using AI tools"?

Easy. You have the candidate use AI tools.

Yeah, I know. Scaaaary.

What if they cheat?
What if I hire someone who can't write code without AI?
What if we don't have interviewers who know how to use AI tools?

Change sucks, but changing with the times keeps you in the loop with top talent and believe it or not, there are ways to design your tests to solidify those answers. If you are technical enough to spot them.

So if you're asking those questions, it's time to change things up.

Let's take a look at how traditional assessment rubrics compare to AI assessment rubrics for a second here:

What if they cheat?

What does "cheating" mean to you? What does it mean to your hiring managers?

You're not testing their ability to remember stuff. You're going to see how much they think and critique while they use AI tools.

That's a completely different ball game than checking for correctness.

Of course, you'll want to have someone in the room to evaluate the code quality, but what you're most interested in is how much the candidate is paying attention to what is being generated, and what opinions they have about the AI's output as they're working.

It's harder to cheat at having good taste than it is to cheat at remembering the correct answers.

What if they're incompetent without AI?

You'll figure this out when they go to prune the AI fluff.

It's possible to ask AI to "just clean it up", but it takes longer to do that than manually editing the specific things you dislike about AI's output.

Besides, the point about "having good taste" in the earlier point about cheating applies here too. You can be reassured that the candidate knows how to write good code if they are aware of what bad code looks like and what to replace it with.

What if the interviewer doesn't have XP with AI tools?

Well...time to get some! Or bring in someone from Engineering who does have this experience.

Or worst case scenario, demand that the Engineer on the call does some learning on YouTube before the call.

There's no excuse nowadays for not having this experience. It's been 2 years since Cursor got popular.

The assumption is that if your team is hiring someone who needs to know how to use agentic coding tools, the team is already using them.

If you're really without a properly experienced auditor, find one externally or hire one.

Let's look at the rubric again

First of all, it's not a perfect rubric. I had AI whip one up to save time writing this newsletter. For example, the note about "writing complex Xpaths and CSS selectors" is not a good use of AI, since those selector strategies are typically garbage/last resorts for a QA Engineer.

That said...what can we learn?

Traditional tech assessments tend to see how skilled a candidate is at generating "the right answer".

AI-augmented tech assessments go further.

In my teacher education studies, we learned about a framework called Bloom's Taxonomy, where Remember is the most primitive thinking skill in the hierarchy.

You can think of traditional assessments as living more in the "basement" of the pyramid, while AI-augmented assessments allow you spend more time testing higher-order thinking skills.

With AI handling the minutia of generating the correct syntax for the task at hand, you can focus on how the candidate evaluates the code in the context of the overall system they're building.

You can see their opinions about the system and interview them about the tradeoffs of their architectural choices.

In essence, your conversations become far more interesting with fewer "hard-and-fast" rules about what's "correct".

Your team can hire based on a candidate's architectural judgement and cultural alignment, because the AI handles the syntactic burden that used to consume the interview.

What this means for you

Is your team already assessing engineering candidates this way? If not, why not?

The goal is to align your technical evaluations to streamline your interview process and help guide decisions.
The way you go about that depends on the type of projects you oversee.

Many companies have AI initiatives and are trying to select and adopt this technology so it might be time to have this conversation with your team, because top talent is easier to spot when you can have more quality conversations in your interviews.

It's much harder to know who you're talking to if the code itself takes up most of the interview time.

And with the current market the way it is, being able to cut your vetting time down will be key.

-Steven

The Better Vetter Letter

Helping tech recruiters vet client requirements and job candidates for technical roles by blending 20+ years of Engineering & Recruiting experience.

Read more from The Better Vetter Letter

Reader, I've saved the best for last, truly. This one was a software testing services firm. For the last several emails I've written about interviews from my job hunt, I provided the req. Well, this place didn't have reqs. You didn't find them -- they found you. It's a cool story, but it makes it harder to write about the interview process objectively. Subjectively? It was awesome. This firm wanted to hire Quality Engineers who are mature enough to make solid judgment calls about testing and...

Reader, In Part 1, you saw how a facilities management SaaS had to take a leap of faith because they didn't assess the technical ability of their candidates. Quick recap on the technical skills the req asked for: fwiw, not all of this seemed necessary from the conversations I had These were the skills the role actually required based on the conversations with the hiring team: Maestro or Appium (mobile app test automation scripting & debugging) Mobile testing TypeScript CI/CD for running tests...

Reader, Wow, are we halfway through February already? That was fast. Hope you had a great V Day over the weekend 🫶 Let's talk about another interview I had back in December.This was with a facilities management SaaS company for an SDET role. A friend from a past job referred me, so I thought it would be a walk in the park.(This is a terrible posture to take when getting referred, by the way. Great recruiters often speak with referrals so make sure you educate your candidate on mindset before...