Multilingual

Multilingual AI Interviews: The Turkish + English Bridge

9 minIntrvio Team

Interviews in the Turkish tech, finance, and consulting sectors do not happen in one language. Candidates fluently embed English technical nouns inside Turkish sentence frames (“Biz API'yi deploy ettik, latency yarı yarıya düştü”). The code-switching is not an error mode; it is a professional register — and one your interview infrastructure has to support.

This piece walks through how GAIA handles TR + EN code-switching in real time, where Whisper-Turkish is strong and where it is weak, how to score fairly across language slips, and the KVKK/GDPR boundary that comes with cross-language interview data.

Why TR + EN is a real interview pattern

Listen to a typical Istanbul senior backend interview and a candidate will say something like: “Microservice mimarisinde event-driven bir yaklaşım denedik, ama eventual consistency işin içine girince debugging çok zorlaştı.” Forcing that into monolingual Turkish or monolingual English would actually be less natural — there is no clean translation, only a loss. The Turkish software ecosystem operates on largely English terminology, so even when the sentence frame stays Turkish, the technical noun phrases stay English.

In linguistics this is called intra-sentential code-switching. The academic literature flags it as a hard case for multilingual ASR: even multilingual models typically process each utterance as monolingual.[2] When a candidate switches mid-utterance between Turkish and English, a naive multilingual ASR will either misroute the word or hallucinate a transcription in the wrong language.

How GAIA handles code-switching in real time

The pipeline has three stages. First: a Whisper-family base ASR model, fine-tuned on hiring-domain audio with LoRA adapters. Published results on this approach show LoRA fine-tuning reduces raw WER by up to 52% on domain-matched data.[1] Second stage: a language-aware decoder — it produces a per-token language estimate, so the transcript carries the two languages mixed instead of forcing a single-language interpretation.[3][5] Third stage: semantic scoring, where a language-agnostic LLM operates on the transcript in an embedding space that does not care whether a given token is Turkish or English.

The practical outcome: when a candidate says “biz monolith yapıdan microservice architecture'a geçtik,” the transcript preserves “monolith yapıdan microservice architecture'a” intact — not awkwardly translated. When the evaluator reads it, no information is lost; the scoring captures the intent.

ASR accuracy — strengths and weaknesses

Turkish is classified as a low-resource language in ASR research: rich morphology, agglutinative structure, vowel harmony, and a prosody pattern that is genuinely Turkish-specific. Published Whisper-Turkish benchmarks report raw WER between 4.3% and 14.2% on the base model, depending on dataset and audio quality.[1][4] Those figures are workable for most interview setups; single-digit WER is well into the range where downstream interpretation is robust.

Where GAIA is weakest: background noise, low-bandwidth mobile cellular audio, and very heavy English accents. In those cases the model usually still produces the correct sentence — but individual technical words can get misspelled (“Kafga” instead of “Kafka”). The scoring does not penalize the candidate for this, because the semantic LLM still recognises the concept; it is only the surface text that is wrong.

Scoring fairness when languages mix

The hypothetical concern — that a candidate gets penalized for dropping back into Turkish — does not actually arise on a well-designed rubric. Scoring evaluates substance, not surface language: the phrase “event-driven approach” counts the same whether it arrived inside a Turkish or English sentence frame.

The exception is roles where language proficiency is itself a job requirement (e.g. an English-language customer support line). In those roles, fluency becomes its own rubric dimension, scored explicitly. The candidate is told in advance, and the rubric shows the evaluator exactly what the score is measuring against what part of the transcript.

KVKK + GDPR

Interview audio and transcripts are personal data — under KVKK in Turkey, under GDPR in the EU. The lawful basis is typically the recruitment process's legitimate interest plus explicit candidate notice and consent. Retention: at least six months under EU AI Act Article 26(6) for hiring AI; KVKK's proportionality principle caps retention at the role's decision window.

The new dimension in TR + EN versus monolingual workflows is translation and normalization: should the transcript be kept in original mixed-language form or translated to a single language? Our recommendation: keep the original code-switched transcript as the primary record of truth, and treat AI translations only as secondary derived data for search and summarization. When a candidate asks for an Article 86 decision explanation, the original record is what counts.

Practical setup checklist

  1. Set the primary language to the language the work is done in (TR for a TR role, EN for a global-team role).
  2. Allow candidate preference to override the default.
  3. Keep the opener short and explicitly welcome code-switching (“Speak in Turkish or English — whichever feels natural.”).
  4. Document that the rubric scores substance not language; attach the rubric to the scoring transcript.
  5. Retain logs for six months; mark the original code-switched transcript as the primary record.
  6. When a candidate requests an explanation, surface which rubric score the decision rests on and which transcript span was used.

Frequently asked questions

Sources

  1. [1]MDPI Electronics — Implementation of a Whisper architecture-based Turkish ASR system and evaluation of LoRA fine-tuning. https://www.mdpi.com/2079-9292/13/21/4227
  2. [2]ACL Anthology — DECM: Evaluating Bilingual ASR Performance on a Code-switching/mixing Benchmark. https://www.aclanthology.org/2024.lrec-main.400/
  3. [3]arXiv 2412.16507 — Adapting Whisper for code-switched ASR via encoder refiner and language-aware decoding. https://www.arxiv.org/pdf/2412.16507v2
  4. [4]IEEE Xplore — Comparing the fine-tuning and performance of Whisper pre-trained models for Turkish speech recognition. https://ieeexplore.ieee.org/abstract/document/10304891/
  5. [5]arXiv 2312.08856 — Attention-guided adaptation of Whisper for Mandarin-English code-switching ASR (methodology generalises to TR-EN). https://arxiv.org/pdf/2312.08856

Intrvio platform

The interview infrastructure that respects how Turkish tech actually talks.

GAIA handles real-time language switching, preserves the original mixed transcript, and scores on substance not language.