If you cannot show how NSS comments were included, categorised, checked, and governed, your findings will be hard to defend when faculty teams, boards, or TEF reviewers ask questions. A defensible NSS open-text workflow has five non-negotiables: clear scope, high coverage (ideally all comments), repeatable categorisation, documented QA, and governance (data protection, redaction, retention, access). If you need a fast decision guide, start with Best NSS comment analysis (2025). If you need a governed operational approach, see Student Voice Analytics.
What "open-text analysis" means (in practice)
Open-text analysis turns NSS free-text comments into evidence teams can use. In practice, that usually means:
Priorities (what matters most by volume and by negativity)
Evidence packs (what you can reasonably say to boards, TEF panels, and programme teams)
A defensible workflow (step-by-step)
1) Define scope and inclusion rules
Set the rules before you look at the outputs. That prevents avoidable disputes later.
Which survey(s): NSS only, or NSS + module evaluations + PTES/PRES/UKES?
Which populations: UG only, or include PGT/PGR where relevant?
What counts as "in scope": duplicates, empty strings, sarcasm/jokes, multi-issue comments.
2) Prepare data (minimal spec)
If the input table is inconsistent, every downstream chart becomes harder to trust.
You should be able to produce a table with:
comment_id, comment_text, survey, survey_year
organisation/unit fields (school/faculty/department) where permitted
discipline fields (CAH/HECoS) where available
cohort fields (level, mode, domicile group, etc.) where policy allows
3) Apply redaction and privacy controls
These controls are what let you move from exploratory analysis to reporting that can be shared safely. For a practical control list, use the student comment analysis governance checklist.
Decide what personal data is in scope to remove (names, emails, phone numbers, identifiers).
Document retention and access policies (least privilege).
4) Categorise comments (repeatably)
Repeatability is what separates governed reporting from a one-off interpretation. Defensibility comes from being able to repeat the analysis and explain what changed.
Track coverage (% assigned to a meaningful theme).
Track drift: if your categories change year to year, you need a mapping and change log.
5) QA and traceability
Even strong theme labels are hard to use in practice if nobody can verify them later.
Human QA: sample checks, edge cases, disagreement review.
Traceability: every headline claim should link back to supporting comments (anonymised).
Versioning: record model/prompt/version so results are reproducible.
Reporting: what good outputs look like
Good reporting should help teams decide what to fix next, not just describe what students said.
A small set of headline themes (top volume + top negative)
"What changed vs last year" (avoiding cohort-mix artefacts)
Benchmarked views where possible (by discipline and cohort)
A short actions section (what to change next term vs next year)
Where tools usually fail (and what to validate)
This is where promising demos often break down in operational use. Validate these issues before you commit to a workflow.
If you are comparing platforms rather than building a workflow in-house, our guide to text analysis software for education sets out where desktop, cloud, and HE-specific tools fit.
Low coverage ("too many uncategorised")
Generic categories that don’t map to HE reality
No benchmarking (or unclear methodology)
Weak governance (no audit trail, unclear data pathways)