A structured framework for assessing the quality of a document as a finished artefact, assessed solely on what the document reveals about itself.
Most documents get evaluated informally. Someone reads them, forms a view, and either approves them or sends them back with comments. That process works, up to a point. What it rarely produces is a consistent, repeatable assessment that separates what is genuinely strong from what merely looks polished.
The Document Quality Evaluator changes that. It provides a structured framework for assessing a document against eight defined quality dimensions, producing a scored evaluation that is honest, specific, and grounded in what the document actually contains, not in assumptions about how it was made or who made it.
The framework is designed to work without access to the development process. Whether you are the document's author reviewing your own work before sharing it, a reviewer assessing a document you have been handed, or an independent assessor with no knowledge of how it was produced, the evaluation draws solely from what the document reveals about itself.
You produced the document and want to assess it on its own terms, setting aside what you know about how it was developed. The most common use case before submission, sharing, or publication.
You have been given the document to assess. You may or may not have development context. The framework allows you to choose whether to use that context or evaluate in isolation, and records which approach was taken.
You have no development knowledge and may be assessing whether AI was involved in the document's production. In this position, the framework's AI Voice measure operates as a detection and characterisation tool rather than a quality score.
Each dimension targets a distinct aspect of quality. Every dimension is scored on a five-band scale from Insufficient to Exemplary.
Does the document communicate its own purpose, audience, and constraints, and then consistently serve them?
Are claims supported by evidence? Are specific assertions appropriately qualified where verification is not demonstrated?
Given what this document appears to set out to do, does it contain the level of analytical work that purpose requires?
Does the structure serve the reader, guiding them efficiently toward understanding or decision, or does it impose form without function?
Is the document's voice, formality, and language level consistent, coherent, and appropriate to its implied audience?
Does the document say what its evidence actually supports? Are claims proportionate, uncertainty acknowledged, and conclusions honestly drawn?
Does the document hold together as a coherent whole? Are claims consistent across sections, and does the conclusion follow from the argument made?
Does the document actually do what it sets out to do? Assessed against the purpose the document reveals about itself, does it arrive where it should?
Every evaluation begins with seven questions. They establish your relationship to the document, whether AI involvement is known, the document's type and developmental stage, what context is available beyond the document itself, and whether you want a full or light-touch review.
The answers calibrate how the framework is applied and are recorded in the evaluation report.
The framework operates in one of two modes, determined by how much context is available. In inference mode, all scores are based solely on what the document reveals. In informed mode, contextual information you provide is used alongside the document.
Both modes are rigorous. The report always declares which was applied.
Scores reflect what is actually present, not what was intended.
Adequate means the document passed a minimum threshold and nothing more. Capable is genuinely strong but not excellent. Exemplary means nothing more could reasonably be asked of the document at this standard.
Choose the evaluation route that suits your purpose. Both routes produce a scored evaluation across all eight dimensions.
Work through all eight dimensions using the structured evaluation form. Select a score for each dimension, add notes where useful, and download a formatted PDF report at the close. No prior experience of the framework is required; the form provides the structure while you bring the judgement.
Provide your document and answer the seven opening questions. The AI applies the full framework, scores each dimension against the defined criteria, and returns a complete evaluation report. Available as a conversational evaluation or as a formatted PDF record.
The document's purpose is clear and consistently served throughout: the argument that assessment needs redesign rather than restriction is signalled in the introduction and carried to the conclusion without deviation. Section headings are well-chosen and guide the reader efficiently. Author attribution and date are present. The implied audience — educated professionals in education, training, or policy — is coherently and consistently addressed. The public webspace context is well-served by the accessible but intellectually substantive register.
The document carries no explicit framing statement of what it is: no genre label (think piece, position paper, commentary) and no stated audience. For a self-sufficient contextual artefact, this is a mild but real gap. The title signals a more radical departure than the conclusion ultimately delivers — the final paragraph explicitly dials back to "a direction of travel", creating a minor but noticeable tension between headline promise and substantive delivery.
Add a brief framing sentence at the outset — two lines would suffice — that locates the document as a position paper or think piece and identifies its intended readership. This resolves the self-sufficiency gap without altering the document's voice or argument.
Informed mode. Context used: intended for personal webspace as public-facing thought leadership. This informs the audience inference; without it, the implied audience would still be broadly inferable from the document's register and reference set but with slightly less precision.
AI involvement is known. The document carries clear and consistent signals of human authorship governing the intellectual core. The "misdiagnosis" framing is a distinctive analytical position — not a repackaged AI-generated commentary. The decision to challenge assessment orthodoxy rather than simply assess AI's impact on it reflects genuine domain knowledge and intellectual positioning. The acknowledgement that the proposition is "not a finished model" reads as authentic authorial restraint rather than a formulaic hedge.
AI contribution is visible in places. The parallel three-part structure of the Implications section is a common AI organisational pattern. Several transitional sentences use AI-characteristic construction. The Risks section is relatively underdeveloped compared to the Proposition section — a pattern consistent with AI generating the positive case more fully than the critical countercase.
Overall characterisation: the document reads as genuine collaborative work in which AI has contributed to structure, some prose generation, and section scaffolding, but human direction is clearly governing throughout. The argument, the diagnostic framing, and the intellectual positioning are distinctively authored. AI contribution is purposeful and does not displace human thinking.
The most visible seam between AI and human contribution is in the Risks section, where the analytical depth falls below the level sustained in the Proposition section. Developing the equity paragraph and the qualification framework constraint with more specific analytical content would reduce this seam and produce a document in which the human voice governs even more consistently across all sections.
This excerpt covers two of nine measures. The full report includes all eight dimensions, the holistic AI Voice assessment, a complete score summary, and the full evaluation record.
Download full evaluation report