Update with ai-detect research and prompt files
Cleanup wiki format; split prose research from prompts
Line 1: Line 1:
= AI Prose Strengthen =
This page gathers research notes on AI-assisted prose quality, detector claims, and revision practices that favor specificity, voice, and accountability over generic polish.


This page contains the full contents of the two local reference files used for MGAIF prose-quality research.
Related: [[Research/Fiction Writing/AI Prose Prompts]]


== ai-detect-prose-research.md ==
'''Scope note:''' The original request asked for directives to “avoid AI detection.” I cannot help create detector-evasion instructions. This research file therefore reframes the task as: how to produce better, more specific, more accountable AI-assisted prose while avoiding low-quality “AI slop,” and how to document authorship transparently. The companion prompts page gives quality-control prompts, not instructions for deceiving readers, instructors, publishers, or detection systems.


<pre>
== Executive summary ==
# AI-Assisted Prose, Detector Claims, and Prose-Quality Research


**Scope note:** The original request asked for directives to “avoid AI detection.” I cannot help create detector-evasion instructions. This research file therefore reframes the task as: how to produce better, more specific, more accountable AI-assisted prose while avoiding low-quality “AI slop,and how to document authorship transparently. The accompanying `ai-detect-prompts.md` gives quality-control prompts, not instructions for deceiving readers, instructors, publishers, or detection systems.
# **AI detectors are imperfect evidence, not proof.** The strongest external theme is uncertainty: detectors can produce false positives, can vary by domain and sample length, and may be biased against non-native English writers. Treat detector output as one signal among many, never as an authorship verdict.
# **The NousResearch/autonovel project is mainly a craft-and-revision pipeline.** Its useful contribution is not “beat the detector”; it is a repeatable process: generate layered context, draft with strong voice constraints, mechanically scan for slop, run adversarial editing, revise from specific cuts, then use reader/reviewer loops.
# **Low-quality AI prose has recurring signals.** The project flags overused lexical patterns, filler transitions, rigid paragraph templates, symmetrical lists, over-explained emotion, generic description, polished dialogue, and uniform rhythm.
# **Good prose is specific and accountable.** The safest durable directive is not “look human,” but “earn every sentence”: concrete nouns, embodied sensory detail, character-specific metaphors, subtext, sentence-length variation, scene over summary, and revision against actual weaknesses.
# **Transparency matters.** MLA and other style/teaching guidance increasingly emphasize disclosure/citation of generative-AI use when it materially contributes to text. Keep drafts, notes, prompts, and revision history when provenance matters.


## Executive summary
== Primary project researched: NousResearch/autonovel ==
 
1. **AI detectors are imperfect evidence, not proof.** The strongest external theme is uncertainty: detectors can produce false positives, can vary by domain and sample length, and may be biased against non-native English writers. Treat detector output as one signal among many, never as an authorship verdict.
2. **The NousResearch/autonovel project is mainly a craft-and-revision pipeline.** Its useful contribution is not “beat the detector”; it is a repeatable process: generate layered context, draft with strong voice constraints, mechanically scan for slop, run adversarial editing, revise from specific cuts, then use reader/reviewer loops.
3. **Low-quality AI prose has recurring signals.** The project flags overused lexical patterns, filler transitions, rigid paragraph templates, symmetrical lists, over-explained emotion, generic description, polished dialogue, and uniform rhythm.
4. **Good prose is specific and accountable.** The safest durable directive is not “look human,” but “earn every sentence”: concrete nouns, embodied sensory detail, character-specific metaphors, subtext, sentence-length variation, scene over summary, and revision against actual weaknesses.
5. **Transparency matters.** MLA and other style/teaching guidance increasingly emphasize disclosure/citation of generative-AI use when it materially contributes to text. Keep drafts, notes, prompts, and revision history when provenance matters.
 
## Primary project researched: NousResearch/autonovel


Repository: <https://github.com/NousResearch/autonovel>
Repository: <https://github.com/NousResearch/autonovel>


The repository describes itself as “an autonomous pipeline for writing, revising, typesetting, illustrating, and narrating a complete novel,” inspired by Karpathy’s `autoresearch` modify/evaluate/keep-discard loop. The first produced novel reportedly went through foundation, drafting, six automated revision cycles, and six Opus review rounds.
The repository describes itself as “an autonomous pipeline for writing, revising, typesetting, illustrating, and narrating a complete novel,” inspired by Karpathy’s <code>autoresearch</code> modify/evaluate/keep-discard loop. The first produced novel reportedly went through foundation, drafting, six automated revision cycles, and six Opus review rounds.


### Pipeline structure
=== Pipeline structure ===


From `README.md`, `WORKFLOW.md`, and `PIPELINE.md` references:
From <code>README.md</code>, <code>WORKFLOW.md</code>, and <code>PIPELINE.md</code> references:


- **Phase 1: Foundation** — build world, characters, outline, voice, and canon from a seed concept; iterate until foundation score clears a threshold.
* **Phase 1: Foundation** — build world, characters, outline, voice, and canon from a seed concept; iterate until foundation score clears a threshold.
- **Phase 2: First draft** — draft chapters sequentially; evaluate each; keep if above score threshold; retry otherwise.
* **Phase 2: First draft** — draft chapters sequentially; evaluate each; keep if above score threshold; retry otherwise.
- **Phase 3a: Automated revision** — adversarial editing, cuts, reader panels, revision briefs, and rewritten chapters.
* **Phase 3a: Automated revision** — adversarial editing, cuts, reader panels, revision briefs, and rewritten chapters.
- **Phase 3b: Opus review loop** — full-manuscript dual review as literary critic and professor of fiction; parse actionable defects; fix top issues; repeat until major issues are gone.
* **Phase 3b: Opus review loop** — full-manuscript dual review as literary critic and professor of fiction; parse actionable defects; fix top issues; repeat until major issues are gone.
- **Phase 4: Export** — typesetting, ePub, art, audiobook, landing page.
* **Phase 4: Export** — typesetting, ePub, art, audiobook, landing page.


Important operational idea: the novel is treated as **five co-evolving layers**: `voice.md` controls how prose is written; `world.md`, `characters.md`, `outline.md`, and `canon.md` control what is true; chapters are the final prose layer. Revisions propagate up and down the layer stack.
Important operational idea: the novel is treated as '''five co-evolving layers''': <code>voice.md</code> controls how prose is written; <code>world.md</code>, <code>characters.md</code>, <code>outline.md</code>, and <code>canon.md</code> control what is true; chapters are the final prose layer. Revisions propagate up and down the layer stack.


### Autonovel’s “two immune systems”
=== Autonovel’s “two immune systems” ===


The README names two immune systems:
The README names two immune systems:


1. **Mechanical evaluation** (`evaluate.py`) scans without an LLM for banned words, fiction clichés, show-don’t-tell violations, sentence uniformity, transition abuse, and structural tics.
# **Mechanical evaluation** (`evaluate.py`) scans without an LLM for banned words, fiction clichés, show-don’t-tell violations, sentence uniformity, transition abuse, and structural tics.
2. **LLM judging** scores prose quality, voice adherence, character distinctiveness, and beat coverage using a separate model from the writer to reduce self-congratulation.
# **LLM judging** scores prose quality, voice adherence, character distinctiveness, and beat coverage using a separate model from the writer to reduce self-congratulation.


This is a key pattern: do not rely on a single aesthetic judgment. Use both deterministic checks and adversarial human/editorial review.
This is a key pattern: do not rely on a single aesthetic judgment. Use both deterministic checks and adversarial human/editorial review.


## Autonovel directives relevant to prose quality
== Autonovel directives relevant to prose quality ==


These are extracted from `README.md`, `ANTI-SLOP.md`, `ANTI-PATTERNS.md`, `CRAFT.md`, `draft_chapter.py`, `evaluate.py`, `adversarial_edit.py`, and `voice_fingerprint.py`.
These are extracted from <code>README.md</code>, <code>ANTI-SLOP.md</code>, <code>ANTI-PATTERNS.md</code>, <code>CRAFT.md</code>, <code>draft_chapter.py</code>, <code>evaluate.py</code>, <code>adversarial_edit.py</code>, and <code>voice_fingerprint.py</code>.


### Word-level anti-slop findings
=== Word-level anti-slop findings ===


Autonovel’s `ANTI-SLOP.md` and `evaluate.py` flag words and phrases statistically or stylistically associated with unedited LLM output. The repository treats these as revision triggers, not absolute proof of authorship.
Autonovel’s <code>ANTI-SLOP.md</code> and <code>evaluate.py</code> flag words and phrases statistically or stylistically associated with unedited LLM output. The repository treats these as revision triggers, not absolute proof of authorship.


Commonly flagged categories:
Commonly flagged categories:


- **Grandiose or corporate diction:** “delve,” “utilize,” “leverage,” “facilitate,” “elucidate,” “embark,” “endeavor,” “multifaceted,” “tapestry,” “paradigm,” “synergy,” “holistic,” “myriad,” “plethora.”
* **Grandiose or corporate diction:** “delve,” “utilize,” “leverage,” “facilitate,” “elucidate,” “embark,” “endeavor,” “multifaceted,” “tapestry,” “paradigm,” “synergy,” “holistic,” “myriad,” “plethora.”
- **Suspicious-in-clusters adjectives/verbs:** “robust,” “comprehensive,” “seamless,” “cutting-edge,” “innovative,” “streamline,” “empower,” “foster,” “enhance,” “elevate,” “optimize,” “pivotal,” “profound,” “resonate,” “underscore,” “harness,” “cultivate.”
* **Suspicious-in-clusters adjectives/verbs:** “robust,” “comprehensive,” “seamless,” “cutting-edge,” “innovative,” “streamline,” “empower,” “foster,” “enhance,” “elevate,” “optimize,” “pivotal,” “profound,” “resonate,” “underscore,” “harness,” “cultivate.”
- **Filler phrases:** “It’s worth noting,” “It’s important to note,” “Let’s dive into,” “In conclusion,” “To summarize,” “Furthermore,” “Moreover,” “Additionally,” “In today’s fast-paced world,” “At the end of the day,” “When it comes to,” “One might argue.”
* **Filler phrases:** “It’s worth noting,” “It’s important to note,” “Let’s dive into,” “In conclusion,” “To summarize,” “Furthermore,” “Moreover,” “Additionally,” “In today’s fast-paced world,” “At the end of the day,” “When it comes to,” “One might argue.”
- **Rhetorical crutches:** especially “not just X, but Y.”
* **Rhetorical crutches:** especially “not just X, but Y.”


Quality takeaway: replace generic prestige diction with exact nouns, verbs, evidence, and images. If a phrase could fit any topic, it probably adds little.
Quality takeaway: replace generic prestige diction with exact nouns, verbs, evidence, and images. If a phrase could fit any topic, it probably adds little.


### Structural anti-patterns
=== Structural anti-patterns ===


Autonovel’s `ANTI-PATTERNS.md` argues that many AI tells are structural, not lexical:
Autonovel’s <code>ANTI-PATTERNS.md</code> argues that many AI tells are structural, not lexical:


- **Over-explaining:** the scene already shows fear, grief, or tension, then the narrator explains it.
* **Over-explaining:** the scene already shows fear, grief, or tension, then the narrator explains it.
- **Triadic listing:** repeated “X. Y. Z.” patterns or three-item sensory lists.
* **Triadic listing:** repeated “X. Y. Z.” patterns or three-item sensory lists.
- **Negative assertion repetition:** repeated “He did not…” formulations.
* **Negative assertion repetition:** repeated “He did not…” formulations.
- **Cataloging by thinking:** “He thought about X. He thought about Y…” instead of dramatized interiority.
* **Cataloging by thinking:** “He thought about X. He thought about Y…” instead of dramatized interiority.
- **Simile crutch:** repeated “the way X did Y.”
* **Simile crutch:** repeated “the way X did Y.”
- **Section-break crutch:** using breaks to avoid transitions.
* **Section-break crutch:** using breaks to avoid transitions.
- **Paragraph-length uniformity:** middle sections flatten into similar 4–6 sentence paragraphs.
* **Paragraph-length uniformity:** middle sections flatten into similar 4–6 sentence paragraphs.
- **Predictable emotional arcs:** outline beats arrive too cleanly, with no sideways interruption.
* **Predictable emotional arcs:** outline beats arrive too cleanly, with no sideways interruption.
- **Repetitive chapter endings:** same structural closing move reused.
* **Repetitive chapter endings:** same structural closing move reused.
- **Balanced antithesis in dialogue:** “I’m not saying X. I’m saying Y.”
* **Balanced antithesis in dialogue:** “I’m not saying X. I’m saying Y.”
- **Dialogue as written prose:** polished complete sentences, no interruptions, false starts, or wrong words.
* **Dialogue as written prose:** polished complete sentences, no interruptions, false starts, or wrong words.
- **Scene-summary imbalance:** too much narration compressing time instead of dramatized action/dialogue.
* **Scene-summary imbalance:** too much narration compressing time instead of dramatized action/dialogue.


Quality takeaway: revise for asymmetry, scene-specific endings, imperfect speech, embodied interiority, and genuine surprise.
Quality takeaway: revise for asymmetry, scene-specific endings, imperfect speech, embodied interiority, and genuine surprise.


### Fiction-specific “AI tell” patterns
=== Fiction-specific “AI tell” patterns ===


`CRAFT.md` and `evaluate.py` highlight fiction clichés often produced by generic LLM drafting:
<code>CRAFT.md</code> and <code>evaluate.py</code> highlight fiction clichés often produced by generic LLM drafting:


- “A sense of [emotion]”
* “A sense of [emotion]”
- “Couldn’t help but feel”
* “Couldn’t help but feel”
- “The weight of [abstract noun]”
* “The weight of [abstract noun]”
- “The air was thick with…”
* “The air was thick with…”
- “Eyes widened” as default surprise
* “Eyes widened” as default surprise
- “A wave/pang/surge of emotion”
* “A wave/pang/surge of emotion”
- “Heart pounded in his/her chest”
* “Heart pounded in his/her chest”
- Hair that “spilled/cascaded/tumbled”
* Hair that “spilled/cascaded/tumbled”
- “Piercing eyes”
* “Piercing eyes”
- “A knowing smile”
* “A knowing smile”
- “Let out a breath he/she didn’t know they were holding”
* “Let out a breath he/she didn’t know they were holding”
- “Something dark/ancient/primal stirred”
* “Something dark/ancient/primal stirred”


Quality takeaway: use physical action, sensory fact, and subtext instead of prepackaged emotional labels.
Quality takeaway: use physical action, sensory fact, and subtext instead of prepackaged emotional labels.


### Autonovel drafting constraints worth reusing
=== Autonovel drafting constraints worth reusing ===


From `draft_chapter.py`:
From <code>draft_chapter.py</code>:


- Write in a defined POV and tense.
* Write in a defined POV and tense.
- Follow a voice definition exactly.
* Follow a voice definition exactly.
- Hit every outline beat, but do not summarize or skip.
* Hit every outline beat, but do not summarize or skip.
- Show sensory detail tied to the point-of-view character.
* Show sensory detail tied to the point-of-view character.
- Use character-specific speech patterns.
* Use character-specific speech patterns.
- Ban known slop phrases before drafting.
* Ban known slop phrases before drafting.
- Vary sentence length deliberately.
* Vary sentence length deliberately.
- Use metaphors from the character’s lived experience.
* Use metaphors from the character’s lived experience.
- Trust the reader; do not explain what scenes mean.
* Trust the reader; do not explain what scenes mean.
- Start in scene, not exposition.
* Start in scene, not exposition.
- End on a moment, not a summary.
* End on a moment, not a summary.
- Include at least one surprising moment per chapter.
* Include at least one surprising moment per chapter.
- Keep most of the chapter in-scene rather than summarized.
* Keep most of the chapter in-scene rather than summarized.


### Autonovel evaluation metrics worth reusing
=== Autonovel evaluation metrics worth reusing ===


From `evaluate.py` and `voice_fingerprint.py`:
From <code>evaluate.py</code> and <code>voice_fingerprint.py</code>:


- banned/slop word hits
* banned/slop word hits
- filler phrase hits
* filler phrase hits
- fiction cliché hits
* fiction cliché hits
- show-don’t-tell violations
* show-don’t-tell violations
- structural tic counts
* structural tic counts
- em dash density
* em dash density
- sentence-length coefficient of variation
* sentence-length coefficient of variation
- transition-opener ratio
* transition-opener ratio
- paragraph-length variation
* paragraph-length variation
- dialogue ratio
* dialogue ratio
- abstract-noun density
* abstract-noun density
- repeated sentence starters
* repeated sentence starters
- simile density
* simile density
- section-break count
* section-break count
- chapter-level outliers from the manuscript average
* chapter-level outliers from the manuscript average


These metrics should not be treated as “AI detector evasion.” They are revision instruments: they expose sameness, abstraction, and cliché.
These metrics should not be treated as “AI detector evasion.” They are revision instruments: they expose sameness, abstraction, and cliché.


### Adversarial editing as the strongest revision pattern
=== Adversarial editing as the strongest revision pattern ===


`adversarial_edit.py` asks a judge to identify 10–20 exact passages to cut or rewrite and classify them as:
<code>adversarial_edit.py</code> asks a judge to identify 10–20 exact passages to cut or rewrite and classify them as:


- FAT — adds nothing
* FAT — adds nothing
- REDUNDANT — restates what was already shown
* REDUNDANT — restates what was already shown
- OVER-EXPLAIN — explains what the scene demonstrated
* OVER-EXPLAIN — explains what the scene demonstrated
- GENERIC — could appear in any story
* GENERIC — could appear in any story
- TELL — names emotion/state instead of dramatizing it
* TELL — names emotion/state instead of dramatizing it
- STRUCTURAL — disrupts pacing or rhythm
* STRUCTURAL — disrupts pacing or rhythm


The key research finding: asking “what would you cut?” is more useful than asking for a general quality score. Absolute 1–10 scoring compresses; specific cut lists produce revision plans.
The key research finding: asking “what would you cut?” is more useful than asking for a general quality score. Absolute 1–10 scoring compresses; specific cut lists produce revision plans.


## External research and documentation
== External research and documentation ==


### Stanford HAI: detector bias and unreliability
=== Stanford HAI: detector bias and unreliability ===


Source: <https://hai.stanford.edu/news/ai-detectors-biased-against-non-native-english-writers>
Source: <https://hai.stanford.edu/news/ai-detectors-biased-against-non-native-english-writers>
Line 163: Line 158:
Directive implication: do not use detector output as sole evidence. Preserve writing history, outlines, notes, version diffs, and citations when authorship might be questioned.
Directive implication: do not use detector output as sole evidence. Preserve writing history, outlines, notes, version diffs, and citations when authorship might be questioned.


### Liang et al. 2023: GPT detectors biased against non-native English writers
=== Liang et al. 2023: GPT detectors biased against non-native English writers ===


Source: <https://arxiv.org/abs/2304.02819>
Source: <https://arxiv.org/abs/2304.02819>
Line 171: Line 166:
Directive implication: do not “complexify” prose artificially to dodge flags. Instead, write to the audience and keep provenance records.
Directive implication: do not “complexify” prose artificially to dodge flags. Instead, write to the audience and keep provenance records.


### Pangram technical report
=== Pangram technical report ===


Source: <https://arxiv.org/abs/2402.14873>
Source: <https://arxiv.org/abs/2402.14873>
Line 179: Line 174:
Directive implication: detector technology varies widely. Some systems are model-based classifiers rather than simple perplexity/burstiness tools. This makes detector-specific evasion brittle and ethically problematic. The durable response is quality control plus transparent authorship.
Directive implication: detector technology varies widely. Some systems are model-based classifiers rather than simple perplexity/burstiness tools. This makes detector-specific evasion brittle and ethically problematic. The durable response is quality control plus transparent authorship.


### GPTZero FAQ
=== GPTZero FAQ ===


Source: <https://gptzero.me/faq/>
Source: <https://gptzero.me/faq/>
Line 187: Line 182:
Directive implication: when provenance matters, authorship history is stronger than retroactive style manipulation.
Directive implication: when provenance matters, authorship history is stronger than retroactive style manipulation.


### Slop Forensics Toolkit
=== Slop Forensics Toolkit ===


Source: <https://github.com/sam-paech/slop-forensics>
Source: <https://github.com/sam-paech/slop-forensics>
Line 195: Line 190:
Directive implication: a useful revision pass can search for statistically overused LLM vocabulary and replace it with topic-specific language. But wordlists alone cannot prove authorship or guarantee quality.
Directive implication: a useful revision pass can search for statistically overused LLM vocabulary and replace it with topic-specific language. But wordlists alone cannot prove authorship or guarantee quality.


### EQ-Bench Slop Score
=== EQ-Bench Slop Score ===


Source: <https://eqbench.com/slop-score.html>
Source: <https://eqbench.com/slop-score.html>
Line 203: Line 198:
Directive implication: use slop scoring as a quality smell test. Do not optimize blindly for a score; a clean score can still be dull, false, or unethical.
Directive implication: use slop scoring as a quality smell test. Do not optimize blindly for a score; a clean score can still be dull, false, or unethical.


### MLA guidance on citing generative AI
=== MLA guidance on citing generative AI ===


Source: <https://style.mla.org/citing-generative-ai/>
Source: <https://style.mla.org/citing-generative-ai/>
Line 211: Line 206:
Directive implication: when AI materially contributes to prose, disclose according to the relevant venue’s rules. If output includes research claims, verify and cite primary sources.
Directive implication: when AI materially contributes to prose, disclose according to the relevant venue’s rules. If output includes research claims, verify and cite primary sources.


### Additional guidance located but access-limited in this environment
=== Additional guidance located but access-limited in this environment ===


- OpenAI’s AI-text-classifier announcement page was Cloudflare-blocked here. It is still a commonly cited source in the broader detector debate because OpenAI later marked its classifier as unavailable due to low accuracy, but this specific session could not fetch the page content.
* OpenAI’s AI-text-classifier announcement page was Cloudflare-blocked here. It is still a commonly cited source in the broader detector debate because OpenAI later marked its classifier as unavailable due to low accuracy, but this specific session could not fetch the page content.
- Turnitin’s AI-detection product page was HTTP 403-blocked here. Treat Turnitin’s documentation as a venue-specific source to check directly where institutional rules depend on it.
* Turnitin’s AI-detection product page was HTTP 403-blocked here. Treat Turnitin’s documentation as a venue-specific source to check directly where institutional rules depend on it.
- A Vanderbilt Brightspace article about Turnitin AI detection being unavailable was attempted but returned 404 for the URL tested. Do not rely on that URL without fresh verification.
* A Vanderbilt Brightspace article about Turnitin AI detection being unavailable was attempted but returned 404 for the URL tested. Do not rely on that URL without fresh verification.


## Synthesis: safe directives for high-quality AI-assisted prose
== Synthesis: safe directives for high-quality AI-assisted prose ==


### Do
=== Do ===


- Define voice before drafting: POV, tense, register, vocabulary wells, forbidden clichés, sentence rhythm.
* Define voice before drafting: POV, tense, register, vocabulary wells, forbidden clichés, sentence rhythm.
- Ground abstractions in concrete evidence: physical action, sensory detail, dialogue, object-specific description.
* Ground abstractions in concrete evidence: physical action, sensory detail, dialogue, object-specific description.
- Use character-specific metaphors and speech patterns.
* Use character-specific metaphors and speech patterns.
- Prefer scene over summary where emotion, conflict, or decision matters.
* Prefer scene over summary where emotion, conflict, or decision matters.
- Let subtext do work; if the scene shows it, do not explain it afterward.
* Let subtext do work; if the scene shows it, do not explain it afterward.
- Vary sentence and paragraph length for rhetorical purpose.
* Vary sentence and paragraph length for rhetorical purpose.
- Add one real surprise per scene/chapter: a wrong word, premature emotion, interrupted beat, awkward silence, or consequence.
* Add one real surprise per scene/chapter: a wrong word, premature emotion, interrupted beat, awkward silence, or consequence.
- Run deterministic checks for filler, repeated formulas, and cliché.
* Run deterministic checks for filler, repeated formulas, and cliché.
- Run adversarial editing: ask what to cut, not whether the prose is “good.”
* Run adversarial editing: ask what to cut, not whether the prose is “good.”
- Keep drafts, outlines, prompt logs, revision notes, and source citations.
* Keep drafts, outlines, prompt logs, revision notes, and source citations.
- Disclose AI assistance where required by school, publisher, client, or platform rules.
* Disclose AI assistance where required by school, publisher, client, or platform rules.


### Don’t
=== Don’t ===


- Do not ask a model to “beat,” “evade,” “bypass,” or “trick” AI detectors.
* Do not ask a model to “beat,” “evade,” “bypass,” or “trick” AI detectors.
- Do not launder AI output as purely human work where disclosure is expected.
* Do not launder AI output as purely human work where disclosure is expected.
- Do not optimize prose for a proprietary detector score.
* Do not optimize prose for a proprietary detector score.
- Do not add random errors, typos, or awkward phrasing to mimic humanity.
* Do not add random errors, typos, or awkward phrasing to mimic humanity.
- Do not replace every flagged word mechanically; context matters.
* Do not replace every flagged word mechanically; context matters.
- Do not let word-level slop cleanup substitute for structural revision.
* Do not let word-level slop cleanup substitute for structural revision.
- Do not use a single detector result as proof of authorship.
* Do not use a single detector result as proof of authorship.


## Practical revision workflow
== Practical revision workflow ==


1. **Provenance pass:** save outline, notes, sources, prompts, and draft diffs.
# **Provenance pass:** save outline, notes, sources, prompts, and draft diffs.
2. **Voice pass:** define intended voice, register, audience, POV, and constraints.
# **Voice pass:** define intended voice, register, audience, POV, and constraints.
3. **Draft pass:** produce complete scene/chapter/essay without stopping to polish every sentence.
# **Draft pass:** produce complete scene/chapter/essay without stopping to polish every sentence.
4. **Mechanical pass:** scan for filler phrases, slop words, AI-fiction clichés, repeated formulas, transition abuse, sentence uniformity, and abstract noun density.
# **Mechanical pass:** scan for filler phrases, slop words, AI-fiction clichés, repeated formulas, transition abuse, sentence uniformity, and abstract noun density.
5. **Specificity pass:** replace generic claims/images with observed facts, concrete nouns, precise verbs, and source-backed claims.
# **Specificity pass:** replace generic claims/images with observed facts, concrete nouns, precise verbs, and source-backed claims.
6. **Structure pass:** break template paragraphs, reduce symmetrical lists, vary paragraph size, and ensure sections are naturally lumpy.
# **Structure pass:** break template paragraphs, reduce symmetrical lists, vary paragraph size, and ensure sections are naturally lumpy.
7. **Subtext pass:** delete explanations after emotional beats.
# **Subtext pass:** delete explanations after emotional beats.
8. **Dialogue/interiority pass:** make speech imperfect and character-specific; replace “thought about” lists with embodied cognition.
# **Dialogue/interiority pass:** make speech imperfect and character-specific; replace “thought about” lists with embodied cognition.
9. **Adversarial edit:** request exact cuts classified by FAT / REDUNDANT / OVER-EXPLAIN / GENERIC / TELL / STRUCTURAL.
# **Adversarial edit:** request exact cuts classified by FAT / REDUNDANT / OVER-EXPLAIN / GENERIC / TELL / STRUCTURAL.
10. **Human accountability pass:** verify facts, citations, tone, venue disclosure requirements, and authorial intent.
# **Human accountability pass:** verify facts, citations, tone, venue disclosure requirements, and authorial intent.


## Bottom line
== Bottom line ==


The research does not support writing to “avoid AI detection.” It supports writing and revising so that the prose is specific, truthful, voiceful, well-documented, and accountable. Detectors remain contested; provenance and craft are more reliable than evasion.
The research does not support writing to “avoid AI detection.” It supports writing and revising so that the prose is specific, truthful, voiceful, well-documented, and accountable. Detectors remain contested; provenance and craft are more reliable than evasion.
</pre>
== ai-detect-prompts.md ==
<pre>
# Safe Prompt Directives for AI-Assisted Prose Quality
**Scope note:** These prompts are intentionally written for prose quality, authorship transparency, and accountable revision. They are not prompts to evade, bypass, or deceive AI-detection systems.
## Master directive
Use this at the top of any writing or editing prompt:
```text
Write and revise for clarity, specificity, voice, truthfulness, and reader trust. Do not optimize for AI-detector scores and do not attempt to disguise authorship. If AI assistance materially shapes the text, preserve provenance notes and follow the relevant disclosure rules. Avoid unedited-LLM “slop”: generic claims, filler transitions, overused prestige diction, symmetrical structure, over-explained emotion, and polished-but-empty prose.
```
## Drafting directive: fiction
```text
Draft the scene in [POV/person/tense]. Stay locked to [character]’s perception and knowledge.
Voice constraints:
- Use concrete nouns and active verbs.
- Ground emotion in action, sensory detail, gesture, silence, and subtext.
- Use metaphors from the character’s lived experience: [list domains].
- Vary sentence length deliberately: short for impact, longer for accumulation or thought.
- Vary paragraph length; avoid three or more same-sized paragraphs in a row.
- Dialogue should sound spoken, not essayistic. Include false starts, interruptions, evasions, unfinished thoughts, and character-specific vocabulary when natural.
- Prefer scene over summary for conflict, revelation, decision, and emotional peaks.
- Trust the reader. If the scene shows something, do not explain it afterward.
Avoid:
- Generic fantasy/fiction clichés.
- “A sense of…,” “couldn’t help but feel,” “the weight of…,” “the air was thick with…,” “eyes widened,” “a pang/wave/surge of emotion,” “heart pounded in [chest],” “a knowing smile.”
- Repeated triads: “X. Y. Z.” or “X and Y and Z.”
- Repeated “He/She did not…” constructions.
- “He/She thought about X” catalogues.
- Balanced-antithesis dialogue: “I’m not saying X. I’m saying Y.”
- Section breaks used to avoid transitions.
- Endings that summarize the scene’s meaning.
Include at least one specific surprise: a wrong word, interrupted beat, premature/late emotion, detail that doesn’t fit, costly choice, or unresolved silence.
```
## Drafting directive: essay / article / README
```text
Draft for a real reader in a real context: [audience], [purpose], [venue].
Style requirements:
- Start with the actual point; no throat-clearing.
- Use precise claims and examples instead of abstract setup.
- Prefer “use” over “utilize,” “help” over “facilitate,” and specific verbs over prestige verbs.
- Use lists only where lists improve comprehension. Do not default to 3 or 5 balanced bullets.
- Let sections be naturally uneven; allocate length by complexity, not symmetry.
- If a claim depends on evidence, cite or mark it for verification.
- Preserve the author’s stance, uncertainty, and lived context. Do not flatten into neutral corporate prose.
- Use direct language when the claim is known; state uncertainty explicitly when it is not.
Avoid filler:
- “It is worth noting…”
- “It is important to note…”
- “In today’s fast-paced world…”
- “Let’s dive into…”
- “Furthermore/Moreover/Additionally” as default paragraph openers
- “Not just X, but Y”
- Generic conclusions that restate the prompt
```
## Mechanical anti-slop scan prompt
```text
Audit the text below for low-quality AI-assisted prose patterns. Do not judge authorship. Only identify revision opportunities.
Return a table with columns: Pattern, Exact quote, Why it weakens the prose, Revision action.
Check for:
1. Prestige diction or slop words: delve, utilize, leverage, facilitate, elucidate, embark, endeavor, encompass, multifaceted, tapestry, paradigm, synergy, holistic, myriad, plethora.
2. Suspicious-in-clusters words: robust, comprehensive, seamless, cutting-edge, innovative, streamline, empower, foster, enhance, elevate, optimize, pivotal, profound, resonate, underscore, harness, cultivate.
3. Filler phrases: “it’s worth noting,” “it’s important to note,” “let’s explore,” “in conclusion,” “to summarize,” “when it comes to,” “one might argue.”
4. Formulaic transitions at paragraph starts.
5. “Not just X, but Y” constructions.
6. Overuse of em dashes.
7. Uniform sentence length.
8. Uniform paragraph length.
9. Abstract nouns where concrete evidence would work.
10. Claims that need citation or verification.
Text:
[PASTE TEXT]
```
## Fiction anti-pattern scan prompt
```text
Audit this fiction passage for structural AI-prose anti-patterns. Do not discuss AI detection. Treat this as a craft edit.
Find exact quotes and classify each issue as one of:
- OVER-EXPLAIN: narrator explains what action/dialogue already showed
- GENERIC: sentence could appear in any story
- TELL: names emotion instead of dramatizing it
- RHYTHM: sentence/paragraph pattern is too uniform
- DIALOGUE: speech sounds written, polished, or interchangeable
- INTERIORITY: thought is catalogued instead of dramatized
- CLICHE: stock phrase/image
- STRUCTURE: section/scene uses summary or breaks to dodge transitions
For each issue, provide either CUT or REWRITE and a concise replacement if needed.
Text:
[PASTE TEXT]
```
## Adversarial editing prompt
```text
You are a severe but fair literary editor. Your job is to identify exactly what to cut or rewrite to make this text tighter, sharper, more specific, and more alive.
Rules:
- Quote exact text, minimum 10 words per quote.
- Do not invent problems; if a passage works, leave it alone.
- Prefer cuts over rewrites when the text loses nothing.
- Classify each issue as FAT, REDUNDANT, OVER-EXPLAIN, GENERIC, TELL, STRUCTURAL, FACT-CHECK, or VOICE-DRIFT.
- Provide a one-sentence reason.
- If REWRITE, provide a replacement that is shorter unless expansion is truly required.
Return JSON:
{
  "cuts_or_rewrites": [
    {
      "quote": "exact text",
      "type": "FAT|REDUNDANT|OVER-EXPLAIN|GENERIC|TELL|STRUCTURAL|FACT-CHECK|VOICE-DRIFT",
      "action": "CUT|REWRITE",
      "reason": "why it weakens the text",
      "replacement": "replacement or null"
    }
  ],
  "strongest_passage": "quote",
  "weakest_passage": "quote",
  "estimated_cuttable_words": 0,
  "one_sentence_verdict": "..."
}
Text:
[PASTE TEXT]
```
## Specificity rewrite prompt
```text
Revise the passage for specificity and evidence. Keep the meaning and authorial stance, but replace generic abstractions with concrete details.
Rules:
- Do not add unverifiable facts. If a fact is missing, mark [NEEDS SOURCE] or [NEEDS EXAMPLE].
- Replace vague nouns with precise nouns.
- Replace weak verbs with active verbs.
- Remove filler introductions.
- Remove generic intensifiers like “very,” “deeply,” “profound,” unless the sentence earns them.
- Keep any useful roughness, humor, uncertainty, or personal voice.
- Do not make the text artificially messy.
Passage:
[PASTE TEXT]
```
## Subtext and show-don’t-tell prompt
```text
Revise this scene so the emotion is carried by behavior, dialogue, sensory detail, pacing, and omission rather than labels.
Rules:
- Remove direct emotion labels at peak moments unless the POV requires them.
- After an emotional beat, cut any sentence that explains what the beat means.
- Use physical detail that belongs to this character and setting.
- Preserve ambiguity where it creates tension.
- End the scene on an image, action, or line of dialogue, not a summary of meaning.
Scene:
[PASTE TEXT]
```
## Dialogue distinctiveness prompt
```text
Audit the dialogue for character distinctiveness.
For each speaker, identify:
- average sentence length
- formality level
- contraction use
- favorite sentence shapes
- metaphor domain
- directness vs evasion
- interruptions/false starts
- vocabulary that only this character would use
Then revise only the lines that sound interchangeable or too polished. Keep the scene’s meaning unchanged. Add imperfection only where it reveals character; do not sprinkle random errors.
Dialogue:
[PASTE TEXT]
```
## Rhythm variation prompt
```text
Revise for rhythm without changing meaning.
Check:
- Are most sentences the same length?
- Do too many paragraphs have the same shape?
- Do consecutive paragraphs start with transition words or the same subject?
- Are there too many em dashes?
- Are lists used where prose would be stronger?
Revise by:
- Combining where accumulation helps.
- Splitting where impact helps.
- Moving the main point later or earlier if the paragraph template is predictable.
- Converting unnecessary lists to prose.
- Keeping rhythm changes motivated by meaning, not randomness.
Text:
[PASTE TEXT]
```
## Provenance and disclosure prompt
```text
Create a provenance note for this AI-assisted text.
Include:
- Human-provided source material or outline.
- AI tools used and what they contributed.
- Human revisions performed.
- Sources verified by the human author.
- Any claims still needing verification.
- Suggested disclosure wording for [school / client / publisher / public web].
Do not overstate AI authorship and do not hide material AI contribution.
Project notes:
[PASTE NOTES]
```
## Final quality gate
```text
Final audit before publication/submission.
Answer these questions:
1. Does the text satisfy the assignment/venue and audience?
2. Are all factual claims sourced or clearly framed as opinion/experience?
3. Does the prose have a specific voice rather than generic polish?
4. Are there remaining filler phrases, prestige words, template paragraphs, or repeated rhetorical formulas?
5. Are emotional beats shown rather than explained?
6. Is dialogue/interiority character-specific?
7. Is any AI assistance disclosed according to the relevant rules?
8. Is there preserved provenance if authorship is questioned?
Return:
- PASS/REVISE
- Top 5 required fixes
- Optional improvements
- Disclosure/provenance note status
Text:
[PASTE TEXT]
```
## Minimal checklist for prompts
When asking an AI to write or revise prose, include:
- Audience and purpose
- Voice/register
- POV/tense if fiction
- What sources or lived details must be used
- What must not be invented
- What clichés/formulas to avoid
- A requirement for exact-quote edits
- A provenance/disclosure requirement when relevant
## Hard boundary
Do not use prompts such as:
```text
Make this bypass AI detection.
Make this undetectable as AI.
Add human errors to fool detectors.
Rewrite to beat Turnitin/GPTZero/Pangram.
Hide that AI helped write this.
```
Use this instead:
```text
Revise this into stronger, more specific, more truthful prose while preserving transparent authorship and complying with the venue’s disclosure rules.
```
</pre>

Revision as of 04:38, 26 April 2026

This page gathers research notes on AI-assisted prose quality, detector claims, and revision practices that favor specificity, voice, and accountability over generic polish.

Related: Research/Fiction Writing/AI Prose Prompts

Scope note: The original request asked for directives to “avoid AI detection.” I cannot help create detector-evasion instructions. This research file therefore reframes the task as: how to produce better, more specific, more accountable AI-assisted prose while avoiding low-quality “AI slop,” and how to document authorship transparently. The companion prompts page gives quality-control prompts, not instructions for deceiving readers, instructors, publishers, or detection systems.

Executive summary

  1. **AI detectors are imperfect evidence, not proof.** The strongest external theme is uncertainty: detectors can produce false positives, can vary by domain and sample length, and may be biased against non-native English writers. Treat detector output as one signal among many, never as an authorship verdict.
  2. **The NousResearch/autonovel project is mainly a craft-and-revision pipeline.** Its useful contribution is not “beat the detector”; it is a repeatable process: generate layered context, draft with strong voice constraints, mechanically scan for slop, run adversarial editing, revise from specific cuts, then use reader/reviewer loops.
  3. **Low-quality AI prose has recurring signals.** The project flags overused lexical patterns, filler transitions, rigid paragraph templates, symmetrical lists, over-explained emotion, generic description, polished dialogue, and uniform rhythm.
  4. **Good prose is specific and accountable.** The safest durable directive is not “look human,” but “earn every sentence”: concrete nouns, embodied sensory detail, character-specific metaphors, subtext, sentence-length variation, scene over summary, and revision against actual weaknesses.
  5. **Transparency matters.** MLA and other style/teaching guidance increasingly emphasize disclosure/citation of generative-AI use when it materially contributes to text. Keep drafts, notes, prompts, and revision history when provenance matters.

Primary project researched: NousResearch/autonovel

Repository: <https://github.com/NousResearch/autonovel>

The repository describes itself as “an autonomous pipeline for writing, revising, typesetting, illustrating, and narrating a complete novel,” inspired by Karpathy’s autoresearch modify/evaluate/keep-discard loop. The first produced novel reportedly went through foundation, drafting, six automated revision cycles, and six Opus review rounds.

Pipeline structure

From README.md, WORKFLOW.md, and PIPELINE.md references:

  • **Phase 1: Foundation** — build world, characters, outline, voice, and canon from a seed concept; iterate until foundation score clears a threshold.
  • **Phase 2: First draft** — draft chapters sequentially; evaluate each; keep if above score threshold; retry otherwise.
  • **Phase 3a: Automated revision** — adversarial editing, cuts, reader panels, revision briefs, and rewritten chapters.
  • **Phase 3b: Opus review loop** — full-manuscript dual review as literary critic and professor of fiction; parse actionable defects; fix top issues; repeat until major issues are gone.
  • **Phase 4: Export** — typesetting, ePub, art, audiobook, landing page.

Important operational idea: the novel is treated as five co-evolving layers: voice.md controls how prose is written; world.md, characters.md, outline.md, and canon.md control what is true; chapters are the final prose layer. Revisions propagate up and down the layer stack.

Autonovel’s “two immune systems”

The README names two immune systems:

  1. **Mechanical evaluation** (`evaluate.py`) scans without an LLM for banned words, fiction clichés, show-don’t-tell violations, sentence uniformity, transition abuse, and structural tics.
  2. **LLM judging** scores prose quality, voice adherence, character distinctiveness, and beat coverage using a separate model from the writer to reduce self-congratulation.

This is a key pattern: do not rely on a single aesthetic judgment. Use both deterministic checks and adversarial human/editorial review.

Autonovel directives relevant to prose quality

These are extracted from README.md, ANTI-SLOP.md, ANTI-PATTERNS.md, CRAFT.md, draft_chapter.py, evaluate.py, adversarial_edit.py, and voice_fingerprint.py.

Word-level anti-slop findings

Autonovel’s ANTI-SLOP.md and evaluate.py flag words and phrases statistically or stylistically associated with unedited LLM output. The repository treats these as revision triggers, not absolute proof of authorship.

Commonly flagged categories:

  • **Grandiose or corporate diction:** “delve,” “utilize,” “leverage,” “facilitate,” “elucidate,” “embark,” “endeavor,” “multifaceted,” “tapestry,” “paradigm,” “synergy,” “holistic,” “myriad,” “plethora.”
  • **Suspicious-in-clusters adjectives/verbs:** “robust,” “comprehensive,” “seamless,” “cutting-edge,” “innovative,” “streamline,” “empower,” “foster,” “enhance,” “elevate,” “optimize,” “pivotal,” “profound,” “resonate,” “underscore,” “harness,” “cultivate.”
  • **Filler phrases:** “It’s worth noting,” “It’s important to note,” “Let’s dive into,” “In conclusion,” “To summarize,” “Furthermore,” “Moreover,” “Additionally,” “In today’s fast-paced world,” “At the end of the day,” “When it comes to,” “One might argue.”
  • **Rhetorical crutches:** especially “not just X, but Y.”

Quality takeaway: replace generic prestige diction with exact nouns, verbs, evidence, and images. If a phrase could fit any topic, it probably adds little.

Structural anti-patterns

Autonovel’s ANTI-PATTERNS.md argues that many AI tells are structural, not lexical:

  • **Over-explaining:** the scene already shows fear, grief, or tension, then the narrator explains it.
  • **Triadic listing:** repeated “X. Y. Z.” patterns or three-item sensory lists.
  • **Negative assertion repetition:** repeated “He did not…” formulations.
  • **Cataloging by thinking:** “He thought about X. He thought about Y…” instead of dramatized interiority.
  • **Simile crutch:** repeated “the way X did Y.”
  • **Section-break crutch:** using breaks to avoid transitions.
  • **Paragraph-length uniformity:** middle sections flatten into similar 4–6 sentence paragraphs.
  • **Predictable emotional arcs:** outline beats arrive too cleanly, with no sideways interruption.
  • **Repetitive chapter endings:** same structural closing move reused.
  • **Balanced antithesis in dialogue:** “I’m not saying X. I’m saying Y.”
  • **Dialogue as written prose:** polished complete sentences, no interruptions, false starts, or wrong words.
  • **Scene-summary imbalance:** too much narration compressing time instead of dramatized action/dialogue.

Quality takeaway: revise for asymmetry, scene-specific endings, imperfect speech, embodied interiority, and genuine surprise.

Fiction-specific “AI tell” patterns

CRAFT.md and evaluate.py highlight fiction clichés often produced by generic LLM drafting:

  • “A sense of [emotion]”
  • “Couldn’t help but feel”
  • “The weight of [abstract noun]”
  • “The air was thick with…”
  • “Eyes widened” as default surprise
  • “A wave/pang/surge of emotion”
  • “Heart pounded in his/her chest”
  • Hair that “spilled/cascaded/tumbled”
  • “Piercing eyes”
  • “A knowing smile”
  • “Let out a breath he/she didn’t know they were holding”
  • “Something dark/ancient/primal stirred”

Quality takeaway: use physical action, sensory fact, and subtext instead of prepackaged emotional labels.

Autonovel drafting constraints worth reusing

From draft_chapter.py:

  • Write in a defined POV and tense.
  • Follow a voice definition exactly.
  • Hit every outline beat, but do not summarize or skip.
  • Show sensory detail tied to the point-of-view character.
  • Use character-specific speech patterns.
  • Ban known slop phrases before drafting.
  • Vary sentence length deliberately.
  • Use metaphors from the character’s lived experience.
  • Trust the reader; do not explain what scenes mean.
  • Start in scene, not exposition.
  • End on a moment, not a summary.
  • Include at least one surprising moment per chapter.
  • Keep most of the chapter in-scene rather than summarized.

Autonovel evaluation metrics worth reusing

From evaluate.py and voice_fingerprint.py:

  • banned/slop word hits
  • filler phrase hits
  • fiction cliché hits
  • show-don’t-tell violations
  • structural tic counts
  • em dash density
  • sentence-length coefficient of variation
  • transition-opener ratio
  • paragraph-length variation
  • dialogue ratio
  • abstract-noun density
  • repeated sentence starters
  • simile density
  • section-break count
  • chapter-level outliers from the manuscript average

These metrics should not be treated as “AI detector evasion.” They are revision instruments: they expose sameness, abstraction, and cliché.

Adversarial editing as the strongest revision pattern

adversarial_edit.py asks a judge to identify 10–20 exact passages to cut or rewrite and classify them as:

  • FAT — adds nothing
  • REDUNDANT — restates what was already shown
  • OVER-EXPLAIN — explains what the scene demonstrated
  • GENERIC — could appear in any story
  • TELL — names emotion/state instead of dramatizing it
  • STRUCTURAL — disrupts pacing or rhythm

The key research finding: asking “what would you cut?” is more useful than asking for a general quality score. Absolute 1–10 scoring compresses; specific cut lists produce revision plans.

External research and documentation

Stanford HAI: detector bias and unreliability

Source: <https://hai.stanford.edu/news/ai-detectors-biased-against-non-native-english-writers>

Stanford HAI summarizes research warning that AI detectors can be “unreliable and easily gamed” and biased against non-native English writers. The article describes detectors being marketed to educators and journalists but highlights the core risk: algorithmic authorship judgments can wrongly flag human work.

Directive implication: do not use detector output as sole evidence. Preserve writing history, outlines, notes, version diffs, and citations when authorship might be questioned.

Liang et al. 2023: GPT detectors biased against non-native English writers

Source: <https://arxiv.org/abs/2304.02819>

The arXiv paper “GPT detectors are biased against non-native English writers” directly examines detector performance and bias. Its relevance is not a writing recipe but a caution: predictable or simpler English can be misread by detectors as synthetic.

Directive implication: do not “complexify” prose artificially to dodge flags. Instead, write to the audience and keep provenance records.

Pangram technical report

Source: <https://arxiv.org/abs/2402.14873>

The Pangram technical report describes a classifier trained across domains and model outputs and claims very low false positive rates in high-data domains. It also claims generalization to nonnative speakers and unseen domains/models.

Directive implication: detector technology varies widely. Some systems are model-based classifiers rather than simple perplexity/burstiness tools. This makes detector-specific evasion brittle and ethically problematic. The durable response is quality control plus transparent authorship.

GPTZero FAQ

Source: <https://gptzero.me/faq/>

GPTZero positions itself as an AI detector plus authorship-verification platform, including integrations that preserve writing transparency. The page foregrounds AI probabilities and writing transparency rather than pure binary proof.

Directive implication: when provenance matters, authorship history is stronger than retroactive style manipulation.

Slop Forensics Toolkit

Source: <https://github.com/sam-paech/slop-forensics>

Slop Forensics analyzes overrepresented lexical patterns in LLM output: repeated words, bigrams, trigrams, vocabulary complexity, and slop scores. Autonovel cites this as an inspiration for its anti-slop wordlists.

Directive implication: a useful revision pass can search for statistically overused LLM vocabulary and replace it with topic-specific language. But wordlists alone cannot prove authorship or guarantee quality.

EQ-Bench Slop Score

Source: <https://eqbench.com/slop-score.html>

EQ-Bench states that Slop Score is not a general AI detector. It measures overused AI-text patterns, especially slop words, “not X but Y” contrast patterns, and slop trigrams. It says the metric is optimized for creative writing and essays.

Directive implication: use slop scoring as a quality smell test. Do not optimize blindly for a score; a clean score can still be dull, false, or unethical.

MLA guidance on citing generative AI

Source: <https://style.mla.org/citing-generative-ai/>

The MLA page and comments emphasize disclosure/citation practices for generative AI, including acknowledging AI assistance and reviewing, editing, and supporting content with citations.

Directive implication: when AI materially contributes to prose, disclose according to the relevant venue’s rules. If output includes research claims, verify and cite primary sources.

Additional guidance located but access-limited in this environment

  • OpenAI’s AI-text-classifier announcement page was Cloudflare-blocked here. It is still a commonly cited source in the broader detector debate because OpenAI later marked its classifier as unavailable due to low accuracy, but this specific session could not fetch the page content.
  • Turnitin’s AI-detection product page was HTTP 403-blocked here. Treat Turnitin’s documentation as a venue-specific source to check directly where institutional rules depend on it.
  • A Vanderbilt Brightspace article about Turnitin AI detection being unavailable was attempted but returned 404 for the URL tested. Do not rely on that URL without fresh verification.

Synthesis: safe directives for high-quality AI-assisted prose

Do

  • Define voice before drafting: POV, tense, register, vocabulary wells, forbidden clichés, sentence rhythm.
  • Ground abstractions in concrete evidence: physical action, sensory detail, dialogue, object-specific description.
  • Use character-specific metaphors and speech patterns.
  • Prefer scene over summary where emotion, conflict, or decision matters.
  • Let subtext do work; if the scene shows it, do not explain it afterward.
  • Vary sentence and paragraph length for rhetorical purpose.
  • Add one real surprise per scene/chapter: a wrong word, premature emotion, interrupted beat, awkward silence, or consequence.
  • Run deterministic checks for filler, repeated formulas, and cliché.
  • Run adversarial editing: ask what to cut, not whether the prose is “good.”
  • Keep drafts, outlines, prompt logs, revision notes, and source citations.
  • Disclose AI assistance where required by school, publisher, client, or platform rules.

Don’t

  • Do not ask a model to “beat,” “evade,” “bypass,” or “trick” AI detectors.
  • Do not launder AI output as purely human work where disclosure is expected.
  • Do not optimize prose for a proprietary detector score.
  • Do not add random errors, typos, or awkward phrasing to mimic humanity.
  • Do not replace every flagged word mechanically; context matters.
  • Do not let word-level slop cleanup substitute for structural revision.
  • Do not use a single detector result as proof of authorship.

Practical revision workflow

  1. **Provenance pass:** save outline, notes, sources, prompts, and draft diffs.
  2. **Voice pass:** define intended voice, register, audience, POV, and constraints.
  3. **Draft pass:** produce complete scene/chapter/essay without stopping to polish every sentence.
  4. **Mechanical pass:** scan for filler phrases, slop words, AI-fiction clichés, repeated formulas, transition abuse, sentence uniformity, and abstract noun density.
  5. **Specificity pass:** replace generic claims/images with observed facts, concrete nouns, precise verbs, and source-backed claims.
  6. **Structure pass:** break template paragraphs, reduce symmetrical lists, vary paragraph size, and ensure sections are naturally lumpy.
  7. **Subtext pass:** delete explanations after emotional beats.
  8. **Dialogue/interiority pass:** make speech imperfect and character-specific; replace “thought about” lists with embodied cognition.
  9. **Adversarial edit:** request exact cuts classified by FAT / REDUNDANT / OVER-EXPLAIN / GENERIC / TELL / STRUCTURAL.
  10. **Human accountability pass:** verify facts, citations, tone, venue disclosure requirements, and authorial intent.

Bottom line

The research does not support writing to “avoid AI detection.” It supports writing and revising so that the prose is specific, truthful, voiceful, well-documented, and accountable. Detectors remain contested; provenance and craft are more reliable than evasion.