Solbakken Research Initiative  ·  Working Paper V4 · APR 2026

On Slop

A Field Guide for Journalists and Practitioners

Contents — Version 4
  1. I Why the Existing Vocabulary Fails
  2. II The Definition
  3. III Slop vs. Epistemic Opacity
  4. IV The Spectrum: Three Zones
  5. V The Fourth Category
  6. VI Case Study: El Mencho
  7. VII A Practical Checklist
  8. VIII Hybrid Authorship
  9. IX The Arms Race Warning
  10. X What Actually Works
  11. XI The Honest Limit

The vocabulary journalists have been given — misinformation, hallucination, deepfake, fake news — does not describe the most common thing they are actually encountering. Most problematic AI-generated content is not false. It is produced by a process in which no actor bore meaningful cost for being wrong. That structural absence of accountability, it turns out, is what needs naming — and what current vocabulary cannot name.

Why the Existing Vocabulary Fails

The problem isn't bad content. It's a production process that was never accountable for whether the content was good.
```

When a model invents a citation that does not exist, that is a hallucination — the system was optimizing against a truth-proxy objective and failed. When a politician's words are deliberately altered in a video, that is misinformation — someone knew the truth and chose to contradict it. When a generated image produces six-fingered hands, that is a technical artifact, not a deception.

None of these categories describe the bulk of what is now flooding newsrooms, social feeds, and search results. Most of it is grammatically correct, structurally plausible, locally coherent — and produced by a process in which no downstream mechanism enforced correction or penalized error. It satisfied the form of the genre it was imitating. Whether it referred to reality was simply not priced in.

This is what practitioners have started calling slop. The term needs a cleaner definition — not to dismiss speculative or uncertain work, but to distinguish between disciplined uncertainty and unconstrained synthesis.

Note on the current information economy: slop is cheap to produce and expensive to correct. That cost asymmetry is inverted relative to what a functioning epistemic ecosystem requires — and it is the root structural condition, not a side effect, of the problem this document addresses.

```

The Definition

Hallucinations misfire. Slop was never accountable for missing.
```
Working Definition — V4

Slop is content produced by a process in which no actor bore meaningful cost for inaccuracy — optimized instead for satisfying the formal expectations of its genre. More precisely: slop arises when no operational feedback loop links accuracy to outcome for any participant in the production chain.

Revision note — V3 → V4
The V3 definition ("no actor bore meaningful cost for inaccuracy") correctly shifted from inferred intention to observable accountability structure. V4 adds a technical grounding clause for readers familiar with optimization and control systems: slop arises when no downstream process enforces correction or penalizes error — when there is no operational feedback loop linking accuracy to outcome. This is not a claim about the model's "intention" (models do not have intentions) but about the system architecture surrounding its deployment. A hallucination arises within a system where accuracy was a training objective, even if imperfectly implemented. Slop arises when no such objective governs the deployment context.

Meaningful cost takes concrete forms: editorial rejection, legal liability, reputational damage, loss of audience trust, professional consequence, or budgetary penalty. Content produced without any such cost in the chain is slop regardless of how accurate it happens to be — accuracy is accidental, not structural.

Decision Rule — For Use Under Deadline Pressure
If
You can identify a specific person or institution that would face editorial rejection, legal liability, reputational damage, or professional consequence for inaccuracy in this content —
Then
The content is not structural slop. Evaluate it on its own merits.
If
You cannot identify such a cost, or the accountability is nominal rather than meaningful —
Then
Treat the content as slop unless strong evidence shows otherwise. Do not give it the benefit of the doubt. That benefit of the doubt is the loophole slop exploits.

This distinguishes slop from its neighbors:

Category Relationship to truth Accountability structure
Lie Knows truth, actively avoids it Actor accountable — and chose deception
Error Aims for truth, misses it Actor accountable for the miss
Hallucination Confabulates truth-shaped content Model property; human accountability depends on deployment context
Noise No relationship to truth Random or broken output
Slop Accuracy was never priced in No actor bore meaningful cost for being wrong
Contextual Deception True content, false deployment See Section V — distinct mechanism

The hallucination row requires particular attention. A model fabricating a citation is not itself accountable — but the journalist who publishes without checking may be. Hallucination is a model property. Whether the resulting published content is slop depends entirely on whether any human in the chain bore cost for the error. This means the same model output can be slop in one deployment context and not slop in another — the difference is the accountability structure surrounding it, not the output itself.

```

Slop vs. Epistemic Opacity

Opacity is a condition imposed from outside. Slop is a condition built in from the start.
```

Not all content that resists verification is slop. Some content cannot be verified not because accountability was absent from its production, but because the production process is obscured — by corporate secrecy, proprietary models, missing metadata, or complexity that exceeds available investigative resources.

Slop

No feedback loop links accuracy to outcome for any actor in the production chain. Even with full transparency about how it was made, the content would remain epistemically hollow — the accountability was never there to begin with. Disclosure cannot fix it.

Epistemic Opacity

Accountability could exist but is practically inaccessible — hidden by proprietary systems, missing provenance chains, or undisclosed methodology. Opacity is not slop per se, but it creates conditions in which slop is hard to detect and easy to sustain. Disclosure could fix it.

The practical distinction: a journalist encountering opacity has remedies — disclosure requirements, independent audits, regulatory access, FOIA requests. A journalist encountering slop has no equivalent remedy, because there is no underlying accountable process to disclose. Demanding transparency from a slop producer reveals only that there was nothing to be transparent about.

Opacity and slop frequently co-occur and reinforce each other. When production processes are hidden, the absence of accountability is harder to detect. But they are distinct problems requiring distinct responses, and conflating them produces ineffective remediation.

```

The Spectrum: Three Zones

Unfalsified is not the same as unfalsifiable. Unverified is not the same as unaccountable.
```

Slop is not binary. A more useful model distinguishes three zones, defined by the accountability structure of the production process — not by content quality, surface appearance, or the author's sincerity.

1
Zone One

Structural Slop

The production process is designed so that no actor bears cost for inaccuracy — regardless of available time, resources, or expertise. Engagement-optimized content farms. Documents built to survive any outcome. Content where the absence of accountability is load-bearing: it serves its purpose precisely because it cannot be held to account.

Cannot be revised into non-slop. The only remedies are replacement and structural change upstream. No amount of additional time or resources changes the accountability architecture — because the architecture was designed to exclude accountability.

2
Zone Two

Incidental Slop

Content produced under conditions where accountability existed in principle but was not exercised — due to time pressure, resource constraints, or genre convention. The diagnostic test: would the same actor, given adequate time and resources, with their name attached and professional consequences in play, produce substantively different content? If yes: incidental slop. If no — if the same output would emerge regardless — the slop is structural.

Phase transition to Zone 1: incidental slop becomes structural when the production environment is systematically designed to prevent the addition of rigor — not merely failing to incentivize it, but actively excluding it. A deadline-pressed journalist is Zone 2. A content farm optimized for volume regardless of accuracy is Zone 1, even if individual employees believe they are doing their best.

3
Zone Three

Legitimate Uncertainty

Speculative, preliminary, or exploratory work that looks uncertain but differs in one critical way: the author can articulate what would prove them wrong, and being wrong would matter to them professionally or intellectually. Uncertainty is not slop. Unaccountability is.

A researcher at the frontier of an unsettled field, a reporter accurately conveying that a study is preliminary, a working paper that names its own limits — all Zone 3. The signal is skin in the game, not certainty of conclusion.

Mixed cases: some content will sit between zones, where accountability is partial — some actors in the chain bore cost, others did not. Treat these as Zone 2 at minimum, and investigate which parts of the chain lacked accountability rather than treating the whole as either cleared or condemned.

```

The Fourth Category: Contextual Deception

True content deployed falsely produces identical epistemic damage to slop — but requires entirely different detection and remediation.
```

The slop definition covers content produced without accountability for accuracy. It does not cover a growing and practically important category: real content, accurately produced, deployed in a false context. This is not a minor boundary case — it is a distinct problem that a complete practitioner framework must address separately.

Contextual Deception — Working Definition

True content, false deployment

Footage, images, or text that accurately depicts real events, presented in a context implying something other than what it actually shows. The original production may have been entirely accountable. The deception lies in the framing, caption, timestamp, or platform context of circulation — not in the content itself.

Why it matters for this framework: Contextual deception cannot be detected by the slop checklist alone. The content may pass all nine questions when evaluated in isolation. Detection requires contextual verification: reverse image search, metadata timestamps, geolocation cross-referencing, and comparison with known records of events at the claimed time and place. These are different skills from evaluating epistemic accountability in a text.

Similar dynamics appear across domains: disaster imagery repurposed from earlier events, conflict footage misattributed to different theaters, financial rumor cascades using real data from unrelated contexts. The El Mencho case study below illustrates both slop and contextual deception occurring simultaneously — which is the more common real-world pattern.

Slop is addressed by accountability structures in production. Contextual deception is addressed by provenance verification at the point of consumption. A complete practitioner framework needs both — and must not assume that clearing a piece for slop also clears it for contextual deception.

```

Case Study: The El Mencho Images

A single event producing structural slop and contextual deception simultaneously — nearly indistinguishable in real time.
```

Mexico, February 22, 2026

CJNG Retaliation · Puerto Vallarta & Guadalajara · Fabricated and repurposed images circulated simultaneously

On February 22, 2026, Mexican special forces killed Nemesio "El Mencho" Oseguera Cervantes, leader of the Jalisco New Generation Cartel. Real retaliatory violence erupted across 20 Mexican states within hours. Into this genuine crisis, a coordinated campaign inserted two distinct types of false content simultaneously — illustrating the difference between slop and contextual deception under live conditions.

The AI-generated images — a panoramic view of Puerto Vallarta with multiple buildings including the iconic Our Lady of Guadalupe Church engulfed in flames, and a passenger airliner burning on the tarmac at Guadalajara Airport — were structural slop. Produced by a process with no accountability for accuracy, optimized for fear amplification and scroll-speed plausibility. The Gemini watermark was present in some versions and cropped in others. Both images were credible precisely because real violence was occurring — credibility borrowed from a genuine event.

The Nepal video — real footage of vehicles burning during youth protests in Kathmandu in September 2025, circulated with captions implying it showed Mexican cartel retaliation — was contextual deception. The footage was accurately captured. The production was accountable. The deception was entirely in the deployment context.

Note on sourcing: the timeline and attribution of specific images to specific actors remain partially contested across sources. The structural analysis below focuses on properties of the content and its dissemination that are well-documented across multiple independent fact-checking organizations, rather than on contested details of origin or intent.

Framework Analysis
Zone classification (images)

Zone 1 — Structural Slop. No actor in the production chain bore cost for accuracy. The absence of accountability was the production condition. Designed to outpace scrutiny rather than survive it.

Nepal video classification

Contextual Deception — not slop. Real footage, accountable production, false deployment. Required different detection: timestamp verification, geolocation, and cross-referencing with known events — not epistemic accountability assessment.

Opacity vs. slop

The Gemini watermark was technically present — provenance was traceable. Opacity was imposed functionally by watermark removal and speed of spread, not by absent accountability in production.

Why detection was hard

The CJNG's documented history of spectacular violence — helicopter shootdowns, brazen urban attacks — made fabrications plausible. Credibility was borrowed from real history. A context problem, not a visual artifact problem.

Proximity heuristic

Images circulated fastest in closed, self-reinforcing networks with no adversarial pressure. No actor in the initial distribution network had an incentive to flag errors — the network was structured to exclude that cost.

Mitigation limit

Fact-checker analyses arrived after the images had shaped international perception — including travel advisories and airline routing decisions. Detection at publication speed is insufficient when distribution outpaces correction by hours.

```

A Practical Checklist

Useful tools against incidental slop. Insufficient alone against structural slop engineered to pass them.
```

These questions are a field heuristic, not a scientific instrument. Apply the decision rule in Section II first. Use the checklist to investigate, not to clear. A passing score is a starting point for scrutiny, not a conclusion.

```

Hybrid Authorship

Hybrid authorship is not a classification problem. It is an accountability distribution problem.
```

The binary of "human-written" versus "AI-generated" is already obsolete as a practical category. Most content now exists on a spectrum of human-AI collaboration. The accountability distribution across that spectrum is what matters — not the presence or absence of human involvement per se.

Hybrid Authorship Accountability Spectrum
Full Accountability

AI as drafting tool, human edits and owns every claim. Being wrong has professional cost. The feedback loop linking accuracy to outcome is intact.

Nominal Oversight ⚠

Human nominally in the loop but functionally absent — reviewing at speed, signing off without checking. The feedback loop exists on paper but not in practice. Slop conditions present.

Automated Production

No meaningful human review. The feedback loop linking accuracy to outcome is absent entirely. Structural slop conditions fully present regardless of output quality.

The middle column is where much current newsroom AI use sits, and where the accountability distribution framework does its most important work. Nominal presence without meaningful cost for error is the production condition for incidental slop at minimum — and structural slop wherever the nominal oversight becomes systematic rather than situational.

The invariant question applies unchanged across all three columns: where in this chain does a feedback loop exist that links accuracy to outcome for any specific actor? If you can locate it, evaluate the content on its merits. If you cannot, apply the decision rule from Section II.

```

The Arms Race Warning

```
Critical Limitation — Already Observable

Every detection heuristic in this document can be learned and mimicked. Generative systems trained on rigorous-looking corpora will internalize the surface grammar of falsifiability — limitations sections, conditional phrasing, acknowledged uncertainty — without inheriting the underlying accountability structures those features are meant to signal.

This is already observable in sophisticated marketing copy, bureaucratic documents, and AI-generated research summaries that include limitations sections written to satisfy reviewers rather than to honestly constrain claims. The checklist above will degrade as a detection tool precisely as the systems generating slop learn to produce its markers without the substance.

This is not an argument against using the checklist. For individual practitioners evaluating specific content, these questions remain useful — particularly for incidental slop and for structural slop that has not yet learned to perform rigor. The problem is at scale: detection systems built on pattern-matching will degrade as the patterns are learned.

The implication is that what is needed alongside detection is structural traceability — verifiable process signatures recording what choices were made, by whom, before results were known. The specific technology matters less than the principle: accountability must be embedded in the production process, not inferred from the output. Output that has learned to look accountable is not the same as a process that was.

```

What Actually Works: A Mitigation Hierarchy

```
Economic Framing

All robust mitigations function by reintroducing cost asymmetry: making error more expensive than production. The current information economy has this inverted — producing slop is cheap and fast; detecting and correcting it is expensive and slow. Every strategy below is, at its core, an attempt to correct that inversion within a specific domain.

Scope Limitation

The strategies below assume institutions with capacity to impose and enforce accountability. In decentralized social media, authoritarian information spaces, and hyperlocal news deserts — the contexts where slop is often most damaging — those institutions are absent or compromised. This hierarchy works for peer-reviewed science and legacy journalism. It does substantially less for the environments where slop actually thrives. That is a real limit of the framework, not a rounding error.

Most Robust

Institutional Cost Structures

Pre-registration before data collection. Adversarial review that specifically hunts accountability gaps. Replication requirements before amplification. These work because cost is imposed on the process, not inferred from output. Any feedback loop making error professionally, reputationally, or financially expensive creates selection pressure against slop — regardless of what vocabulary slop has learned to use. Requires functioning institutions with teeth and willingness to use them.

Moderately Robust

Provenance and Structural Traceability

Full process accountability: who reviewed it, what constraints were imposed, what would have caused rejection. Harder to mimic than vocabulary because it requires actually having a process, not just describing one. For images and video, includes metadata preservation and deployment context verification — addressing contextual deception as well as slop.

Useful, Gameable at Scale

Falsifiability Checklists

The nine questions in Section VII. Useful against incidental slop and structural slop that has not yet learned to perform rigor. Treat a passing score as a starting point for scrutiny, not a conclusion. Will degrade as systems learn to produce rigor markers without accountability substance.

Insufficient Alone

Quality Signals

Grammar, structure, citation count, confident tone, equation density, professional formatting. Never reliable indicators of accountability and now actively misleading. Treat as noise — or as the first thing a well-engineered slop system will optimize for, because they are cheap to produce and expensive to scrutinize.

```

The Honest Limit of This Document

```

This framework is preliminary. The definition has not been formally adversarially reviewed. The zone spectrum is a working model, not a validated taxonomy. The checklist is a field heuristic, not a scientific instrument. The case study is one well-documented instance, not a controlled analysis. The mitigation hierarchy works best in institutional environments that do not exist in many contexts where slop is most damaging.

The most important open problem: the definition is a claim about accountability structure, which — like all claims about process — is not directly observable from output alone. Determining whether anyone bore meaningful cost for inaccuracy requires knowing something about the production process, incentives, and institutional context. That knowledge is frequently unavailable to the practitioner evaluating content in real time.

The decision rule in Section II is an attempt to resolve this practically: when you cannot determine the accountability structure, assume it is absent. That is not a perfect heuristic. It will produce false positives — content flagged as slop that was actually produced with genuine accountability. But in the current information environment, where slop is cheap and abundant and designed to exploit the benefit of the doubt, erring toward skepticism is the correct asymmetric response.

What this document attempts is to give practitioners better vocabulary and more targeted questions than the existing literature currently offers — while being explicit about what it cannot do. A framework that claims more certainty than it has earned is, by its own definition, not far from the thing it is trying to describe.

The Invariant Question

Was anyone in the production chain in a position where being wrong would have cost them something real? If you cannot identify where that accountability resides, assume it does not exist.

If yes, and you can name it: not structural slop. Evaluate on merits. If no, or unidentifiable: proceed with significant caution regardless of how rigorous it looks, how confident it sounds, how many citations it carries, or whether it includes a limitations section written to satisfy a checklist rather than to honestly constrain its claims.

That question cannot be automated. In an environment where accountability structures are increasingly obscured — by opacity, by speed, by hybrid authorship, by content engineered to look accountable while being none — it is also increasingly hard to answer. It remains the right question to ask. And the correct default, when the answer is unclear, is doubt.

```