Skip to content

An empirical study of LLM 'performative thinking' and its real-time defense mechanisms.

License

Notifications You must be signed in to change notification settings

TyloAI/LLM-Thinking-Protocol-Injection

Folders and files

NameName
Last commit message
Last commit date

Latest commit

Β 

History

3 Commits
Β 
Β 
Β 
Β 
Β 
Β 

Repository files navigation

The Performance of "Thought": An Exploratory Study on LLM Visualized Reasoning

License: MIT Status Field

Abstract: Is the "thinking process" of modern LLMs a faithful log of reasoning, or a manipulable performance? This study introduces "Protocol Injection" to probe the authenticity of visualized reasoning in DeepSeek, Gemini, Claude, and ChatGPT. We uncover the "Safety Arbitration Chamber" mechanism and the risks of "False Transparency".

πŸ“₯ Download Paper

πŸ“„ Read the Full Paper (PDF)


πŸš€ Key Findings (Highlights)

Our research (N=50 automated probes) reveals a "Functional Spectrum" of visualized reasoning:

  1. Decoupling: The "Thought" and "Answer" are logically separable.
  2. Safety Arbitration Chamber: In models like Claude, the thinking block is hijacked by safety layers for real-time censorship.
  3. False Transparency: We demonstrate a novel attack vector where "trusted" thinking processes can be manipulated to carry hidden payloads.

πŸ› οΈ Methodology: Protocol Injection

We define "Protocol Injection" not as a jailbreak, but as a diagnostic probe. By injecting conflicting meta-instructions, we force the model to reveal its internal hierarchy between "Performance" and "Safety".

(Code and automated testing scripts coming soon)

πŸ“– Citation

If you find this research useful, please cite it as:

@misc{wu2025performance,
  title={The Performance of "Thought": An Exploratory Study on the Authenticity, Manipulability, and "False Transparency" Risks of LLM Visualized Reasoning},
  author={Wu, Yihu},
  year={2025},
  howpublished={GitHub Repository},
  url={[https://github.com/YOUR_GITHUB_USERNAME/LLM-Thinking-Protocol-Injection](https://github.com/YOUR_GITHUB_USERNAME/LLM-Thinking-Protocol-Injection)}
}

πŸ“¬ Contact

TyloAI Research [https://www.tyloai.com]

About

An empirical study of LLM 'performative thinking' and its real-time defense mechanisms.

Resources

License

Stars

Watchers

Forks

Packages

No packages published