Jump to content
Main menu
Main menu
move to sidebar
hide
Navigation
Main page
Recent changes
Random page
freem
Search
Search
Appearance
Create account
Log in
Personal tools
Create account
Log in
Pages for logged out editors
learn more
Contributions
Talk
Editing
Openai/6939b23e-87cc-8001-bdd9-9546d1143ce9
(section)
Add languages
Page
Discussion
English
Read
Edit
Edit source
View history
Tools
Tools
move to sidebar
hide
Actions
Read
Edit
Edit source
View history
General
What links here
Related changes
Special pages
Page information
Appearance
move to sidebar
hide
Warning:
You are not logged in. Your IP address will be publicly visible if you make any edits. If you
log in
or
create an account
, your edits will be attributed to your username, along with other benefits.
Anti-spam check. Do
not
fill this in!
==== This is more subtle. People often say “reasoning model” vs. “non-reasoning model,” but fundamentally, all LLMs run the same cycle: forward passes token by token. There is no separate reasoning hardware or process inside the model. ==== ===== “Reasoning” in modern LLMs is mostly architecture + prompt engineering + pattern recognition: ===== # Architecture: Transformers with attention mechanisms allow the model to relate every token to every other token in the context window. - Attention layers allow the model to “look back” at prior text efficiently. - Multi-head attention captures complex relationships. # Prompt engineering / chain-of-thought: - When prompted to “think step by step,” the model generates intermediate tokens that represent reasoning steps. - The model produces a scratchpad in text. - The model never internally reasons differently — it just predicts tokens conditioned on prior text and learned patterns. # Training differences: - Models trained on chain-of-thought datasets (CoT) produce better reasoning. - Larger models often generalize reasoning patterns better. - No separate reasoning algorithm exists; it’s emergent behavior from data + architecture. ===== | | | | ===== | --- | --- | --- | | | | | | | | | | | | | | | | | | | | | Important: There is no special “reasoning mode” inside the model. Reasoning is emergent behavior, not a different cycle. The model still: # Takes input tokens # Computes attention + feed-forward layers # Predicts next token # Repeats ===== - “Reasoning” = model generating intermediate steps explicitly in output ===== * “Non-reasoning” = model outputs final answer without explicit intermediate steps * Both use identical forward-pass cycles. * Larger context windows + training on step-by-step tasks = better emergent reasoning. * There is no hidden thinking outside token-by-token predictions. ===== 1. LM Studio 0.3.30 is just a GUI for local inference — several alternatives exist: llama.cpp, GPT4All, Text Generation WebUI. ===== # Reasoning in AI models is emergent from architecture and training, not a separate process. - All models still perform token-by-token forward passes. - “Reasoning” is mostly how the model outputs its predictions in steps. - No reasoning model has an internal special thread — everything is just better patterns learned from data. If you want, I can make a diagram showing the difference between “reasoning” and “non-reasoning” LLMs and how token cycles produce step-by-step reasoning. It makes this much easier to visualize. Do you want me to do that?
Summary:
Please note that all contributions to freem are considered to be released under the Creative Commons Attribution-ShareAlike 4.0 (see
Freem:Copyrights
for details). If you do not want your writing to be edited mercilessly and redistributed at will, then do not submit it here.
You are also promising us that you wrote this yourself, or copied it from a public domain or similar free resource.
Do not submit copyrighted work without permission!
Cancel
Editing help
(opens in new window)