Building a Thinking Model from a Non-Thinking Model Using Chain-of-Thought

Artificial Intelligence has made remarkable progress in generating human-like responses, but not all AI models naturally “think” before answering. Many models — especially smaller or base versions — simply map input to output without following a reasoning process. While this works for simple questions, it often fails for tasks requiring multi-step reasoning.
This is where Chain-of-Thought (CoT) prompting comes into play. By strategically guiding a model to reason step-by-step, we can transform a non-thinking model into one that shows its reasoning process and arrives at more accurate answers.
What is Chain-of-Thought (CoT)?
Chain-of-Thought prompting is a technique where we explicitly tell the model to “think step by step” before giving the final answer. Instead of producing an immediate output, the AI generates intermediate reasoning steps, which helps it arrive at a more accurate conclusion.
Example:
Normal prompt:
"What is 17 × 23?" → Model might guess: "354"With Chain-of-Thought:
"What is 17 × 23? Let's think step by step." →
"17 × 20 = 340, 17 × 3 = 51, 340 + 51 = 391" → 391 ✅
Why Non-Thinking Models Struggle
A “non-thinking” model:
Often memorizes patterns from training data rather than reasoning.
Gives confident but wrong answers on multi-step tasks.
Struggles with logical puzzles, math, and multi-hop questions.
By encouraging reasoning through CoT, we help the model break down the problem instead of jumping to an answer.
How to Turn a Non-Thinking Model into a Thinking Model
1. Use Explicit Step-by-Step Instructions
Example:
Q: If Alice has 4 apples and buys 3 more, then eats 2, how many apples does she have left?
A: Let's solve step by step:
- Start with 4 apples
- Buy 3 more → total = 7
- Eat 2 → total = 5
Answer: 5
2. Self-Ask Prompting
Encourage the model to ask itself sub-questions:
Q: How many sides does a shape with twice as many sides as a triangle have?
A: First, a triangle has 3 sides.
Twice 3 is 6.
So, the answer is 6.
3. Few-Shot CoT Prompting
Show examples of step-by-step reasoning before asking the actual question:
Example 1:
Q: What is 12 × 15?
A: Let's think step by step:
12 × 10 = 120, 12 × 5 = 60, total = 180.
Example 2:
Q: What is 9 × 14?
A: Let's think step by step:
9 × 10 = 90, 9 × 4 = 36, total = 126.
Now your turn:
Q: What is 8 × 17?
4. Ask for Intermediate Reasoning + Final Answer
Make it clear that the reasoning is separate from the final answer:
Reasoning:
Step 1: …
Step 2: …
Final Answer: …
Benefits of Chain-of-Thought Prompting
Higher Accuracy — Reduces guesswork in math, logic, and planning tasks.
Transparency — You can see exactly how the AI reached its answer.
Debugging Friendly — Mistakes are easier to identify in reasoning steps.
Scalable — Works even with smaller or cheaper models.
Potential Risks & Considerations
Longer Responses — CoT outputs are more verbose, which may not be ideal for every use case.
Hallucination in Reasoning — If the reasoning is wrong, the final answer will still be wrong.
Prompt Sensitivity — The exact phrasing of “think step by step” can impact results.
When to Use CoT Prompting
Math and Arithmetic problems.
Logic puzzles and riddles.
Multi-hop reasoning in QA systems.
Planning tasks like scheduling or workflow generation.
Conclusion
A non-thinking model can be taught to reason with the right prompting strategy. By using Chain-of-Thought prompting, you enable the model to break down problems into smaller steps, improving accuracy and transparency.
In the age of AI, the ability to prompt for thinking is just as important as having a powerful model — and with CoT, even smaller models can think like bigger ones.
Subscribe to my newsletter
Read articles from Rajesh directly inside your inbox. Subscribe to the newsletter, and don't miss out.
Written by
Rajesh
Rajesh
Hi, I’m Rajesh 👋 I’m learning Web Development 🌐 and Generative AI 🤖. I enjoy creating simple websites, and improving my coding skills every day. Excited to grow, learn, and connect with like-minded people!