Critical thinking is often listed as a twenty-first-century skill, but it is one of the oldest capabilities studied in psychology and philosophy. At its core, critical thinking is the ability to evaluate information rigorously -- to separate valid arguments from weak ones, identify hidden assumptions, and recognise when your own mind is leading you astray.

Strong critical thinkers tend to score higher on IQ tests, make better decisions under uncertainty, and show greater resistance to misinformation. This article explores what critical thinking actually is, why cognitive biases undermine it, and how to develop the skill systematically.


What Critical Thinking Really Is

Critical thinking is not scepticism for its own sake. It is not distrust of authority, contrarianism, or the habit of "doing your own research" on social media. Cognitive scientists define critical thinking more precisely: it is the disciplined process of actively conceptualising, analysing, synthesising, and evaluating information gathered from observation, experience, reflection, reasoning, or communication [1].

The Core Components

Researchers typically identify five components of critical thinking:

ComponentWhat It MeansExample
AnalysisBreaking complex arguments into partsIdentifying the claim, evidence, and assumptions
EvaluationJudging the quality of evidence and reasoningAssessing whether evidence actually supports the claim
InferenceDrawing justified conclusionsDistinguishing what is supported from what is merely suggested
InterpretationUnderstanding meaning in contextRecognising how framing affects perception
Self-regulationMonitoring your own reasoningNoticing when emotions or biases influence conclusions
"Critical thinking is thinking about thinking in order to improve thinking."
-- Richard Paul & Linda Elder, The Miniature Guide to Critical Thinking [2]

Critical Thinking and IQ

Critical thinking and IQ are related but not identical. IQ tests primarily measure raw cognitive capacity -- how quickly and accurately you can solve novel problems. Critical thinking tests (like the Watson-Glaser Critical Thinking Appraisal) measure applied reasoning -- how you use that capacity when evaluating real-world arguments.

A person can have a high IQ and poor critical thinking skills. Research by Stanovich and West has documented this gap extensively, showing that intelligent people can be just as susceptible to cognitive biases as less intelligent people -- sometimes more so, because they are better at constructing elaborate justifications for biased conclusions [3].


The Anatomy of an Argument

Before you can evaluate an argument, you need to understand its structure. Every argument has three components:

  1. Premises: The starting facts, evidence, or assumptions.
  2. Inference: The logical connection between premises and conclusion.
  3. Conclusion: What the argument is trying to establish.

Valid vs. Sound Arguments

Logicians distinguish between validity and soundness:

  • A valid argument is one where if the premises are true, the conclusion must be true. Validity is about logical structure.
  • A sound argument is both valid and has true premises.

For example:

  • Premise 1: All birds can fly.
  • Premise 2: Penguins are birds.
  • Conclusion: Therefore, penguins can fly.

This argument is valid (the logic is correct) but unsound (Premise 1 is false). Evaluating arguments requires checking both the logic and the facts.

Deductive vs. Inductive Reasoning

Deductive reasoning moves from general principles to specific conclusions. If the premises are true and the logic is valid, the conclusion is guaranteed.

Inductive reasoning moves from specific observations to general conclusions. It produces probable conclusions, not certain ones. Most real-world reasoning is inductive, and understanding its limits is essential.

TypeDirectionCertaintyExample
DeductiveGeneral to specificCertain if valid and sound"All humans are mortal. Socrates is human. Therefore Socrates is mortal."
InductiveSpecific to generalProbable"Every swan I've seen is white. Therefore all swans are white." (False -- black swans exist)
AbductiveObservation to best explanationProbable"The grass is wet. It probably rained."

Cognitive Biases: The Enemies of Critical Thinking

Cognitive biases are systematic errors in thinking that affect decisions and judgments. They are not random mistakes -- they are predictable patterns that occur because of how the human brain evolved to make quick decisions with limited information.

Daniel Kahneman's work, documented in Thinking, Fast and Slow, identifies two systems of thinking [4]:

  • System 1 (fast, intuitive, automatic): Pattern recognition, gut feelings, rapid judgment.
  • System 2 (slow, deliberate, effortful): Logical analysis, mathematical calculation, complex reasoning.

Cognitive biases mostly originate in System 1 and are difficult to override even when you are aware of them. Below are the biases that most commonly corrupt reasoning.

Confirmation Bias

The tendency to seek, interpret, and remember information that confirms existing beliefs while ignoring or dismissing contradictory evidence.

Example: A person who believes a particular diet is healthy notices articles that support it and forgets those that don't. On social media, they follow accounts that reinforce their view and mute those that don't. Over time, they feel increasingly certain -- because they've been selectively exposed to agreement.

How to counter it:

  • Actively seek out the best arguments against your position.
  • Ask: "What evidence would change my mind?" If you can't answer, your belief is not evidence-based.
  • Use the "steel man" technique: describe your opponent's position in its strongest form before criticising it.

Availability Heuristic

Judging the probability of an event by how easily examples come to mind.

Example: After seeing news coverage of a plane crash, people overestimate the danger of flying. Statistically, flying is far safer than driving, but memorable events distort perception.

How to counter it:

  • Ask: "Am I relying on vivid examples or representative data?"
  • Seek base-rate statistics rather than anecdotes.
  • Recognise that news coverage is biased toward rare, dramatic events.

Anchoring Bias

The tendency to rely too heavily on the first piece of information encountered (the "anchor") when making decisions.

Example: In a negotiation, the first number mentioned strongly influences the final outcome. Research by Tversky and Kahneman (1974) demonstrated this effect repeatedly [5].

How to counter it:

  • Before seeing any figures, estimate a reasonable range yourself.
  • Ask: "Would I reach the same conclusion if I had started from a different number?"
  • In negotiations, anchor first to influence the outcome in your favour.

The Dunning-Kruger Effect

People with limited knowledge in a domain tend to overestimate their competence, while experts tend to underestimate theirs.

The original 1999 study by Dunning and Kruger found that the lowest-performing participants estimated their performance as above average, while the highest performers believed others would perform at least as well as they did [6].

How to counter it:

  • Recognise that confidence is not a reliable indicator of competence.
  • When you feel certain, ask: "What would change my mind?"
  • Seek feedback from people more knowledgeable than yourself.

"The trouble with the world is that the stupid are cocksure, and the intelligent are full of doubt."
-- Bertrand Russell

Sunk Cost Fallacy

Continuing an endeavour because of previously invested resources (time, money, effort), even when doing so is irrational.

Example: Continuing to watch a bad film because you've already watched an hour; staying in a failing project because you've already invested six months.

How to counter it:

  • Ask: "If I were starting from scratch today, would I choose this course?"
  • Previously spent resources are gone regardless of what you do now. Only future costs and benefits matter.
  • Recognise that quitting is often rational, not a failure.

Survivorship Bias

Focusing on entities that passed a selection process while overlooking those that did not, leading to false conclusions.

Example: Studies of "successful people" often conclude that their habits caused their success -- but the same habits may have been present in people who failed, and the research ignored them.

How to counter it:

  • Ask: "Where are the cases that failed? Did they have the same characteristics?"
  • Be sceptical of success-based advice; pair it with failure data.
  • In research, consider selection effects explicitly.

The Halo Effect

Allowing one positive characteristic of a person or thing to influence perceptions of unrelated characteristics.

Example: Attractive people are perceived as more intelligent, honest, and competent -- traits unrelated to appearance. Products from prestigious brands are judged as higher quality even when identical to cheaper alternatives.

How to counter it:

  • Evaluate each attribute independently.
  • Ask: "Am I letting my overall impression affect my judgment of this specific aspect?"
  • Use structured evaluation (checklists, rubrics) rather than holistic impressions.

Base Rate Neglect

Ignoring statistical base rates in favour of specific, anecdotal information.

Classic example: A person is described as shy, introverted, and detail-oriented. Is she more likely to be a librarian or a salesperson? Many people say librarian, because the description fits the stereotype. But salespeople outnumber librarians by many times, so the base rate favours salesperson [4].

How to counter it:

  • Start any probability judgment with the base rate.
  • Ask: "How common is this category in the general population?"
  • Only then adjust for specific information.


A Framework for Evaluating Arguments

When you encounter an argument -- in an article, a debate, or a conversation -- apply this five-step framework:

Step 1: Identify the Claim

What exactly is being claimed? Strip away emotional language and find the core assertion. Be specific.

Weak framing: "Social media is bad for kids."
Specific claim: "Daily use of Instagram by teenage girls correlates with a 12% increase in self-reported depression symptoms over a 3-year period."

Specific claims can be evaluated. Vague claims cannot.

Step 2: Identify the Premises

What evidence or assumptions does the argument rest on? List them explicitly. Ask:

  • Are these premises stated, or are some hidden?
  • Are the premises facts, opinions, or assumptions?
  • What would happen to the argument if a premise turned out to be false?

Step 3: Evaluate the Logic

Does the conclusion actually follow from the premises?

Look for common logical fallacies:

FallacyWhat It IsExample
Ad hominemAttacking the person rather than the argument"You can't trust her research; she has a bias."
Straw manMisrepresenting someone's argument to attack it"You want gun control? So you want to disarm everyone?"
False dichotomyPresenting two options when more exist"Either you're with us or against us."
Slippery slopeClaiming one event will cause extreme consequences"If we allow A, then B, C, and D will inevitably follow."
Appeal to authorityAccepting a claim because an expert says so"Einstein said it, so it must be true." (even if outside his field)
Circular reasoningThe conclusion is used as a premise"The Bible is true because it says it is."
Post hocAssuming causation from sequence"I started taking this vitamin and my cold went away, so it worked."

Step 4: Evaluate the Evidence

Even if the logic is valid, weak evidence undermines the argument.

  • Source: Is the evidence from a credible source? Peer-reviewed research? Primary source?
  • Sample size: How many data points? Small samples are unreliable.
  • Methodology: How was the evidence gathered? Randomised? Controlled? Replicated?
  • Representativeness: Does the evidence apply to the case being discussed?
  • Age: Is the evidence current, or has the field moved on?

Step 5: Consider Alternative Explanations

Before accepting a conclusion, ask: "What else could explain this?"

If a study shows that people who drink coffee live longer, possible explanations include:

  • Coffee causes longer life (the implied conclusion).
  • Coffee drinkers are wealthier and have better healthcare.
  • Sicker people are told to stop coffee, leaving healthier drinkers.
  • Coffee drinkers are more social, and social connection extends life.

Good critical thinkers generate alternatives before accepting any single explanation.


Training Critical Thinking

Critical thinking is a skill that improves with practice. Here are evidence-based techniques for developing it.

Read Widely, Including Sources You Disagree With

Exposure to diverse arguments builds the mental models needed to evaluate new arguments. Reading only sources you agree with strengthens confirmation bias.

Write Out Your Reasoning

Writing forces clarity. When you write out an argument, hidden premises and weak links become visible. Journaling about difficult decisions can reveal the biases shaping your thinking.

Play the Devil's Advocate

For any belief you hold strongly, construct the best possible argument against it. If you cannot, your belief may be weaker than you think.

Keep a Bias Journal

Note moments when you catch yourself in a biased judgment. Over time, patterns emerge. Awareness is the first step to correction.

Learn the Fallacies

Studying logical fallacies makes them easier to spot in others -- and, more importantly, in yourself.

Practice Probabilistic Thinking

Instead of "X is true/false," think "There is a 70% chance X is true." Probabilistic thinking captures uncertainty honestly and avoids the binary thinking that fuels most biases.

"Nothing in life is as important as you think it is while you are thinking about it."
-- Daniel Kahneman, Thinking, Fast and Slow [4]

Engage in Formal Reasoning Practice

Logic puzzles, debate clubs, mock trials, and structured writing courses all train critical thinking. Online platforms offer courses in formal logic, argumentation theory, and statistical reasoning.


The Hardest Part: Applying Critical Thinking to Yourself

The greatest failure mode of critical thinkers is applying the skill only to others. You scrutinise opposing arguments rigorously while giving your own beliefs a free pass.

Cognitive psychologists call this the bias blind spot: the tendency to recognise biases in others more easily than in yourself [7]. Research shows this blind spot is stronger in people with high IQs and high critical thinking training.

The counter is epistemic humility: the recognition that your own reasoning is also vulnerable to the errors you criticise in others.

Practical applications:

  • When you feel certain, be most cautious.
  • When you dismiss an argument quickly, consider why.
  • When you feel emotional about a topic, expect your reasoning to be compromised.
  • Invite criticism of your views from people you respect.


Summary

Critical thinking is the ability to evaluate arguments, evidence, and your own reasoning with discipline. It is distinct from raw intelligence (though correlated with it), and it can be learned and practiced.

The five-step framework -- identify the claim, examine the premises, evaluate the logic, assess the evidence, and consider alternatives -- provides a reliable method for analysing any argument. The major cognitive biases -- confirmation bias, anchoring, Dunning-Kruger, sunk cost, survivorship, halo effect, and base rate neglect -- represent predictable patterns to watch for.

The hardest part is applying critical thinking to your own reasoning. Humans are much better at spotting bias in others than in themselves. But with deliberate practice -- reading widely, writing out your reasoning, playing devil's advocate, and keeping track of your own mistakes -- critical thinking becomes a durable habit rather than an occasional effort.

In an information environment flooded with low-quality content, strategic misinformation, and algorithmic amplification of emotional appeals, critical thinking is no longer optional. It is one of the most valuable cognitive skills a person can develop.


References

[1] Facione, P. A. (1990). Critical Thinking: A Statement of Expert Consensus for Purposes of Educational Assessment and Instruction. The Delphi Report. American Philosophical Association.

[2] Paul, R., & Elder, L. (2019). The Miniature Guide to Critical Thinking: Concepts and Tools (8th ed.). Rowman & Littlefield.

[3] Stanovich, K. E., & West, R. F. (2008). On the relative independence of thinking biases and cognitive ability. Journal of Personality and Social Psychology, 94(4), 672-695. doi:10.1037/0022-3514.94.4.672

[4] Kahneman, D. (2011). Thinking, Fast and Slow. Farrar, Straus and Giroux.

[5] Tversky, A., & Kahneman, D. (1974). Judgment under uncertainty: Heuristics and biases. Science, 185(4157), 1124-1131. doi:10.1126/science.185.4157.1124

[6] Kruger, J., & Dunning, D. (1999). Unskilled and unaware of it: How difficulties in recognizing one's own incompetence lead to inflated self-assessments. Journal of Personality and Social Psychology, 77(6), 1121-1134. doi:10.1037/0022-3514.77.6.1121

[7] Pronin, E., Lin, D. Y., & Ross, L. (2002). The bias blind spot: Perceptions of bias in self versus others. Personality and Social Psychology Bulletin, 28(3), 369-381. doi:10.1177/0146167202286008