Blabberbox » Society » Data Does Not Equal FactShare on Twitter

Data Does Not Equal Fact

January 24th, 2018 | Posted by pftq in Society | #
"Data!=Fact: Inductive vs Deductive Reasoning"

     Part of my frustration with the emphasis on big data, statistics, data driven, and numbers in general is that these are all forms of inductive reasoning.  Many seem to not understand that inductive reasoning only provides an estimate of world.  It does not and cannot prove anything.  Evidence alone does not provide truth.  Data does not equal fact.

     There are two main forms of reasoning: inductive and deductive.  Inductive reasoning is the process of generalizing from a few observations to a broad conclusion.  Deductive reasoning is the process of understanding what must be true and apply to something more specific given a premise or broader assumption.  The former emphasizes correct observation while the latter emphasizes actual logical flow and reasoning, the how and why as opposed to just what it is you are seeing.  Only deductive reasoning is actually logical, but the phrase logic is often misused to describe both.

Inductive Reasoning

     Inductive reasoning is backwards looking and by definition says nothing of the future.  You start from seeing one thing and then try to generalize to something you haven't seen before, but you can never be fully sure.  The easiest way to identify it is it almost always depends on evidence, concrete numbers, quantitative data, statistical sampling - stuff that you can measure.  It is useful for describing many things together, giving a big picture from high up of what's happening on the ground, but people seem to forget that no matter how large the sample size is, it never is absolute.  Everything always boils down to a probability as opposed to a cause and reaction.  Even in the most introductory statistics class, the teacher tells you correlation is not causation.

     Yet we have phrases like "social validation," "empirically proven," or "statistically impossible."  In Silicon Valley, the thought is that if the data suggests a conclusion with high enough statistical significance, then it must be true unconditionally.  They won't even look at you if you don't have data to back your conclusion, if your ideas are not "data driven."  Similarly, in Wall Street and academia, the thought is that if the biggest names or majority of people are doing things one way, it must be the right way.  One of the most frustrating things in college for me was having actual professors tell students that a theory doesn't need explaining because it's always worked.  Things are true just because they work and have always worked.

Deductive Reasoning

     Deductive reasoning is more abstract (harder) and less about the numbers, if at all.  It drills down and keeps asking why at every level, on the ground as opposed to from a distance.  It is forward-looking, in that you pro-actively think out how something *must* work rather than just observe that something *does* work - an objective truth independent of observation.  Tesla, Einstein, and many other scientists are famous for having come to their ideas through sheer mental simulations and thought experiments rather than physically observing - General Relativity, for example, is still being confirmed decades after the theory came out (and was largely dismissed until there was evidence).  The clearest example of how this differs from empirical / observation-based reasoning might be one from my own life on how I compare different routes through a city grid; an inductive reasoning approach would try all possible paths, but if you visualize and manipulate the grid in your mind, you realize all paths are the same.


Some will ask, how do you know something *must* work or *why* something works? How do you know what's logical? A lot of people simply do not believe there is such thing as objective truth in this manner; I've had people from college literally try to argue that even math only works because everyone agrees on it, that 1+1 only equals 2 because everyone believes it does, and that it's just as susceptible to being "right until it's wrong." Much of it comes down to understanding logical fallacies, but the gist of it is that any reason you come up with should not create a circle back to itself (aka infinite loops).  Pretty much any logical fallacy can be identified by asking why and realizing if the reason leads you in a circle.  This is what happens with inductive reasoning because at the end of the day, you're only describing what you see; it's observation piled upon observation without explanation or looking beneath the surface.  With deductive reasoning, you keep going deeper and deeper down the cause-and-effect chain, which may not necessarily ever end but being able to move from point to point without going in circles, without contradicting yourself, is what logic is about.  It doesn't automatically mean something is true, but if you're wrong, you know it's from the observations/evidence, which then is the right place for inductive reasoning to come in.

Chart

Inductive vs Deductive Reasoning
Inductive reasoning describes what *is*. Deductive reasoning explores what *if*, the how and why.  Put in simple terms, you use inductive reasoning to gather your information, but you use deductive reasoning to understand and interpret it.  The data gives you eyes, but you've got to use your brain to make sense of it. Unfortunately, more and more it seems we blindly lump them together as equal approaches to logic, almost as if it's just a matter of preference for what the user is comfortable with when one is actually completely useless without the other.

Misuse of inductive reasoning essentially amount to misuse of statistics.  Especially in tech and finance, there is this bizarre assumption that an anomaly will never happen, and if it does, it somehow does not count.  They'll point to the 99 other times things succeed, but they don't seem to fully grasp that they only get to live in one life, that the outcome is discrete.  You see decisions based on "expected value," where they'll take the decision with the highest weighted average of outcomes, even if it includes the potential outcome of death - but they don't care because the probability is so small that it’s "statistically impossible." There is no back-up plan, no acknowledgement whatsoever that the improbable could happen.  The entirety of Blockchain, for example, is based on assumptions like no two people ever drawing the same random number for a wallet, even though there's nothing that says they can't and nothing that checks if they do.  They say that the odds are so low that it would take a billion years to happen, but the last I checked, that's not how probability works.  It could happen tomorrow, and that would be the end of it.  Again, they ignore that you only get to live with one discrete outcome, and once something happens, it's done, it's over.  You don't get to live in the weighted average of the possibilities.  That's not a real number.  It's literally made up.

On the opposite side of that are people who simply can't recognize what's right in front of them unless they get a thousand other opinions or datapoints to confirm it.  Their own house could be on fire, but they'd sooner check Twitter than look with their own eyes.  The purpose of inductive reasoning and statistics overall is to make it easier to think about large amounts of data as an estimate.  It's for zooming out to the big picture and thinking about what a hundred things will do overall.  It's wrong when you try to drill down and apply it to just that one thing in front of you.  Formally, it's the fallacy of affirming the consequent.

A great quote that illustrates all this comes from the movie "I Origins":
"If I drop a phone 1000 times, and one of those times it doesn't fall, that it just hovers in the air, then that is the one datapoint that matters." (Paraphrased)

If you only used inductive reasoning, you'd need to be hit by a car 20 times before you know it hurts.  If you used deductive reasoning, you'd know not to be hit in the first place, not because you magically predict the future but because you are trying to think and imagine what could happen.  It's not luck or selection bias.  People live this everyday.  You don't get to redraw your life from a hat.

===================

Also published on Substack: https://pftq.substack.com/p/data-does-not-equal-fact
1193 unique view(s)

Responses

  1. Aaron said,
    Jan-25-2018, 02:54pm

    Hm. Do you know of anything in current AI that's deductive? Nothing in machine learning, I believe.

    Have been deeply studying reinforcement learning lately. I wonder what learning a deductive policy would look like.

  2. pftq said,
    Jan-27-2018, 09:49am

    I believe my approach here would lead to one: https://www.pftq.com/blabberbox/?page=Creating_Sentient_Artificial_Intelligence

    But it would be a significant undertaking to build and implement.  I haven't found the time nor resource to attempt it.  And it's been difficult to find anyone to help that actually understands, can figure out the details/intricacies, and isn't dismissive of it from the outset (because it's not empirical).

  3. Aaron said,
    Jan-29-2018, 09:07am

    Interesting ideas... lot of useful distinctions. Seems like it's moving in the right direction, at a minimum.

    Thanks for pointing that article out. Am pondering on it, and how to apply it.

  4. bx549 said,
    Mar-30-2018, 06:50pm

    Nice thoughts and summary of trends in data science and the requirement that all decisions be data-driven. I have one quibble about your statements that outcomes are discrete and that we don't get to live in a world where the possible outcomes are weighted averages. What you say is true, but the whole of utility theory says that it's OK to make decisions based on expected value. The outcome is like a lottery; you get one of the "prizes". Your preference for a "prize" is measured by the risk you are willing to take to receive it. I want to use my crypto-currency wallet to store my coins and sometimes buy things. A collision is possible, but as a practical matter I'm willing to take that risk. Similarly, an asteroid could hit my car while I'm driving to work, but I make the decision to go anyway. Well, I agree with you that deductive thinking is (unfortunately) losing favor.

  5. pftq said,
    Jun-19-2018, 09:45am

    It's less about whether you still make the decision and more that you understand you are guessing and should have a backup plan.  My criticism is mainly of people who put everything on the line because the "expected value" is so high, then have no fallback plan when the "statistically impossible" happens.  It doesn't matter that anyone says it's "ok" to make decisions based on expected value.  It's still not a "real" value, in that you will never take home that expected value as an outcome because it doesn't actually exist (it's all or nothing). 

    In your crypto example, the consideration should be what happens if your wallet does get a collision, not that your value of using it is some expected amount.  Whoever happens to get your wallet key is not going to politely leave you the expected value of your usage; they'll take everything.

    The asteroid example, meanwhile is a fallacious argument, equivalent to "we cannot do everything, so we should do nothing."  It's kind of like saying we can die any time so why bother to care about the future at all.  In the real world though, NASA does put quite a bit of effort into tracking and monitoring asteroids, as no matter how small the chance is, it only takes one.

Leave a Comment

Name: (Have an account? Login or Register)
Email: (Won't be published)
Website: (Optional)
Comment:
Enter the code from image: