Book cover of Big Data by Viktor Mayer-Schönberger

Big Data

by Viktor Mayer-Schönberger

16 min readRating: 3.7 (8,592 ratings)
Genres
Buy full book on Amazon

Introduction

In the digital age, we are surrounded by an unprecedented amount of information. From our online searches to our social media interactions, from our shopping habits to our physical movements, every aspect of our lives is being captured and stored as data. This explosion of data, combined with powerful computing technologies, has given rise to what we now call "big data."

Viktor Mayer-Schönberger's book "Big Data" explores this phenomenon and its far-reaching implications for society, business, and our daily lives. The author takes us on a journey through the world of big data, explaining its potential to revolutionize how we understand and interact with the world around us.

The Rise of Big Data

From Scarcity to Abundance

Not too long ago, collecting and analyzing data was a time-consuming and expensive process. The author illustrates this point with the example of the 1880 US census, which took over eight years to complete and publish. By the time the results were available, they were already outdated.

Fast forward to today, and the picture has changed dramatically. The advent of computers, digitization, and the Internet has transformed the landscape of data collection and analysis. Information can now be gathered passively or with minimal effort, and the cost of data storage has plummeted.

This shift from data scarcity to data abundance marks the beginning of the big-data era. While there's no formal definition, "big data" refers to the massive scale of data being captured and the unprecedented opportunities for insights that these large datasets offer.

Google Flu Trends: A Big Data Success Story

To illustrate the power of big data, the author presents the case of Google Flu Trends. In 2009, Google published a research paper demonstrating how they could predict flu outbreaks by analyzing users' search terms. By comparing historical search data with official flu spread statistics, they identified 45 search terms that could accurately predict flu outbreaks.

This breakthrough came just in time for the H1N1 flu pandemic. Google's system provided more timely and useful information to public health officials than traditional government statistics. This example showcases how big data can offer insights and predictions that were previously impossible, potentially revolutionizing fields like public health.

The Power of Datafication

Turning Everything into Data

One of the key concepts introduced in the book is "datafication" – the process of capturing information about the world in the form of data. This trend extends far beyond our online activities and is increasingly being applied to aspects of our lives we never thought could be quantified.

The author provides several fascinating examples of this trend:

  1. Buttprint Recognition: Japan's Advanced Institute of Industrial Technology has developed pressure sensors that can identify individuals based on the weight distribution of their backsides on a car seat. This technology could potentially be used as a security device for cars.

  2. Health Monitoring Earbuds: Apple has patented technology to measure blood oxygenation, heart rate, and body temperature through earbuds.

  3. Smart Floors: IBM has patented touch-sensitive floor surfaces that can track people's movements across a space.

These examples demonstrate how researchers and companies are finding new ways to capture data from unexpected sources. The goal is to gain valuable insights into human behavior and create innovative products and services.

Beyond Sampling: The Whole Picture

The Limitations of Small Samples

Traditionally, when we wanted to understand a large population or phenomenon, we relied on sampling – taking a small subset of data and extrapolating it to represent the whole. The author uses the example of a telephone survey for a local election to illustrate this point.

While sampling can provide useful insights, it has inherent limitations. As you try to analyze smaller subgroups within your sample, you quickly run out of data points, making it impossible to draw reliable conclusions. For instance, in our election survey example, you might have enough data to make general predictions about the entire population, but not enough to say anything meaningful about specific subgroups like public servants under 30.

The Big Data Advantage

Big data changes this paradigm. Instead of relying on small samples, we now have access to vast amounts of data – sometimes even all of it. This allows us to "zoom in" on subgroups almost endlessly without losing statistical significance.

In the context of our election survey example, a big data approach might involve analyzing the voting preferences of tens of thousands of people, or even everyone in the town. This wealth of data allows for much more detailed and reliable analysis, even for very specific subgroups.

Embracing Messiness: Quantity over Quality

The IBM Translation Experiment

The author shares an interesting anecdote about IBM's attempt to develop a language translation program in the 1980s. Instead of using traditional methods based on grammar rules and dictionaries, IBM engineers tried a novel statistical approach. They fed the computer with three million sentence pairs from official Canadian parliamentary documents, hoping the system would learn to translate based on statistical probabilities.

Despite initial promise, the project ultimately failed. The system could reliably translate common words and phrases but struggled with less frequent ones. The problem wasn't the quality of the data – it was the quantity. There simply wasn't enough data for the system to learn from.

Google's Success with Messy Data

Less than a decade later, Google tackled the same problem with a different approach. Instead of using a limited set of high-quality data, they used the entire global Internet – billions of pages of text in various languages. Despite the questionable quality of much of this data, the sheer volume made Google's translations more accurate than any rival system.

This example illustrates a key principle of big data: sometimes, having vast amounts of messy data can be more valuable than having a smaller amount of clean, high-quality data. When working with big data, we can afford to be more forgiving of inaccuracies because the sheer volume of data tends to minimize their impact.

Correlation vs. Causation: Letting the Data Speak

The Orange Car Mystery

The author presents an intriguing example from a data analysis competition. Contestants were given the task of identifying factors that predict whether a used car is likely to be a "lemon" (a defective car). Surprisingly, the analysis revealed that orange cars were half as likely to have defects as the average car.

This finding raises an obvious question: why? As humans, we naturally want to understand the reasons behind such correlations. However, one of the key insights of big data is that we don't always need to know why two things are related – sometimes, it's enough to know that they are.

The Power of Unexpected Correlations

The author argues that automatic analyses of large datasets can reveal correlations we never even thought to look for. While we may not always understand the underlying causes, these correlations can still be incredibly useful.

He illustrates this point with a study conducted by IBM and the University of Ontario on premature babies. By analyzing vast amounts of data on babies' vital signs, they discovered that very stable vital signs often preceded serious infections – a counterintuitive finding that doctors could use to provide better care.

This example shows how big data can uncover valuable insights that go beyond our intuitive understanding or preconceived notions. While it doesn't tell us why things are related, it can tell us that they are – and often, that's enough to drive meaningful improvements.

The Hidden Value of Data

Primary and Secondary Uses of Data

Most data is collected with a specific purpose in mind. Stores collect sales data for accounting, factories monitor output for productivity tracking, and websites analyze user behavior to improve user experience. However, the author argues that the real value of data often lies in its secondary uses – applications that weren't originally intended but can prove even more valuable than the primary use.

He provides several examples:

  1. Swift's GDP Forecasts: The interbank payment system Swift collects data on financial transactions for record-keeping. They later discovered this data correlates well with global economic activity, allowing them to offer accurate GDP forecasts as a new service.

  2. Search Term Mining: Companies like Experian allow clients to analyze old Internet search terms to gain insights into consumer tastes and market trends.

  3. Mobile Phone Location Data: While primarily collected for routing calls, this data can be used for traffic monitoring or location-based advertising.

These examples demonstrate how data collected for one purpose can find valuable secondary applications. This realization has led forward-thinking companies to design their products and systems with potential secondary uses of data in mind.

The Big Data Mindset

Spotting Opportunities in Data

The author introduces the concept of the "big data mindset" – the ability to recognize where available data can be mined for valuable information. He argues that anyone can spot new opportunities to create value from data, even without owning vast amounts of data or possessing advanced analytical skills.

He provides two examples of individuals who successfully leveraged this mindset:

  1. FlightCaster: Bradford Cross and his friends combined publicly available data on flight times and historical weather records to predict flight delays across the US. Their predictions became so accurate that even airline employees began using their site.

  2. Decide.com: This company records billions of price quotes for millions of products from e-commerce sites. By analyzing this data, they not only provide users with the cheapest price but also advise on the best time to buy, predicting future price changes.

These examples show how individuals with a big data mindset can spot opportunities to extract value from available data, creating innovative services and products in the process.

The Power of Data Combination

Synergy in Data Sets

The author draws an analogy to the board game Clue (Cluedo) to illustrate how combining different pieces of information can reveal insights that aren't apparent when looking at each piece in isolation. This principle applies to big data as well – combining different datasets can often create greater value than the sum of their parts.

He provides two examples:

  1. Danish Cancer Study: A research group combined mobile phone user data with cancer patient records to conduct one of the largest studies on the potential link between mobile phone use and cancer. The comprehensive nature of the data allowed them to control for factors like education and income without compromising reliability.

  2. Inrix Traffic Analysis: This company gathers real-time location data from various sources (car manufacturers, commercial fleets, and their own smartphone app) to create accurate, timely data on traffic flows and jams.

These examples demonstrate how combining different datasets can reveal trends and insights that weren't discoverable from the individual datasets alone.

The Data Exhaust Gold Mine

Recycling User Interactions

The author introduces the concept of "data exhaust" – the trail of data we leave behind as we interact with online services. Smart companies are increasingly capturing and analyzing this data to improve their products and services.

He provides several examples:

  1. Google's Innovations: Google uses data from users' search queries and typos to improve its spell-checker and autocomplete systems.

  2. Facebook's Layout Optimization: By analyzing user behavior, Facebook discovered that users were more likely to post content or reply to posts if they had just seen a friend do so. They used this insight to adjust their layout.

  3. Zynga's Game Refinement: The online gaming company analyzes player behavior to identify and address points where players tend to give up, improving the overall gaming experience.

These examples show how companies can leverage the wealth of data generated by user interactions to continuously refine and enhance their offerings.

The Privacy Paradox

Outdated Privacy Protection

The author argues that current privacy laws and anonymization methods are becoming increasingly ineffective in the age of big data. He points out two main issues:

  1. Consent and Purpose Limitation: Current laws require companies to inform users about what data is being collected and for what purpose, and to obtain consent. This approach hinders the discovery of valuable secondary uses for data, as companies would need to seek approval from every user before adopting data for a new purpose.

  2. Re-identification Risk: The detailed nature of big data makes it possible to re-identify individuals from anonymized data sets. The author cites the example of AOL releasing anonymized search terms in 2006, which led to the New York Times successfully identifying one of the users within days.

These issues highlight the need for new approaches to privacy protection that are better suited to the realities of big data.

The Ethics of Prediction

The Minority Report Dilemma

The author draws a parallel between the science fiction movie "Minority Report" and the increasing use of predictive analytics in law enforcement and criminal justice. He points out that while big data facilitates the prediction of criminal behavior, we must be cautious about how we use these predictions.

He provides examples of how predictive analytics are already being used:

  1. Parole Decisions: Many US states use data-analysis-based predictions of a prisoner's chance of re-offending when deciding on parole.

  2. Predictive Policing: Police departments are increasingly using data analysis to allocate resources, often based on profiling of individuals, groups, and neighborhoods.

While these methods can be useful, the author warns against taking them to extremes. He argues that we must never judge or punish someone for what they are predicted to do, only for what they have actually done. To do otherwise would deny individuals their free will and the possibility of moral choice.

The Perils of Data-Driven Decision Making

When Data Leads Us Astray

While big data offers powerful tools for decision-making, the author cautions against becoming overly reliant on data. He identifies several potential pitfalls:

  1. Measuring the Wrong Thing: Quantifying complex phenomena can lead us to focus on metrics that don't truly capture what we intend to measure. The author uses the example of standardized tests in education, which may not fully reflect the range of qualities we expect education to provide.

  2. Unintended Incentives: Misuse of data can incentivize behavior we never intended. Again, standardized tests serve as an example, as their importance has led teachers and students to focus on improving test scores rather than overall education quality.

  3. Relying on Inaccurate Data: Being overly data-driven can lead us to base decisions on biased or unreliable data. The author cites the example of Robert McNamara during the Vietnam War, who became fixated on enemy body count as a measure of progress, despite the unreliability of this data in wartime conditions.

These examples highlight the importance of maintaining perspective and critical thinking when using big data, rather than blindly following what the data seems to tell us.

Conclusion: Navigating the Big Data Revolution

As we move further into the age of big data, it's clear that we're dealing with a fundamentally new phenomenon that requires us to adjust our thinking and approaches. The vast amounts of data being collected, shared, and combined offer unprecedented opportunities for creating value, enhancing products and services, and gaining new insights into the world around us.

However, as the author has shown throughout the book, big data also comes with significant challenges and potential pitfalls. We need to be mindful of privacy concerns, ethical considerations, and the limitations of data-driven decision-making.

The key to navigating this new landscape lies in developing a nuanced understanding of big data – its potential and its pitfalls. We need to embrace the opportunities it offers while remaining critical and thoughtful about how we use it.

As individuals and organizations, we should strive to cultivate a big data mindset – the ability to recognize the potential value in the data around us. At the same time, we must guard against becoming overly reliant on data, always remembering that data should inform our decisions, not make them for us.

The big data revolution is already transforming how we live, work, and think. By understanding its principles and implications, we can harness its power to create a better future while avoiding the potential pitfalls along the way.

Actionable Idea: Extracting Hidden Value from Data

One of the most powerful takeaways from the book is that anyone can create value from big data – you just need to identify the right data and users. Here's how you can put this idea into action:

  1. Inventory Available Data: Start by considering what data you have access to, both in your personal life and professional context. Don't forget about publicly available data, especially online sources.

  2. Think Beyond Primary Uses: For each dataset, try to think of uses that are different from the reason it was initially collected. How could this data be valuable in a completely different context?

  3. Consider Combinations: Think about how different datasets could be combined to reveal new insights. Remember, the value often lies in the unexpected connections between different types of data.

  4. Adopt Different Perspectives: Try to look at the data from the viewpoint of different industries or businesses. How could they benefit from this information?

  5. Identify Potential Users: For each potential use you've identified, think about who would find this information valuable. Could it help consumers make better decisions? Could it help businesses optimize their operations?

  6. Prototype and Test: If you've identified a promising idea, consider creating a simple prototype or proof of concept. This could be as simple as a spreadsheet analysis or a basic web application.

By following these steps, you might discover an innovative way to turn the data around you into a valuable resource. Remember, some of the most successful big data applications came from individuals who simply saw potential where others didn't. With creativity and persistence, you too could uncover the hidden value in the data that surrounds us all.

Books like Big Data