Follow Slashdot blog updates by subscribing to our blog RSS feed

 



Forgot your password?
typodupeerror
×
Data Storage Microsoft Supercomputing

Too Much Data? Then 'Good Enough' Is Good Enough 56

ChelleChelle writes "While classic systems could offer crisp answers due to the relatively small amount of data they contained, today's systems hold humongous amounts of data content — thus, the data quality and meaning is often fuzzy. In this article, Microsoft's Pat Helland examines the ways in which today's answers differ from what we used to expect, before moving on to state the criteria for a new theory and taxonomy of data."
This discussion has been archived. No new comments can be posted.

Too Much Data? Then 'Good Enough' Is Good Enough

Comments Filter:
  • by Lifyre ( 960576 ) on Thursday June 02, 2011 @06:20PM (#36326350)

    Obviously 640k was "Good Enough"

    Seriously though he makes a good point. If you have so much information that it isn't stored consistently, with varying standards, or is an open field to be populated by an individuals perceptions. The example of all of the different colors of green (Green, emerald, asparagus, chartreuse, olive, pear, shamrock) is a great example of how one piece of information can be expressed in multiple ways. While you can define the color by using the hex code for it that isn't exactly an elegant or user friendly method of input or output.

    He talks about various ways to handle these types of information from limiting input options to finding patterns and using those to "correct" the data.

    • Re:Obligatory (Score:4, Insightful)

      by Fluffeh ( 1273756 ) on Thursday June 02, 2011 @06:36PM (#36326466)

      It's not that there is too much data. That's not a problem at all. From my own experience (I work as a senior analyst for a multinational retailer employing around 200,000 people) it is rather that there isn't a single plan to utilize all the data we have available. Every time we introduce a new system or change the way we do something, the project inevitably drops a new table into our data warehouse. Now, this may seem like an acceptable way to do things, but after this has happened twenty times, it is nigh impossible to run a query that will return data from all these tables in any sort of reasonable time.

      Would it cost more time, effort and money to properly introduce the new data to proper fact tables each time? Of course. However, the benefits would be that we could stop pretending that "we have too much data these days..." - because we don't. We just have too much mess with our data and it becomes unusable.

      In the example above (different descriptions for green) the base system may need these particular terms, but if the data needs to be aggregated or used in another system, then the jobs that pass this to your data repository need to make those changes to adapt the data to work with the rest of your data warehouse. Having said that, if the new system is being developed inhouse, then during development the question should be asked "Can we store the color information in RGB right off the bat and adapt our own system to mask these values behind pretty descriptions?" rather than having to later do it via an ETL.

      • Re:Obligatory (Score:4, Insightful)

        by icebike ( 68054 ) on Thursday June 02, 2011 @07:07PM (#36326698)

        It's not that there is too much data. That's not a problem at all.

        Often, (more often then not, I contend), there is indeed just too much data.

        Because we have all these marvelous computerized data capture system doesn't mean the data is necessary, useful, or worth keeping. However, someone always comes along in the project design stage and insists the millisecond by millisecond weight of a bag of popcorn weighed in real time as it is being filled is going to provide a wealth of data for the design of future bagging systems and materials handling in general.

        The scale was only there to assure that 10 pounds were in the sack and to shut the hopper. Then some fool found out it measured ever few milliseconds and recorded the data.

        So the project manager gets brow beaten into recording this trash which invariably never gets used for anyone for any purpose at any time, as those who lobbied for it wander off to sabotage other projects and never revisit the cesspool they created.

        This happens way way more than you might imagine in the real world these days.

        It used to be projects had to fight for every byte of data collected, there were useful sinks identified for every field. But with falling storage costs the tendency is to simply keep shoveling it in because its easier than dealing with the demands by those "researchers" looking for another horse to ride.

        • Re:Obligatory (Score:5, Insightful)

          by StuartHankins ( 1020819 ) on Thursday June 02, 2011 @07:30PM (#36326850)
          +1 Insightful. I would argue that -- just like you have a lifecycle for software development -- you have a lifecycle for nontrivial amounts of data. Some data is useful in detail for a short term, but wherever possible it should be more coarsely aggregated as time progresses, and you should get sign-in from executives that it can be dumped after a period of time.

          Where I work, I estimated the cost to upgrade our SAN to continue to store a set of large tables which helped everyone understand the cost in real terms. People tend to think once the data is imported or created that it's a small incremental cost to house it from that point forward, but backup times and storage along with execution plan costs increase with size. There is a performance benefit to this trimming; partitioning and check constraints will only get you so far.

          What is difficult to gauge in advance sometimes is how the data will be used -- some things are obvious in the short-term, but as the company looks to different metrics or to shine some light on an aberration, you really need to be able to determine how quickly you can dump the detail. Get signoff then add some padding so you are conservative when you destroy. Make a backup "just in case" and delete it after a few months. The good news in my work is that changing your mind later to adapt to the new requirements means expectations are already set to change the way it works "from this point forward". There are many fields of work that do not have that luxury, because of the time or cost to gather detail again.
          • I think there's a Monty Python episode like that, called "Fade to Black". Data ought to have a half-life. Otherwise, it will conquer the world. No one throws stuff out. I wonder how many SAN "accidents" are just frustrated DBAs.

            • "---- Teach Peace. It's Cheaper Than War."

              Nice sig. Now the Republicans who spent too much are trying to blame it on Obama.

              • Peace is easy as long as everyone lives under my rules.

              • Both major parties are complicit in the spending and war disasters. Obama may not have started the wars in Afghanistan and Iraq, but he's certainly done little to end them. He has added his own excursion in Libya with vague threats to Syria and others, so who knows.

      • isn't a single plan to utilize all the data we have available.

        Fluffeh, if you haven't done so already, I think you would enjoy taking John Zachman's course in Enterprise Architecture. If one were building an addition on one's house, would one just start hammering things in place or would one look at the existing plans first? Too many IT projects just look at their own plans and don't look at the larger plans they should fit into. And in most IT shops those larger plans don't exist anyway. So we just hammer things into place and wonder why the data doesn't work out.

        --
        .

      • by guruevi ( 827432 )

        With good design or dumping all your data daily to a platform designed for such data analysis your problems would be solved (of course, given enough money, time and brain resources).

        The problem imho is that we collect too much cruft and we're simply unsure what to do with it. The systems for good data queries have been designed be they SQL, NoSQL or some specific BI solution. The problem is that most DBA's don't know how to use the collection of them correctly, there are in many cases no "data architects" o

    • It's not a new idea, it's been explored before and it only works in certain cases. Take a look at Ontologies are overrated [shirky.com]. From the section called "Mind Reading":

      You can't do it. You can't collapse these categorizations without some signal loss. The problem is, because the cataloguers assume their classification should have force on the world, they underestimate the difficulty of understanding what users are thinking, and they overestimate the amount to which users will agree, either with one another or wi

  • The data quality and meaning of this summary is rather fuzzy. I have no clue what exactly they're talking about. No, i haven't RTFA yet, but the summary isn't making it very clear if TFA is something i'd be interested in or not.
    • From a quick skim, it seemed to be yet another "SQL/RDBMS is dying because we have too much heterogeneous data to handle", and a rather ambling and long-winded one at that.
      • by Anonymous Coward

        Can you recommend other articles that have a better take on it? I'm fascinated by this stuff, but I find it difficult to explain to people why this is useful, and why it's a growing field.

        This is one from O'Reilly [oreilly.com] that is decent.

        • by tqk ( 413719 )

          Can you recommend other articles that have a better take on it?

          Or, for the darker side, grep /. for "NSA whisleblower".

      • Really. This seemed to be one of the better articles I've seen featured on /. in the last year.
      • I agree with your accessment .
        This was a long ambling caffienated rambling, about too much data to handle. To some degree I think it does a disservice when the author doesn't state up front in what context he prescribes for this data agglomeration. RDBMS is not going away anytime soon. There are too many people who are jumping on this as a hot topic and start speaking/thinking metaphorically about data, when they have very little information about true data and understanding of relationships.
        It's kind
    • I skimmed the article, and I can say this much: they mean a fairly specific type and use of data.

      Too much data from scientific results? Only the researcher him or herself would ever say there's "too much data." Everyone else says "not enough data." Everyone. At all times. Especially his committee and reviewers. Even when I've worked so hard for so long for so little money. After all, THEY'RE not the ones who are sacrificing their happiness, time, effort, hairline, and relationships to...

      Uh, I m
  • by billrp ( 1530055 ) on Thursday June 02, 2011 @06:29PM (#36326410)
    SQL DBs are not appropriate for storing, processing, querying, and browsing unstructured documents.
    • I think the important insight is that Parkinson's Law applies not just to the quantity of data, but also to the varieties of data.

  • A bunch of rambling self-evident or speculative statements, followed by conclusion:

    Conclusion

    NoSQL systems are emerging because the world of data is changing. The size and heterogeneity of data means that the old guarantees simply cannot be met. Fortunately, we are learning how to meet the needs of business in ways outside of the old and classic database.

    Which was apparent to everyone, and missed the real point: We have lots of data, and we're too impatient to wait for it to be aggregated, synchronized and processed. There goes 10 minutes of my life I'll never get back.

    Here's a hint: People working on the solutions to this problem work in the financial sector and in quantum physics.

    • by Anonymous Coward

      Here's a hint: People working on the solutions to this problem work in the financial sector and in quantum physics.

      Or journalists [wikipedia.org]. Or intelligence agencies [wikipedia.org]. Or any business that's large enough to have information silos [wikipedia.org]. Or transportation departments [dot.gov]. Or internet startups [nytimes.com].

    • by Anonymous Coward

      Which was apparent to everyone, and missed the real point: We have lots of data, and we're too impatient to wait for it to be aggregated, synchronized and processed.

      And that the management don't do their jobs, the business process engineering consultant is too expensive, the company is no longer a monolithic, singular entity with a clear hierarchical organization and the number of collaborations between small organizations and organizational units in a merged company increases as the most efficient way of performing a task is sought after. The article tries to push a problem of human organization to the fields of technology.

  • by Comrade Ogilvy ( 1719488 ) on Thursday June 02, 2011 @06:35PM (#36326456)

    The researcher is just throwing together a bunch of problems that have existed, in some fashion, for a very long time, and concludes with open questions rather than even vague proposals for solutions. So I would say this article is both too detailed, and not detailed enough.

  • From TFA:

    As a transaction geek, I've spent decades working on systems that provide a guaranteed form of consistency.

    Uh...so you spent decades working on systems which are not needed for many problems (many problems don't need transactions, especially mostly read web publishing problems, which is a strength of no-sql), and now you are upset that people are not using your systems?

    • by Anonymous Coward

      You are talking about removing the C of ACID. Remove any one of those 4 and you get speed. The guys who came up with ACID knew it.

      http://en.wikipedia.org/wiki/ACID [wikipedia.org]

      However some problems lend themselves to not having ACID involved. Such things include Data Warehousing, and web views of transient data.

      Other problems you *want* it to work perfect and every time. Such as when you are a business and printing out bills. Some of the data missing not so good...

      Now in say a discussion forum. You say you want th

  • Should be 'Too Many Data'. Morans.
  • The article makes an assumption that all data in the world consists of marketing surveys and transcripts of phone wiretaps.

  • by lucm ( 889690 ) on Thursday June 02, 2011 @07:28PM (#36326838)

    This article is confusing because most of the verbiage is made up by the author (such as "inside" or "locked" data). It is also misleading because it seems to indicate that structured and unstructured data usage is the same. Well it's not - a very large proportion of unstructured data is blog posts and emails but the amount of search and aggregation that is performed on this type of information outside of a few major companies (such as Google) is very low, which makes this usage a niche and not a trend maker.

    The reality is that there are three categories of data that are relevant for databases: numbers, text and spatial. Everything else, which falls under the umbrella of "binary", is very unlikely to benefit from a database engine; only the metada can be manipulated and this metadata falls under one of the other categories and is a very good target for ETL. And so far nobody came up with a reliable way to search binary, such as video or audio, without relying on heavy indexing, metadata or any kind of transformation that takes binary and make it text data.

    If a piece of data cannot be searched or aggregated, it does not belong in a database, it belongs on a filesystem. Anything can be done with blob columns but performance is usually not very good because the database engine cache is not designed for large objects. NoSql or not.

    Also there is so much happening with storage infrastructure, such as sub-volume tiering or block-level replication, any analysis of data that does not take a look at storage is flawed.

    • by tqk ( 413719 )

      The reality is that there are three categories of data that are relevant for databases: numbers, text and spatial. Everything else, which falls under the umbrella of "binary", is very unlikely to benefit from a database engine; only the metada can be manipulated ...

      Ya know, my email client, via its ~/.mailcap assignments, manages to handle blobs fairly well. What's wrong with your tech?

      Never blame the technology. Blame the bum who's misusing it. Not saying that's you. But if mutt can do it, why can't La

      • You can search inside video files and pictures with your email client? Where do I sign up?

        • by tqk ( 413719 )

          You can search inside video files and pictures with your email client?

          You've video and picture files your db can't open up correctly? Why?

    • by dkf ( 304284 )

      The reality is that there are three categories of data that are relevant for databases: numbers, text and spatial. Everything else, which falls under the umbrella of "binary", is very unlikely to benefit from a database engine; only the metada can be manipulated and this metadata falls under one of the other categories and is a very good target for ETL.

      Actually, it depends on whether you can define relations over the data. The set of relevant relations will vary with the data type. For example, I can imagine it being possible to do searches over images, sounds or movies; there is fundamental structure there, relations are definable. That's not to say it is easy, or that we know the right set of relations, or that implementations are good yet, but to dismiss it as impossible? You jump too far.

      • by lucm ( 889690 )

        > For example, I can imagine it being possible to do searches over images, sounds or movies; there is fundamental structure there, relations are definable.

        What you talk about is metadata. Defining an index of sorts to store patterns and checksum does allow one to establish relations between images - but the search is then performed on the said metadata, not on the binaries. A rule of thumb: if you must index data before you can search it, then you cannot search the data, you can only search the metadata

  • by Anonymous Coward on Thursday June 02, 2011 @07:41PM (#36326908)

    We don't read articles, just skim the headline, maybe the submittal, and then a few top ranked posts.

    That's Good Enough! (tm)

  • This is why Statistics will become more and more important over time--it allows you to make inferences about populations that you couldn't possibly count. If you already know Comp Sci or or learned how to program on your own, go for a couple of Stats degrees. Along with your programming skills Stats will do you very well as the information age unfolds.
  • Using impure data to make real world decisions is just plain wrong. This is how 5 year olds end up on the "do not fly" list, how credit scores get incorrectly reported - add your own examples of how idiots read more from data than it contains.

    So-called scientists saying it's OK to just take a guess only shows what scientists have become in this modern world. Once you get to that point, you may as well throw out the data and base your guess on whatever floats your boat. It wouldn't be any less valid - and

    • It's the problem of Significant Figures for verbal data sets.

      Last I recalled, you can only keep he number of significant figures equal to the fuzziest of the inputs. So you have 45.236 + 12.877 + "one million" ... means your answer can only keep the one significant figure of "a little over a million".

      So for these non-verbal data sets, you get too many data fields, and misc people forget to put the stuff in ref1, someone puts a date instead of an invoice number in ref2, the vendor code is wrong in ID1, some

    • by N1AK ( 864906 )

      Using impure data to make real world decisions is just plain wrong.

      To call someone an idiot because you're too blinkered to comprehend that the importance of accuracy can vary depending on the decision you appear like an ignorant idiot yourself.
      There is nothing wrong with a search engine making an educated guess about the colour of shoes based on an expert system or similar methodology. It might be annoying when it gets it wrong, but most users would prefer to have the option than not. His point, and it w

  • Nothing new (Score:4, Insightful)

    by Whuffo ( 1043790 ) on Friday June 03, 2011 @01:34AM (#36328706) Homepage Journal

    If the people that write these stories would familiarize themselves with Information Theory (Claude Shannon, in the 1940's) then they'd understand that you still can't make silk purses from sow's ears.

    Yes, it's a lot of records. Yes, the data entry people made mistakes. All this really means is that there's more noise in the data. As the signal to noise ratio declines, the value of the results also declines. Making decisions based on noisy data isn't science, it's only guesswork. That's fine for weather forecasting (a similar problem) but expecting the results from the described data to be more accurate than weather forecasts is foolish. Remember: garbage in, garbage out.

  • by Anonymous Coward

    Since when has MS ever had any OTHER opinion on ANYTHING!

  • by AtomicSnarl ( 549626 ) on Friday June 03, 2011 @07:24AM (#36329884) Homepage
    The problem being encountered is one I've faced often in 30 years of weather forecasting: Ambiguity Management.

    The weather business deals with reams of data from thousands of sources and all the complexity of trying to follow a single swirl within a flowing river to figure out where it will be tomorrow. Decades of research and modeling have evolved into dozens of primary rule-based tools available to forecasters which are applicable to most situations. Objectively, you should be able to follow the rules, weed out the conflicting or contradictory ones, and get a reliable answer. Realistically, you don't. Why? Two reasons:

    1. The dataset is incomplete.
    2. The tools are imperfect.

    You simply can't have perfect knowledge of all the relevant details in the atmosphere to feed a completely objective tool (computerized model or whatever) to get your perfect prediction. Like Rosanne Rosannadana's mother said, "It's Always Something!"

    The trick then in being a good (aka reliable) weather forecaster then is how you manage the ambiguity of incomplete data filtered through inherently biased tools. Some weather stations run hot or cold, have local effects enhancing or reducing pressure or winds, etc, etc, etc. Good models account for this, but that's a static adjustment, not a dynamic one. Models run hot or cold, fast or slow, depending on their structure and assumptions, and they reval their strengths and weakness over time compared to other models and reality at verification time.

    The basic forecasting questions are - Where is it, Where is it going, an what will happen when it gets there? Because the models are perfect (100% replication of output from identical starting states), but are always wrong (inherent model and data limitations), you make your money examining the consistency. The model(s) are running slow and cold recently due to the whatever event going on? Ok -- warm it up a few degrees and expecting things a few hours earlier than it forecasts tomorrow. Some models handle well in winter but get klutzy with large thunderstorm events. One model I worked with covered the world in clouds if you waited long enough. Solution? Don't trust it past X number of hours. And so on for the family of models through the decades and to today. Some models have high skill up to a certain point then it drops off quickly. Others show less skill, but are decent for the long haul. You get the idea. You can make a forecast using only one tool, but you can make a better one using several and sorting out their differences by using ambiguity management.

    Needless to say, you needed a solid understanding of the physics and dynamics of the atmosphere to help make good decisions to do all this effectively. The modelers and users now data mining these huge collections of information likewise need a solid understanding of Statistics and the event mechanics they're examining to make any good sense of it all. At the very minimum, a large poster announcing "Coincidence is not Causation" needs to be in every office, otherwise you start getting breathless announcements about how underarm deodorant "causes" cancer because people eating hamburgers had a lower incidence rate by comparison.

    Your Mileage May Vary -- a lot. That's the point.
  • Doesn't "how much is too much" depend more on what sort of data you are talking about than the systems used to record and analyse it? Aircraft risk analysts would surely argue that they need all the data they can get to help prevent every instance of catastrophic failure. Biologists on the other hand are used to working with extraordinarily fuzzy data and still drawing valid conclusions

If all else fails, lower your standards.

Working...