Data Intelligence, Business Analytics
Much of the explosion of big data has been driven by increased efficiency in sever performance, memory cost, distributed architecture improvements (cloud, and truly parallel databases, e.g. noSQL) and essentially, by how much it costs to process a terabyte of data, both in terms of memory and bandwidth resources.
However, most of the very big data is very sparse, from an information point of view : big data is essentially made of noise or redundant information (think about videos or tweet data where information redundancy is huge) and can be compacted by 90-95% without any significant information loss. Storing and processing the entire data is a very inefficient process. I believe we can do much better by smartly sampling and smartly summarizing very big data (particularly stuff that is more than 4 week old) - a process known as data reduction or signal processing - rather than storing everything. The sampling / summarizing process is a task that should be left to expert, very senior statisticians, not to computer scientists.
At the end of the day, you should answer the following questions: