This is such a cool topic, if not insanely complicated for me specifically to try to figure out. 5 exabytes of data encompasses everything we know up until 2003.
That number is swiftly going to climb. In fact, as covered in the intro to the white paper, it already has. We double that number exponentially, it seems. With the ability we have now to digitize, in high quality, almost everything that could ever happen from a billion different angles, how do we mine that data for the important information?
We don’t. Currently. The processing power would be insane. But I really like the idea that this could happen some day, using our crowdsourced data collection and using it to press all of that information together in a way that benefits us so well.
In the meantime….Search engines! Selective processing! We can take a guestimate of the data out there and work to provide a little thumbnail view of that data. The problem is we can generate hundreds of times more data these days than we could ever process as a human being, so we really need to start generating ideas for how to handle all of this data.
1. What big issues do you think could be solved just by combining and collecting all the data that’s available out there, if we could do it in the right way?
2. Random guestimate: How much data do you produce daily? Weekly? Yearly?