The challenge, however, lies in how to quickly analyse increasing volume of data (often
loosely structured). Big data usually refers to very large data sets in the pretabyte and exabyte
range, i.e. billions to trillions of information from different resources. Traditional DSSs would
often not be capable of managing and analysing those nontraditional data. One of the
resolutions industries are taking up is to use Hadoop, an open source software framework for
working with various big data sets. It breaks a big data set into smaller clusters , processing
them distributedly, and then combines the results into a smaller data set that is easier to
analyse (further information can be obtained via http://www-
01.ibm.com/software/data/infosphere/hadoop/).
The challenge, however, lies in how to quickly analyse increasing volume of data (oftenloosely structured). Big data usually refers to very large data sets in the pretabyte and exabyterange, i.e. billions to trillions of information from different resources. Traditional DSSs wouldoften not be capable of managing and analysing those nontraditional data. One of theresolutions industries are taking up is to use Hadoop, an open source software framework forworking with various big data sets. It breaks a big data set into smaller clusters , processingthem distributedly, and then combines the results into a smaller data set that is easier toanalyse (further information can be obtained via http://www-01.ibm.com/software/data/infosphere/hadoop/).
การแปล กรุณารอสักครู่..
