Cookie policy: This site uses cookies (small files stored on your computer) to simplify and improve your experience of this website. Cookies are small text files stored on the device you are using to access this website. For more information on how we use and manage cookies please take a look at our privacy and cookie policies. Some parts of the site may not work properly if you choose not to accept cookies.

Home > General Atomics > Tackling the Big Data Deluge in Science with Metadata

Tackling the Big Data Deluge in Science with Metadata

White Paper Published By: General Atomics
General Atomics
Published:  Jan 13, 2015
Type:  White Paper
Length:  12 pages

The term “Big Data” has become virtually synonymous with “schema on read” (where data is applied to a plan or schema as it is ingested or pulled out of a stored location) unstructured data analysis and handling techniques like Hadoop. These “schema on read” techniques have been most famously exploited on relatively ephemeral human-readable data like retail trends, twitter sentiment, social network mining, log files, etc.

But what if you have unstructured data that, on its own, is hugely valuable, enduring, and created at great expense? Data that may not immediately be human readable or indexable on search? Exactly the kind of data most commonly created and analyzed in science and HPC. Research institutions are awash with such data from large-scale experiments and extreme-scale computing that is used for high-consequence.  

Click here to read more!

Tags : 
general atomics, big data, metadata, nirvana, hadoop, data center design and management, colocation and web hosting