资讯

Quantcast, an internet audience measurement and ad targeting service, processes over 20 petabytes of data per day using Apache Hadoop and its own custom file system called Quantcast File System (QFS).
In this whitepaper, Yahoo engineers Konstantin Shvachko, Hairong Kuang, Sanjay Radia, and Robert Chansle look at HDFS, the file system component of Hadoop. While the interface to HDFS is patterned ...
Cloud computing is a new technology which comes from distributed computing, parallel computing, grid computing and other computing technologies. In cloud computing, the data storage and computing are ...
The rapid growth of data and the changing nature of data applications is challenging established architectural concepts for how to store big data. Where once organizations may have first looked to ...
The Hadoop Distributed File System (HDFS) is considered a core component of Hadoop, but it’s not an essential one. Lately, IBM has been talking up the benefits of hooking Hadoop up to the General ...
Big data can mean big threats to security, thanks to the tempting volumes of information that may sit waiting for hackers to peruse. BlueTalon hopes to tackle that problem with what it calls the first ...
This paper provides a high-level overview of how Apache Cassandra™ can be used to replace HDFS, with no programming changes required from a developer perspective, and how a number of compelling ...
Seagate this week introduced a ClusterStor Hadoop Workflow Accelerator that introduces the Hadoop on Lustre Connector, which allows clusters based on Hadoop and the open source Lustre file system “to ...