When.com Web Search

Search results

  1. Results From The WOW.Com Content Network
  2. Apache Hadoop - Wikipedia

    en.wikipedia.org/wiki/Apache_Hadoop

    Apache Hadoop (/ h ə ˈ d uː p /) is a collection of open-source software utilities for reliable, scalable, distributed computing.It provides a software framework for distributed storage and processing of big data using the MapReduce programming model.

  3. File:Log analysis using Splunk Hadoop Connect (IA ...

    en.wikipedia.org/wiki/File:Log_analysis_using_S...

    Main page; Contents; Current events; Random article; About Wikipedia; Contact us

  4. File:Hadoop-Hdfs.pdf - Wikipedia

    en.wikipedia.org/wiki/File:Hadoop-Hdfs.pdf

    Original file (1,666 × 1,250 pixels, file size: 133 KB, MIME type: application/pdf, 15 pages) This is a file from the Wikimedia Commons . Information from its description page there is shown below.

  5. Apache Hive - Wikipedia

    en.wikipedia.org/wiki/Apache_Hive

    Apache Hive is a data warehouse software project. It is built on top of Apache Hadoop for providing data query and analysis. [3] [4] Hive gives an SQL-like interface to query data stored in various databases and file systems that integrate with Hadoop.

  6. Mike Cafarella - Wikipedia

    en.wikipedia.org/wiki/Mike_Cafarella

    Mike Cafarella is a computer scientist specializing in database management systems.He is a principal research scientist of computer science at MIT Computer Science and Artificial Intelligence Laboratory. [1]

  7. Apache Pig - Wikipedia

    en.wikipedia.org/wiki/Apache_Pig

    Apache Pig [1] is a high-level platform for creating programs that run on Apache Hadoop. The language for this platform is called Pig Latin. [1] Pig can execute its Hadoop jobs in MapReduce, Apache Tez, or Apache Spark. [2]

  8. Apache Parquet - Wikipedia

    en.wikipedia.org/wiki/Apache_Parquet

    Apache Parquet is a free and open-source column-oriented data storage format in the Apache Hadoop ecosystem. It is similar to RCFile and ORC, the other columnar-storage file formats in Hadoop, and is compatible with most of the data processing frameworks around Hadoop.

  9. Doug Cutting - Wikipedia

    en.wikipedia.org/wiki/Doug_Cutting

    Cutting and Mike Cafarella, realizing the importance of this paper to extending Lucene into the realm of extremely large search problems, created the open-source Hadoop framework. This framework allows applications based on the MapReduce paradigm to be run on large clusters of commodity hardware.