When.com Web Search

Search results

  1. Results From The WOW.Com Content Network
  2. List of datasets for machine-learning research - Wikipedia

    en.wikipedia.org/wiki/List_of_datasets_for...

    Information about this dataset's format is available in the HuggingFace dataset card and the project's website. The dataset can be downloaded here, and the rejected data here. 2016 [343] Paperno et al. FLAN A re-preprocessed version of the FLAN dataset with updates since the original FLAN dataset was released is available in Hugging Face: test data

  3. The Pile (dataset) - Wikipedia

    en.wikipedia.org/wiki/The_Pile_(dataset)

    The Pile is an 886.03 GB diverse, open-source dataset of English text created as a training dataset for large language models (LLMs). It was constructed by EleutherAI in 2020 and publicly released on December 31 of that year. [1] [2] It is composed of 22 smaller datasets, including 14 new ones. [1]

  4. Hugging Face - Wikipedia

    en.wikipedia.org/wiki/Hugging_Face

    huggingface.co Hugging Face is a French-American company that develops computation tools for building applications using machine learning . It is known for its transformers library built for natural language processing applications.

  5. GPT-2 - Wikipedia

    en.wikipedia.org/wiki/GPT-2

    GPT-2 was pre-trained on a dataset of 8 million web pages. [2] It was partially released in February 2019, followed by full release of the 1.5-billion-parameter model on November 5, 2019. [3] [4] [5] GPT-2 was created as a "direct scale-up" of GPT-1 [6] with a ten-fold increase in both its parameter count and the size of its training dataset. [5]

  6. BLOOM (language model) - Wikipedia

    en.wikipedia.org/wiki/BLOOM_(language_model)

    BigScience was led by HuggingFace and involved several hundreds of researchers and engineers from France and abroad representing both the academia and the private sector. BigScience was supported by a large-scale public compute grant on the French public supercomputer Jean Zay, managed by GENCI and IDRIS ( CNRS ), on which it was trained.

  7. IBM Granite - Wikipedia

    en.wikipedia.org/wiki/IBM_Granite

    IBM Granite is a series of decoder-only AI foundation models created by IBM. [3] It was announced on September 7, 2023, [4] [5] and an initial paper was published 4 days later. [6]

  8. List of datasets in computer vision and image processing

    en.wikipedia.org/wiki/List_of_datasets_in...

    KIT AIS Data Set Multiple labeled training and evaluation datasets of aerial images of crowds. Images manually labeled to show paths of individuals through crowds. ~ 150 Images with paths People tracking, aerial tracking 2012 [158] [159] M. Butenuth et al. Wilt Dataset Remote sensing data of diseased trees and other land cover.

  9. GPT-J - Wikipedia

    en.wikipedia.org/wiki/GPT-J

    GPT-J or GPT-J-6B is an open-source large language model (LLM) developed by EleutherAI in 2021. [1] As the name suggests, it is a generative pre-trained transformer model designed to produce human-like text that continues from a prompt.