Search results
Results From The WOW.Com Content Network
Information about this dataset's format is available in the HuggingFace dataset card and the project's website. The dataset can be downloaded here, and the rejected data here. 2016 [343] Paperno et al. FLAN A re-preprocessed version of the FLAN dataset with updates since the original FLAN dataset was released is available in Hugging Face: test data
RAWPED is a dataset for detection of pedestrians in the context of railways. The dataset is labeled box-wise. 26000 Images Object recognition and classification 2020 [70] [71] Tugce Toprak, Burak Belenlioglu, Burak Aydın, Cuneyt Guzelis, M. Alper Selver OSDaR23 OSDaR23 is a multi-sensory dataset for detection of objects in the context of railways.
huggingface.co Hugging Face is a French-American company that develops computation tools for building applications using machine learning . It is known for its transformers library built for natural language processing applications.
GPT-2 was pre-trained on a dataset of 8 million web pages. [2] It was partially released in February 2019, followed by full release of the 1.5-billion-parameter model on November 5, 2019. [3] [4] [5] GPT-2 was created as a "direct scale-up" of GPT-1 [6] with a ten-fold increase in both its parameter count and the size of its training dataset. [5]
BigScience Large Open-science Open-access Multilingual Language Model (BLOOM) [1] [2] is a 176-billion-parameter transformer-based autoregressive large language model (LLM). The model, as well as the code base and the data used to train it, are distributed under free licences. [3]
The Pile is an 886.03 GB diverse, open-source dataset of English text created as a training dataset for large language models (LLMs). It was constructed by EleutherAI in 2020 and publicly released on December 31 of that year. [1] [2] It is composed of 22 smaller datasets, including 14 new ones. [1]
The dataset consists of around 985 million words, and the books that comprise it span a range of genres, including romance, science fiction, and fantasy. [ 3 ] The corpus was introduced in a 2015 paper by researchers from the University of Toronto and MIT titled "Aligning Books and Movies: Towards Story-like Visual Explanations by Watching ...
363 billion token dataset based on Bloomberg's data sources, plus 345 billion tokens from general purpose datasets [66] Proprietary Trained on financial data from proprietary sources, for financial tasks. PanGu-Σ: March 2023: Huawei: 1085: 329 billion tokens [67] Proprietary OpenAssistant [68] March 2023: LAION: 17: 1.5 trillion tokens Apache 2.0