When.com Web Search

Search results

  1. Results From The WOW.Com Content Network
  2. List of large language models - Wikipedia

    en.wikipedia.org/wiki/List_of_large_language_models

    A large language model (LLM) is a type of machine learning model designed for natural language processing tasks such as language generation. LLMs are language models with many parameters, and are trained with self-supervised learning on a vast amount of text. This page lists notable large language models.

  3. Large language model - Wikipedia

    en.wikipedia.org/wiki/Large_language_model

    A large language model (LLM) is a type of machine learning model designed for natural language processing tasks such as language generation. LLMs are language models with many parameters, and are trained with self-supervised learning on a vast amount of text. The largest and most capable LLMs are generative pretrained transformers (GPTs).

  4. T5 (language model) - Wikipedia

    en.wikipedia.org/wiki/T5_(language_model)

    T5 (Text-to-Text Transfer Transformer) is a series of large language models developed by Google AI introduced in 2019. [1] [2] Like the original Transformer model, [3] T5 models are encoder-decoder Transformers, where the encoder processes the input text, and the decoder generates the output text.

  5. Principles and parameters - Wikipedia

    en.wikipedia.org/wiki/Principles_and_parameters

    Principles and parameters is a framework within generative linguistics in which the syntax of a natural language is described in accordance with general principles (i.e. abstract rules or grammars) and specific parameters (i.e. markers, switches) that for particular languages are either turned on or off.

  6. Llama (language model) - Wikipedia

    en.wikipedia.org/wiki/Llama_(language_model)

    Llama (Large Language Model Meta AI, formerly stylized as LLaMA) is a family of large language models (LLMs) released by Meta AI starting in February 2023. [2] [3] The latest version is Llama 3.3, released in December 2024. [4] Llama models are trained at different parameter sizes, ranging between 1B and 405B. [5]

  7. Generative pre-trained transformer - Wikipedia

    en.wikipedia.org/wiki/Generative_pre-trained...

    The semi-supervised approach OpenAI employed to make a large-scale generative system—and was first to do with a transformer model—involved two stages: an unsupervised generative "pretraining" stage to set initial parameters using a language modeling objective, and a supervised discriminative "fine-tuning" stage to adapt these parameters to ...

  8. BERT (language model) - Wikipedia

    en.wikipedia.org/wiki/BERT_(language_model)

    It is notable for its dramatic improvement over previous state-of-the-art models, and as an early example of a large language model. As of 2020, BERT is a ubiquitous baseline in natural language processing (NLP) experiments. [3] BERT is trained by masked token prediction and next sentence prediction.

  9. Minimalist program - Wikipedia

    en.wikipedia.org/wiki/Minimalist_program

    The language faculty consists of a computational system (C HL) whose initial state (S 0) contains invariant principles and parameters. Language acquisition consists of acquiring a lexicon and fixing the parameter values of the target language. Language generates an infinite set of expressions given as a sound-meaning pair (π, λ).