Abstract is: A language model is a probability distribution over sequences of words. Given such a sequence of length m, a language model assigns a probability to the whole sequence. Language models generate probabilities by training on text corpora in one or many languages. Given that languages can be used to express an infinite variety of valid sentences (the property of digital infinity), language modeling faces the problem of assigning non-zero probabilities to linguistically valid sequences that may never be encountered in the training data. Several modelling approaches have been designed to surmount this problem, such as applying the Markov assumption or using neural architectures such as recurrent neural networks or transformers. Language models are useful for a variety of problems in computational linguistics; from initial applications in speech recognition to ensure nonsensical (i.e. low-probability) word sequences are not predicted, to wider use in machine translation (e.g. scoring candidate translations), natural language generation (generating more human-like text), part-of-speech tagging, parsing, Optical Character Recognition, handwriting recognition, grammar induction, information retrieval, and other applications. Language models are used in information retrieval in the query likelihood model. There, a separate language model is associated with each document in a collection. Documents are ranked based on the probability of the query Q in the document's language model : . Commonly, the unigram language model is used for this purpose.
statistical model | Q3284399 |
P1269 | facet of | natural language processing | Q30642 |
P366 | has use | synthetic media | Q96407327 |
P910 | topic's main category | Category:Language modeling | Q16500316 |
Q115936636 | AI Novelist |
Q107031872 | ALBERT |
Q108212505 | BART |
Q113013356 | BLOOM |
Q112063555 | BioBERT |
Q75720911 | CamemBERT |
Q107031978 | DeBERTa |
Q107032059 | DistilBERT |
Q107032020 | ELECTRA |
Q117861376 | FinanceGPT |
Q117449859 | InstructGPT |
Q116894231 | LLaMA |
Q113636409 | NLLB-200 |
Q114450428 | NovelAI |
Q112939824 | Pathways Language Model |
Q85124095 | RoBERTa |
Q107031912 | StructBERT |
Q108570419 | Wu Dao |
Q109615839 | XLM-RoBERTa |
Q107031747 | XLNet |
Q5428734 | factored language model |
Q115305900 | large language model |
Q24868018 | masked language model |
Q107615525 | multilingual language model |
Q123759530 | small language model |
Q124149738 | statistical language model |
Q123784954 | 1st workshop on Knowledge Base Construction from Pre-Trained Language Models (KBC-LM) |
Q125487302 | 2nd Workshop on Knowledge Base Construction from Pre-Trained Language Models |
Q125567741 | 2nd challenge on Language Models for Knowledge Base Construction |
Q113957468 | A Neural Knowledge Language Model |
Q113957091 | A Review on Language Models as Knowledge Bases |
Q37934780 | A Study on Neural Network Language Modeling |
Q54825628 | A Tri-Partite Neural Document Language Model for Semantic Information Retrieval |
Q33039506 | A neural probabilistic language model |
Q113156620 | Abolitionist Networks: Modeling Language Change in Nineteenth-Century Activist Newspapers |
Q118893726 | Activity Recommendation for Business Process Modeling with Pre-trained Language Models |
Q126911602 | An Empirical Study of Mamba-based Language Models |
Q62118614 | An Empirical Study of Smoothing Techniques for Language Modeling |
Q62118749 | An empirical study of smoothing techniques for language modeling |
Q124169015 | AnyMAL: An Efficient and Scalable Any-Modality Augmented Language Model |
Q57267388 | BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding |
Q118602000 | BRENT: Bidirectional Retrieval Enhanced Norwegian Transformer |
Q66774193 | Barack's Wife Hillary: Using Knowledge-Graphs for Fact-Aware Language Modeling |
Q112325337 | Beyond the Imitation Game: Quantifying and extrapolating the capabilities of language models |
Q72683074 | CTRL: A Conditional Transformer Language Model for Controllable Generation |
Q100695104 | Creative Storytelling with Language Models and Knowledge Graphs |
Q125468442 | Data Augmented Knowledge Graph Completion via Pre-trained Language Models |
Q112040352 | Do As I Can, Not As I Say: Grounding Language in Robotic Affordances |
Q121077898 | Do Large Language Models Know What Humans Know? |
Q113771703 | Don’t Stop Pretraining: Adapt Language Models to Domains and Tasks |
Q112063300 | Efficient Self-Supervised Metric Information Retrieval: A Bibliography Based Method Applied to COVID Literature |
Q125526131 | Enhancing Knowledge Base Construction from Pre-trained Language Models using Prompt Ensembles |
Q118869009 | Evaluating Language Models for Knowledge Base Completion |
Q29519258 | Exploring the Limits of Language Modeling |
Q120970016 | Extracting Multi-valued Relations from Language Models |
Q105405670 | Extracting Training Data from Large Language Models |
Q47488653 | Fine-tuned Language Models for Text Classification |
Q111939498 | Flamingo: a Visual Language Model for Few-Shot Learning |
Q126028910 | GPT-SW3: An Autoregressive Language Model for the Scandinavian Languages |
Q59483227 | GroupReduce: Block-Wise Low-Rank Approximation for Neural Language Model Shrinking |
Q113957463 | HTLM: Hyper-Text Pre-Training and Prompting of Language Models |
Q113531353 | Integrating Knowledge Graph embedding and pretrained Language Models in Hypercomplex Spaces |
Q109956946 | Knowledge Based Multilingual Language Model |
Q107009138 | Knowledge Graph Based Synthetic Corpus Generation for Knowledge-Enhanced Language Model Pre-training |
Q114766679 | Knowledge Prompts: Injecting World Knowledge into Language Models through Soft Prompts |
Q101086640 | KnowlyBERT - Hybrid Query Answering over Language Models and Knowledge Graphs |
Q125567785 | LM-KBC 2023: 2nd Challenge on Knowledge Base Construction from Pre-trained Language Models |
Q115265833 | LM-KBC: Knowledge Base Construction from Pre-trained Language Models |
Q101086667 | LM4KG: Improving Common Sense Knowledge Graphs with Language Models |
Q113121693 | Language Modelling with Pixels |
Q109286381 | Language Models As or For Knowledge Bases |
Q95727440 | Language Models are Few-Shot Learners |
Q105836040 | Language Models are Open Knowledge Graphs |
Q67482086 | Language Models as Knowledge Bases? |
Q64852346 | Language resources extracted from Wikipedia |
Q29180889 | Learning to Generate Reviews and Discovering Sentiment |
Q118603872 | Low-resource Bilingual Dialect Lexicon Induction with Large Language Models |
Q116297402 | Measuring Harmful Sentence Completion in Language Models for LGBTQIA+ Individuals |
Q112593974 | Memory-Based Model Editing at Scale |
Q114740676 | Mind the Labels: Describing Relations in Knowledge Graphs With Pretrained Models |
Q107764438 | Multilingual Language Models Predict Human Reading Behavior |
Q116161646 | Neural Codec Language Models are Zero-Shot Text to Speech Synthesizers |
Q111942481 | OPT: Open Pre-trained Transformer Language Models |
Q44658761 | Offline recognition of unconstrained handwritten texts using HMMs and statistical language models |
Q85691222 | On structuring probabilistic dependences in stochastic language modelling |
Q105943036 | On the Dangers of Stochastic Parrots: Can Language Models Be Too Big? 🦜 |
Q103980531 | On the Dangers of Stochastic Parrots: Can Language Models Be Too Big?🦜 |
Q124079235 | PHD: Pixel-Based Language Modeling of Historical Documents |
Q111971083 | PaLM: Scaling Language Modeling with Pathways |
Q108750211 | Playing with Words at the National Library of Sweden -- Making a Swedish BERT |
Q123258567 | Pre-trained Language Model Representations for Language Generation |
Q58959855 | Predicting reading difficulty with statistical language models |
Q118289384 | Probing Pre-Trained Language Models for Cross-Cultural Differences in Values |
Q118602378 | Probing structural constraints of negation in Pretrained Language Models |
Q115250769 | Proceedings of the Semantic Web Challenge on Knowledge Base Construction from Pre-trained Language Models 2022 |
Q115265834 | Prompting as Probing: Using Language Models for Knowledge Base Construction |
Q107553960 | QA-GNN: Reasoning with Language Models and Knowledge Graphs for Question Answering |
Q116449100 | Quantifying Memorization Across Neural Language Models |
Q125710723 | Quantifying gender bias towards politicians in cross-lingual language models |
Q84876883 | REALM: Retrieval-Augmented Language Model Pre-Training |
Q64276703 | Real-time Controlling Dynamics Sensing in Air Traffic System |
Q110929311 | Recurrent neural network based language model |
Q126009181 | Revealing the Parametric Knowledge of Language Models: A Unified Framework for Attribution Methods |
Q44651706 | SVD-Softmax: Fast Softmax Approximation on Large Vocabulary Neural Networks |
Q104433073 | SciBERT: A Pretrained Language Model for Scientific Text |
Q115250771 | Semantic Web Challenge on Knowledge Base Construction from Pre-trained Language Models 2022 |
Q99340957 | Somun: entity-centric summarization incorporating pre-trained language models |
Q29519472 | Strategies for Training Large Vocabulary Neural Language Models |
Q122931977 | Tailoring an Interpretable Neural Language Model |
Q124340251 | TinyGSM: achieving >80% on GSM8k with small language models |
Q124251384 | TinyStories: How Small Can Language Models Be and Still Speak Coherent English? |
Q125526125 | Towards Ontology Construction with Language Models |
Q113662049 | Training a T5 Using Lab-sized Resources |
Q86226450 | Transformer-XL: Attentive Language Models Beyond a Fixed-Length Context |
Q77226772 | Universal Language Model Fine-tuning for Text Classification |
Q125556494 | Utilizing Language Model Probes for Knowledge Graph Repair |
Q105430726 | Wiki-40B: Multilingual Language Model Dataset |
Q11702366 | сегментація за наміром | product or material produced or service provided | P1056 |
Q125563942 | paramètre d'un modèle de langage | facet of | P1269 |
Q34770 | language | has characteristic | P1552 |
Q16500316 | Category:Language modeling | category's main topic | P301 |
Q118640062 | Automatic Closed Captioning for Estonian Live Broadcasts | describes a project that uses | P4510 |
uri / http://www.wikidata.org/entity/L1319437-S1 | L1319437-S1 | item for this sense | P5137 |
Taalmodel | wikipedia | |
Arabic (ar / Q13955) | نموذج اللغة | wikipedia |
Моўная мадэль | wikipedia | |
Езиков модел | wikipedia | |
Catalan (ca / Q7026) | Model de llenguatge | wikipedia |
Jazykový model | wikipedia | |
Sprachmodell | wikipedia | |
Language model | wikipedia | |
Modelación del lenguaje | wikipedia | |
Keelemudel | wikipedia | |
Basque language (eu / Q8752) | Hizkuntza eredu | wikipedia |
Persian (fa / Q9168) | مدل زبانی | wikipedia |
Kielimalli | wikipedia | |
Modèle de langage | wikipedia | |
מודל שפה | wikipedia | |
hyw | Լեզուի Կաղապար | wikipedia |
言語モデル | wikipedia | |
언어 모델 | wikipedia | |
Valodas modelis | wikipedia | |
Taalmodel | wikipedia | |
Norwegian, Nynorsk (nn / Q25164) | Språkmodell | wikipedia |
Modelo de linguagem | wikipedia | |
Языковая модель | wikipedia | |
Språkmodell | wikipedia | |
Dil modeli | wikipedia | |
Модель мови | wikipedia | |
uz | Til modeli | wikipedia |
Mô hình ngôn ngữ | wikipedia | |
yue | 語言模型 | wikipedia |
語言模型 | wikipedia | |
zu | UNongo lolimi | wikipedia |
Search more.