Show newer

Several categories of Large Language Models (LLMs): A Short Survey. (arXiv:2307.10188v1 [cs.CL]) 

A comparative analysis of SRGAN models. (arXiv:2307.09456v2 [cs.CV] UPDATED) 

Pseudo Outlier Exposure for Out-of-Distribution Detection using Pretrained Transformers. (arXiv:2307.09455v2 [cs.CL] UPDATED) 

Let's ViCE! Mimicking Human Cognitive Behavior in Image Generation Evaluation. (arXiv:2307.09416v2 [cs.CV] UPDATED) 

Llama 2: Open Foundation and Fine-Tuned Chat Models. (arXiv:2307.09288v2 [cs.CL] UPDATED) 

Retentive Network: A Successor to Transformer for Large Language Models. (arXiv:2307.08621v2 [cs.CL] UPDATED) 

Distilling Large Vision-Language Model with Out-of-Distribution Generalizability. (arXiv:2307.03135v2 [cs.CV] UPDATED) 

LongNet: Scaling Transformers to 1,000,000,000 Tokens. (arXiv:2307.02486v2 [cs.CL] UPDATED) 

Iterated Piecewise Affine (IPA) Approximation for Language Modeling. (arXiv:2306.12317v2 [cs.CL] UPDATED) 

GEmo-CLAP: Gender-Attribute-Enhanced Contrastive Language-Audio Pretraining for Speech Emotion Recognition. (arXiv:2306.07848v5 [cs.CL] UPDATED) 

ChatGPT Outperforms Crowd-Workers for Text-Annotation Tasks. (arXiv:2303.15056v2 [cs.CL] UPDATED) 

Understand Legal Documents with Contextualized Large Language Models. (arXiv:2303.12135v4 [cs.CL] UPDATED) 

ThoughtSource: A central hub for large language model reasoning data. (arXiv:2301.11596v3 [cs.CL] UPDATED) 

Lego-MT: Learning Detachable Models for Massively Multilingual Machine Translation. (arXiv:2212.10551v3 [cs.CL] UPDATED) 

Can In-context Learners Learn a Reasoning Concept from Demonstrations?. (arXiv:2212.01692v4 [cs.CL] UPDATED) 

Revisiting Softmax for Uncertainty Approximation in Text Classification. (arXiv:2210.14037v2 [cs.LG] UPDATED) 

DialogStudio: Towards Richest and Most Diverse Unified Dataset Collection for Conversational AI. (arXiv:2307.10172v1 [cs.CL]) 

Challenges and Applications of Large Language Models. (arXiv:2307.10169v1 [cs.CL]) 

LLMs as Workers in Human-Computational Algorithms? Replicating Crowdsourcing Pipelines with LLMs. (arXiv:2307.10168v1 [cs.CL]) 

Exploring Transformer Extrapolation. (arXiv:2307.10156v1 [cs.CL]) 

Show older
Qoto Mastodon

QOTO: Question Others to Teach Ourselves
An inclusive, Academic Freedom, instance
All cultures welcome.
Hate speech and harassment strictly forbidden.