Show newer

QA-LoRA: Quantization-Aware Low-Rank Adaptation of Large Language Models. (arXiv:2309.14717v1 [cs.LG] CROSS LISTED) 

Enhancing Keyphrase Generation by BART Finetuning with Splitting and Shuffling. (arXiv:2309.06726v1 [cs.CL] CROSS LISTED) 

Seeing and hearing what has not been said; A multimodal client behavior classifier in Motivational Interviewing with interpretable fusion. (arXiv:2309.14398v2 [cs.LG] UPDATED) 

ALLURE: Auditing and Improving LLM-based Evaluation of Text using Iterative In-Context-Learning. (arXiv:2309.13701v2 [cs.CL] UPDATED) 

SLHCat: Mapping Wikipedia Categories and Lists to DBpedia by Leveraging Semantic, Lexical, and Hierarchical Features. (arXiv:2309.11791v2 [cs.DL] UPDATED) 

Are Large Language Models Really Robust to Word-Level Perturbations?. (arXiv:2309.11166v2 [cs.CL] UPDATED) 

MBR and QE Finetuning: Training-time Distillation of the Best and Most Expensive Decoding Methods. (arXiv:2309.10966v3 [cs.CL] UPDATED) 

Estimating Contamination via Perplexity: Quantifying Memorisation in Language Model Evaluation. (arXiv:2309.10677v2 [cs.CL] UPDATED) 

Sudden Drops in the Loss: Syntax Acquisition, Phase Transitions, and Simplicity Bias in MLMs. (arXiv:2309.07311v2 [cs.CL] UPDATED) 

Cognitive Architectures for Language Agents. (arXiv:2309.02427v2 [cs.AI] UPDATED) 

Learning Speech Representation From Contrastive Token-Acoustic Pretraining. (arXiv:2309.00424v4 [eess.AS] UPDATED) 

DS4DH at #SMM4H 2023: Zero-Shot Adverse Drug Events Normalization using Sentence Transformers and Reciprocal-Rank Fusion. (arXiv:2308.12877v2 [cs.CL] UPDATED) 

Advancing Beyond Identification: Multi-bit Watermark for Large Language Models. (arXiv:2308.00221v2 [cs.CL] UPDATED) 

Robust Distortion-free Watermarks for Language Models. (arXiv:2307.15593v2 [cs.LG] UPDATED) 

Single and Multi-Speaker Cloned Voice Detection: From Perceptual to Learned Features. (arXiv:2307.07683v2 [cs.SD] UPDATED) 

Beyond Scale: the Diversity Coefficient as a Data Quality Metric Demonstrates LLMs are Pre-trained on Formally Diverse Data. (arXiv:2306.13840v2 [cs.CL] UPDATED) 

AutoTAMP: Autoregressive Task and Motion Planning with LLMs as Translators and Checkers. (arXiv:2306.06531v2 [cs.RO] UPDATED) 

Large Language Models Can Be Used to Estimate the Latent Positions of Politicians. (arXiv:2303.12057v4 [cs.CY] UPDATED) 

A Deep Learning System for Domain-specific Speech Recognition. (arXiv:2303.10510v2 [cs.CL] UPDATED) 

GPT-Neo for commonsense reasoning -- a theoretical and practical lens. (arXiv:2211.15593v2 [cs.CL] UPDATED) 

Show older
Qoto Mastodon

QOTO: Question Others to Teach Ourselves
An inclusive, Academic Freedom, instance
All cultures welcome.
Hate speech and harassment strictly forbidden.