TheoremQA: A Theorem-driven Question Answering dataset. (arXiv:2305.12524v3 [cs.CL] UPDATED)
HalOmi: A Manually Annotated Benchmark for Multilingual Hallucination and Omission Detection in Machine Translation. (arXiv:2305.11746v2 [cs.CL] UPDATED)
Completeness, Recall, and Negation in Open-World Knowledge Bases: A Survey. (arXiv:2305.05403v2 [cs.AI] UPDATED)
Exploring Distributional Shifts in Large Language Models for Code Analysis. (arXiv:2303.09128v2 [cs.CL] UPDATED)
Model-tuning Via Prompts Makes NLP Models Adversarially Robust. (arXiv:2303.07320v2 [cs.CL] UPDATED)
BiasTestGPT: Using ChatGPT for Social Bias Testing of Language Models. (arXiv:2302.07371v3 [cs.CL] UPDATED)
Continual Learning for Instruction Following from Realtime Feedback. (arXiv:2212.09710v2 [cs.CL] UPDATED)
Entailment Semantics Can Be Extracted from an Ideal Language Model. (arXiv:2209.12407v2 [cs.CL] UPDATED)
A Comprehensive Review of Visual-Textual Sentiment Analysis from Social Media Networks. (arXiv:2207.02160v2 [cs.CL] UPDATED)
TraSE: Towards Tackling Authorial Style from a Cognitive Science Perspective. (arXiv:2206.10706v2 [cs.CL] UPDATED)
All the World's a (Hyper)Graph: A Data Drama. (arXiv:2206.08225v3 [cs.LG] UPDATED)
OneLLM: One Framework to Align All Modalities with Language. (arXiv:2312.03700v1 [cs.CV])
PROMISE: A Framework for Model-Driven Stateful Prompt Orchestration. (arXiv:2312.03699v1 [cs.CL])
Evaluating and Mitigating Discrimination in Language Model Decisions. (arXiv:2312.03689v1 [cs.CL])
An Integration of Pre-Trained Speech and Language Models for End-to-End Speech Recognition. (arXiv:2312.03668v1 [eess.AS])
Generative agent-based modeling with actions grounded in physical, social, or digital space using Concordia. (arXiv:2312.03664v1 [cs.AI])
Interpretability Illusions in the Generalization of Simplified Models. (arXiv:2312.03656v1 [cs.LG])
Not All Large Language Models (LLMs) Succumb to the "Reversal Curse": A Comparative Study of Deductive Logical Reasoning in BERT and GPT Models. (arXiv:2312.03633v1 [cs.CL])
Improving Bias Mitigation through Bias Experts in Natural Language Understanding. (arXiv:2312.03577v1 [cs.CL])
XAIQA: Explainer-Based Data Augmentation for Extractive Question Answering. (arXiv:2312.03567v1 [cs.CL])
All recent Computation and Language articles on arXiv.org for the Fediverse
Inspired by https://twitter.com/arxiv_cscl