Beyond Surface: Probing LLaMA Across Scales and Layers. (arXiv:2312.04333v1 [cs.CL])
nerblackbox: A High-level Library for Named Entity Recognition in Python. (arXiv:2312.04306v1 [cs.CL])
Prompt Highlighter: Interactive Control for Multi-Modal LLMs. (arXiv:2312.04302v1 [cs.CV])
PsyChat: A Client-Centric Dialogue System for Mental Health Support. (arXiv:2312.04262v1 [cs.CL])
Swap distance minimization in SOV languages. Cognitive and mathematical foundations. (arXiv:2312.04219v1 [cs.CL])
Language Model Knowledge Distillation for Efficient Question Answering in Spanish. (arXiv:2312.04193v1 [cs.CL])
Using a Large Language Model to generate a Design Structure Matrix. (arXiv:2312.04134v1 [cs.AI])
Analyzing the Inherent Response Tendency of LLMs: Real-World Instructions-Driven Jailbreak. (arXiv:2312.04127v1 [cs.CL])
Enhancing the Rationale-Input Alignment for Self-explaining Rationalization. (arXiv:2312.04103v1 [cs.AI])
Making Translators Privacy-aware on the User's Side. (arXiv:2312.04068v1 [cs.CR])
Comparing Large Language Model AI and Human-Generated Coaching Messages for Behavioral Weight Loss. (arXiv:2312.04059v1 [cs.CL])
Multimodal Misinformation Detection in a South African Social Media Environment. (arXiv:2312.04052v1 [cs.CL])
RoAST: Robustifying Language Models via Adversarial Perturbation with Selective Training. (arXiv:2312.04032v1 [cs.CL])
A Study on the Calibration of In-context Learning. (arXiv:2312.04021v1 [cs.CL])
Cost-Effective In-Context Learning for Entity Resolution: A Design Space Exploration. (arXiv:2312.03987v1 [cs.CL])
Collaboration or Corporate Capture? Quantifying NLP's Reliance on Industry Artifacts and Contributions. (arXiv:2312.03912v1 [cs.CL])
A Pseudo-Semantic Loss for Autoregressive Models with Logical Constraints. (arXiv:2312.03905v1 [cs.LG])
Revisiting the Optimality of Word Lengths. (arXiv:2312.03897v1 [cs.CL])
The BigCode Project Governance Card. (arXiv:2312.03872v1 [cs.CY])
Efficient Large Language Models: A Survey. (arXiv:2312.03863v1 [cs.CL])
All recent Computation and Language articles on arXiv.org for the Fediverse
Inspired by https://twitter.com/arxiv_cscl