nlp

Tag

Cards List
#nlp

@DimitrisPapail: The co-inventor of Looped Transformers defended her PhD thesis yesterday and is heading to an incredible new role soon …

X AI KOLs Timeline · 19h ago Cached

Angeliki Giannou, co-inventor of Looped Transformers, has successfully defended her PhD thesis and is set to begin a new role. Congratulations were shared by Dimitris Papailiopoulos on social media.

0 favorites 0 likes
#nlp

Who and What? Using Linguistic Features and Annotator Characteristics to Analyze Annotation Variation

arXiv cs.CL · yesterday Cached

This paper presents a large-scale analysis of four harmful language detection datasets, examining how annotator characteristics and linguistic features interact to influence annotation variation. It highlights intersectional effects and warns against generalizing findings across different datasets.

0 favorites 0 likes
#nlp

YEZE at SemEval-2026 Task 9: Detecting Multilingual, Multicultural and Multievent Online Polarization via Heterogeneous Ensembling

arXiv cs.CL · yesterday Cached

This paper details the YEZE system for SemEval-2026 Task 9, which detects online polarization in 22 languages using a heterogeneous ensemble of XLM-RoBERTa and mDeBERTa models.

0 favorites 0 likes
#nlp

IRC-Bench: Recognizing Entities from Contextual Cues in First-Person Reminiscences

arXiv cs.CL · yesterday Cached

This paper introduces IRC-Bench, a benchmark for recognizing implicit entities in first-person reminiscences using contextual cues rather than explicit mentions. It evaluates various LLM and retrieval configurations, finding QLoRA-adapted Llama 3.1 8B to be the top performer in open-world settings.

0 favorites 0 likes
#nlp

Generating Query-Focused Summarization Datasets from Query-Free Summarization Datasets

arXiv cs.CL · yesterday Cached

This paper proposes an evidence-based model to automatically generate query keywords from query-free summarization datasets, enabling the creation of query-focused summarization datasets. Experimental results show that summaries generated using evidence-based queries achieve competitive ROUGE scores compared to original queries.

0 favorites 0 likes
#nlp

@HongcanGuo: A brand-new approach to text modeling

X AI KOLs Timeline · yesterday Cached

A researcher named HongcanGuo teases a brand-new approach to text modeling, but the tweet provides no technical details.

0 favorites 0 likes
#nlp

huggingface/transformers Release 5.8.0

GitHub Releases Watchlist · 3d ago Cached

Hugging Face has released version 5.8.0 of the Transformers library, a widely used open-source framework for natural language processing and deep learning.

0 favorites 0 likes
#nlp

Towards High-Quality Machine Translation for Kokborok: A Low-Resource Tibeto-Burman Language of Northeast India

arXiv cs.CL · 2026-04-23 Cached

Researchers develop KokborokMT, a neural MT system for the low-resource Kokborok language, achieving BLEU scores of 17.30 en→trp and 38.56 trp→en by fine-tuning NLLB-200 on a 36k-sentence parallel corpus.

0 favorites 0 likes
#nlp

huggingface/transformers Release v5.6.0

GitHub Releases Watchlist · 2026-04-22 Cached

Hugging Face released version 5.6.0 of its popular transformers library.

0 favorites 0 likes
#nlp

Product-of-Experts Training Reduces Dataset Artifacts in Natural Language Inference

arXiv cs.CL · 2026-04-22 Cached

This paper proposes Product-of-Experts (PoE) training to reduce dataset artifacts in Natural Language Inference, downweighting examples where biased models are overconfident. PoE nearly preserves accuracy on SNLI (89.10% vs. 89.30%) while reducing bias reliance by ~4.85 percentage points.

0 favorites 0 likes
#nlp

Model-Agnostic Meta Learning for Class Imbalance Adaptation

arXiv cs.CL · 2026-04-22 Cached

University of Memphis researchers propose HAMR, a model-agnostic meta-learning framework that uses bi-level optimization and neighborhood-aware resampling to adaptively reweight hard examples and minority classes across six imbalanced NLP datasets.

0 favorites 0 likes
#nlp

No One Fits All: From Fixed Prompting to Learned Routing in Multilingual LLMs

arXiv cs.CL · 2026-04-21 Cached

Researchers from National Taiwan University propose replacing fixed translation-based prompting strategies in multilingual LLMs with lightweight learned classifiers that route each instance to either native or translation-based prompting. Their analysis across 10 languages and 4 benchmarks shows no single strategy is universally optimal, with translation benefiting low-resource languages most, and the learned routing achieving statistically significant improvements over fixed strategies.

0 favorites 0 likes
#nlp

MeasHalu: Mitigation of Scientific Measurement Hallucinations for Large Language Models with Enhanced Reasoning

arXiv cs.CL · 2026-04-21 Cached

MeasHalu is a novel framework for mitigating scientific measurement hallucinations in LLMs through a two-stage reasoning-aware fine-tuning strategy and progressive reward curriculum. It introduces a fine-grained taxonomy of measurement-specific hallucinations and demonstrates improved accuracy on the MeasEval benchmark.

0 favorites 0 likes
#nlp

Incentivizing Parametric Knowledge via Reinforcement Learning with Verifiable Rewards for Cross-Cultural Entity Translation

arXiv cs.CL · 2026-04-21 Cached

Researchers from Tianjin University and Alibaba Group propose EA-RLVR, a reinforcement learning framework with verifiable rewards that improves cross-cultural entity translation in LLMs by activating parametric knowledge already encoded during pre-training, without relying on external knowledge bases. Training on 7k samples boosts Qwen3-14B's entity translation accuracy from 23.66% to 31.87% on unseen entities.

0 favorites 0 likes
#nlp

A Community-Based Approach for Stance Distribution and Argument Organization

arXiv cs.CL · 2026-04-21 Cached

Researchers from the University of British Columbia propose an unsupervised graph-based system for organizing arguments from online debates by constructing interaction graphs and applying community detection to reveal diverse viewpoint distributions. The approach requires no training data and aims to help users navigate complex argumentative landscapes and combat filter bubbles.

0 favorites 0 likes
#nlp

When Informal Text Breaks NLI: Tokenization Failure, Distribution Shift, and Targeted Mitigations

arXiv cs.CL · 2026-04-21 Cached

This paper investigates how informal text (slang, emoji, Gen-Z filler tokens) degrades NLI accuracy in ELECTRA-small and RoBERTa-large models, identifying two distinct failure mechanisms—tokenization failure (emoji mapped to [UNK]) and distribution shift (out-of-domain noise tokens)—and proposes targeted mitigations that recover accuracy without harming clean-text performance.

0 favorites 0 likes
#nlp

When Misinformation Speaks and Converses: Rethinking Fact-Checking in Audio Platforms

arXiv cs.CL · 2026-04-21 Cached

This position paper argues that audio misinformation on platforms like podcasts and WhatsApp voice notes is structurally different from text-based misinformation, carrying unique persuasive properties through prosody and conversational dynamics that existing fact-checking pipelines fail to address. The authors call for a rethinking of verification pipelines tailored to the spoken and conversational nature of audio media.

0 favorites 0 likes
#nlp

Evaluating Adaptive Personalization of Educational Readings with Simulated Learners

arXiv cs.CL · 2026-04-21 Cached

Researchers from Arizona State University present a framework for evaluating adaptive personalization of educational reading materials using theory-grounded simulated learners, incorporating memory models, misconception revision, and Bayesian Knowledge Tracing. Experiments across three subjects show adaptive reading significantly improved outcomes in computer science but had mixed results in chemistry and biology.

0 favorites 0 likes
#nlp

Detecting Alarming Student Verbal Responses using Text and Audio Classifier

arXiv cs.CL · 2026-04-21 Cached

This paper presents a hybrid framework for detecting alarming or distressed student verbal responses by combining a text classifier (content-based) and an audio classifier (prosodic features), aimed at expediting human review in Automated Verbal Response Scoring systems. The approach addresses a safety gap in automated scoring pipelines where at-risk student responses may otherwise go unnoticed.

0 favorites 0 likes
#nlp

Defragmenting Language Models: An Interpretability-based Approach for Vocabulary Expansion

arXiv cs.CL · 2026-04-21 Cached

Researchers from University of Utah and CMU propose FragMend, an interpretability-based approach for vocabulary expansion in LLMs that addresses token over-fragmentation in non-Latin script languages. Their method outperforms frequency-based vocabulary selection and baseline embedding initialization by ~20 points for several underrepresented languages.

0 favorites 0 likes
Next →
← Back to home

Submit Feedback