Current research in natural language processing is increasingly focused on enhancing the efficiency and effectiveness of language models, particularly in specialized applications like word sense disambiguation and critical thinking. Recent work demonstrates that low-parameter models can rival high-parameter counterparts in tasks such as disambiguating rare terms and understanding complex arguments, reducing computational costs significantly. Additionally, innovations like multilingual reference assessment systems for Wikipedia aim to streamline content verification, addressing the labor-intensive nature of manual editing. The exploration of personalized debunking strategies using personality traits highlights a growing interest in tailoring communication for better engagement. Furthermore, advancements in masked diffusion models and long-context encoders are pushing the boundaries of how language models can process and generate text, particularly in low-resource languages and nuanced contexts. These developments suggest a shift toward more accessible, efficient, and context-aware NLP solutions, with potential applications in content moderation, education, and public health monitoring.
Word Sense Disambiguation (WSD) remains a key challenge in Natural Language Processing (NLP), especially when dealing with rare or domain-specific senses that are often misinterpreted. While modern hi...
To think critically about arguments, human learners are trained to identify, reconstruct, and evaluate arguments. Argument reconstruction is especially important because it makes an argument's underly...
Wikipedia is a critical source of information for millions of users across the Web. It serves as a key resource for large language models, search engines, question-answering systems, and other Web-bas...
Masked diffusion models (MDM) exhibit superior generalization when learned using a Partial masking scheme (Prime). This approach converts tokens into sub-tokens and models the diffusion process at the...
Word sense plausibility rating requires predicting the human-perceived plausibility of a given word sense on a 1--5 scale in the context of short narrative stories containing ambiguous homonyms. This ...
Social media text shows promise for monitoring trends in the opioid overdose crisis; however, the overwhelming majority of social media text is unrelated to opioids. When leveraging social media text ...
Interlinear glossed text (IGT) is a standard notation for language documentation which is linguistically rich but laborious to produce manually. Recent automated IGT methods treat glosses as character...
Large language models (LLMs) can produce chains of thought (CoT) that do not accurately reflect the actual factors driving their answers. In multiple-choice settings with an injected hint favoring a p...
While decoder-only Large Language Models (LLMs) have recently dominated the NLP landscape, encoder-only architectures remain a cost-effective and parameter-efficient standard for discriminative tasks....
Large language models are known to often exhibit inconsistent knowledge. This is particularly problematic in multilingual scenarios, where models are likely to be asked similar questions in different ...