Natural Language Processing
CulturePark: Boosting Cross-cultural Understanding in Large Language Models
·2738 words·13 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Microsoft Research
CulturePark, a novel multi-agent communication framework, generates high-quality cross-cultural data to fine-tune LLMs, significantly reducing cultural bias and boosting cross-cultural understanding.
CultureLLM: Incorporating Cultural Differences into Large Language Models
·2507 words·12 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Microsoft Research
CultureLLM, a new approach, effectively incorporates cultural nuances into LLMs using semantic data augmentation, significantly outperforming existing models.
Cross-model Control: Improving Multiple Large Language Models in One-time Training
·1811 words·9 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 East China Normal University
One-time training improves multiple LLMs using a tiny portable model, drastically reducing costs and resource needs for model enhancement.
CriticEval: Evaluating Large-scale Language Model as Critic
·4755 words·23 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Beijing Institute of Technology
CRITICEVAL: A new benchmark reliably evaluates LLMs’ ability to identify and correct flaws in their responses, addressing limitations of existing methods by offering comprehensive and reliable evaluat…
Crafting Interpretable Embeddings for Language Neuroscience by Asking LLMs Questions
·1981 words·10 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 UC Berkeley
LLM-based text embeddings are powerful but lack interpretability. This paper introduces QA-Emb, a novel method that uses an LLM to answer yes/no questions about a text, thereby producing an interpreta…
CoVoMix: Advancing Zero-Shot Speech Generation for Human-like Multi-talker Conversations
·2583 words·13 mins·
loading
·
loading
Natural Language Processing
Dialogue Systems
🏢 Shanghai Jiao Tong University
CoVoMix: Generating human-like, multi-speaker conversations with zero-shot speech synthesis.
Cost-efficient Knowledge-based Question Answering with Large Language Models
·1874 words·9 mins·
loading
·
loading
AI Generated
Natural Language Processing
Question Answering
🏢 Hong Kong Polytechnic University
Coke: A cost-efficient KBQA strategy using LLMs and KGMs, maximizing accuracy while minimizing GPT-4 fees by up to 20.89%
CorDA: Context-Oriented Decomposition Adaptation of Large Language Models for Task-Aware Parameter-Efficient Fine-tuning
·2973 words·14 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
🏢 King Abdullah University of Science and Technology
CorDA: Context-oriented weight decomposition enhances large language model fine-tuning by integrating task context, improving efficiency and mitigating catastrophic forgetting.
Cooperate or Collapse: Emergence of Sustainable Cooperation in a Society of LLM Agents
·6245 words·30 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 ETH Zurich
LLMs struggle to cooperate sustainably; GOVSIM reveals this, showing communication and ‘universalization’ reasoning improve outcomes.
Continual Learning with Global Alignment
·1784 words·9 mins·
loading
·
loading
Natural Language Processing
Text Classification
🏢 Stony Brook University
Researchers developed a novel continual learning method achieving state-of-the-art performance by aligning data representations across tasks using pre-trained tokens, eliminating the need for experien…
ContextCite: Attributing Model Generation to Context
·7666 words·36 mins·
loading
·
loading
AI Generated
Natural Language Processing
Text Generation
🏢 MIT
CONTEXTCITE pinpoints which parts of a given context led a language model to generate a specific statement, improving model verification and response quality.
Construction and Application of Materials Knowledge Graph in Multidisciplinary Materials Science via Large Language Model
·1941 words·10 mins·
loading
·
loading
AI Generated
Natural Language Processing
Information Extraction
🏢 School of Computer Science and Engineering, University of New South Wales
Researchers created a Materials Knowledge Graph (MKG) using large language models to efficiently organize and integrate knowledge from a decade of high-quality materials science research, enhancing da…
ConStat: Performance-Based Contamination Detection in Large Language Models
·4433 words·21 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
🏢 ETH Zurich
ConStat: Exposing hidden LLM contamination!
Connecting the Dots: LLMs can Infer and Verbalize Latent Structure from Disparate Training Data
·13838 words·65 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
🏢 UC Berkeley
LLMs surprisingly infer censored knowledge from implicit training data hints, posing safety challenges.
Conformal Alignment: Knowing When to Trust Foundation Models with Guarantees
·3577 words·17 mins·
loading
·
loading
Natural Language Processing
Question Answering
🏢 Department of Statistics, University of Chicago
Conformal Alignment certifies trustworthy foundation model outputs by guaranteeing a user-specified fraction meet alignment criteria, regardless of the model or data.
Confidence Regulation Neurons in Language Models
·3393 words·16 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
🏢 ETH Zurich
LLMs regulate uncertainty via specialized ’entropy’ and ’token frequency’ neurons, impacting prediction confidence without directly altering logits.
Concentrate Attention: Towards Domain-Generalizable Prompt Optimization for Language Models
·3084 words·15 mins·
loading
·
loading
Natural Language Processing
Text Classification
🏢 Xi'an Jiaotong University
Boost language model performance across domains with ‘Concentration’: a new prompt optimization objective that prioritizes stable, deep-layer attention.
Compressing Large Language Models using Low Rank and Low Precision Decomposition
·2393 words·12 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
🏢 Stanford University
CALDERA: a new post-training LLM compression algorithm achieving state-of-the-art zero-shot performance using low-rank, low-precision decomposition.
Compositional 3D-aware Video Generation with LLM Director
·2894 words·14 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Microsoft Research
LLM-directed compositional 3D-aware video generation (C3V) achieves high-fidelity video generation with diverse motion and flexible concept control by decomposing prompts, generating 3D concepts, and …
Compact Language Models via Pruning and Knowledge Distillation
·4214 words·20 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
🏢 NVIDIA
MINITRON: Efficiently creating smaller, high-performing LLMs via pruning & distillation, slashing training costs by up to 40x!