Natural Language Processing
Leveraging Environment Interaction for Automated PDDL Translation and Planning with Large Language Models
·1918 words·10 mins·
loading
·
loading
Natural Language Processing
Large Language Models
π’ University of British Columbia
This paper presents a fully automated method for PDDL translation and planning using LLMs and environment interaction, achieving a 66% success rate on challenging PDDL domains.
LESS: Label-Efficient and Single-Stage Referring 3D Segmentation
·2019 words·10 mins·
loading
·
loading
Natural Language Processing
Vision-Language Models
π’ College of Computer Science and Software Engineering, Shenzhen University
LESS achieves state-of-the-art Referring 3D Segmentation using only binary masks, significantly reducing labeling effort and improving efficiency with a novel single-stage pipeline.
LeDex: Training LLMs to Better Self-Debug and Explain Code
·3820 words·18 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
π’ Purdue University
LEDEX: A novel training framework significantly boosts LLMs’ code self-debugging by using automated data collection, supervised fine-tuning, and reinforcement learning, leading to more accurate code a…
Learning to Reason via Program Generation, Emulation, and Search
·1757 words·9 mins·
loading
·
loading
Natural Language Processing
Large Language Models
π’ Johns Hopkins University
Language models excel at generating programs for algorithmic tasks, but struggle with soft reasoning. COGEX leverages pseudo-programs and program emulation to tackle these tasks, while COTACS searches…
Learning to Discuss Strategically: A Case Study on One Night Ultimate Werewolf
·2358 words·12 mins·
loading
·
loading
Natural Language Processing
Large Language Models
π’ Institute of Automation, Chinese Academy of Sciences
RL-instructed language models excel at strategic communication in One Night Ultimate Werewolf, demonstrating the importance of discussion tactics in complex games.
Learning Goal-Conditioned Representations for Language Reward Models
·3372 words·16 mins·
loading
·
loading
Natural Language Processing
Large Language Models
π’ Scale AI
Goal-conditioned contrastive learning boosts language reward model performance and enables better control of language model generation.
Learning Complete Protein Representation by Dynamically Coupling of Sequence and Structure
·2792 words·14 mins·
loading
·
loading
AI Generated
Natural Language Processing
Representation Learning
π’ Zhejiang University
CoupleNet dynamically links protein sequences and structures for improved representations, surpassing state-of-the-art methods in function prediction, particularly for uncommon proteins.
Learning and Transferring Sparse Contextual Bigrams with Linear Transformers
·1445 words·7 mins·
loading
·
loading
Natural Language Processing
Text Generation
π’ Princeton University
Linear transformers efficiently learn sparse contextual bigrams by leveraging both in-context and global information, achieving polynomial sample complexity.
Learnability Matters: Active Learning for Video Captioning
·2406 words·12 mins·
loading
·
loading
Natural Language Processing
Text Generation
π’ Hangzhou Dianzi University
Active learning for video captioning is enhanced by a novel algorithm that prioritizes ’learnability’, diversity, and uncertainty to address annotation inconsistency.
Learn more, but bother less: parameter efficient continual learning
·2442 words·12 mins·
loading
·
loading
Natural Language Processing
Large Language Models
π’ Pennsylvania State University
LB-CL: A novel parameter-efficient continual learning method for LLMs that boosts performance and reduces forgetting by leveraging parametric knowledge transfer and maintaining orthogonal low-rank sub…
LCGen: Mining in Low-Certainty Generation for View-consistent Text-to-3D
·2307 words·11 mins·
loading
·
loading
Natural Language Processing
Text Generation
π’ Shanghai Engineering Research Center of AI & Robotics, Academy for Engineering & Technology, Fudan University
LCGen: A novel method for view-consistent text-to-3D generation, resolving the ‘Janus Problem’ by strategically using low-certainty priors to align viewpoints and optimize the generation process.
Latent Paraphrasing: Perturbation on Layers Improves Knowledge Injection in Language Models
·2300 words·11 mins·
loading
·
loading
Natural Language Processing
Large Language Models
π’ KRAFTON
LaPael improves LLM knowledge injection by applying learned noise to early layers, enabling diverse and efficient knowledge updates without repeated external model usage.
Large Language Models-guided Dynamic Adaptation for Temporal Knowledge Graph Reasoning
·2160 words·11 mins·
loading
·
loading
Natural Language Processing
Large Language Models
π’ Beijing University of Technology
LLM-DA dynamically adapts LLM-generated rules for accurate, interpretable temporal knowledge graph reasoning, significantly improving accuracy without fine-tuning.
Large Language Models Must Be Taught to Know What They Donβt Know
·3020 words·15 mins·
loading
·
loading
Natural Language Processing
Large Language Models
π’ New York University
Teach LLMs uncertainty for reliable high-stakes predictions: Fine-tuning with graded examples significantly improves LLM’s uncertainty calibration and generalizes well.
Large language model validity via enhanced conformal prediction methods
·2089 words·10 mins·
loading
·
loading
Natural Language Processing
Large Language Models
π’ Stanford University
New conformal inference methods enhance LLM validity by providing adaptive validity guarantees and improving the quality of LLM outputs, addressing prior methods’ limitations.
Large Language Model Unlearning via Embedding-Corrupted Prompts
·7618 words·36 mins·
loading
·
loading
Natural Language Processing
Large Language Models
π’ UC Santa Cruz
ECO prompts enable efficient LLM unlearning by corrupting prompts flagged for forgetting, achieving promising results across various LLMs and tasks with minimal side effects.
Large Language Model Unlearning
·6002 words·29 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
π’ Meta GenAI
This paper presents a novel method for large language model (LLM) unlearning, enabling LLMs to ‘forget’ undesirable behaviors by using only negative examples. This computationally efficient approach o…
Language Models as Zero-shot Lossless Gradient Compressors: Towards General Neural Parameter Prior Models
·2064 words·10 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
π’ CISPA Helmholtz Center for Information Security
Large language models (LLMs) achieve lossless gradient compression, surpassing existing methods by up to 17.2%, thereby advancing distributed learning efficiency.
Language Models as Hierarchy Encoders
·2232 words·11 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
π’ University of Oxford
Language models struggle with hierarchical information. This work introduces Hierarchy Transformer Encoders (HITs), a novel method to retrain transformer encoders using hyperbolic geometry and special…
Language Grounded Multi-agent Reinforcement Learning with Human-interpretable Communication
·2019 words·10 mins·
loading
·
loading
Natural Language Processing
Human-AI Interaction
π’ University of Pittsburgh
LangGround: MARL agents learn human-interpretable communication via LLM-grounded training, enabling effective human-agent collaboration.