🏢 Renmin University of China
Uncovering Safety Risks of Large Language Models through Concept Activation Vector
·4605 words·22 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
🏢 Renmin University of China
Researchers developed SCAV, a novel framework to effectively reveal safety risks in LLMs by accurately interpreting their safety mechanisms. SCAV-guided attacks significantly improve attack success r…
Towards Understanding the Working Mechanism of Text-to-Image Diffusion Model
·3139 words·15 mins·
loading
·
loading
Multimodal Learning
Vision-Language Models
🏢 Renmin University of China
Stable Diffusion’s text-to-image generation is sped up by 25% by removing text guidance after the initial shape generation, revealing that the [EOS] token is key to early-stage image construction.
Towards Understanding How Transformers Learn In-context Through a Representation Learning Lens
·2618 words·13 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Renmin University of China
Transformers’ in-context learning (ICL) is explained using representation learning, revealing its ICL process as gradient descent on a dual model and offering modifiable attention layers for enhanced …
S-MolSearch: 3D Semi-supervised Contrastive Learning for Bioactive Molecule Search
·1909 words·9 mins·
loading
·
loading
Machine Learning
Semi-Supervised Learning
🏢 Renmin University of China
S-MolSearch: a novel semi-supervised framework using 3D molecular data and contrastive learning achieves state-of-the-art in bioactive molecule search, outperforming existing methods.
Intruding with Words: Towards Understanding Graph Injection Attacks at the Text Level
·5345 words·26 mins·
loading
·
loading
AI Theory
Robustness
🏢 Renmin University of China
Researchers unveil text-level graph injection attacks, revealing a new vulnerability in GNNs and highlighting the importance of text interpretability in attack success.
Enhancing In-Context Learning Performance with just SVD-Based Weight Pruning: A Theoretical Perspective
·2209 words·11 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Renmin University of China
SVD-based weight pruning surprisingly boosts in-context learning in large language models, especially when applied to deeper layers, offering a novel approach to model compression and efficiency.
Conjugate Bayesian Two-step Change Point Detection for Hawkes Process
·1942 words·10 mins·
loading
·
loading
AI Applications
Security
🏢 Renmin University of China
A novel conjugate Bayesian two-step change point detection method for Hawkes processes, CoBay-CPD, achieves higher accuracy and efficiency by employing data augmentation for improved dynamic event mod…
CausalStock: Deep End-to-end Causal Discovery for News-driven Multi-stock Movement Prediction
·1729 words·9 mins·
loading
·
loading
AI Applications
Finance
🏢 Renmin University of China
CausalStock: A novel framework for accurate news-driven multi-stock movement prediction, using lag-dependent causal discovery and LLMs for enhanced noise reduction and explainability.
Bridging The Gap between Low-rank and Orthogonal Adaptation via Householder Reflection Adaptation
·2524 words·12 mins·
loading
·
loading
Large Language Models
🏢 Renmin University of China
Householder Reflection Adaptation (HRA) bridges low-rank and orthogonal LLM adaptation, achieving superior performance with fewer parameters than existing methods. By using a chain of Householder refl…