AI Theory
Discovering highly efficient low-weight quantum error-correcting codes with reinforcement learning
·6573 words·31 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
AI Theory
Optimization
🏢 University of Texas at Austin
RL optimizes quantum error-correcting codes, slashing physical qubit overhead for fault-tolerant quantum computing.
Why Safeguarded Ships Run Aground? Aligned Large Language Models' Safety Mechanisms Tend to Be Anchored in The Template Region
·2482 words·12 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
AI Theory
Safety
🏢 Hong Kong Polytechnic University
Aligned LLMs’ safety often anchors in the template region, creating vulnerabilities. Detaching safety mechanisms shows promise in mitigation.
Presumed Cultural Identity: How Names Shape LLM Responses
·2724 words·13 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
AI Theory
Fairness
🏢 University of Copenhagen
LLMs personalize based on user names, but this study reveals that cultural presumptions in LLM responses risk reinforcing stereotypes.
o3-mini vs DeepSeek-R1: Which One is Safer?
·578 words·3 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
AI Theory
Safety
🏢 Mondragon University
ASTRAL, a novel automated safety testing tool, reveals DeepSeek-R1’s significantly higher unsafe response rate compared to OpenAI’s o3-mini, highlighting critical safety concerns in advanced LLMs.
Early External Safety Testing of OpenAI's o3-mini: Insights from the Pre-Deployment Evaluation
·1678 words·8 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
AI Theory
Safety
🏢 Mondragon University
Researchers used ASTRAL to systematically test OpenAI’s 03-mini LLM’s safety, revealing key vulnerabilities and highlighting the need for continuous, robust safety mechanisms in large language models.
Evolution and The Knightian Blindspot of Machine Learning
·2850 words·14 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
AI Theory
Robustness
🏢 Second Nature AI
Machine learning overlooks robustness to an unknowable future; this paper contrasts reinforcement learning with biological evolution, revealing that ML’s formalisms limit engagement with unknown unkno…
Trusted Machine Learning Models Unlock Private Inference for Problems Currently Infeasible with Cryptography
·1464 words·7 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
AI Theory
Privacy
🏢 Google DeepMind
Machine learning models can enable secure computations previously impossible with cryptography, achieving privacy and efficiency in Trusted Capable Model Environments (TCMEs).
Game-theoretic LLM: Agent Workflow for Negotiation Games
·4966 words·24 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
AI Theory
Optimization
🏢 UC Santa Barbara
Game-theoretic LLMs: Agent Workflow for Negotiation Games enhances large language model (LLM) rationality in strategic decision-making through novel game-theoretic workflows.
Minimum Entropy Coupling with Bottleneck
·2581 words·13 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
AI Theory
Optimization
🏢 University of Toronto
A new lossy compression framework handles reconstruction distribution divergence by integrating a bottleneck, extending minimum entropy coupling and offering guaranteed performance.