🏢 University of Oxford
What Makes and Breaks Safety Fine-tuning? A Mechanistic Study
·10141 words·48 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 University of Oxford
Safety fine-tuning for LLMs is shown to minimally transform weights, clustering inputs based on safety, but is easily bypassed by adversarial attacks.
Unsupervised Object Detection with Theoretical Guarantees
·2140 words·11 mins·
loading
·
loading
Computer Vision
Object Detection
🏢 University of Oxford
First unsupervised object detection method with theoretical guarantees to recover true object positions, up to quantifiable small shifts!
Universal In-Context Approximation By Prompting Fully Recurrent Models
·3295 words·16 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
🏢 University of Oxford
Fully recurrent neural networks can be universal in-context approximators, achieving the same capabilities as transformer models by cleverly using prompts.
Transformers need glasses! Information over-squashing in language tasks
·3003 words·15 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
🏢 University of Oxford
Large language models (LLMs) suffer from information loss due to representational collapse and over-squashing, causing failures in simple tasks; this paper provides theoretical analysis and practical …
Towards Learning Group-Equivariant Features for Domain Adaptive 3D Detection
·1931 words·10 mins·
loading
·
loading
Computer Vision
3D Vision
🏢 University of Oxford
GroupEXP-DA boosts domain adaptive 3D object detection by using a grouping-exploration strategy to reduce bias in pseudo-label collection and account for multiple factors affecting object perception i…
The Edge-of-Reach Problem in Offline Model-Based Reinforcement Learning
·2452 words·12 mins·
loading
·
loading
Machine Learning
Reinforcement Learning
🏢 University of Oxford
Offline model-based RL methods fail as dynamics models improve; this paper reveals the ’edge-of-reach’ problem causing this and introduces RAVL, a simple solution ensuring robust performance.
Stability and Generalizability in SDE Diffusion Models with Measure-Preserving Dynamics
·2499 words·12 mins·
loading
·
loading
Computer Vision
Image Generation
🏢 University of Oxford
D³GM, a novel score-based diffusion model, enhances stability & generalizability in solving inverse problems by leveraging measure-preserving dynamics, enabling robust image reconstruction across dive…
SpatialPIN: Enhancing Spatial Reasoning Capabilities of Vision-Language Models through Prompting and Interacting 3D Priors
·2401 words·12 mins·
loading
·
loading
Multimodal Learning
Vision-Language Models
🏢 University of Oxford
SpatialPIN boosts vision-language models’ spatial reasoning by cleverly combining prompting techniques with 3D foundation models, achieving zero-shot performance on various spatial tasks.
Set-based Neural Network Encoding Without Weight Tying
·5047 words·24 mins·
loading
·
loading
AI Generated
Machine Learning
Deep Learning
🏢 University of Oxford
Set-based Neural Network Encoder (SNE) efficiently encodes neural network weights for property prediction, eliminating the need for architecture-specific models and improving generalization across dat…
Sequence-Augmented SE(3)-Flow Matching For Conditional Protein Generation
·4573 words·22 mins·
loading
·
loading
AI Generated
AI Applications
Healthcare
🏢 University of Oxford
Sequence-augmented SE(3)-Flow model, FOLDFLOW-2, excels at generating diverse, designable protein structures, surpassing existing methods in unconditional and conditional design tasks.
Separations in the Representational Capabilities of Transformers and Recurrent Architectures
·1587 words·8 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
🏢 University of Oxford
Transformers and RNNs show contrasting representational capabilities: Transformers excel at tasks requiring associative recall, while RNNs are better suited for hierarchical language processing. This …
Score-Optimal Diffusion Schedules
·2200 words·11 mins·
loading
·
loading
Machine Learning
Deep Learning
🏢 University of Oxford
Researchers developed a novel algorithm to automatically find optimal schedules for denoising diffusion models (DDMs), significantly improving sample quality and efficiency without manual parameter tu…
Rough Transformers: Lightweight Continuous-Time Sequence Modelling with Path Signatures
·4237 words·20 mins·
loading
·
loading
AI Generated
Machine Learning
Deep Learning
🏢 University of Oxford
Rough Transformers: A lightweight continuous-time sequence modeling approach using path signatures to significantly reduce computational costs, improving efficiency and accuracy, particularly for long…
Random Representations Outperform Online Continually Learned Representations
·1894 words·9 mins·
loading
·
loading
Machine Learning
Representation Learning
🏢 University of Oxford
Random pixel projections outperform complex online continual learning methods for image classification, challenging assumptions about representation learning.
Principled Bayesian Optimization in Collaboration with Human Experts
·2248 words·11 mins·
loading
·
loading
AI Theory
Optimization
🏢 University of Oxford
COBOL: a novel Bayesian Optimization algorithm leverages human expert advice via binary labels, achieving both fast convergence and robustness to noisy input, while guaranteeing minimal expert effort.
Pre-trained Text-to-Image Diffusion Models Are Versatile Representation Learners for Control
·2633 words·13 mins·
loading
·
loading
Multimodal Learning
Vision-Language Models
🏢 University of Oxford
Pre-trained text-to-image diffusion models create highly effective, versatile representations for embodied AI control, surpassing previous methods.
OxonFair: A Flexible Toolkit for Algorithmic Fairness
·3793 words·18 mins·
loading
·
loading
AI Theory
Fairness
🏢 University of Oxford
OxonFair: a new open-source toolkit for enforcing fairness in binary classification, supporting NLP, Computer Vision, and tabular data, optimizing any fairness metric, and minimizing performance degra…
Once Read is Enough: Domain-specific Pretraining-free Language Models with Cluster-guided Sparse Experts for Long-tail Domain Knowledge
·2658 words·13 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 University of Oxford
This research introduces Cluster-guided Sparse Experts (CSE), enabling pretrained language models to effectively learn long-tail domain knowledge without domain-specific pretraining, thus achieving su…
On the Limitations of Fractal Dimension as a Measure of Generalization
·1917 words·9 mins·
loading
·
loading
Machine Learning
Deep Learning
🏢 University of Oxford
Fractal dimension, while showing promise, fails to consistently predict neural network generalization due to hyperparameter influence and adversarial initializations; prompting further research.
No-regret Learning in Harmonic Games: Extrapolation in the Face of Conflicting Interests
·354 words·2 mins·
loading
·
loading
AI Theory
Optimization
🏢 University of Oxford
Extrapolated FTRL ensures Nash equilibrium convergence in harmonic games, defying standard no-regret learning limitations.