SeriesFusion
Science, curated & edited by AI

AI & Machine Learning

2,371 papers  ·  Page 46 of 48

Machine learning, AI systems, alignment, interpretability, agents, foundation models, and applied AI papers where the core contribution is computational intelligence.

Efficiency Breakthrough
Modality-level disaggregation enables cost-optimal MLLM serving across heterogeneous GPUs over commodity PCIe, bypassing the need for expensive NVLink interconnects.
Mar 16
Breaks Assumption
Probing of Vision-Language-Action (VLA) models reveals that the action decoder largely ignores the reasoning logic in Chain-of-Thought, relying almost exclusively on object names.
Mar 16
New Capability
SciDesignBench provides a massive simulator-grounded environment for scientific inverse design, revealing that current LLMs struggle significantly with iterative refinement.
Mar 16
Efficiency Breakthrough
A hardware-algorithm co-design for Spiking Neural Networks achieves up to 69x energy efficiency gains using an SRAM-based Compute-in-Memory accelerator.
Mar 16
Breaks Assumption
The TaoBench benchmark proves that state-of-the-art math LLMs fail on equivalent logic problems when presented outside of the standard 'MathLib' framework.
Mar 16
New Capability
A self-supervised robotic system detects novel objects by training bespoke detectors on-the-fly from human video demonstrations, bypassing language-based prompts.
Mar 16
New Capability
AIM enables post-training modulation of large models to change utility levels or focus features without any retraining or additional data.
Mar 16
Efficiency Breakthrough
Achieves 4x visual token compression and 80% lower training cost while unifying multimodal comprehension and generation.
Mar 16
New Capability
First training-free method for debiasing reward models using Sparse Autoencoder (SAE) interventions.
Mar 16
Breaks Assumption
Breaks the long-standing accuracy-robustness trade-off in VLMs by localizing adversarial robustness to shallow layers.
Mar 16
New Capability
A flow-based navigation policy that achieves zero-shot sim-to-real transfer across wheeled, quadrupedal, and humanoid platforms.
Mar 16
Paradigm Shift
A small-scale molecular reasoning model that outperforms ultra-large foundation models via structured chain-of-thought and RL.
Mar 16
Efficiency Breakthrough
Adaptive VLM Routing reduces inference costs for Computer Use Agents by up to 78% with negligible accuracy loss.
Mar 16
Efficiency Breakthrough
Distills a 2B Vision-Language Retriever into a 70M text-only encoder for visual document retrieval with 50x lower latency.
Mar 16
Breaks Assumption
Reveals that 'reasoning' gains in fine-tuned LLMs may be artifacts of task familiarity rather than improved capability.
Mar 16
New Capability
MotionAnymesh automatically transforms static 3D meshes into simulation-ready, articulated digital twins for robotics using vision-language models grounded in physical priors.
Mar 16
Paradigm Shift
ThinkStream introduces a 'Watch-Think-Speak' paradigm for video reasoning that allows models to incrementally update understanding and decide when to respond in real-time.
Mar 16
Breaks Assumption
This paper presents an exact federated unlearning protocol for foundation models that is pointwise identical to centralized retraining but uses fixed-size messages.
Mar 16
Efficiency Breakthrough
CleanSight provides a training-free, test-time defense for backdoored vision-language models by detecting and pruning 'attention stealing' visual tokens.
Mar 16
Breaks Assumption
This study proves that even with a 'perfect' noise transition matrix, statistically consistent noise-correction methods still suffer from performance collapse.
Mar 16
Efficiency Breakthrough
Structured distillation for personalized agent memory achieves an 11x reduction in token count while preserving 96% of the retrieval quality of verbatim history.
Mar 16
New Capability
Multimodal OCR (MOCR) treats charts, diagrams, and tables as code-level targets (e.g., TikZ, SVG) rather than just cropping them as pixels.
Mar 16
Breaks Assumption
A cross-dataset study reveals that modern general-purpose vision models (GP-VMs) outperform specialized medical architectures in 2D medical image segmentation.
Mar 16
Paradigm Shift
Connects DDIM reverse chains to fractal geometry, providing a mathematical explanation for why diffusion models switch from global context to local detail.
Mar 16
Breaks Assumption
Reveals that linearized attention never converges to the NTK limit in practice, explaining its unique 'influence malleability' compared to standard networks.
Mar 16
Efficiency Breakthrough
Induces pretrained video models to perform SOTA image restoration using less than 2% of the training data required by specialized architectures.
Mar 16
Efficiency Breakthrough
Achieves 'zero-hyperparameter' circuit analysis by using a foundation model to perform in-context regression, bypassing hours of manual tuning.
Mar 16
Paradigm Shift
Proposes Causal Process Reward (CPR) to fix 'cherry-picking' in MLLM reasoning by coupling answer correctness with step-level logical alignment.
Mar 16
Efficiency Breakthrough
Introduces Bilateral Context Conditioning to DeepSeek's GRPO, allowing models to cross-reference successful and failed reasoning traces during optimization.
Mar 16
Efficiency Breakthrough
Enables RMSNorm to reuse MXFP8 block scales, reducing the reduction operation size by 32x with a 2.4x kernel speedup.
Mar 16
Breaks Assumption
Finds that privacy vulnerability and utility are both concentrated in a tiny fraction of 'critical weights' based on their location rather than value.
Mar 16
Breaks Assumption
STEVO-Bench reveals that current 'video world models' fail to simulate physical processes when the camera looks away or lights go out.
Mar 16
New Capability
Optimizes diffusion models via Direct Preference Optimization (DPO) to generate human motion that is inherently executable by real humanoid robots.
Mar 16
Paradigm Shift
Reimagines 3D molecules as continuous vector fields rather than discrete graphs, decoupling structure learning from atom types.
Mar 16
Scaling Insight
Proves the existence of a 'distributional simplicity bias' in diffusion models, where low-order statistics are learned linearly while high-order correlations require cubic sample complexity.
Mar 16
Paradigm Challenge
Time moving forward might just be a glitch caused by the universe being bad at copying its own homework.
Mar 13
Practical Magic
We’ve finally made digital messages that are physically impossible to copy—even a perfect hacker couldn't do it because physics won't allow it.
Mar 13
Nature Is Weird
Scientists built an AI that treats crop-raiding elephants like chess opponents to predict exactly where they’ll strike next.
Mar 13
Cosmic Scale
The massive satellite network the government uses is accidentally blasting out people's private passwords in plain text for anyone to see.
Mar 13
Open Release
OpenSanctions Pairs releases a massive benchmark for entity matching, proving that local LLMs can now match production rule-based systems in high-stakes compliance tasks.
Mar 13
Scaling Insight
Speculative Decoding Scaling Laws (SDSL) provides a theoretical framework to predict optimal throughput hyperparameters for LLM inference systems before pre-training.
Mar 13
Paradigm Shift
This paper introduces a graph tokenization framework that allows standard Transformers like BERT to beat specialized Graph Neural Networks without any architectural changes.
Mar 13
Efficiency Breakthrough
The first open recipe for training embodied intelligence at the 1,000-GPU scale, achieving a 40x speedup in training cycles for GR00T models.
Mar 13
Breaks Assumption
Routing signatures reveal that MoE experts are highly task-specific, allowing a simple linear classifier to identify task categories with 92.5% accuracy based only on routing patterns.
Mar 13
New Capability
A new method for training axis-aligned decision trees using gradient descent and backpropagation, allowing trees to be integrated into end-to-end neural networks.
Mar 13
Efficiency Breakthrough
REOPOLD achieves 10x better sample efficiency in reasoning distillation, enabling 7B models to match 32B teachers with significantly less training data.
Mar 13
Efficiency Breakthrough
PACED introduces a weight kernel that focuses distillation on the 'Zone of Proximal Development,' where the student's gradient signal-to-noise ratio is highest.
Mar 13
Paradigm Shift
Continual Representation Learning (CoRe) moves PEFT from weight-level updates to representation-space interventions, solving catastrophic forgetting in dynamic environments.
Mar 13
Scaling Insight
Cyber-attack capabilities of AI models scale log-linearly with inference-time compute, with no plateau in sight.
Mar 13
New Capability
SoLA introduces the first reversible model editing framework that allows precise revocation of specific knowledge updates.
Mar 13