SeriesFusion
Science, curated & edited by AI

AI & Machine Learning

2,371 papers  ·  Page 41 of 48

Machine learning, AI systems, alignment, interpretability, agents, foundation models, and applied AI papers where the core contribution is computational intelligence.

Scaling Insight
Provides a geometric 'manifold envelopment' framework to explain why unsupervised RL for mathematical reasoning often collapses and how to stabilize it.
Mar 18
Paradigm Shift
Formalizes AI agent governance as 'policies on paths,' moving from static prompts to runtime enforcement of complex legal and safety constraints.
Mar 18
Efficiency Breakthrough
Enables stable 4-bit microscaling (MXFP4) quantization for Multi-modal LLMs, which previously suffered from performance collapse.
Mar 18
New Capability
Introduces a way to train Reward Models that generate 'transferable rubrics'—explicit scoring criteria that improve performance across different tasks and models.
Mar 18
New Capability
OmniSONAR scales cross-lingual sentence embeddings to over 1,500 languages across text, speech, code, and math in a single semantic space.
Mar 18
Paradigm Shift
Aligns a base model to a target model's behavior by optimizing the 'data mixture' weights instead of using RLHF or DPO.
Mar 18
Breaks Assumption
Achieves high-bandwidth, precise Cartesian control of a fully soft continuum robot, breaking the assumption that softness and precision are incompatible.
Mar 18
New Capability
Fine-tuning language models on journal publication records allows them to match or exceed human experts in judging 'scientific taste'—the ability to identify which research ideas are worth pursuing.
Mar 18
Paradigm Shift
This paper introduces a Markov-based discrete reasoning model that learns its own stopping criterion and can re-mask and correct its own mistakes.
Mar 18
Breaks Assumption
Fast-WAM proves that World Action Models do not actually need to generate future 'imagination' frames at test-time to achieve state-of-the-art performance in embodied control.
Mar 18
Scaling Insight
The study provides a formal link showing that internal 'world model' representations in transformers are a direct byproduct of the predictive geometry of the training data.
Mar 18
Breaks Assumption
Chain-of-thought (CoT) reasoning in Vision-Language Models systematically degrades the reliability of uncertainty estimates, making models dangerously overconfident.
Mar 18
Efficiency Breakthrough
Low-precision optimizer states cause 'state staleness' where updates round back to stored values, but scheduled resets can fully recover performance loss.
Mar 18
Open Release
IQuest-Coder-V1 introduces a series of high-performance code models including a unique 'Loop' variant with a recurrent mechanism for efficiency.
Mar 18
New Capability
This method non-rigidly aligns inconsistent video diffusion frames into globally-consistent 3D pointclouds to enable high-quality environment reconstruction.
Mar 18
Paradigm Shift
Infrastructure-taught 3D perception uses static roadside sensors as unsupervised teachers for moving vehicles, eliminating the need for manual labels.
Mar 18
New Capability
pADAM is a unified generative framework that learns shared priors across heterogeneous multi-physics families (e.g., scalar diffusion to Navier-Stokes).
Mar 18
Breaks Assumption
The SOMP attack demonstrates that private training text can be reconstructed from shared gradients even at high batch sizes (up to B=128).
Mar 18
Paradigm Shift
TraceR1 uses a two-stage reinforcement learning framework to train multimodal agents to forecast entire trajectories before execution, rather than acting reactively.
Mar 18
Breaks Assumption
Zero-shot sim-to-real transfer for complex robotic manipulation is achievable using only synthetic simulated data at scale.
Mar 18
Paradigm Shift
Video models perform reasoning during the diffusion denoising steps rather than sequentially across video frames.
Mar 18
Breaks Assumption
Using the best-performing models as anchors for 'LLM-as-a-judge' evaluations significantly reduces the reliability of human ranking correlations.
Mar 18
Efficiency Breakthrough
GIST achieves O(N) complexity for Graph Transformers while maintaining gauge invariance, enabling scaling to meshes with 750K nodes.
Mar 18
Paradigm Shift
Intermittently resetting an agent to a fixed state significantly accelerates policy convergence in Reinforcement Learning.
Mar 18
New Capability
SOMA provides a unified, differentiable layer that bridges incompatible human body models like SMPL and SMPL-X in a single closed-form pass.
Mar 18
New Capability
LEAFE allows LLM agents to internalize feedback as actionable experience, enabling them to backtrack and recover from failures autonomously.
Mar 18
Efficiency Breakthrough
Pretrained 3D generative models can be repurposed for high-quality part segmentation using less than 1% of the typical labeled data.
Mar 18
Breaks Assumption
Neural PDE solvers are not learning general operators, but rather a family of solutions specifically indexed to the boundary conditions seen during training.
Mar 18
Paradigm Shift
DreamPlan fine-tunes Vision-Language planners entirely within the 'imagination' of a video world model, bypassing costly physical robot trials.
Mar 18
Open Release
SurgΣ is a massive open-source release of 5.98M multimodal conversations and foundation models for surgical intelligence.
Mar 18
Paradigm Challenge
Turns out the math for how things cool down or rot works just fine even if time doesn't move forward.
Mar 17
Practical Magic
Our computers are way slower than they should be because they're hardwired to think time only goes one way.
Mar 17
Nature Is Weird
Your satellite internet doesn't actually care about clouds—it’s just the hidden liquid water inside them that’s killing your signal.
Mar 17
Practical Magic
If someone hacks a self-driving car, the way it steers leaves a 'fingerprint' that's so weird the car can actually tell it's being hijacked.
Mar 17
Paradigm Challenge
An AI just started cracking math problems about the laws of physics that have basically been bullying scientists for centuries.
Mar 17
Nature Is Weird
There’s this 'impossible' crystal structure that lets you squeeze data down as small as you want without it ever breaking.
Mar 17
Nature Is Weird
There's this one weird number—the natural log of 3—that basically decides if a group will work together or descend into total chaos.
Mar 17
Nature Is Weird
When vanilla prices skyrocketed, farmers in Madagascar actually cleared *more* forest, killing the idea that getting richer helps the environment.
Mar 17
Paradigm Challenge
The main tool we use to decide if science is 'true' was actually just a lazy shortcut invented to deal with all the new scientists after WWII.
Mar 17
Paradigm Shift
Diffusion LLMs can match autoregressive (AR) reasoning performance by using AR-generated plans as globally visible scaffolds.
Mar 17
Breaks Assumption
Researchers identified just three specific attention heads that govern persona and style, enabling precise steering without degrading model coherence.
Mar 17
Scaling Insight
Factual selection in LLMs is driven by rotational dynamics on a hypersphere rather than scalar magnitude shifts, with the behavior emerging suddenly at the 1.6B parameter mark.
Mar 17
Paradigm Shift
The Spherical Kernel Operator (SKO) replaces dot-product attention with ultraspherical polynomials to bypass the saturation phenomenon that bottlenecks world models.
Mar 17
Efficiency Breakthrough
Truncated-Reasoning Self-Distillation (TRSD) allows models to maintain accuracy even when their chain-of-thought traces are heavily shortened.
Mar 17
Paradigm Shift
Sparse Autoencoders (SAEs) can be used to build retrieval models that outperform traditional vocabulary-based sparse retrieval in multilingual settings.
Mar 17
Efficiency Breakthrough
The ICaRus architecture allows multiple different models to share a single, frozen KV cache for the same prompt.
Mar 17
Efficiency Breakthrough
Using parallel associative scans achieves a 44x speedup in training continuous-time Spiking Neural Networks (SNNs).
Mar 17
Efficiency Breakthrough
RelayCaching eliminates redundant prefill computation in multi-agent systems by reusing the decoding-phase KV cache from previous agents.
Mar 17
Paradigm Shift
ICPRL enables vision-language models to acquire physical intuition and adapt their policies in-context through trial-and-error interaction.
Mar 17
New Capability
Prism prevents 'diversity collapse' in self-evolving reasoning systems by using semantic partitioning to guide the generation of new problems.
Mar 17