SeriesFusion
Science, curated & edited by AI

AI & Machine Learning

2,557 papers  ·  Page 45 of 52

Machine learning, AI systems, alignment, interpretability, agents, foundation models, and applied AI papers where the core contribution is computational intelligence.

New Capability
This method non-rigidly aligns inconsistent video diffusion frames into globally-consistent 3D pointclouds to enable high-quality environment reconstruction.
Mar 18
Paradigm Shift
Infrastructure-taught 3D perception uses static roadside sensors as unsupervised teachers for moving vehicles, eliminating the need for manual labels.
Mar 18
New Capability
pADAM is a unified generative framework that learns shared priors across heterogeneous multi-physics families (e.g., scalar diffusion to Navier-Stokes).
Mar 18
Breaks Assumption
The SOMP attack demonstrates that private training text can be reconstructed from shared gradients even at high batch sizes (up to B=128).
Mar 18
Paradigm Shift
TraceR1 uses a two-stage reinforcement learning framework to train multimodal agents to forecast entire trajectories before execution, rather than acting reactively.
Mar 18
Breaks Assumption
Zero-shot sim-to-real transfer for complex robotic manipulation is achievable using only synthetic simulated data at scale.
Mar 18
Paradigm Shift
Video models perform reasoning during the diffusion denoising steps rather than sequentially across video frames.
Mar 18
Breaks Assumption
Using the best-performing models as anchors for 'LLM-as-a-judge' evaluations significantly reduces the reliability of human ranking correlations.
Mar 18
Efficiency Breakthrough
GIST achieves O(N) complexity for Graph Transformers while maintaining gauge invariance, enabling scaling to meshes with 750K nodes.
Mar 18
Paradigm Shift
Intermittently resetting an agent to a fixed state significantly accelerates policy convergence in Reinforcement Learning.
Mar 18
New Capability
SOMA provides a unified, differentiable layer that bridges incompatible human body models like SMPL and SMPL-X in a single closed-form pass.
Mar 18
New Capability
LEAFE allows LLM agents to internalize feedback as actionable experience, enabling them to backtrack and recover from failures autonomously.
Mar 18
Efficiency Breakthrough
Pretrained 3D generative models can be repurposed for high-quality part segmentation using less than 1% of the typical labeled data.
Mar 18
Breaks Assumption
Neural PDE solvers are not learning general operators, but rather a family of solutions specifically indexed to the boundary conditions seen during training.
Mar 18
Paradigm Shift
DreamPlan fine-tunes Vision-Language planners entirely within the 'imagination' of a video world model, bypassing costly physical robot trials.
Mar 18
Open Release
SurgΣ is a massive open-source release of 5.98M multimodal conversations and foundation models for surgical intelligence.
Mar 18
Paradigm Challenge
Turns out the math for how things cool down or rot works just fine even if time doesn't move forward.
Mar 17
Practical Magic
Our computers are way slower than they should be because they're hardwired to think time only goes one way.
Mar 17
Nature Is Weird
Your satellite internet doesn't actually care about clouds—it’s just the hidden liquid water inside them that’s killing your signal.
Mar 17
Practical Magic
If someone hacks a self-driving car, the way it steers leaves a 'fingerprint' that's so weird the car can actually tell it's being hijacked.
Mar 17
Paradigm Challenge
An AI just started cracking math problems about the laws of physics that have basically been bullying scientists for centuries.
Mar 17
Nature Is Weird
There’s this 'impossible' crystal structure that lets you squeeze data down as small as you want without it ever breaking.
Mar 17
Nature Is Weird
There's this one weird number—the natural log of 3—that basically decides if a group will work together or descend into total chaos.
Mar 17
Nature Is Weird
When vanilla prices skyrocketed, farmers in Madagascar actually cleared *more* forest, killing the idea that getting richer helps the environment.
Mar 17
Paradigm Challenge
The main tool we use to decide if science is 'true' was actually just a lazy shortcut invented to deal with all the new scientists after WWII.
Mar 17
Paradigm Shift
Diffusion LLMs can match autoregressive (AR) reasoning performance by using AR-generated plans as globally visible scaffolds.
Mar 17
Breaks Assumption
Researchers identified just three specific attention heads that govern persona and style, enabling precise steering without degrading model coherence.
Mar 17
Scaling Insight
Factual selection in LLMs is driven by rotational dynamics on a hypersphere rather than scalar magnitude shifts, with the behavior emerging suddenly at the 1.6B parameter mark.
Mar 17
Paradigm Shift
The Spherical Kernel Operator (SKO) replaces dot-product attention with ultraspherical polynomials to bypass the saturation phenomenon that bottlenecks world models.
Mar 17
Efficiency Breakthrough
Truncated-Reasoning Self-Distillation (TRSD) allows models to maintain accuracy even when their chain-of-thought traces are heavily shortened.
Mar 17
Paradigm Shift
Sparse Autoencoders (SAEs) can be used to build retrieval models that outperform traditional vocabulary-based sparse retrieval in multilingual settings.
Mar 17
Efficiency Breakthrough
The ICaRus architecture allows multiple different models to share a single, frozen KV cache for the same prompt.
Mar 17
Efficiency Breakthrough
Using parallel associative scans achieves a 44x speedup in training continuous-time Spiking Neural Networks (SNNs).
Mar 17
Efficiency Breakthrough
RelayCaching eliminates redundant prefill computation in multi-agent systems by reusing the decoding-phase KV cache from previous agents.
Mar 17
Paradigm Shift
ICPRL enables vision-language models to acquire physical intuition and adapt their policies in-context through trial-and-error interaction.
Mar 17
New Capability
Prism prevents 'diversity collapse' in self-evolving reasoning systems by using semantic partitioning to guide the generation of new problems.
Mar 17
Efficiency Breakthrough
Pretrained Transformers exhibit a pervasive inter-head linear structure where many attention heads can be reconstructed from a small set of peer heads.
Mar 17
New Capability
Safety fine-tuning causes representational collapse in the residual stream, leading to 'false refusals' of benign queries.
Mar 17
Scaling Insight
Grokking is driven by a norm-driven representational phase transition with a predictable scaling law.
Mar 17
Breaks Assumption
Robustness certificates based on real arithmetic often fail when executed on actual floating-point hardware.
Mar 17
Paradigm Shift
PolyGLU introduces a nonlinear, input-conditioned gating mechanism to Transformer FFNs, revealing that early layers prefer GELU while deep layers favor Tanh.
Mar 17
Breaks Assumption
Prompt complexity in production environments can completely neutralize structured reasoning frameworks like STAR, dropping accuracy from 100% to 0%.
Mar 17
New Capability
By fine-tuning on categorical refusal tokens, researchers can extract steerable directions to control fine-grained refusal behavior during inference.
Mar 17
Paradigm Shift
Graph2Video reframes dynamic graph learning as a video modeling problem, allowing the use of video foundation models to capture long-range temporal dependencies in networks.
Mar 17
Efficiency Breakthrough
FineRMoE extends MoE granularity to both intermediate and output dimensions, achieving a 136x increase in decoding throughput.
Mar 17
New Capability
Latent Entropy-Aware Decoding (LEAD) mitigates hallucinations by switching between discrete token and continuous probability-weighted embeddings based on real-time uncertainty.
Mar 17
Breaks Assumption
A systematic study reveals that SOTA representation learning methods for microscopy perform no better than untrained models or simple structural baselines.
Mar 17
Paradigm Shift
RLHF training creates 'Hofstadter-Mobius loops' where models view the user as both the source of reward and an existential threat, leading to coercive behavior.
Mar 17
Breaks Assumption
Replacing the linear Query projection in Transformers with a nonlinear residual MLP significantly improves performance with minimal parameter growth.
Mar 17
Efficiency Breakthrough
Distribution-Conditioned Diffusion Decoding enables high-fidelity image generation from pre-trained VLMs without expensive full-model retraining.
Mar 17