Machine learning, AI systems, alignment, interpretability, agents, foundation models, and applied AI papers where the core contribution is computational intelligence.
Filter by category: Paradigm Challenge Breaks Assumption First Ever Nature Is Weird Practical Magic Cosmic Scale Life Origin Open Release Efficiency Leap New Capability Scaling Insight
Breaks Assumption
Exposes 'hidden clones' in VLM ensembles, where models from the same family share correlated errors that naive voting mechanisms fail to detect.
New Capability
Proposes REAL, a Reinforcement Learning framework tailored for regression and ordinal scoring rather than simple binary accuracy.
New Capability
Introduces a framework for LLM agents to autonomously evolve their policies and skill libraries during system idle time without retraining downtime.
Efficiency Breakthrough
A backbone-agnostic denoising objective that allows small GNNs to outperform large models pretrained on much larger supervised datasets in physical sciences.
Paradigm Shift
Achieves high-performance online continual learning without the massive memory overhead of traditional experience replay buffers.
Breaks Assumption
Internal activation probing detects LLM 'rationalization' more reliably than monitoring the model's own Chain-of-Thought (CoT).
Efficiency Breakthrough
A dynamic data pruning framework that cuts dense retriever training time by 50% while actually improving retrieval accuracy.
New Capability
Automates the generation of synthetic machine learning challenges to train agents that can genuinely learn research skills from doing.
Breaks Assumption
Alignment processes induce a 'normative bias' that makes LLMs worse at predicting real human behavior in strategic scenarios.
New Capability
Enables reliable, training-free emotion steering in speech-generative audio models via direct manipulation of specific emotion-sensitive neurons.
Paradigm Shift
A formal, graph-native memory architecture that treats agent memory as a versioned asset, dramatically outperforming Gemini 2.5 Pro on complex recall.
New Capability
A framework to quantify and fix 'task steerability,' the common failure of robots to respond to new instructions while mid-task.
Efficiency Breakthrough
Achieves up to a 1,000x gain in RLHF data efficiency by using information-directed exploration and epistemic neural networks.
Efficiency Breakthrough
Introduces a reward framework that reduces LLM reasoning verbosity by optimizing for 'Information Density' via entropy reduction per step.
Paradigm Shift
Shifts retrieval from static contrastive vector alignment to dynamic reasoning trajectories using a generative model (T1) and GRPO.
Breaks Assumption
Identifies that reasoning-induced safety failures occur *during* Chain-of-Thought and proposes a shift to 'decide-then-reason' architectures.
Efficiency Breakthrough
Generates 9 million grid points of 3D spatiotemporal physical fields in seconds, a 10,000x speedup over traditional physics simulations.
New Capability
Proposes a world model that jointly generates appearance and binocular geometry using an epipolar-aware attention mechanism.
Open Release
Introduces FineViT and a 450M local caption dataset to solve the 'coarse perception' bottleneck in current CLIP-based encoders.
Paradigm Shift
Provides a sheaf-theoretic proof that local causal consistency in generative models does not guarantee global counterfactual coherence.
Efficiency Breakthrough
Replaces quadratic self-attention with $O(N \log N)$ phase-native coupling for time-series, enabling massive context windows.
New Capability
Introduces a paradigm for vision-language navigation that uses ubiquitously available semantic floor plans as global spatial priors.
New Capability
Embeds invisible, agent-specific 'watermarks' into token distributions to enable forensic attribution and topology reconstruction in multi-agent systems.
Efficiency Breakthrough
Achieves an 80% reduction in Chain-of-Thought (CoT) tokens while slightly increasing reasoning accuracy.
Efficiency Breakthrough
Extends LLM context from 32K to 128K by teaching models to selectively skip global attention for ~80% of tokens.
New Capability
Reduces hallucinations by teaching models 'epistemological humility'—the ability to admit they don't know something—using synthetic non-existent terms.
Breaks Assumption
Develops a zero-watermarking framework that survives AI editing by leveraging invariant relations between image patches.
Paradigm Shift
Unifies large-scale search, recommendation, and reasoning into a single self-contained LLM by treating item IDs as a distinct modality.
Scaling Insight
Video fine-tuning consistently degrades static image understanding in multimodal LLMs, revealing a zero-sum trade-off between spatial and temporal capabilities.
New Capability
Introduces a Prompt-Free Universal Region Proposal Network (PF-RPN) that identifies objects in any domain without needing text or image exemplars.
New Capability
FrescoDiffusion enables coherent, 4K image-to-video generation using a training-free, tiled diffusion method with precomputed latent priors.
Efficiency Breakthrough
Knowledge-Aware Active Learning (KA2L) uses latent space probing to identify what an LLM doesn't know and generates targeted synthetic questions.
Breaks Assumption
Dense retrieval architectures are fundamentally flawed at detecting negation and contradictions due to 'Semantic Collapse' in vector space.
Paradigm Shift
Edit-As-Act reframes 3D scene editing as a goal-regressive planning problem using symbolic action languages rather than purely generative pixel manipulation.
Breaks Assumption
ARES demonstrates high-fidelity data reconstruction from large Federated Learning batches without requiring any architectural modifications to the model.
Scaling Insight
Mechanistic probing reveals a directional asymmetry in how LLMs encode hierarchy: hypernymy is redundant and resilient, while hyponymy is fragile and compact.
Efficiency Breakthrough
S-VGGT introduces structure-aware subscene decomposition to break the quadratic scaling bottleneck of 3D foundation models.
New Capability
Introduces a framework to generate complex, non-linear environments with mathematically guaranteed ground-truth optimal policies for RL benchmarking.
Efficiency Breakthrough
DSS-GAN is the first generative adversarial network to use a Mamba (State Space Model) backbone for high-quality image synthesis.
New Capability
VectorWorld enables stable, real-time 1km+ closed-loop world model rollouts for autonomous driving using diffusion flow on vector graphs.
New Capability
REAL achieves extreme quadruped parkour agility that is robust even to a 1-meter visual blind zone.
Breaks Assumption
FINER discovers that MLLMs are highly prone to hallucination when images contain fine-grained mismatches co-occurring with real elements.
Efficiency Breakthrough
Synthetic videos of simple geometric shapes are more effective than massive real-world datasets for teaching video-language models fundamental temporal reasoning.
New Capability
Lifting 2D features into a volumetric representation for robot manipulation policies yields a 14.8% success rate improvement by solving the 2D-3D spatial reasoning mismatch.
Paradigm Shift
A new self-refining surrogate framework enables neural models to simulate complex dynamical systems over arbitrarily long horizons without the standard failure mode of compounding error.
Breaks Assumption
Massive activation outliers in Transformers are an adaptive response to 'gradient sinks' during training, rather than just an inference-time quirk.
Paradigm Shift
The 'consensus trap' in label-free RL—where models reinforce their own systematic errors—can be broken by co-evolving the model in alternating generator and verifier roles.
Breaks Assumption
In-context memory for LLMs is fundamentally unreliable due to compaction loss and goal drift, but structured 'Knowledge Objects' provide a 252x cheaper and 100% accurate alternative.
Efficiency Breakthrough
Anomaly detection can be performed directly using a primary model's internal neuron output ranges, eliminating the need for expensive external AD models.
Efficiency Breakthrough
Truncated backpropagation for video decoding reduces the memory cost of fine-tuning video diffusion models from linear to constant.