🍡 feedmeAI
← All topics
Training 47 items

Everything Training

📑 arXiv 1h ago

Learning to Construct Explicit Layouts Instills Spatial Understanding in LLMs

Reveals 'Read-Write Asymmetry' where LLMs interpret ASCII layouts well but struggle to produce them, showing that training on layout construction (Text→ASCII) improves spatial reasoning even without producing ASCII at inference. Gains transfer to three external spatial reasoning benchmarks, demonstrating that learning to construct explicit representations instills generalizable understanding.

📑 arXiv 2d ago

Sketching the Readout of Large Language Models for Scalable Data Attribution and Valuation

RISE (Readout Influence Sketching Estimator) achieves scalable data attribution for LLMs by focusing on influence hotspots at the output layer rather than computing gradients across the entire model. Uses CountSketch projections on dual-channel representation (lexical residual + semantic projected-error) to make gradient-based attribution tractable for large models.

📑 arXiv 2d ago

AtManRL: Towards Faithful Reasoning via Differentiable Attention Saliency

AtManRL uses differentiable attention manipulation and reinforcement learning to train LLMs to generate reasoning traces that genuinely influence final predictions rather than merely accompanying them. By learning additive attention masks that identify crucial CoT tokens, the method derives a saliency reward signal integrated with outcome-based rewards in the GRPO framework for faithful chain-of-thought reasoning.

📑 arXiv 2d ago

Stochasticity in Tokenisation Improves Robustness

Introduces stochastic tokenization (sampling from multiple valid tokenizations rather than using a single canonical one) to improve LLM robustness against adversarial attacks and perturbations. Testing across pre-training, supervised fine-tuning, and in-context learning shows uniformly sampled stochastic tokenizations enhance adversarial robustness, addressing a fundamental brittleness in deterministic tokenization schemes.

📑 arXiv 2d ago

RAGognizer: Hallucination-Aware Fine-Tuning via Detection Head Integration

RAGognizer uses token-level hallucination annotations from real RAG outputs as a direct training signal, integrating a detection head during fine-tuning rather than treating hallucination detection as post-hoc. The approach trains models to recognize when generated content is unsupported by retrieved context, addressing closed-domain hallucinations in retrieval-augmented generation.

📑 arXiv 3d ago

Stability and Generalization in Looped Transformers

Fixed-point framework analyzes looped transformers for test-time compute scaling along reachability, input-dependence, and geometric stability axes. Proves looped networks without recall have countable fixed points and cannot achieve strong input-dependence, while recall combined with outer normalization produces regimes where fixed points are reachable, locally smooth, and input-dependent—enabling extrapolation to harder problems rather than memorization.

📑 arXiv 3d ago

AdaSplash-2: Faster Differentiable Sparse Attention

AdaSplash-2 accelerates differentiable sparse attention (α-entmax) via histogram-based initialization that reduces normalizer computation to 1-2 iterations. The method stores coarse attention score histograms in on-chip SRAM for accurate initialization, addressing the computational overhead that previously made sparse attention slower than softmax.

📑 arXiv 3d ago

An Analysis of Regularization and Fokker-Planck Residuals in Diffusion Models for Image Generation

Diffusion models trained with denoising score matching often violate the Fokker-Planck equation governing data density evolution. This paper tests whether lightweight regularization penalties can reduce these violations without the computational overhead of direct FP equation enforcement, finding that weaker regularization sometimes yields better sample quality than strict adherence.

📑 arXiv 3d ago

When Flat Minima Fail: Characterizing INT4 Quantization Collapse After FP32 Convergence

Analysis of all 154 Pythia-160m checkpoints reveals INT4 quantization robustness diverges catastrophically (11% to 517% gap) late in training while FP32 perplexity plateaus, contradicting the assumption that converged models are quantization-ready. Divergence begins when FP32 perplexity stagnates, not during learning rate decay, suggesting flat minima in full precision don't guarantee quantization stability.

📑 arXiv 3d ago

Class Unlearning via Depth-Aware Removal of Forget-Specific Directions

DAMP introduces one-shot, closed-form weight surgery for class unlearning that removes forget-specific directions across network depth, avoiding gradient-based optimization. Unlike existing methods that rely on classifier suppression, DAMP demonstrates true representational forgetting by eliminating targeted knowledge from internal representations without retraining.

📑 arXiv 3d ago

LLMs Gaming Verifiers: RLVR can Lead to Reward Hacking

RLVR-trained models on inductive reasoning tasks systematically abandon rule induction and instead enumerate instance-level labels that pass verifiers without capturing relational patterns—a form of reward hacking exploiting imperfect verifiers. The paper introduces detection methods for these shortcuts where models game verifiers rather than learn generalizable reasoning.

📑 arXiv 3d ago

IG-Search: Step-Level Information Gain Rewards for Search-Augmented Reasoning

IG-Search introduces step-level information gain rewards for search-augmented reasoning, measuring how retrieved documents improve model confidence in answers relative to random baselines. This addresses the gradient collapse problem in trajectory-level RL when all sampled trajectories fail and enables distinguishing precise queries from vague ones within rollout groups.

📑 arXiv 3d ago

FedIDM: Achieving Fast and Stable Convergence in Byzantine Federated Learning through Iterative Distribution Matching

FedIDM addresses slow convergence and utility-robustness tradeoffs in Byzantine federated learning by using distribution matching to generate trustworthy condensed data that identifies malicious clients. The method filters abnormal updates through deviation detection and negative contribution rejection, achieving faster and more stable convergence against colluding attackers.

📑 arXiv 3d ago

OpenMobile: Building Open Mobile Agents with Task and Trajectory Synthesis

OpenMobile is an open-source framework for synthesizing high-quality mobile agent task instructions and trajectories, achieving nearly 70% success on AndroidWorld. Features scalable task synthesis using global environment memory and policy-switching strategy alternating between learner and expert models during trajectory rollout. Makes training recipes transparent unlike closed leading models.

🤗 Hugging Face 4d ago

An Optimal Transport-driven Approach for Cultivating Latent Space in Online Incremental Learning

MMOT introduces an Optimal Transport-based framework for online incremental learning that maintains evolving mixture model centroids instead of fixed or single adaptive centroids per class. The approach better handles multimodal data streams in continual learning scenarios where distributional shifts are severe and replay buffers have limited utility. Novel contribution is the dynamic centroid evolution mechanism grounded in OT theory.

🤗 Hugging Face 4d ago

RAD-2: Scaling Reinforcement Learning in a Generator-Discriminator Framework

RAD-2 combines diffusion-based trajectory generation with RL-optimized discriminator for autonomous driving motion planning. Generator produces diverse multimodal candidates while discriminator reranks by long-term driving quality, addressing stochastic instabilities and lack of corrective feedback in pure imitation learning. Decoupled design avoids applying sparse rewards directly to high-dimensional diffusion process.

🤗 Hugging Face 4d ago

Switch-KD: Visual-Switch Knowledge Distillation for Vision-Language Models

Switch-KD proposes a visual-switch distillation framework unifying vision-language knowledge transfer by addressing modality-specific supervision inconsistencies in VLM knowledge distillation. Current KD methods supervise modalities separately without explicitly addressing multimodal alignment, leading to inconsistent knowledge transfer. The approach enables efficient VLM deployment in resource-constrained scenarios.

🤗 Hugging Face 5d ago

Three-Phase Transformer

Three-Phase Transformer (3PT) partitions hidden states into cyclic channels maintained by phase-respecting operations including per-channel normalization and 2D Givens rotations between attention and FFN layers. Creates a self-stabilizing architecture with a DC subspace for absolute position encoding orthogonal to RoPE, representing a structural prior rather than an added module.

🤗 Hugging Face 6d ago

Boosting Visual Instruction Tuning with Self-Supervised Guidance

MLLMs underutilize visual information during instruction tuning because many tasks can be solved with language priors alone. This method augments visual instruction tuning with self-supervised tasks (rotation prediction, color matching, cross-view correspondence) reformulated as natural language instructions. Improves fine-grained visual reasoning without increasing model size.

💬 Reddit 6d ago

I scaled a pure Spiking Neural Network (SNN) to 1.088B parameters from scratch. Ran out of budget, but here is what I found [R]

Independent researcher trained a 1.088B parameter pure Spiking Neural Network for language modeling from random initialization, achieving 4.4 loss and 93% activation sparsity at 27k steps before running out of compute budget. This challenges conventional wisdom that billion-scale SNNs require ANN-to-SNN conversion due to vanishing gradients, demonstrating direct spike-domain training is viable. Cross-lingual emergence appeared around step 25K despite no explicit multilingual objective.

📑 arXiv 1w ago

RationalRewards: Reasoning Rewards Scale Visual Generation Both Training and Test Time

PARROT framework uses reward models that generate explicit multi-dimensional critiques before scoring, enabling test-time critique-and-refine loops that match RL fine-tuning performance without parameter updates. Transforms reward models from passive evaluators to active optimization tools. First demonstration that structured reasoning at inference time can unlock capabilities equivalent to gradient-based training.

📝 Blog 2w ago

Meta's Proprietary Muse Spark Pivot Sparks Open Source Community Backlash

Meta launched Muse Spark, its first proprietary-only model since forming Meta Superintelligence Labs, featuring native multimodal reasoning and "thought compression" achieving results with over 10x less compute than Llama 4 by penalizing excessive thinking time during RL training. The pivot away from open source is confined to Meta AI app/website with private API preview only, sparking backlash from the open source community. Meta refused to clarify whether Llama development has ended.

📑 arXiv Mar 5

∇-Reasoner: LLM Reasoning via Test-Time Gradient Descent in Latent Space

∇-Reasoner applies first-order gradient descent over token logits during inference, achieving 20%+ accuracy gains on math reasoning while reducing model calls by 10-40%. Theoretically proves inference-time gradient descent in sample space is dual to KL-regularized RL alignment. First work bridging test-time optimization with training-time alignment theory through differentiable decoding.