🔹 Title: R-Zero: Self-Evolving Reasoning LLM from Zero Data
🔹 Publication Date: Published on Aug 7
🔹 Abstract: R-Zero is a self-evolving framework that autonomously generates and learns from its own training data, improving reasoning capabilities in LLMs without human-curated tasks. AI-generated summary Self-evolving Large Language Models ( LLMs ) offer a scalable path toward super-intelligence by autonomously generating, refining, and learning from their own experiences. However, existing methods for training such models still rely heavily on vast human-curated tasks and labels, typically via fine-tuning or reinforcement learning, which poses a fundamental bottleneck to advancing AI systems toward capabilities beyond human intelligence. To overcome this limitation, we introduce R-Zero , a fully autonomous framework that generates its own training data from scratch. Starting from a single base LLM, R-Zero initializes two independent models with distinct roles, a Challenger and a Solver . These models are optimized separately and co-evolve through interaction: the Challenger is rewarded for proposing tasks near the edge of the Solver capability, and the Solver is rewarded for solving increasingly challenging tasks posed by the Challenger . This process yields a targeted, self-improving curriculum without any pre-existing tasks and labels. Empirically, R-Zero substantially improves reasoning capability across different backbone LLMs , e.g., boosting the Qwen3-4B-Base by +6.49 on math-reasoning benchmarks and +7.54 on general-domain reasoning benchmarks .
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2508.05004
• PDF: https://arxiv.org/pdf/2508.05004
• Project Page: https://chengsong-huang.github.io/R-Zero.github.io/
• Github: https://github.com/Chengsong-Huang/R-Zero
🔹 Datasets citing this paper:
No datasets found
🔹 Spaces citing this paper:
No spaces found
==================================
For more data science resources:
✓ https://news.1rj.ru/str/DataScienceT
🔹 Publication Date: Published on Aug 7
🔹 Abstract: R-Zero is a self-evolving framework that autonomously generates and learns from its own training data, improving reasoning capabilities in LLMs without human-curated tasks. AI-generated summary Self-evolving Large Language Models ( LLMs ) offer a scalable path toward super-intelligence by autonomously generating, refining, and learning from their own experiences. However, existing methods for training such models still rely heavily on vast human-curated tasks and labels, typically via fine-tuning or reinforcement learning, which poses a fundamental bottleneck to advancing AI systems toward capabilities beyond human intelligence. To overcome this limitation, we introduce R-Zero , a fully autonomous framework that generates its own training data from scratch. Starting from a single base LLM, R-Zero initializes two independent models with distinct roles, a Challenger and a Solver . These models are optimized separately and co-evolve through interaction: the Challenger is rewarded for proposing tasks near the edge of the Solver capability, and the Solver is rewarded for solving increasingly challenging tasks posed by the Challenger . This process yields a targeted, self-improving curriculum without any pre-existing tasks and labels. Empirically, R-Zero substantially improves reasoning capability across different backbone LLMs , e.g., boosting the Qwen3-4B-Base by +6.49 on math-reasoning benchmarks and +7.54 on general-domain reasoning benchmarks .
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2508.05004
• PDF: https://arxiv.org/pdf/2508.05004
• Project Page: https://chengsong-huang.github.io/R-Zero.github.io/
• Github: https://github.com/Chengsong-Huang/R-Zero
🔹 Datasets citing this paper:
No datasets found
🔹 Spaces citing this paper:
No spaces found
==================================
For more data science resources:
✓ https://news.1rj.ru/str/DataScienceT
❤3
🔹 Title: Are We on the Right Way for Assessing Document Retrieval-Augmented Generation?
🔹 Publication Date: Published on Aug 5
🔹 Abstract: Double-Bench is a large-scale, multilingual, and multimodal evaluation system for document Retrieval-Augmented Generation (RAG) systems, addressing limitations in current benchmarks and providing comprehensive assessments of system components. AI-generated summary Retrieval-Augmented Generation (RAG) systems using Multimodal Large Language Models (MLLMs) show great promise for complex document understanding, yet their development is critically hampered by inadequate evaluation. Current benchmarks often focus on specific part of document RAG system and use synthetic data with incomplete ground truth and evidence labels, therefore failing to reflect real-world bottlenecks and challenges. To overcome these limitations, we introduce Double-Bench : a new large-scale , multilingual , and multimodal evaluation system that is able to produce fine-grained assessment to each component within document RAG system s. It comprises 3,276 documents (72,880 pages) and 5,168 single- and multi-hop queries across 6 languages and 4 document types with streamlined dynamic update support for potential data contamination issues. Queries are grounded in exhaustively scanned evidence pages and verified by human experts to ensure maximum quality and completeness. Our comprehensive experiments across 9 state-of-the-art embedding models, 4 MLLMs and 4 end-to-end document RAG frameworks demonstrate the gap between text and visual embedding models is narrowing, highlighting the need in building stronger document retrieval models . Our findings also reveal the over-confidence dilemma within current document RAG frameworks that tend to provide answer even without evidence support. We hope our fully open-source Double-Bench provide a rigorous foundation for future research in advanced document RAG system s. We plan to retrieve timely corpus and release new benchmarks on an annual basis.
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2508.03644
• PDF: https://arxiv.org/pdf/2508.03644
• Project Page: https://double-bench.github.io/
• Github: https://github.com/Episoode/Double-Bench
🔹 Datasets citing this paper:
No datasets found
🔹 Spaces citing this paper:
No spaces found
==================================
For more data science resources:
✓ https://news.1rj.ru/str/DataScienceT
🔹 Publication Date: Published on Aug 5
🔹 Abstract: Double-Bench is a large-scale, multilingual, and multimodal evaluation system for document Retrieval-Augmented Generation (RAG) systems, addressing limitations in current benchmarks and providing comprehensive assessments of system components. AI-generated summary Retrieval-Augmented Generation (RAG) systems using Multimodal Large Language Models (MLLMs) show great promise for complex document understanding, yet their development is critically hampered by inadequate evaluation. Current benchmarks often focus on specific part of document RAG system and use synthetic data with incomplete ground truth and evidence labels, therefore failing to reflect real-world bottlenecks and challenges. To overcome these limitations, we introduce Double-Bench : a new large-scale , multilingual , and multimodal evaluation system that is able to produce fine-grained assessment to each component within document RAG system s. It comprises 3,276 documents (72,880 pages) and 5,168 single- and multi-hop queries across 6 languages and 4 document types with streamlined dynamic update support for potential data contamination issues. Queries are grounded in exhaustively scanned evidence pages and verified by human experts to ensure maximum quality and completeness. Our comprehensive experiments across 9 state-of-the-art embedding models, 4 MLLMs and 4 end-to-end document RAG frameworks demonstrate the gap between text and visual embedding models is narrowing, highlighting the need in building stronger document retrieval models . Our findings also reveal the over-confidence dilemma within current document RAG frameworks that tend to provide answer even without evidence support. We hope our fully open-source Double-Bench provide a rigorous foundation for future research in advanced document RAG system s. We plan to retrieve timely corpus and release new benchmarks on an annual basis.
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2508.03644
• PDF: https://arxiv.org/pdf/2508.03644
• Project Page: https://double-bench.github.io/
• Github: https://github.com/Episoode/Double-Bench
🔹 Datasets citing this paper:
No datasets found
🔹 Spaces citing this paper:
No spaces found
==================================
For more data science resources:
✓ https://news.1rj.ru/str/DataScienceT
❤2
🔹 Title: RoboMemory: A Brain-inspired Multi-memory Agentic Framework for Lifelong Learning in Physical Embodied Systems
🔹 Publication Date: Published on Aug 2
🔹 Abstract: RoboMemory, a brain-inspired multi-memory framework, enhances lifelong learning in physical robots by integrating cognitive neuroscience principles and achieving state-of-the-art performance in real-world tasks. AI-generated summary We present RoboMemory, a brain-inspired multi-memory framework for lifelong learning in physical embodied systems, addressing critic al challenges in real-world environments: continuous learning, multi-module memory latency, task correlation capture, and infinite-loop mitigation in closed-loop planning. Grounded in cognitive neuroscience, it integrates four core modules: the Information Preprocessor (thalamus-like), the Lifelong Embodied Memory System (hippocampus-like), the Closed-Loop Planning Module (prefrontal lobe-like), and the Low-Level Executer (cerebellum-like) to enable long-term planning and cumulative learning. The Lifelong Embodied Memory System , central to the framework, alleviates inference speed issues in complex memory frameworks via parallelized updates/retrieval across Spatial, Temporal, Episodic, and Semantic submodules. It incorporates a dynamic Knowledge Graph (KG) and consistent architectural design to enhance memory consistency and scalability. Evaluations on EmbodiedBench show RoboMemory outperforms the open-source baseline ( Qwen2.5-VL-72B-Ins ) by 25% in average success rate and surpasses the closed-source State-of-the-Art ( SOTA ) ( Claude3.5-Sonnet ) by 5%, establishing new SOTA . Ablation studies validate key components ( critic , spatial memory , long-term memory ), while real-world deployment confirms its lifelong learning capability with significantly improved success rates across repeated tasks. RoboMemory alleviates high latency challenges with scalability, serving as a foundational reference for integrating multi-modal memory systems in physical robots.
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2508.01415
• PDF: https://arxiv.org/pdf/2508.01415
🔹 Datasets citing this paper:
No datasets found
🔹 Spaces citing this paper:
No spaces found
==================================
For more data science resources:
✓ https://news.1rj.ru/str/DataScienceT
🔹 Publication Date: Published on Aug 2
🔹 Abstract: RoboMemory, a brain-inspired multi-memory framework, enhances lifelong learning in physical robots by integrating cognitive neuroscience principles and achieving state-of-the-art performance in real-world tasks. AI-generated summary We present RoboMemory, a brain-inspired multi-memory framework for lifelong learning in physical embodied systems, addressing critic al challenges in real-world environments: continuous learning, multi-module memory latency, task correlation capture, and infinite-loop mitigation in closed-loop planning. Grounded in cognitive neuroscience, it integrates four core modules: the Information Preprocessor (thalamus-like), the Lifelong Embodied Memory System (hippocampus-like), the Closed-Loop Planning Module (prefrontal lobe-like), and the Low-Level Executer (cerebellum-like) to enable long-term planning and cumulative learning. The Lifelong Embodied Memory System , central to the framework, alleviates inference speed issues in complex memory frameworks via parallelized updates/retrieval across Spatial, Temporal, Episodic, and Semantic submodules. It incorporates a dynamic Knowledge Graph (KG) and consistent architectural design to enhance memory consistency and scalability. Evaluations on EmbodiedBench show RoboMemory outperforms the open-source baseline ( Qwen2.5-VL-72B-Ins ) by 25% in average success rate and surpasses the closed-source State-of-the-Art ( SOTA ) ( Claude3.5-Sonnet ) by 5%, establishing new SOTA . Ablation studies validate key components ( critic , spatial memory , long-term memory ), while real-world deployment confirms its lifelong learning capability with significantly improved success rates across repeated tasks. RoboMemory alleviates high latency challenges with scalability, serving as a foundational reference for integrating multi-modal memory systems in physical robots.
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2508.01415
• PDF: https://arxiv.org/pdf/2508.01415
🔹 Datasets citing this paper:
No datasets found
🔹 Spaces citing this paper:
No spaces found
==================================
For more data science resources:
✓ https://news.1rj.ru/str/DataScienceT
❤1
🔹 Title: On the Expressiveness of Softmax Attention: A Recurrent Neural Network Perspective
🔹 Publication Date: Published on Jul 31
🔹 Abstract: Softmax attention is more expressive than linear attention due to its recurrent form, which can be analyzed using RNN components. AI-generated summary Since its introduction, softmax attention has become the backbone of modern transformer architectures due to its expressiveness and scalability across a wide range of tasks. However, the main drawback of softmax attention is the quadratic memory requirement and computational complexity with respect to the sequence length . By replacing the softmax nonlinearity, linear attention and similar methods have been introduced to avoid the quadratic bottleneck of softmax attention . Despite these linear forms of attention being derived from the original softmax formulation, they typically lag in terms of downstream accuracy. While strong intuition of the softmax nonlinearity on the query and key inner product suggests that it has desirable properties compared to other nonlinearities, the question of why this discrepancy exists still remains unanswered. This work demonstrates that linear attention is an approximation of softmax attention by deriving the recurrent form of softmax attention . Using this form, each part of softmax attention can be described in the language of recurrent neural networks (RNNs) . Describing softmax attention as an RNN allows for the ablation of the components of softmax attention to understand the importance of each part and how they interact. In this way, our work helps explain why softmax attention is more expressive than its counterparts.
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2507.23632
• PDF: https://arxiv.org/pdf/2507.23632
• Github: https://github.com/gmongaras/On-the-Expressiveness-of-Softmax-Attention-A-Recurrent-Neural-Network-Perspective
🔹 Datasets citing this paper:
No datasets found
🔹 Spaces citing this paper:
No spaces found
==================================
For more data science resources:
✓ https://news.1rj.ru/str/DataScienceT
🔹 Publication Date: Published on Jul 31
🔹 Abstract: Softmax attention is more expressive than linear attention due to its recurrent form, which can be analyzed using RNN components. AI-generated summary Since its introduction, softmax attention has become the backbone of modern transformer architectures due to its expressiveness and scalability across a wide range of tasks. However, the main drawback of softmax attention is the quadratic memory requirement and computational complexity with respect to the sequence length . By replacing the softmax nonlinearity, linear attention and similar methods have been introduced to avoid the quadratic bottleneck of softmax attention . Despite these linear forms of attention being derived from the original softmax formulation, they typically lag in terms of downstream accuracy. While strong intuition of the softmax nonlinearity on the query and key inner product suggests that it has desirable properties compared to other nonlinearities, the question of why this discrepancy exists still remains unanswered. This work demonstrates that linear attention is an approximation of softmax attention by deriving the recurrent form of softmax attention . Using this form, each part of softmax attention can be described in the language of recurrent neural networks (RNNs) . Describing softmax attention as an RNN allows for the ablation of the components of softmax attention to understand the importance of each part and how they interact. In this way, our work helps explain why softmax attention is more expressive than its counterparts.
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2507.23632
• PDF: https://arxiv.org/pdf/2507.23632
• Github: https://github.com/gmongaras/On-the-Expressiveness-of-Softmax-Attention-A-Recurrent-Neural-Network-Perspective
🔹 Datasets citing this paper:
No datasets found
🔹 Spaces citing this paper:
No spaces found
==================================
For more data science resources:
✓ https://news.1rj.ru/str/DataScienceT
❤2
🔹 Title: ReasonRank: Empowering Passage Ranking with Strong Reasoning Ability
🔹 Publication Date: Published on Aug 9
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2508.07050
• PDF: https://arxiv.org/pdf/2508.07050
• Project Page: https://github.com/8421BCD/ReasonRank
• Github: https://github.com/8421BCD/ReasonRank
🔹 Datasets citing this paper:
• https://huggingface.co/datasets/liuwenhan/reasonrank_data_13k
• https://huggingface.co/datasets/liuwenhan/reasonrank_data_rl
• https://huggingface.co/datasets/liuwenhan/reasonrank_data_sft
🔹 Spaces citing this paper:
No spaces found
==================================
For more data science resources:
✓ https://news.1rj.ru/str/DataScienceT
🔹 Publication Date: Published on Aug 9
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2508.07050
• PDF: https://arxiv.org/pdf/2508.07050
• Project Page: https://github.com/8421BCD/ReasonRank
• Github: https://github.com/8421BCD/ReasonRank
🔹 Datasets citing this paper:
• https://huggingface.co/datasets/liuwenhan/reasonrank_data_13k
• https://huggingface.co/datasets/liuwenhan/reasonrank_data_rl
• https://huggingface.co/datasets/liuwenhan/reasonrank_data_sft
🔹 Spaces citing this paper:
No spaces found
==================================
For more data science resources:
✓ https://news.1rj.ru/str/DataScienceT
🔹 Title: WideSearch: Benchmarking Agentic Broad Info-Seeking
🔹 Publication Date: Published on Aug 11
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2508.07999
• PDF: https://arxiv.org/pdf/2508.07999
• Project Page: https://widesearch-seed.github.io/
• Github: https://widesearch-seed.github.io/
🔹 Datasets citing this paper:
• https://huggingface.co/datasets/ByteDance-Seed/WideSearch
🔹 Spaces citing this paper:
No spaces found
==================================
For more data science resources:
✓ https://news.1rj.ru/str/DataScienceT
🔹 Publication Date: Published on Aug 11
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2508.07999
• PDF: https://arxiv.org/pdf/2508.07999
• Project Page: https://widesearch-seed.github.io/
• Github: https://widesearch-seed.github.io/
🔹 Datasets citing this paper:
• https://huggingface.co/datasets/ByteDance-Seed/WideSearch
🔹 Spaces citing this paper:
No spaces found
==================================
For more data science resources:
✓ https://news.1rj.ru/str/DataScienceT
❤1
🔹 Title: Omni-Effects: Unified and Spatially-Controllable Visual Effects Generation
🔹 Publication Date: Published on Aug 11
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2508.07981
• PDF: https://arxiv.org/pdf/2508.07981
• Project Page: https://amap-ml.github.io/Omni-Effects.github.io/
• Github: https://github.com/AMAP-ML/Omni-Effects
🔹 Datasets citing this paper:
• https://huggingface.co/datasets/GD-ML/Omni-VFX
🔹 Spaces citing this paper:
No spaces found
==================================
For more data science resources:
✓ https://news.1rj.ru/str/DataScienceT
🔹 Publication Date: Published on Aug 11
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2508.07981
• PDF: https://arxiv.org/pdf/2508.07981
• Project Page: https://amap-ml.github.io/Omni-Effects.github.io/
• Github: https://github.com/AMAP-ML/Omni-Effects
🔹 Datasets citing this paper:
• https://huggingface.co/datasets/GD-ML/Omni-VFX
🔹 Spaces citing this paper:
No spaces found
==================================
For more data science resources:
✓ https://news.1rj.ru/str/DataScienceT
❤1
🔹 Title: Klear-Reasoner: Advancing Reasoning Capability via Gradient-Preserving Clipping Policy Optimization
🔹 Publication Date: Published on Aug 11
🔹 Paper Links:
• arXiv Page: https://arxiv.org/pdf/2508.07629
• PDF: https://arxiv.org/pdf/2508.07629
• Project Page: https://github.com/suu990901/KlearReasoner
• Github: https://github.com/suu990901/KlearReasoner
🔹 Datasets citing this paper:
No datasets found
🔹 Spaces citing this paper:
No spaces found
==================================
For more data science resources:
✓ https://news.1rj.ru/str/DataScienceT
🔹 Publication Date: Published on Aug 11
🔹 Paper Links:
• arXiv Page: https://arxiv.org/pdf/2508.07629
• PDF: https://arxiv.org/pdf/2508.07629
• Project Page: https://github.com/suu990901/KlearReasoner
• Github: https://github.com/suu990901/KlearReasoner
🔹 Datasets citing this paper:
No datasets found
🔹 Spaces citing this paper:
No spaces found
==================================
For more data science resources:
✓ https://news.1rj.ru/str/DataScienceT
❤1
🔹 Title: UserBench: An Interactive Gym Environment for User-Centric Agents
🔹 Publication Date: Published on Jul 29
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2507.22034
• PDF: https://arxiv.org/pdf/2507.22034
• Github: https://github.com/SalesforceAIResearch/UserBench
🔹 Datasets citing this paper:
No datasets found
🔹 Spaces citing this paper:
No spaces found
==================================
For more data science resources:
✓ https://news.1rj.ru/str/DataScienceT
🔹 Publication Date: Published on Jul 29
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2507.22034
• PDF: https://arxiv.org/pdf/2507.22034
• Github: https://github.com/SalesforceAIResearch/UserBench
🔹 Datasets citing this paper:
No datasets found
🔹 Spaces citing this paper:
No spaces found
==================================
For more data science resources:
✓ https://news.1rj.ru/str/DataScienceT
❤1
🔹 Title: BrowseComp-Plus: A More Fair and Transparent Evaluation Benchmark of Deep-Research Agent
🔹 Publication Date: Published on Aug 8
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2508.06600
• PDF: https://arxiv.org/pdf/2508.06600
• Project Page: https://texttron.github.io/BrowseComp-Plus/
• Github: https://github.com/texttron/BrowseComp-Plus
🔹 Datasets citing this paper:
• https://huggingface.co/datasets/Tevatron/browsecomp-plus-corpus
• https://huggingface.co/datasets/Tevatron/browsecomp-plus
🔹 Spaces citing this paper:
• https://huggingface.co/spaces/Tevatron/BrowseComp-Plus
==================================
For more data science resources:
✓ https://news.1rj.ru/str/DataScienceT
🔹 Publication Date: Published on Aug 8
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2508.06600
• PDF: https://arxiv.org/pdf/2508.06600
• Project Page: https://texttron.github.io/BrowseComp-Plus/
• Github: https://github.com/texttron/BrowseComp-Plus
🔹 Datasets citing this paper:
• https://huggingface.co/datasets/Tevatron/browsecomp-plus-corpus
• https://huggingface.co/datasets/Tevatron/browsecomp-plus
🔹 Spaces citing this paper:
• https://huggingface.co/spaces/Tevatron/BrowseComp-Plus
==================================
For more data science resources:
✓ https://news.1rj.ru/str/DataScienceT
❤1
🔹 Title: OmniEAR: Benchmarking Agent Reasoning in Embodied Tasks
🔹 Publication Date: Published on Aug 7
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2508.05614
• PDF: https://arxiv.org/pdf/2508.05614
• Project Page: https://zju-real.github.io/OmniEmbodied/
• Github: https://zju-real.github.io/OmniEmbodied/
🔹 Datasets citing this paper:
• https://huggingface.co/datasets/wangzx1210/OmniEAR
🔹 Spaces citing this paper:
No spaces found
==================================
For more data science resources:
✓ https://news.1rj.ru/str/DataScienceT
🔹 Publication Date: Published on Aug 7
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2508.05614
• PDF: https://arxiv.org/pdf/2508.05614
• Project Page: https://zju-real.github.io/OmniEmbodied/
• Github: https://zju-real.github.io/OmniEmbodied/
🔹 Datasets citing this paper:
• https://huggingface.co/datasets/wangzx1210/OmniEAR
🔹 Spaces citing this paper:
No spaces found
==================================
For more data science resources:
✓ https://news.1rj.ru/str/DataScienceT
❤1
🔹 Title: MolmoAct: Action Reasoning Models that can Reason in Space
🔹 Publication Date: Published on Aug 11
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2508.07917
• PDF: https://arxiv.org/pdf/2508.07917
🔹 Datasets citing this paper:
No datasets found
🔹 Spaces citing this paper:
No spaces found
==================================
For more data science resources:
✓ https://news.1rj.ru/str/DataScienceT
🔹 Publication Date: Published on Aug 11
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2508.07917
• PDF: https://arxiv.org/pdf/2508.07917
🔹 Datasets citing this paper:
No datasets found
🔹 Spaces citing this paper:
No spaces found
==================================
For more data science resources:
✓ https://news.1rj.ru/str/DataScienceT
❤2
🔹 Title: Temporal Self-Rewarding Language Models: Decoupling Chosen-Rejected via Past-Future
🔹 Publication Date: Published on Aug 8
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2508.06026
• PDF: https://arxiv.org/pdf/2508.06026
🔹 Datasets citing this paper:
No datasets found
🔹 Spaces citing this paper:
No spaces found
==================================
For more data science resources:
✓ https://news.1rj.ru/str/DataScienceT
🔹 Publication Date: Published on Aug 8
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2508.06026
• PDF: https://arxiv.org/pdf/2508.06026
🔹 Datasets citing this paper:
No datasets found
🔹 Spaces citing this paper:
No spaces found
==================================
For more data science resources:
✓ https://news.1rj.ru/str/DataScienceT
❤1
🔹 Title: Reinforcement Learning in Vision: A Survey
🔹 Publication Date: Published on Aug 11
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2508.08189
• PDF: https://arxiv.org/pdf/2508.08189
• Github: https://github.com/weijiawu/Awesome-Visual-Reinforcement-Learning
🔹 Datasets citing this paper:
No datasets found
🔹 Spaces citing this paper:
No spaces found
==================================
For more data science resources:
✓ https://news.1rj.ru/str/DataScienceT
🔹 Publication Date: Published on Aug 11
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2508.08189
• PDF: https://arxiv.org/pdf/2508.08189
• Github: https://github.com/weijiawu/Awesome-Visual-Reinforcement-Learning
🔹 Datasets citing this paper:
No datasets found
🔹 Spaces citing this paper:
No spaces found
==================================
For more data science resources:
✓ https://news.1rj.ru/str/DataScienceT
❤1
🔹 Title: SONAR-LLM: Autoregressive Transformer that Thinks in Sentence Embeddings and Speaks in Tokens
🔹 Publication Date: Published on Aug 7
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2508.05305
• PDF: https://arxiv.org/pdf/2508.05305
• Github: https://github.com/FusionBrainLab/SONAR-LLM/tree/main
🔹 Datasets citing this paper:
No datasets found
🔹 Spaces citing this paper:
No spaces found
==================================
For more data science resources:
✓ https://news.1rj.ru/str/DataScienceT
🔹 Publication Date: Published on Aug 7
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2508.05305
• PDF: https://arxiv.org/pdf/2508.05305
• Github: https://github.com/FusionBrainLab/SONAR-LLM/tree/main
🔹 Datasets citing this paper:
No datasets found
🔹 Spaces citing this paper:
No spaces found
==================================
For more data science resources:
✓ https://news.1rj.ru/str/DataScienceT
❤1
🔹 Title: Part I: Tricks or Traps? A Deep Dive into RL for LLM Reasoning
🔹 Publication Date: Published on Aug 11
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2508.08221
• PDF: https://arxiv.org/pdf/2508.08221
🔹 Datasets citing this paper:
No datasets found
🔹 Spaces citing this paper:
No spaces found
==================================
For more data science resources:
✓ https://news.1rj.ru/str/DataScienceT
🔹 Publication Date: Published on Aug 11
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2508.08221
• PDF: https://arxiv.org/pdf/2508.08221
🔹 Datasets citing this paper:
No datasets found
🔹 Spaces citing this paper:
No spaces found
==================================
For more data science resources:
✓ https://news.1rj.ru/str/DataScienceT
❤1
🔹 Title: Follow-Your-Shape: Shape-Aware Image Editing via Trajectory-Guided Region Control
🔹 Publication Date: Published on Aug 11
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2508.08134
• PDF: https://arxiv.org/pdf/2508.08134
• Github: https://github.com/mayuelala/FollowYourShape
🔹 Datasets citing this paper:
No datasets found
🔹 Spaces citing this paper:
No spaces found
==================================
For more data science resources:
✓ https://news.1rj.ru/str/DataScienceT
🔹 Publication Date: Published on Aug 11
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2508.08134
• PDF: https://arxiv.org/pdf/2508.08134
• Github: https://github.com/mayuelala/FollowYourShape
🔹 Datasets citing this paper:
No datasets found
🔹 Spaces citing this paper:
No spaces found
==================================
For more data science resources:
✓ https://news.1rj.ru/str/DataScienceT
❤1
🔹 Title: Less Is More: Training-Free Sparse Attention with Global Locality for Efficient Reasoning
🔹 Publication Date: Published on Aug 9
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2508.07101
• PDF: https://arxiv.org/pdf/2508.07101
• Github: https://github.com/DerrickYLJ/LessIsMore
🔹 Datasets citing this paper:
No datasets found
🔹 Spaces citing this paper:
No spaces found
==================================
For more data science resources:
✓ https://news.1rj.ru/str/DataScienceT
🔹 Publication Date: Published on Aug 9
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2508.07101
• PDF: https://arxiv.org/pdf/2508.07101
• Github: https://github.com/DerrickYLJ/LessIsMore
🔹 Datasets citing this paper:
No datasets found
🔹 Spaces citing this paper:
No spaces found
==================================
For more data science resources:
✓ https://news.1rj.ru/str/DataScienceT
❤1
🔹 Title: VisR-Bench: An Empirical Study on Visual Retrieval-Augmented Generation for Multilingual Long Document Understanding
🔹 Publication Date: Published on Aug 10
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2508.07493
• PDF: https://arxiv.org/pdf/2508.07493
• Github: https://github.com/puar-playground/VisR-Bench
🔹 Datasets citing this paper:
• https://huggingface.co/datasets/puar-playground/VisR-Bench
🔹 Spaces citing this paper:
No spaces found
==================================
For more data science resources:
✓ https://news.1rj.ru/str/DataScienceT
🔹 Publication Date: Published on Aug 10
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2508.07493
• PDF: https://arxiv.org/pdf/2508.07493
• Github: https://github.com/puar-playground/VisR-Bench
🔹 Datasets citing this paper:
• https://huggingface.co/datasets/puar-playground/VisR-Bench
🔹 Spaces citing this paper:
No spaces found
==================================
For more data science resources:
✓ https://news.1rj.ru/str/DataScienceT
❤1
🔹 Title: GLiClass: Generalist Lightweight Model for Sequence Classification Tasks
🔹 Publication Date: Published on Aug 11
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2508.07662
• PDF: https://arxiv.org/pdf/2508.07662
🔹 Datasets citing this paper:
No datasets found
🔹 Spaces citing this paper:
No spaces found
==================================
For more data science resources:
✓ https://news.1rj.ru/str/DataScienceT
🔹 Publication Date: Published on Aug 11
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2508.07662
• PDF: https://arxiv.org/pdf/2508.07662
🔹 Datasets citing this paper:
No datasets found
🔹 Spaces citing this paper:
No spaces found
==================================
For more data science resources:
✓ https://news.1rj.ru/str/DataScienceT
❤1
🔹 Title: Deep Ignorance: Filtering Pretraining Data Builds Tamper-Resistant Safeguards into Open-Weight LLMs
🔹 Publication Date: Published on Aug 8
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2508.06601
• PDF: https://arxiv.org/pdf/2508.06601
🔹 Datasets citing this paper:
• https://huggingface.co/datasets/EleutherAI/deep-ignorance-pretraining-mix
• https://huggingface.co/datasets/EleutherAI/deep-ignorance-annealing-mix
🔹 Spaces citing this paper:
No spaces found
==================================
For more data science resources:
✓ https://news.1rj.ru/str/DataScienceT
🔹 Publication Date: Published on Aug 8
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2508.06601
• PDF: https://arxiv.org/pdf/2508.06601
🔹 Datasets citing this paper:
• https://huggingface.co/datasets/EleutherAI/deep-ignorance-pretraining-mix
• https://huggingface.co/datasets/EleutherAI/deep-ignorance-annealing-mix
🔹 Spaces citing this paper:
No spaces found
==================================
For more data science resources:
✓ https://news.1rj.ru/str/DataScienceT
❤1