ML Research Hub – Telegram
ML Research Hub
32.7K subscribers
4.03K photos
230 videos
23 files
4.34K links
Advancing research in Machine Learning – practical insights, tools, and techniques for researchers.

Admin: @HusseinSheikho || @Hussein_Sheikho
Download Telegram
🔹 Title: A Survey on Diffusion Language Models

🔹 Publication Date: Published on Aug 14

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2508.10875
• PDF: https://arxiv.org/pdf/2508.10875

🔹 Datasets citing this paper:
No datasets found

🔹 Spaces citing this paper:
No spaces found
==================================

For more data science resources:
https://news.1rj.ru/str/DataScienceT
🔹 Title: Processing and acquisition traces in visual encoders: What does CLIP know about your camera?

🔹 Publication Date: Published on Aug 14

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2508.10637
• PDF: https://arxiv.org/pdf/2508.10637
• Github: https://github.com/ryan-caesar-ramos/visual-encoder-traces

🔹 Datasets citing this paper:
No datasets found

🔹 Spaces citing this paper:
No spaces found
==================================

For more data science resources:
https://news.1rj.ru/str/DataScienceT
🔹 Title: STream3R: Scalable Sequential 3D Reconstruction with Causal Transformer

🔹 Publication Date: Published on Aug 14

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2508.10893
• PDF: https://arxiv.org/pdf/2508.10893
• Project Page: https://nirvanalan.github.io/projects/stream3r
• Github: https://github.com/NIRVANALAN/STream3R

🔹 Datasets citing this paper:
No datasets found

🔹 Spaces citing this paper:
No spaces found
==================================

For more data science resources:
https://news.1rj.ru/str/DataScienceT
🔹 Title: When Explainability Meets Privacy: An Investigation at the Intersection of Post-hoc Explainability and Differential Privacy in the Context of Natural Language Processing

🔹 Publication Date: Published on Aug 14

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2508.10482
• PDF: https://arxiv.org/pdf/2508.10482
• Github: https://github.com/dmah10/xpnlp

🔹 Datasets citing this paper:
No datasets found

🔹 Spaces citing this paper:
No spaces found
==================================

For more data science resources:
https://news.1rj.ru/str/DataScienceT
3
🔹 Title: Artificial Intelligence and Misinformation in Art: Can Vision Language Models Judge the Hand or the Machine Behind the Canvas?

🔹 Publication Date: Published on Aug 2

🔹 Abstract: State-of-the-art vision language models struggle with accurately attributing artists and distinguishing AI-generated images, highlighting the need for improvement to prevent misinformation. AI-generated summary The attribution of artworks in general and of paintings in particular has always been an issue in art. The advent of powerful artificial intelligence models that can generate and analyze images creates new challenges for painting attribution. On the one hand, AI models can create images that mimic the style of a painter, which can be incorrectly attributed, for example, by other AI models. On the other hand, AI models may not be able to correctly identify the artist for real paintings, inducing users to incorrectly attribute paintings. In this paper, both problems are experimentally studied using state-of-the-art AI models for image generation and analysis on a large dataset with close to 40,000 paintings from 128 artists. The results show that vision language models have limited capabilities to: 1) perform canvas attribution and 2) to identify AI generated images . As users increasingly rely on queries to AI models to get information, these results show the need to improve the capabilities of VLMs to reliably perform artist attribution and detection of AI generated images to prevent the spread of incorrect information.

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2508.01408

• PDF: https://arxiv.org/pdf/2508.01408

• Github: https://ama2210.github.io/WikiArt_VLM_Web/

🔹 Datasets citing this paper:
No datasets found

🔹 Spaces citing this paper:
No spaces found
==================================

For more data science resources:
https://news.1rj.ru/str/DataScienceT
🔹 Title: Puppeteer: Rig and Animate Your 3D Models

🔹 Publication Date: Published on Aug 14

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2508.10898
• PDF: https://arxiv.org/pdf/2508.10898
• Project Page: https://chaoyuesong.github.io/Puppeteer/
• Github: https://github.com/Seed3D/Puppeteer

🔹 Datasets citing this paper:
No datasets found

🔹 Spaces citing this paper:
No spaces found
==================================

For more data science resources:
https://news.1rj.ru/str/DataScienceT
2
🔹 Title: We-Math 2.0: A Versatile MathBook System for Incentivizing Visual Mathematical Reasoning

🔹 Publication Date: Published on Aug 14

🔹 Paper Links:
• arXiv Page: https://arxivexplained.com/papers/we-math-20-a-versatile-mathbook-system-for-incentivizing-visual-mathematical-reasoning
• PDF: https://arxiv.org/pdf/2508.10433
• Project Page: https://we-math2.github.io/
• Github: https://github.com/We-Math/We-Math2.0

🔹 Datasets citing this paper:
https://huggingface.co/datasets/We-Math/We-Math2.0-Pro
https://huggingface.co/datasets/We-Math/We-Math2.0-Standard

🔹 Spaces citing this paper:
No spaces found
==================================

For more data science resources:
https://news.1rj.ru/str/DataScienceT
4
🔹 Title: ChartCap: Mitigating Hallucination of Dense Chart Captioning

🔹 Publication Date: Published on Aug 5

🔹 Abstract: ChartCap, a large-scale dataset with dense, type-specific captions for real-world charts, improves caption accuracy and reduces hallucinations in vision language models. AI-generated summary Generating accurate, informative, and hallucination-free captions for charts remains challenging for vision language models , primarily due to the lack of large-scale, high-quality datasets of real-world charts . However, existing real-world chart datasets suffer from the inclusion of extraneous information that cannot be inferred from the chart and failure to sufficiently capture structural elements and key insights . Therefore, we introduce ChartCap, a large-scale dataset of 565K real-world chart images paired with type-specific, dense captions that exclude extraneous information and highlight both structural elements and key insights in detail. To build ChartCap, we design a four-stage pipeline that generates captions using only the discernible data from the chart and employ a cycle consistency-based human verification , which accelerates quality control without sacrificing accuracy. Additionally, we propose a novel metric, the Visual Consistency Score , which evaluates caption quality by measuring the similarity between the chart regenerated from a caption and the original chart, independent of reference captions. Extensive experiments confirms that models fine-tuned on ChartCap consistently generate more accurate and informative captions with reduced hallucinations , surpassing both open-source and proprietary models and even human-annotated captions.

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2508.03164

• PDF: https://arxiv.org/pdf/2508.03164

• Project Page: https://junyoung-00.github.io/ChartCap/

• Github: https://junyoung-00.github.io/ChartCap/

🔹 Datasets citing this paper:
https://huggingface.co/datasets/junyoung-00/ChartCap

🔹 Spaces citing this paper:
No spaces found
==================================

For more data science resources:
https://news.1rj.ru/str/DataScienceT
🔹 Title: AlignGuard-LoRA: Alignment-Preserving Fine-Tuning via Fisher-Guided Decomposition and Riemannian-Geodesic Collision Regularization

🔹 Publication Date: Published on Aug 4

🔹 Abstract: AlignGuard-LoRA (AGL) is a framework that preserves alignment during fine-tuning of large language models by introducing regularization techniques and a diagnostic benchmark to mitigate alignment drift. AI-generated summary Low-rank adaptation ( LoRA ) has become a standard tool for efficiently fine-tuning large language models (LLMs). Yet, even minor LoRA updates can induce alignment drift , weakening safety and behavioral constraints through entangled parameter changes. To address this, we propose AlignGuard-LoRA (AGL), a principled framework for preserving alignment during finetuning. AGL introduces several key components: a primary task loss for supervision, Fisher Information Matrix-based regularization to restrict updates in alignment-sensitive subspaces, and task-specific regularization to stabilize the integration of new knowledge. We further introduce collision-aware regularization, blending Riemannian overlap -- which penalizes coordinate-wise interference -- and geodesic separation -- which encourages disjoint update geometry. We curate DriftCaps , a targeted diagnostic benchmark of safe and unsafe prompts designed to quantify alignment drift and safety degradation. Empirical evaluations show that AGL mitigates alignment drift by up to 50% on safety-critical benchmarks without degrading downstream task performance. Comprehensive ablation confirms that each component contributes distinctly to preserving latent safety behaviors. Finally, we derive and validate a scaling law for catastrophic forgetting , revealing that AGL flattens post-finetuning loss escalation while preserving adaptation dynamics . AGL is a structurally grounded refinement of LoRA , ensuring alignment preservation with minimal trade-offs. To encourage further exp lora tion and development, we open-source our implementation.

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2508.02079

• PDF: https://arxiv.org/pdf/2508.02079

🔹 Datasets citing this paper:
No datasets found

🔹 Spaces citing this paper:
No spaces found
==================================

For more data science resources:
https://news.1rj.ru/str/DataScienceT
🔹 Title: Uncertainty-Based Methods for Automated Process Reward Data Construction and Output Aggregation in Mathematical Reasoning

🔹 Publication Date: Published on Aug 3

🔹 Abstract: An uncertainty-driven framework for automated process reward data construction and aggregation methods improves the effectiveness and efficiency of Process-Level Reward Models in mathematical reasoning tasks. AI-generated summary Large language models have demonstrated remarkable capabilities in complex math ematical reasoning tasks, but they inevitably generate errors throughout multi-step solutions. Process-level Reward Models ( PRMs ) have shown great promise by providing supervision and evaluation at each intermediate step, thereby effectively improving the models' reasoning abilities. However, training effective PRMs requires high-quality process reward data, yet existing methods for constructing such data are often labour-intensive or inefficient. In this paper, we propose an uncertainty-driven framework for automated process reward data construction, encompassing both data generation and annotation processes for PRMs . Additionally, we identify the limitations of both majority vote and PRMs , and introduce two generic uncertainty-aware output aggregation methods: Hybrid Majority Reward Vote and Weighted Reward Frequency Vote , which combine the strengths of majority vote with PRMs . Extensive experiments on ProcessBench , MATH , and GSMPlus show the effectiveness and efficiency of the proposed PRM data construction framework, and demonstrate that the two output aggregation methods further improve the math ematical reasoning abilities across diverse PRMs . The code and data will be publicly available at https://github.com/Jiuzhouh/UnPRM.

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2508.01773

• PDF: https://arxiv.org/pdf/2508.01773

• Github: https://github.com/Jiuzhouh/UnPRM

🔹 Datasets citing this paper:
No datasets found

🔹 Spaces citing this paper:
No spaces found
==================================

For more data science resources:
https://news.1rj.ru/str/DataScienceT
1
This channels is for Programmers, Coders, Software Engineers.

0️⃣ Python
1️⃣ Data Science
2️⃣ Machine Learning
3️⃣ Data Visualization
4️⃣ Artificial Intelligence
5️⃣ Data Analysis
6️⃣ Statistics
7️⃣ Deep Learning
8️⃣ programming Languages

https://news.1rj.ru/str/addlist/8_rRW2scgfRhOTc0

https://news.1rj.ru/str/Codeprogrammer
Please open Telegram to view this post
VIEW IN TELEGRAM
🔹 Title: Thyme: Think Beyond Images

🔹 Publication Date: Published on Aug 15

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2508.11630
• PDF: https://arxiv.org/pdf/2508.11630
• Project Page: https://thyme-vl.github.io/
• Github: https://github.com/yfzhang114/Thyme

🔹 Datasets citing this paper:
https://huggingface.co/datasets/Kwai-Keye/Thyme-RL
https://huggingface.co/datasets/Kwai-Keye/Thyme-SFT

🔹 Spaces citing this paper:
No spaces found
==================================

For more data science resources:
https://news.1rj.ru/str/DataScienceT
2
🔹 Title: PaperRegister: Boosting Flexible-grained Paper Search via Hierarchical Register Indexing

🔹 Publication Date: Published on Aug 14

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2508.11116
• PDF: https://arxiv.org/pdf/2508.11116
• Github: https://github.com/Li-Z-Q/PaperRegister

🔹 Datasets citing this paper:
No datasets found

🔹 Spaces citing this paper:
No spaces found
==================================

For more data science resources:
https://news.1rj.ru/str/DataScienceT
2
🔹 Title: StyleMM: Stylized 3D Morphable Face Model via Text-Driven Aligned Image Translation

🔹 Publication Date: Published on Aug 15

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2508.11203
• PDF: https://arxiv.org/pdf/2508.11203
• Github: https://kwanyun.github.io/stylemm_page/

🔹 Datasets citing this paper:
No datasets found

🔹 Spaces citing this paper:
No spaces found
==================================

For more data science resources:
https://news.1rj.ru/str/DataScienceT
2
🔹 Title: SPARSE Data, Rich Results: Few-Shot Semi-Supervised Learning via Class-Conditioned Image Translation

🔹 Publication Date: Published on Aug 8

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2508.06429
• PDF: https://arxiv.org/pdf/2508.06429
• Github: https://github.com/GuidoManni/SPARSE

🔹 Datasets citing this paper:
No datasets found

🔹 Spaces citing this paper:
No spaces found
==================================

For more data science resources:
https://news.1rj.ru/str/DataScienceT
4
🔹 Title: Controlling Multimodal LLMs via Reward-guided Decoding

🔹 Publication Date: Published on Aug 15

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2508.11616
• PDF: https://arxiv.org/pdf/2508.11616

🔹 Datasets citing this paper:
No datasets found

🔹 Spaces citing this paper:
No spaces found
==================================

For more data science resources:
https://news.1rj.ru/str/DataScienceT
2
🔹 Title: FantasyTalking2: Timestep-Layer Adaptive Preference Optimization for Audio-Driven Portrait Animation

🔹 Publication Date: Published on Aug 15

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2508.11255
• PDF: https://arxiv.org/pdf/2508.11255
• Project Page: https://fantasy-amap.github.io/fantasy-talking2/
• Github: https://fantasy-amap.github.io/fantasy-talking2/

🔹 Datasets citing this paper:
No datasets found

🔹 Spaces citing this paper:
No spaces found
==================================

For more data science resources:
https://news.1rj.ru/str/DataScienceT
🔹 Title: TexVerse: A Universe of 3D Objects with High-Resolution Textures

🔹 Publication Date: Published on Aug 14

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2508.10868
• PDF: https://arxiv.org/pdf/2508.10868
• Github: https://github.com/yiboz2001/TexVerse

🔹 Datasets citing this paper:
https://huggingface.co/datasets/YiboZhang2001/TexVerse

🔹 Spaces citing this paper:
No spaces found
==================================

For more data science resources:
https://news.1rj.ru/str/DataScienceT
3
🔹 Title: X-Node: Self-Explanation is All We Need

🔹 Publication Date: Published on Aug 14

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2508.10461
• PDF: https://arxiv.org/pdf/2508.10461
• Github: https://github.com/basiralab/X-Node

🔹 Datasets citing this paper:
No datasets found

🔹 Spaces citing this paper:
No spaces found
==================================

For more data science resources:
https://news.1rj.ru/str/DataScienceT
🔹 Title: MAESTRO: Masked AutoEncoders for Multimodal, Multitemporal, and Multispectral Earth Observation Data

🔹 Publication Date: Published on Aug 14

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2508.10894
• PDF: https://arxiv.org/pdf/2508.10894
• Github: https://github.com/IGNF/MAESTRO

🔹 Datasets citing this paper:
No datasets found

🔹 Spaces citing this paper:
No spaces found
==================================

For more data science resources:
https://news.1rj.ru/str/DataScienceT
🔹 Title: SSRL: Self-Search Reinforcement Learning

🔹 Publication Date: Published on Aug 14

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2508.10874
• PDF: https://arxiv.org/pdf/2508.10874
• Project Page: https://huggingface.co/collections/TsinghuaC3I/ssrl-6899957a64d4a31f7f43bc88
• Github: https://github.com/TsinghuaC3I/SSRL

🔹 Datasets citing this paper:
No datasets found

🔹 Spaces citing this paper:
No spaces found
==================================

For more data science resources:
https://news.1rj.ru/str/DataScienceT
1