ML Research Hub – Telegram
ML Research Hub
32.7K subscribers
4.09K photos
237 videos
23 files
4.41K links
Advancing research in Machine Learning – practical insights, tools, and techniques for researchers.

Admin: @HusseinSheikho || @Hussein_Sheikho
Download Telegram
🔹 Title: STAR-Bench: Probing Deep Spatio-Temporal Reasoning as Audio 4D Intelligence

🔹 Publication Date: Published on Oct 28

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2510.24693
• PDF: https://arxiv.org/pdf/2510.24693
• Project Page: https://internlm.github.io/StarBench/
• Github: https://github.com/InternLM/StarBench

🔹 Datasets citing this paper:
No datasets found

🔹 Spaces citing this paper:
No spaces found
==================================

For more data science resources:
https://news.1rj.ru/str/DataScienceT
🔹 Title: ATLAS: Adaptive Transfer Scaling Laws for Multilingual Pretraining, Finetuning, and Decoding the Curse of Multilinguality

🔹 Publication Date: Published on Oct 24

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2510.22037
• PDF: https://arxiv.org/pdf/2510.22037

🔹 Datasets citing this paper:
No datasets found

🔹 Spaces citing this paper:
No spaces found
==================================

For more data science resources:
https://news.1rj.ru/str/DataScienceT
🔹 Title: Uniform Discrete Diffusion with Metric Path for Video Generation

🔹 Publication Date: Published on Oct 28

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2510.24717
• PDF: https://arxiv.org/pdf/2510.24717
• Github: https://github.com/baaivision/URSA

🔹 Datasets citing this paper:
No datasets found

🔹 Spaces citing this paper:
No spaces found
==================================

For more data science resources:
https://news.1rj.ru/str/DataScienceT
💡 ViT for Fashion MNIST Classification

This lesson demonstrates how to use a pre-trained Vision Transformer (ViT) to classify an image from the Fashion MNIST dataset. ViT treats an image as a sequence of patches, similar to how language models treat sentences, making it a powerful architecture for computer vision tasks. We will use a model from the Hugging Face Hub that is already fine-tuned for this specific dataset.

from transformers import ViTImageProcessor, ViTForImageClassification
from datasets import load_dataset
import torch

# 1. Load a model fine-tuned on Fashion MNIST and its processor
model_name = "abhishek/autotrain-fashion-mnist-283834433"
processor = ViTImageProcessor.from_pretrained(model_name)
model = ViTForImageClassification.from_pretrained(model_name)

# 2. Load the dataset and get a sample image
dataset = load_dataset("fashion_mnist", split="test")
image = dataset[100]['image'] # Get the 100th image

# 3. Preprocess the image and prepare it for the model
inputs = processor(images=image, return_tensors="pt")

# 4. Perform inference to get the classification logits
with torch.no_grad():
outputs = model(**inputs)
logits = outputs.logits

# 5. Get the predicted class and its label
predicted_class_idx = logits.argmax(-1).item()
predicted_class = model.config.id2label[predicted_class_idx]

print(f"Image is a: {dataset[100]['label']}")
print(f"Model predicted: {predicted_class}")


Code explanation: This noscript uses the transformers library to load a ViT model specifically fine-tuned for Fashion MNIST classification. It then loads the dataset, selects a single sample image, and uses the model's processor to convert it into the correct input format. The model performs inference, and the noscript identifies the most likely class from the output logits, printing the final human-readable prediction.

#Python #MachineLearning #ViT #ComputerVision #HuggingFace

━━━━━━━━━━━━━━━
By: @DataScienceT
💡 ViT for Fashion MNIST Classification

This lesson demonstrates how to use a pre-trained Vision Transformer (ViT) to classify an image from the Fashion MNIST dataset. ViT treats an image as a sequence of patches, similar to how language models treat sentences, making it a powerful architecture for computer vision tasks. We will use a model from the Hugging Face Hub that is already fine-tuned for this specific dataset.

from transformers import ViTImageProcessor, ViTForImageClassification
from datasets import load_dataset
import torch

# 1. Load a model fine-tuned on Fashion MNIST and its processor
model_name = "abhishek/autotrain-fashion-mnist-283834433"
processor = ViTImageProcessor.from_pretrained(model_name)
model = ViTForImageClassification.from_pretrained(model_name)

# 2. Load the dataset and get a sample image
dataset = load_dataset("fashion_mnist", split="test")
image = dataset[100]['image'] # Get the 100th image

# 3. Preprocess the image and prepare it for the model
inputs = processor(images=image, return_tensors="pt")

# 4. Perform inference to get the classification logits
with torch.no_grad():
outputs = model(**inputs)
logits = outputs.logits

# 5. Get the predicted class and its label
predicted_class_idx = logits.argmax(-1).item()
predicted_class = model.config.id2label[predicted_class_idx]

print(f"Image is a: {dataset[100]['label']}")
print(f"Model predicted: {predicted_class}")


Code explanation: This noscript uses the transformers library to load a ViT model specifically fine-tuned for Fashion MNIST classification. It then loads the dataset, selects a single sample image, and uses the model's processor to convert it into the correct input format. The model performs inference, and the noscript identifies the most likely class from the output logits, printing the final human-readable prediction.

#Python #MachineLearning #ViT #ComputerVision #HuggingFace

━━━━━━━━━━━━━━━
By: @DataScienceT
🔹 Title: PartNeXt: A Next-Generation Dataset for Fine-Grained and Hierarchical 3D Part Understanding

🔹 Publication Date: Published on Oct 23

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2510.20155
• PDF: https://arxiv.org/pdf/2510.20155
• Project Page: https://authoritywang.github.io/partnext/
• Github: https://github.com/AuthorityWang/PartNeXt

🔹 Datasets citing this paper:
No datasets found

🔹 Spaces citing this paper:
No spaces found
==================================

For more data science resources:
https://news.1rj.ru/str/DataScienceT
🔹 Title: VisJudge-Bench: Aesthetics and Quality Assessment of Visualizations

🔹 Publication Date: Published on Oct 25

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2510.22373
• PDF: https://arxiv.org/pdf/2510.22373
• Github: https://github.com/HKUSTDial/VisJudgeBench

🔹 Datasets citing this paper:
No datasets found

🔹 Spaces citing this paper:
No spaces found
==================================

For more data science resources:
https://news.1rj.ru/str/DataScienceT
🔹 Title: PatenTEB: A Comprehensive Benchmark and Model Family for Patent Text Embedding

🔹 Publication Date: Published on Oct 25

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2510.22264
• PDF: https://arxiv.org/pdf/2510.22264
• Github: https://github.com/iliass-y/patenteb

🔹 Datasets citing this paper:
https://huggingface.co/datasets/datalyes/class_bloom
https://huggingface.co/datasets/datalyes/class_nli_oldnew
https://huggingface.co/datasets/datalyes/clusters_ext_full_ipc
https://huggingface.co/datasets/datalyes/class_text2ipc3

🔹 Spaces citing this paper:
No spaces found
==================================

For more data science resources:
https://news.1rj.ru/str/DataScienceT
🔹 Title: From Spatial to Actions: Grounding Vision-Language-Action Model in Spatial Foundation Priors

🔹 Publication Date: Published on Oct 20

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2510.17439
• PDF: https://arxiv.org/pdf/2510.17439
• Github: https://falcon-vla.github.io/

🔹 Datasets citing this paper:
No datasets found

🔹 Spaces citing this paper:
No spaces found
==================================

For more data science resources:
https://news.1rj.ru/str/DataScienceT
🔹 Title: UltraHR-100K: Enhancing UHR Image Synthesis with A Large-Scale High-Quality Dataset

🔹 Publication Date: Published on Oct 23

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2510.20661
• PDF: https://arxiv.org/pdf/2510.20661

🔹 Datasets citing this paper:
No datasets found

🔹 Spaces citing this paper:
No spaces found
==================================

For more data science resources:
https://news.1rj.ru/str/DataScienceT
⚡️ OpenAI presents gpt-oss-safeguard - open models for flexible and transparent content moderation

Instead of a rigidly trained classifier, the model takes your own security policy as input and reasons whether the message complies with this policy.

The result is not just "safe/unsafe," but a chain of reasoning that you can verify and improve.

The models are available in two sizes: 120B and 20B.
• gpt-oss-safeguard-120B
• gpt-oss-safeguard-20B

💡 Why they are needed:
• Policies can be changed without retraining the model
• Suitable for niche or rapidly changing risks (e.g., cheating in games or fake reviews)
• Does not require thousands of labeled examples
• Ideal when explainability is important rather than minimal latency

Both are available under the Apache 2.0 license - they can be freely used, modified, and deployed.

🔗 Official announcement
🤗 Hugging Face

#openai #chatgpt #opensource
Please open Telegram to view this post
VIEW IN TELEGRAM
Please open Telegram to view this post
VIEW IN TELEGRAM
🔹 Title: Optimize Any Topology: A Foundation Model for Shape- and Resolution-Free Structural Topology Optimization

🔹 Publication Date: Published on Oct 26

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2510.23667
• PDF: https://arxiv.org/pdf/2510.23667

🔹 Datasets citing this paper:
No datasets found

🔹 Spaces citing this paper:
No spaces found
==================================

For more data science resources:
https://news.1rj.ru/str/DataScienceT
🔹 Title: Agent Data Protocol: Unifying Datasets for Diverse, Effective Fine-tuning of LLM Agents

🔹 Publication Date: Published on Oct 28

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2510.24702
• PDF: https://arxiv.org/pdf/2510.24702
• Project Page: https://agentdataprotocol.com/
• Github: https://github.com/neulab/agent-data-protocol

🔹 Datasets citing this paper:
No datasets found

🔹 Spaces citing this paper:
No spaces found
==================================

For more data science resources:
https://news.1rj.ru/str/DataScienceT
🔹 Title: SPICE: Self-Play In Corpus Environments Improves Reasoning

🔹 Publication Date: Published on Oct 28

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2510.24684
• PDF: https://arxiv.org/pdf/2510.24684

🔹 Datasets citing this paper:
No datasets found

🔹 Spaces citing this paper:
No spaces found
==================================

For more data science resources:
https://news.1rj.ru/str/DataScienceT
🔹 Title: SAO-Instruct: Free-form Audio Editing using Natural Language Instructions

🔹 Publication Date: Published on Oct 26

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2510.22795
• PDF: https://arxiv.org/pdf/2510.22795
• Project Page: https://eth-disco.github.io/sao-instruct/
• Github: https://github.com/eth-disco/sao-instruct

🔹 Datasets citing this paper:
No datasets found

🔹 Spaces citing this paper:
No spaces found
==================================

For more data science resources:
https://news.1rj.ru/str/DataScienceT
🔹 Title: GRPO-Guard: Mitigating Implicit Over-Optimization in Flow Matching via Regulated Clipping

🔹 Publication Date: Published on Oct 25

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2510.22319
• PDF: https://arxiv.org/pdf/2510.22319
• Project Page: https://jingw193.github.io/GRPO-Guard

🔹 Datasets citing this paper:
No datasets found

🔹 Spaces citing this paper:
No spaces found
==================================

For more data science resources:
https://news.1rj.ru/str/DataScienceT
🔹 Title: MMPersuade: A Dataset and Evaluation Framework for Multimodal Persuasion

🔹 Publication Date: Published on Oct 26

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2510.22768
• PDF: https://arxiv.org/pdf/2510.22768

🔹 Datasets citing this paper:
No datasets found

🔹 Spaces citing this paper:
No spaces found
==================================

For more data science resources:
https://news.1rj.ru/str/DataScienceT
🔹 Title: Global PIQA: Evaluating Physical Commonsense Reasoning Across 100+ Languages and Cultures

🔹 Publication Date: Published on Oct 28

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2510.24081
• PDF: https://arxiv.org/pdf/2510.24081

🔹 Datasets citing this paper:
https://huggingface.co/datasets/mrlbenchmarks/global-piqa-nonparallel

🔹 Spaces citing this paper:
No spaces found
==================================

For more data science resources:
https://news.1rj.ru/str/DataScienceT
🔹 Title: Beyond Reasoning Gains: Mitigating General Capabilities Forgetting in Large Reasoning Models

🔹 Publication Date: Published on Oct 24

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2510.21978
• PDF: https://arxiv.org/pdf/2510.21978

🔹 Datasets citing this paper:
No datasets found

🔹 Spaces citing this paper:
No spaces found
==================================

For more data science resources:
https://news.1rj.ru/str/DataScienceT