ML Research Hub – Telegram
ML Research Hub
32.7K subscribers
3.96K photos
225 videos
23 files
4.27K links
Advancing research in Machine Learning – practical insights, tools, and techniques for researchers.

Admin: @HusseinSheikho || @Hussein_Sheikho
Download Telegram
WeakTr: Exploring Plain Vision Transformer for Weakly-supervised Semantic Segmentation

🖥 Github: https://github.com/hustvl/weaktr

Paper: https://arxiv.org/abs/2304.01184v1

💨 Dataset: https://paperswithcode.com/dataset/imagenet

https://news.1rj.ru/str/DataScienceT
❤‍🔥2
This media is not supported in your browser
VIEW IN TELEGRAM
Test of Time: Instilling Video-Language Models with a Sense of Time

GPT-5 will likely have video abilities, but will it have a sense of time? Here is answer to this question in #CVPR2023 paper by student of University of Amsterdam to learn how to instil time into video-language foundation models.

Paper:
https://arxiv.org/abs/2301.02074

Code:
https://github.com/bpiyush/TestOfTime

Project Page:
https://bpiyush.github.io/testoftime-website/

https://news.1rj.ru/str/DataScienceT
❤‍🔥4
This media is not supported in your browser
VIEW IN TELEGRAM
Segment Anything

The Segment Anything Model (SAM) produces high quality object masks from input prompts such as points or boxes, and it can be used to generate masks for all objects in an image.

🖥 Github: https://github.com/facebookresearch/segment-anything

⭐️ Project: https://segment-anything.com/

Paper: https://arxiv.org/abs/2304.02643v1

💨 Dataset: https://segment-anything.com/dataset/index.html

https://news.1rj.ru/str/DataScienceT
❤‍🔥5
Media is too big
VIEW IN TELEGRAM
Painter → SegGPT: Vision Foundation Models from BAAI

SegGPT, a generalist model for segmenting everything in context.

🖥 Github: https://github.com/baaivision/painter

Paper: https://arxiv.org/abs/2304.03284v1

Demo: https://huggingface.co/spaces/BAAI/SegGPT

💨 Dataset: https://paperswithcode.com/dataset/youtube-vos

https://news.1rj.ru/str/DataScienceT
❤‍🔥3👍1
To watch paid channel content
All you have to do is subscribe to the paid channel. The paid channel includes multiple and huge programming courses, in addition to very useful books that are not available for free except in the paid channel.

To request a subnoscription: talk to @Hussein_Sheikho

Channel link: https://news.1rj.ru/str/+LnCmAFJO3tNmYjUy

⭐️⭐️⭐️⭐️⭐️⭐️⭐️⭐️⭐️⭐️⭐️

We launched a special bot some time ago to download all scientific, software and mathematics books The bot contains more than thirty million books, and new books are downloaded first, In addition to the possibility of downloading all articles and scientific papers for free

To request a subnoscription: talk to @Hussein_Sheikho
❤‍🔥2
Instruction Tuning with GPT-4

First attempt to use GPT-4 to generate instruction-following data for LLM finetuning.

🖥 Github: https://github.com/Instruction-Tuning-with-GPT-4/GPT-4-LLM

Paper: https://arxiv.org/abs/2304.03277v1

Project: https://instruction-tuning-with-gpt-4.github.io/

https://news.1rj.ru/str/DataScienceT
👍3
⚜️ OpenAGI: When LLM Meets Domain Experts

Reinforcement Learning from Task Feedback (RLTF) mechanism, which uses the task-solving result as feedback to improve the LLM's task-solving ability

git clone https://github.com/agiresearch/OpenAGI.git

🖥 Github: https://github.com/agiresearch/openagi

Paper: https://arxiv.org/pdf/2304.04370.pdf

⭐️ Dataset: https://drive.google.com/drive/folders/1AjT6y7qLIMxcmHhUBG5IE1_5SnCPR57e?usp=share_link

https://news.1rj.ru/str/DataScienceT
❤‍🔥3👍1
⭐️ Hard Patches Mining for Masked Image Modeling

We observe that the reconstruction loss can naturally be the metric of the difficulty of the pre-training task.

🖥 Github: https://github.com/haochen-wang409/hpm

Paper: https://arxiv.org/abs/2304.05919v1

⭐️ Dataset: https://paperswithcode.com/dataset/ade20k

https://news.1rj.ru/str/DataScienceT
👍3❤‍🔥2
👀SEEM: Segment Everything Everywhere All at Once

Universal, interactive multi-modal interface for any types of segmentation with ONE SINGLE MODE.

🖥 Github: https://github.com/ux-decoder/segment-everything-everywhere-all-at-once

Paper: https://arxiv.org/abs/2304.06718v1

⭐️ Dataset: https://paperswithcode.com/dataset/refcoco

https://news.1rj.ru/str/DataScienceT
❤‍🔥2
This media is not supported in your browser
VIEW IN TELEGRAM
🎨 Animated Drawings

A Method for Automatically Animating Children's Drawings of the Human Figure

🖥 Github: https://github.com/facebookresearch/AnimatedDrawings

⭐️Project: https://fairanimateddrawings.com/site/home

Paper: arxiv.org/pdf/2303.12741.pdf

https://news.1rj.ru/str/DataScienceT
❤‍🔥3👍2
​​InceptionNeXt: When Inception Meets ConvNeXt

Large-kernel convolutions, such as those employed in ConvNeXt, can improve model performance but often come at the cost of efficiency due to high memory access costs. Although reducing kernel size may increase speed, it often leads to significant performance degradation.

To address this issue, the authors propose InceptionNeXt, which decomposes large-kernel depthwise convolution into four parallel branches along the channel dimension. This new Inception depthwise convolution results in networks with high throughputs and competitive performance. For example, InceptionNeXt-T achieves 1.6x higher training throughputs than ConvNeX-T and a 0.2% top-1 accuracy improvement on ImageNet-1K. InceptionNeXt has the potential to serve as an economical baseline for future architecture design, helping to reduce carbon footprint.

A detailed unofficial overview of the paper: https://andlukyane.com/blog/paper-review-inceptionnext

Paper link:https://arxiv.org/abs/2303.16900

Code link: https://github.com/sail-sg/inceptionnext

https://news.1rj.ru/str/DataScienceT
❤‍🔥5👍2
💻 Graph classification with Transformers

This notebook shows how to fine-tune the Graphormer model for Graph Classification on a dataset available on the hub.

🤗Hugging face blog: https://huggingface.co/blog/graphml-classification

Intro to Graphs: https://news.1rj.ru/str/ai_machinelearning_big_data/3214

🖥 Github: https://github.com/huggingface/blog/blob/main/notebooks/graphml-classification.ipynb

Paper: https://arxiv.org/abs/2106.05234

⭐️Dataset: https://ogb.stanford.edu/

https://news.1rj.ru/str/DataScienceT
❤‍🔥2👍21
📝 An open, billion-scale corpus of images interleaved with text.

MultimodalC4 is a multimodal extension of c4 that interleaves millions of images with text.

🖥 Github: https://github.com/allenai/mmc4

Paper: https://arxiv.org/abs/2304.06939v1

⭐️ Dataset: https://paperswithcode.com/dataset/c4

https://news.1rj.ru/str/DataScienceT
❤‍🔥3👍1
STU-Net: Scalable and Transferable Medical Image Segmentation Models Empowered by Large-Scale Supervised Pre-training

🖥 Github: https://github.com/ziyan-huang/stu-net

Paper: https://arxiv.org/pdf/2304.06716v1.pdf

💨 Dataset: https://paperswithcode.com/dataset/abdomenct-1k

https://news.1rj.ru/str/DataScienceT
❤‍🔥5
📸 Omni Aggregation Networks for Lightweight Image Super-Resolution

Omni Self-attention paradigm for simultaneous spatial and channel interactions,mining all the potential correlations across omni-axis.

🖥 Github: https://github.com/francis0625/omni-sr

Paper: https://arxiv.org/abs/2304.10244v1

⭐️ Dataset: https://paperswithcode.com/dataset/manga109

https://news.1rj.ru/str/DataScienceT
Best Data Science Channels and groups on Telegram:

https://news.1rj.ru/str/addlist/8_rRW2scgfRhOTc0

Only click on OK and Will automatically add you to all channels

Update telegram version
🔍 Unleashing Infinite-Length Input Capacity for Large-scale Language Models with Self-Controlled Memory System

Self-Controlled Memory (SCM) system to unleash infinite-length input capacity for large-scale language models.

🖥 Github: https://github.com/toufunao/SCM4LLMs

Paper: https://arxiv.org/abs/2304.13343v1

📌 Tasks: https://paperswithcode.com/task/language-modelling

https://news.1rj.ru/str/DataScienceT
👍5❤‍🔥1