ML Research Hub – Telegram
ML Research Hub
32.7K subscribers
3.99K photos
226 videos
23 files
4.29K links
Advancing research in Machine Learning – practical insights, tools, and techniques for researchers.

Admin: @HusseinSheikho || @Hussein_Sheikho
Download Telegram
🐼 PandaLM: ReProducible and Automated Language Model Assessment

Judge large language model, named PandaLM, which is trained to distinguish the superior model given several LLMs. PandaLM's focus extends beyond just the objective correctness of responses, which is the main focus of traditional evaluation datasets.

🖥 Github: https://github.com/weopenml/pandalm

📕 Paper: https://arxiv.org/abs/2306.05087v1

🔗 Dataset: https://github.com/tatsu-lab/stanford_alpaca#data-release

https://news.1rj.ru/str/DataScienceT
❤‍🔥2
This media is not supported in your browser
VIEW IN TELEGRAM
📹 Video-LLaMA: An Instruction-tuned Audio-Visual Language Model for Video Understanding

LLaMA is working on empowering large language models with video and audio understanding capability.

🖥 Github: https://github.com/damo-nlp-sg/video-llama

📕 Paper: https://arxiv.org/abs/2306.02858

Demo: https://huggingface.co/spaces/DAMO-NLP-SG/Video-LLaMA

📌 Model: https://modelscope.cn/studios/damo/video-llama/summary

https://news.1rj.ru/str/DataScienceT
❤‍🔥3👍3🏆1
A list of the best Telegram channels related to data science, programming languages, and artificial intelligence.

Join Quickly:
https://news.1rj.ru/str/addlist/8_rRW2scgfRhOTc0
❤‍🔥3
🏔️ Large Language Model for Geoscience

We introduce K2 (7B), an open-source language model trained by firstly further pretraining LLaMA on collected and cleaned geoscience literature, including geoscience open-access papers and Wikipedia pages, and secondly fine-tuning with knowledge-intensive instruction tuning data (GeoSignal).

git clone https://github.com/davendw49/k2.git
cd k2
conda env create -f k2.yml
conda activate k2


🖥 Github: https://github.com/davendw49/k2

⭐️ Demo: https://huggingface.co/daven3/k2_fp_delta

📕 Paper: https://arxiv.org/abs/2306.05064v1

🔗 Dataset: https://huggingface.co/datasets/daven3/geosignal

https://news.1rj.ru/str/DataScienceT
❤‍🔥4👍2
💲 FinGPT: Open-Source Financial Large Language Models

Unlike proprietary models, FinGPT takes a data-centric approach, providing researchers and practitioners with accessible and transparent resources to develop their FinLLMs.

🖥 Github: https://github.com/ai4finance-foundation/fingpt

⭐️ FinNLP: https://github.com/ai4finance-foundation/finnlp

📕 Paper: https://arxiv.org/abs/2306.06031v1

🔗 Project: https://ai4finance-foundation.github.io/FinNLP/

https://news.1rj.ru/str/DataScienceT
❤‍🔥4👍41
You can now download and watch all paid data science courses for free by subscribing to our new channel

https://news.1rj.ru/str/udemy13
👍2❤‍🔥1
This media is not supported in your browser
VIEW IN TELEGRAM
🧔 4DHumans: Reconstructing and Tracking Humans with Transformers

Fully "transformerized" version of a network for human mesh recovery.

🖥 Github: https://github.com/shubham-goel/4D-Humans

⭐️ Colab: https://colab.research.google.com/drive/1Ex4gE5v1bPR3evfhtG7sDHxQGsWwNwby?usp=sharing

📕 Paper: https://arxiv.org/pdf/2305.20091.pdf

🔗 Project: https://shubham-goel.github.io/4dhumans/

https://news.1rj.ru/str/DataScienceT
❤‍🔥2
Galactic: Scaling End-to-End Reinforcement Learning for Rearrangement
at 100k Steps-Per-Second

🖥 Github: https://github.com/facebookresearch/galactic

Paper: https://arxiv.org/pdf/2306.07552v1.pdf

💨 Dataset: https://paperswithcode.com/dataset/vizdoom

https://news.1rj.ru/str/DataScienceT
❤‍🔥41
Macaw-LLM: Multi-Modal Language Modeling with Image, Audio, Video, and Text Integration

Macaw-LLM is a model of its kind, bringing together state-of-the-art models for processing visual, auditory, and textual information, namely CLIP, Whisper, and LLaMA.

🖥 Github: https://github.com/lyuchenyang/macaw-llm

⭐️ Model: https://tinyurl.com/yem9m4nf

📕 Paper: https://tinyurl.com/4rsexudv

🔗 Dataset: https://github.com/lyuchenyang/Macaw-LLM/blob/main/data

https://news.1rj.ru/str/DataScienceT
👍42❤‍🔥1
Semi-supervised learning made simple with self-supervised clustering [CVPR 2023]

🖥 Github: https://github.com/pietroastolfi/suave-daino

Paper: https://arxiv.org/pdf/2306.07483v1.pdf

💨 Dataset: https://paperswithcode.com/dataset/imagenet

https://news.1rj.ru/str/DataScienceT
❤‍🔥21👍1
How do Transformers work?

All
the Transformer models mentioned above (GPT, BERT, BART, T5, etc.) have been trained as language models. This means they have been trained on large amounts of raw text in a self-supervised fashion. Self-supervised learning is a type of training in which the objective is automatically computed from the inputs of the model. That means that humans are not needed to label the data!

This type of model develops a statistical understanding of the language it has been trained on, but it’s not very useful for specific practical tasks. Because of this, the general pretrained model then goes through a process called transfer learning. During this process, the model is fine-tuned in a supervised way — that is, using human-annotated labels — on a given task

🔗 Read More

🌸 https://news.1rj.ru/str/DataScienceT
👍32❤‍🔥2