ML Research Hub – Telegram
ML Research Hub
32.7K subscribers
4K photos
228 videos
23 files
4.31K links
Advancing research in Machine Learning – practical insights, tools, and techniques for researchers.

Admin: @HusseinSheikho || @Hussein_Sheikho
Download Telegram
Efficiently Democratizing Medical LLMs for 50 Languages via a Mixture of Language Family Experts

💻 Github: https://github.com/freedomintelligence/apollomoe

🔖 Paper: https://arxiv.org/abs/2410.10626v1

🤗 Dataset: https://paperswithcode.com/dataset/mmlu

https://news.1rj.ru/str/DataScienceT 🏵
Please open Telegram to view this post
VIEW IN TELEGRAM
👍21
LOOKING FOR A NEW SOURCE OF INCOME?
Average earnings from 100$ a day

Lisa is looking for people who want to earn money. If you are responsible, motivated and want to change your life. Welcome to her channel.

WHAT YOU NEED TO WORK:
1. phone or computer
2. Free 15-20 minutes a day
3. desire to earn

❗️ Requires 20 people ❗️
Access is available at the link below
👇

https://news.1rj.ru/str/+NhwYZAXFlT8yZDIx
1👍1
Please open Telegram to view this post
VIEW IN TELEGRAM
👍3
Please open Telegram to view this post
VIEW IN TELEGRAM
👍52
Forwarded from Tomas
This media is not supported in your browser
VIEW IN TELEGRAM
🚨With me you will make money! I have made over $20,000 in the last week! 🔥

I don't care where you are and what you can do, I will help absolutely everyone earn money.

My name is Lisa and:
✔️ I will teach you trading for FREE in a short period of time
✔️ I will give you FREE signals every day
✔️ I will help you to get income of 1,000$ in a week

Sounds unbelievable?

You have 2 hours to join our channel.

But it’s true - just look at the results in my channel and JOIN FOR FREE 👉🏻 https://news.1rj.ru/str/+fJ0XM3sZkaxkNjgx
👍1
SAM2Long: Enhancing SAM 2 for Long Video Segmentation with a Training-Free Memory Tree

🖥 Github: https://github.com/mark12ding/sam2long

📕 Paper: https://arxiv.org/abs/2410.16268v1

🤗 HF: https://huggingface.co/papers/2410.16268
👍4
📖 LLM-Agent-Paper-List is a repository of papers on the topic of agents based on large language models (LLM)! The papers are divided into categories such as LLM agent architectures, autonomous LLM agents, reinforcement learning (RL), natural language processing methods, multimodal approaches and tools for developing LLM agents, and more.

🖥 Github

https://news.1rj.ru/str/DataScienceT
Please open Telegram to view this post
VIEW IN TELEGRAM
👍4
Don’t sleep on Vision Language Models (VLMs).

With the releases of Llama 3.2 and ColQwen2, multimodal models are gaining more and more traction.

VLMs are multimodal models that can handle image and text modalities:

Input: Image and text
Output: Text

They can be used for many use cases, including visual question answering or document understanding (as in the case of ColQwen2).

How do they work under the hood?

The main challenge in VLMs is to unify the image and text representations.

For this, a typical VLM architecture consists of the following components:

• image encoder (e.g., CLIP, SigLIP)
• embedding projector to align image and text representations
• text decoder (e.g., Vicuna, Gemma)

huggingface.co/blog/vlms

https://news.1rj.ru/str/DataScienceT
👍62
🔦 Biggest Sale Of The Year NOW ON 🔦Double 11 Shopping Festival Event is live! Check out your most loved for less. 🛍️

Enjoy SPOTO Double 11 Crazy Sale to Join Lucky Draw and win gifts worth up to $1000!💸
🎁⏯️: https://www.spotoexam.com/snsdouble11sale2024/?id=snstxrbzhussein

🔗📝Test Your IT Skills for Free: https://bit.ly/48q8Cb3

🔗📲Contact for 1v1 IT Certs Exam Help: https://wa.link/k0vy3x
🌐📚 JOIN IT Study GROUP to Get Madness Discount 👇: https://chat.whatsapp.com/HqzBlMaOPci0wYvkEtcCDa
👍2
Forwarded from Data Science Library
🎁 Your balance is credited $4,000 , the owner of the channel wants to contact you!

Dear subscriber, we would like to thank you very much for supporting our channel, and as a token of our gratitude we would like to provide you with free access to Lisa's investor channel, with the help of which you can earn today

T.me/Lisainvestor

Be sure to take advantage of our gift, admission is free, don't miss the opportunity, change your life for the better.

You can follow the link :
https://news.1rj.ru/str/+j4-NLonPlWJmZDVh
👍1
Constrained Diffusion Implicit Models!

We use diffusion models to solve noisy inverse problems like inpainting, sparse-recovery, and colorization. 10-50x faster than previous methods!

Paper: arxiv.org/pdf/2411.00359

Demo: https://t.co/m6o9GLnnZF

https://news.1rj.ru/str/DataScienceT
👍6
A promising digital wallet will distribute $40 for free to every user who creates an account on this wallet

Terms of creating an account: Subscribe to their channel only.

https://news.1rj.ru/str/TronKeeperBot/app?startapp=418788114
📌 Practical exercises and additional materials for the book "Build a Large Language Model (From Scratch)"

A Github repository with practical exercises, notebooks with code for developing, pre-training, and fine-tuning a GPT-type LLM model based on one of the best books on building an LLM from scratch.

▶️ About the book:
In this book, you will learn and understand how large language models work from the inside, creating your own LLM step by step, with a detailed explanation of each stage in clear language, diagrams and examples.

The method described in the book demonstrates the approach used to create large fundamental models such as those underlying ChatGPT.

In the repository, each chapter of the book has several (3-4) applied examples in ipynb format or as an executable python noscript. The code is aimed at a wide audience, is designed to run on regular laptops and does not require specialized equipment.

▶️ The main value of the repository is additional practical materials that will help you to study in more depth the subtleties and nuances of the process of setting up and learning LLM:

Setting

🟢 Tips on Setting Up Python
🟢 Installing Python Packages and Libraries
🟢 Docker Environment Setup Guide

Chapter 2: Working with Text Data

🟠 Comparison of different implementations of Byte Pair Encoding (BPE)
🟠 Understanding the difference between embedding and line layers
🟠 Dataloader Intuition with Prime Numbers

Chapter 3: Code of Attention Mechanisms

🟢 Comparison of Effective Implementations of Multi-Head Attention
🟢 PyTorch Buffers

Chapter 4: Implementing the GPT Model from Scratch

🟠 FLOPS Analysis

Chapter 5: Pre-training on unlabeled data

🟢 Alternative Loading of HuggingFace Scales Using Transformers
🟢 Pre-training GPT on the Project Gutenberg dataset
🟢 Adding more features to the learning cycle
🟢 Hyperparameter optimization for pretraining
🟢 Creating a user interface for interacting with LLM
🟢 Convert GPT to Llama
🟢 Llama 3.2 from scratch
🟢 Memory-efficient model loading

Chapter 6: Fine-tuning for Classification

🟠 More experiments on fine-tuning the different layers and using larger models
🟠 Fine-tuning various models based on a 50K row IMDB movie review dataset.
🟠 Building a User Interface for Interacting with a GPT-Based Spam Classifier

Chapter 7: Fine-tuning to Follow Instructions

🟢 Dataset utilities for finding close duplicates and creating passive voice entries
🟢 Evaluating responses to instructions using OpenAI and Ollama APIs
🟢 Creating a dataset for fine-tuning instructions
🟢 Improving the dataset for fine-tuning instructions
🟢 Creating a Preference Dataset with Llama 3.1 70B and Ollama
🟢 DPO for LLM Alignment procedure
🟢 Creating a user interface for interacting with a GPT model with fine-tuning of instructions

🖥 Github

https://news.1rj.ru/str/DataScienceT
Please open Telegram to view this post
VIEW IN TELEGRAM
Please open Telegram to view this post
VIEW IN TELEGRAM
👍10
Please open Telegram to view this post
VIEW IN TELEGRAM
👍1
Please open Telegram to view this post
VIEW IN TELEGRAM
1👍1