ML Research Hub – Telegram
ML Research Hub
32.7K subscribers
4.01K photos
229 videos
23 files
4.32K links
Advancing research in Machine Learning – practical insights, tools, and techniques for researchers.

Admin: @HusseinSheikho || @Hussein_Sheikho
Download Telegram
Forwarded from Data Science Premium (Books & Courses)
🟢 We present to you a list of our paid services:

Paid Channel (Books): It includes a huge encyclopedia of free channels and a huge collection of important and rare books

💵 Price: 7$ - one time payment

Paid channel (courses): includes a complete set of courses downloaded from Udemy, Coursera, and other learning platforms.

💵 Price: 20$ - one time payment

Paid bot 1: A bot that contains ten million books and articles in all fields of programming, data science, life sciences, and medicine, with daily updates and additions.

💵 Price: 10$ - one time payment

Paid bot 2: The bot contains more than 30 million books and 10 million scientific articles, with your own account, your own control panel, and important features.

💵 Price: 17$ - one time payment

Coursera Scholarship: This scholarship gives you free access to all Coursera courses

💵 Price: 30$ - one time payment

🔥 Special offer: Get all four services for only $50

💠💠 Available payment methods:
PayPal - Payeer - Crypto - udst
MasterCard - Credit Card

To request a subnoscription:
t.me/Hussein_Sheikho
Please open Telegram to view this post
VIEW IN TELEGRAM
This media is not supported in your browser
VIEW IN TELEGRAM
🆒 EAGLE is a method that allows you to generate LLM responses faster

Is it possible to generate LLM response on two RTX 3060s faster than on A100 (which is 16+ times more expensive)?
Yes, it is possible with EAGLE (Extrapolation Algorithm for Greater Language-model Efficiency) and the accuracy of the responses is preserved.

EAGLE allows you to extrapolate the context feature vectors of the second top layer of the LLM, which greatly improves the generation efficiency.

EAGLE is 2x faster than Lookahead (13B), and 1.6x faster than Medusa (13B).
And yes, EAGLE can be combined with other acceleration techniques like vLLM, DeepSpeed, Mamba, FlashAttention, quantization and hardware optimization.

🤗 Hugging Face
💻 GitHub
Please open Telegram to view this post
VIEW IN TELEGRAM
👍1
Forwarded from Data Science Premium (Books & Courses)
Building Transformer Models with Attention (2023)

The number one book for learning transformers from beginners to professionalism. The book provides an interesting explanation of transformers. The issue is available to the first twenty people only.

Price: 4$

PayPal / Credit Card: https://www.patreon.com/DataScienceBooks/shop/building-transformer-models-with-2023-253971

Crypto Payment: http://t.me/send?start=IVq3O4aPlSWF
👍2
🏠 MedMNIST-C: benchmark dataset based on the MedMNIST+ collection covering 12 2D datasets and 9 imaging modalities.

pip install medmnistc

🖥 Github: https://github.com/francescodisalvo05/medmnistc-api

📕 Paper: https://arxiv.org/abs/2406.17536v2

🔥 Dataset: https://paperswithcode.com/dataset/imagenet-c

https://news.1rj.ru/str/DataScienceT
Please open Telegram to view this post
VIEW IN TELEGRAM
👍31
تنبيه النموذج المفتوح الجديد من مختبر شنغهاي للذكاء الاصطناعي🚨

مرحبًا بك داخليًا 2.5
❤️‍🔥 موديل 7B عالي الجودة
🚀 ما يصل إلى مليون نافذة سياقية
⚙️ إمكانيات استخدام الأداة

نتطلع للعب معها

https://huggingface.co/collections/internlm/internlm25-66853f32717072d17581bc13

https://news.1rj.ru/str/DataScienceT
Please open Telegram to view this post
VIEW IN TELEGRAM
👍1
Hey community!

We’re sharing insights on how rewarded ads can boost your revenue and retain users. Our guide covers special promotion days and seamless integration.

Join our new community on Discord and get the ultimate monetization guide: https://discord.gg/2FMh2ZjE8w
👍31
🔥 ESPNet XEUS is the new speech recognition SoTA.

A multi-lingual speech recognition and translation model from Carnegie Mellon University that is trained in over 4000 languages! 🔥

> MIT license
> 577 million parameters.
> Superior to MMS 1B and w2v-BERT v2 2.0
> E-Branchformer architecture
> Dataset 8900 hours of audio recordings in over 4023 languages

git lfs install
git clone https://huggingface.co/espnet/XEUS

▪️ HF: https://huggingface.co/espnet/xeus
▪️ Dataset: https://huggingface.co/datasets/espnet/mms_ulab_v2

https://news.1rj.ru/str/DataScienceT
Please open Telegram to view this post
VIEW IN TELEGRAM
🌟 MoMA is an open-source model from ByteDance for generating images from a reference.

MoMA requires no training and can quickly generate image images with high detail accuracy and identity preservation.
The speed of MoMA is achieved by optimizing the attention mechanism, which transfers features from the original image to the diffusion model.
The model is a universal adapter and can be applied to various models without modification.
Today, MoMA outperforms similar existing methods in synthetic tests and allows you to create images with a high level of compliance with the prompt while preserving the style of the reference image as much as possible.

✍️ Recommended parameters for optimizing VRAM consumption:

22 GB or more GPU memory:
args.load_8bit, args.load_4bit = False, False

18 GB or more GPU memory:
args.load_8bit, args.load_4bit = True, False

14 GB or more GPU memory:
args.load_8bit, args.load_4bit = False, True


🟡 MoMA page
🖥 GitHub
🤗 Hugging Face
🟡 Demo

https://news.1rj.ru/str/DataScienceT
Please open Telegram to view this post
VIEW IN TELEGRAM
👍61
This media is not supported in your browser
VIEW IN TELEGRAM
🎁 Lisa has given away over $100,000 in the last 30 days. Every single one of her subscribers is making money.

She is a professional trader and broadcasts her way of making money trading on her channel EVERY subscriber she has helped, and she will help you.

🧠 Do this and she will help you earn :

1. Subscribe to her channel
2. Write “GIFT” to her private messages
3. Follow her channel and trade with her.
Repeat transactions after her = earn a lot of money.

Subscribe 👇🏻
https://news.1rj.ru/str/+DqIxOkOWtVw3ZjYx
👍3
⚡️ InternLM introduced XComposer-2.5 - a multi-modal 7B VLM with increased context for input and output.

InternLM-XComposer-2.5 copes with the tasks of text denoscription of images with complex composition, achieving the capabilities of GPT-4V. Trained with alternating 24 KB image-text contexts, it can easily expand to 96 KB contexts via RoPE extrapolation.

Compared to the previous version 2.0, InternLM-XComposer-2.5 has three major improvements:
- understanding of ultra-high resolution;
- detailed understanding of the video;
- process several images in the context of 1 dialogue.

Using extra Lora, XComposer-2.5 is capable of performing complex tasks:
- creation of web pages;
- creation of high-quality text articles with images.

XComposer-2.5 was evaluated on 28 benchmarks, outperforming existing state-of-the-art open source models in 16 benchmarks . It also closely competes with GPT-4V and Gemini Pro on 16 key tasks.

🖥 GitHub
🟡 Arxiv
🟡 Model
🟡 Demo
📺 Demo video

https://news.1rj.ru/str/DataScienceT
Please open Telegram to view this post
VIEW IN TELEGRAM
Please open Telegram to view this post
VIEW IN TELEGRAM
2
Forwarded from Machine Learning
If you ask: which crypto channels are worth watching?👇🏽

Then I will immediately share with you the CRYPTO BARON channel. This is one of the best channels for spot trading🔥 Daily plan for Bitcoin movement and spot signals for alts.

Don’t pay for channels with analytics, everything is free here: https://news.1rj.ru/str/+_kCQKqG5SNZkODhi
Artificial Intelligence A-Z 2024: Build 7 AI + LLM & ChatGPT

Updated : Version 2024

Price: 30$ - full course [offline]

📖 Combine the power of Data Science, Machine Learning and Deep Learning to create powerful AI for Real-World applications!

🔊 Taught By: Hadelin de Ponteves, Kirill Eremenko

Contact @Husseinsheikho
👍3
🌟 MInference 1.0 by Microsoft pre-release

In anticipation of the upcoming ICML 2024 (Vienna, July 21-27, 2024), Microsoft has published the results of a study from the MInference project. This method allows you to speed up the processing of long sequences due to sparse calculations and the use of unique templates in matrices.
The MInference technique does not require changes in pre-training settings.

Microsoft researchers' synthetic tests of the method on the LLaMA-3-1M, GLM4-1M, Yi-200K, Phi-3-128K, and Qwen2-128K models show up to a 10x reduction in latency and prefill errors on the A100 while maintaining accuracy.

🟡 Discuss at Huggingface
🖥 GitHub
🟡 Arxiv
🟡 MInference 1.0 project page

https://news.1rj.ru/str/DataScienceT
Please open Telegram to view this post
VIEW IN TELEGRAM
Please open Telegram to view this post
VIEW IN TELEGRAM
👍2
🌟 Arcee Agent 7B - a new model based on Qwen2-7B

Arcee Agent 7B is superior to GPT-3.5-Turbo, and many other models in writing and interpreting code.
Arcee Agent 7B is especially suitable for those wishing to implement complex AI solutions without the computational expense of large language models.

And yes, there are also quantized GGUF versions of Arcee Agent 7B.

🤗 Hugging Face

https://news.1rj.ru/str/DataScienceT
Please open Telegram to view this post
VIEW IN TELEGRAM
Please open Telegram to view this post
VIEW IN TELEGRAM
👍2
🆕 Meet Kolors - a diffusion model for image generation with a focus on photorealism

Kolors is a large diffusion model published recently by the Kuaishou Kolors team.

Kolors has been trained on billions of text-to-image pairs and shows excellent results in generating complex photorealistic images.

As evaluated by 50 independent experts, the Kolors model generates more realistic and beautiful images than Midjourney-v6, Stable Diffusion 3, DALL-E 3 and other models

🟡 Kolors page
🟡 Try
🖥 GitHub

https://news.1rj.ru/str/DataScienceT
Please open Telegram to view this post
VIEW IN TELEGRAM
Please open Telegram to view this post
VIEW IN TELEGRAM
👍31