Machine Learning – Telegram
Machine Learning
39.2K subscribers
3.85K photos
32 videos
42 files
1.31K links
Machine learning insights, practical tutorials, and clear explanations for beginners and aspiring data scientists. Follow the channel for models, algorithms, coding guides, and real-world ML applications.

Admin: @HusseinSheikho || @Hussein_Sheikho
Download Telegram
📌 A Deep Dive into In-Context Learning

🗂 Category: NATURAL LANGUAGE PROCESSING

🕒 Date: 2024-05-31 | ⏱️ Read time: 11 min read

Stepping out of the “comfort zone” – part 2/3 of a deep-dive into domain adaptation…
📌 Deep Dive into Anthropic’s Sparse Autoencoders by Hand

🗂 Category: LARGE LANGUAGE MODELS

🕒 Date: 2024-05-31 | ⏱️ Read time: 12 min read

Explore the concepts behind the interpretability quest for LLMs
📌 On-Device Machine Learning in Spatial Computing

🗂 Category: MACHINE LEARNING

🕒 Date: 2025-02-17 | ⏱️ Read time: 18 min read

The landscape of computing is undergoing a profound transformation with the emergence of spatial computing…
📌 Roadmap to Becoming a Data Scientist, Part 4: Advanced Machine Learning

🗂 Category: DATA SCIENCE

🕒 Date: 2025-02-14 | ⏱️ Read time: 15 min read

Introduction Data science is undoubtedly one of the most fascinating fields today. Following significant breakthroughs in…
📌 Building a Data Engineering Center of Excellence

🗂 Category: DATA ENGINEERING

🕒 Date: 2025-02-13 | ⏱️ Read time: 11 min read

As data continues to grow in importance and become more complex, the need for skilled…
🤖🧠 NanoChat: The Best ChatGPT That $100 Can Buy

🗓️ 20 Oct 2025
📚 AI News & Trends

In a world dominated by billion-dollar AI models like GPT-4 and Claude 3, it’s refreshing to see a minimalist, open-source alternative that puts the power of Large Language Models (LLMs) back into the hands of hackers, researchers and enthusiasts. Enter NanoChat – an end-to-end, full-stack implementation of a ChatGPT-style AI chatbot developed by Andrej Karpathy, ...

#NanoChat #ChatGPT #AI #LargeLanguageModels #OpenSource #AndrejKarpathy
1
🤖🧠 NanoChat: The Best ChatGPT That $100 Can Buy

🗓️ 20 Oct 2025
📚 AI News & Trends

In a world dominated by billion-dollar AI models like GPT-4 and Claude 3, it’s refreshing to see a minimalist, open-source alternative that puts the power of Large Language Models (LLMs) back into the hands of hackers, researchers and enthusiasts. Enter NanoChat – an end-to-end, full-stack implementation of a ChatGPT-style AI chatbot developed by Andrej Karpathy, ...

#NanoChat #ChatGPT #AI #LargeLanguageModels #OpenSource #AndrejKarpathy
🤖🧠 PaddleOCR-VL: Redefining Multilingual Document Parsing with a 0.9B Vision-Language Model

🗓️ 20 Oct 2025
📚 AI News & Trends

In an era where information is predominantly digital, the ability to extract, interpret and organize data from documents is crucial. From invoices and research papers to multilingual contracts and handwritten notes, document parsing stands at the intersection of vision and language. Traditional Optical Character Recognition (OCR) systems have made impressive strides but they often fall ...

#PaddleOCR-VL #Multilingual #DocumentParsing #VisionLanguageModel #OCR #AI
🤖🧠 PaddleOCR-VL: Redefining Multilingual Document Parsing with a 0.9B Vision-Language Model

🗓️ 20 Oct 2025
📚 AI News & Trends

In an era where information is predominantly digital, the ability to extract, interpret and organize data from documents is crucial. From invoices and research papers to multilingual contracts and handwritten notes, document parsing stands at the intersection of vision and language. Traditional Optical Character Recognition (OCR) systems have made impressive strides but they often fall ...

#PaddleOCR-VL #Multilingual #DocumentParsing #VisionLanguageModel #OCR #AI
1
🤖🧠 Top 30 More Retro Bollywood Diwali Portrait Prompts for Women Using Gemini AI – Part 2

🗓️ 20 Oct 2025
📚 AI News & Trends

The Diwali celebrations continue and so does the nostalgia! After the huge buzz around our Top 20 Retro Bollywood Diwali Portrait Ideas, we’re back with Part 2 featuring prompts 21 to 50 curated to help you create even more magical, cinematic AI portraits using Google Gemini AI. If you loved the 90s-style Diwali aesthetics shimmering ...
2
Here is a trick to optimize a neural network that gives about 4x speedup when transferring data from CPU to GPU.

Let's consider an image classification task.

We define the model, load and transform the data.

In the training loop, we transfer data to the GPU and train the network.

What's the problem:

If you look into the profiler,

- most of the resources go to the kernel (i.e., the training itself),
- but a noticeable amount of time is also spent on transferring data from CPU to GPU (cudaMemcpyAsync).

This can be easily reduced.

Initially, the dataset consists of pixels as 8-bit integers. We convert them to 32-bit floats.
Then we send these float tensors to the GPU. As a result, the data size becomes 4 times larger, making the transfer heavier.

The solution:

Shift the transformation step after the transfer. That is, first transfer the 8-bit ints, and then convert them to floats on the GPU.

As a result, the data transfer step speeds up significantly.

Of course, this doesn't work everywhere; for example, in NLP we initially deal with float embeddings.
But in cases where it applies, the speedup is very noticeable.

👉  @DataScienceM
Please open Telegram to view this post
VIEW IN TELEGRAM
4
Please open Telegram to view this post
VIEW IN TELEGRAM
2
🤖🧠 Wan 2.1: Alibaba’s Open-Source Revolution in Video Generation

🗓️ 21 Oct 2025
📚 AI News & Trends

The landscape of artificial intelligence has been evolving rapidly, especially in the domain of video generation. Since OpenAI unveiled Sora in 2024, the world has witnessed an explosive surge in research and innovation within generative AI. However, most of these cutting-edge tools remained closed-source limiting transparency and accessibility. Recognizing this gap, Alibaba Group introduced Wan, ...

#Alibaba #Wan2.1 #VideoGeneration #GenerativeAI #OpenSource #ArtificialIntelligence
1
🤖🧠 DeepSeek-OCR: Redefining Document Understanding Through Optical Context Compression

🗓️ 21 Oct 2025
📚 AI News & Trends

In the age of large language models (LLMs) and vision-language models (VLMs), handling long and complex textual data efficiently remains a massive challenge. Traditional models struggle with processing extended contexts because the computational cost increases quadratically with sequence length. To overcome this, researchers from DeepSeek-AI have introduced a groundbreaking approach – DeepSeek-OCR, a model that ...
📌 Implementing the Fourier Transform Numerically in Python: A Step-by-Step Guide

🗂 Category: MATH

🕒 Date: 2025-10-21 | ⏱️ Read time: 19 min read

What if the FFT functions in NumPy and SciPy don’t actually compute the Fourier transform…
3
🤖🧠 The Art of Scaling Reinforcement Learning Compute for LLMs: Top Insights from Meta, UT Austin and Harvard University

🗓️ 21 Oct 2025
📚 AI News & Trends

As Large Language Models (LLMs) continue to redefine artificial intelligence, a new research breakthrough has emerged from Meta, The University of Texas at Austin, University College London, UC Berkeley, Harvard University and Periodic Labs. Their paper, noscriptd “The Art of Scaling Reinforcement Learning Compute for LLMs,” introduces a transformative framework for understanding how reinforcement learning ...

#ReinforcementLearning #LLMs #AIResearch #Meta #UTAustin #HarvardUniversity
🤖🧠 The Art of Scaling Reinforcement Learning Compute for LLMs: Top Insights from Meta, UT Austin and Harvard University

🗓️ 21 Oct 2025
📚 AI News & Trends

As Large Language Models (LLMs) continue to redefine artificial intelligence, a new research breakthrough has emerged from Meta, The University of Texas at Austin, University College London, UC Berkeley, Harvard University and Periodic Labs. Their paper, noscriptd “The Art of Scaling Reinforcement Learning Compute for LLMs,” introduces a transformative framework for understanding how reinforcement learning ...

#ReinforcementLearning #LLMs #AIResearch #Meta #UTAustin #HarvardUniversity
🤖🧠 Master Machine Learning with Stanford’s CS229 Cheatsheets: The Ultimate Learning Resource

🗓️ 21 Oct 2025
📚 AI News & Trends

Machine learning is one of the most transformative fields in technology today. From powering recommendation systems to enabling self-driving cars, machine learning is at the core of modern artificial intelligence. However, mastering its vast concepts, equations and algorithms can be overwhelming especially for beginners and busy professionals. That’s where the Stanford CS229 Machine Learning Cheatsheets ...
🤖🧠 Master Machine Learning with Stanford’s CS229 Cheatsheets: The Ultimate Learning Resource

🗓️ 21 Oct 2025
📚 AI News & Trends

Machine learning is one of the most transformative fields in technology today. From powering recommendation systems to enabling self-driving cars, machine learning is at the core of modern artificial intelligence. However, mastering its vast concepts, equations and algorithms can be overwhelming especially for beginners and busy professionals. That’s where the Stanford CS229 Machine Learning Cheatsheets ...
📌 How to Build An AI Agent with Function Calling and GPT-5

🗂 Category: AGENTIC AI

🕒 Date: 2025-10-20 | ⏱️ Read time: 15 min read

How an AI agent works: a step-by-step guide
📌 How to Use Frontier Vision LLMs: Qwen3-VL

🗂 Category: LARGE LANGUAGE MODELS

🕒 Date: 2025-10-20 | ⏱️ Read time: 11 min read

Learn how to apply VLMs to advanced document understanding tasks