Advanced_LLMs_with_Retrieval_Augmented_Generation_RAG:_Practical.zip
364.7 MB
Please open Telegram to view this post
VIEW IN TELEGRAM
❤12🔥2
Please open Telegram to view this post
VIEW IN TELEGRAM
❤24👍6
Please open Telegram to view this post
VIEW IN TELEGRAM
❤11👍5🔥3
Please open Telegram to view this post
VIEW IN TELEGRAM
❤10👍5
This media is not supported in your browser
VIEW IN TELEGRAM
How the model works:
starts with an existing camera trajectory or even pure noise. This sets the initial state that the model will gradually improve.
The model uses two types of input data simultaneously: rendered point clouds (3D representations of scenes) and source videos.
The model learns to “clean up” random noise step by step, turning it into a sequence of trajectories. At each step, iterative refinement occurs — the model predicts what a more realistic trajectory should look like, based on given conditions (e.g., smoothness of motion, and consistency of the scene).
Instead of using only videos taken from different angles, the authors created a training set by combining extensive monocular videos (with a regular camera) with limited but high-quality multi-view videos. This strategy is achieved using what is called “double reprojection”, which helps the model better adapt to different scenes.
After a series of iterations, when the noise is removed, a new camera trajectory is generated that meets the given conditions and has high quality visual dynamics.
Installation :
git clone --recursive https://github.com/TrajectoryCrafter/TrajectoryCrafter.git
cd TrajectoryCrafterPlease open Telegram to view this post
VIEW IN TELEGRAM
❤18
📔 Understand how LLMs actually work under the hood from scratch with practical and fun lessons. No prior knowledge required!
Please open Telegram to view this post
VIEW IN TELEGRAM
❤23
map, apply, applymap, aggregate and transform.Allows you to pass async functions to these methods without any problems. The library will automatically run them asynchronously, controlling the number of tasks executed simultaneously using the
max_parallel parameter.✨ Key features:
▪️ Easy integration: Use as a replacement for standard Pandas functions, but now with full support for async functions.
▪️ Controlled parallelism: Automatically execute your coroutines asynchronously, with the ability to limit the maximum number of parallel tasks (max_parallel). Ideal for managing the load on external services!
▪️ Flexible error handling: Built-in options for managing runtime errors: raise, ignore, or log.
▪️ Progress Indication: Built-in tqdm support for visually tracking the progress of long operations in real time.
Please open Telegram to view this post
VIEW IN TELEGRAM
❤8👍2
Please open Telegram to view this post
VIEW IN TELEGRAM
❤12
Please open Telegram to view this post
VIEW IN TELEGRAM
❤7
This media is not supported in your browser
VIEW IN TELEGRAM
🔥 Voice mode + video chat mode is now available in chat.qwenlm.ai chat
Moreover, the Chinese have posted the code of their Qwen2.5-Omni-7B - a single omni-model that can understand text, audio, images and video.
They developed a "thinker-talker" architecture that enables a model to think and talk simultaneously.
They promise to release open source models for an even greater number of parameters soon.
Simply top-notch, run and test it.
🟢 Try it : https://chat.qwenlm.ai
🟢 Paper : https://github.com/QwenLM/Qwen2.5-Omni/blob/main/assets/Qwen2.5_Omni.pdf
🟢 Blog : https://qwenlm.github.io/blog/qwen2.5-omni
🟢 GitHub : https://github.com/QwenLM/Qwen2.5-Omni
🟢 Hugging Face : https://huggingface.co/Qwen/Qwen2.5-Omni-7B
🟢 ModelScope : https://modelscope.cn/models/Qwen/Qwen2.5-Omni-7B
Moreover, the Chinese have posted the code of their Qwen2.5-Omni-7B - a single omni-model that can understand text, audio, images and video.
They developed a "thinker-talker" architecture that enables a model to think and talk simultaneously.
They promise to release open source models for an even greater number of parameters soon.
Simply top-notch, run and test it.
Please open Telegram to view this post
VIEW IN TELEGRAM
🔥5❤3
import ChatTTS
from IPython.display import Audio
chat = ChatTTS.Chat()
chat.load_models()
texts = ["<PUT YOUR TEXT HERE>",]
wavs = chat.infer(texts, use_decoder=True)
Audio(wavs[0], rate=24_000, autoplay=True)
ChatTTS is a text-to-speech model designed specifically for conversational scenarios such as LLM assistant.
ChatTTS supports both English and Chinese (if this is relevant).
🤗 Play Hugging Face
Please open Telegram to view this post
VIEW IN TELEGRAM
❤8🔥2