gonzo-обзоры ML статей – Telegram
gonzo-обзоры ML статей
24K subscribers
2.75K photos
2 videos
3 files
1.36K links
Авторы:
Гриша Сапунов, ранее руководитель разработки Яндекс-Новостей, ныне CTO Intento. Области интересов: AI/ML/DL, биоинформатика.
Лёша Тихонов, ранее аналитик в Яндексе, автор Автопоэта, Нейронной Обороны... Области интересов: discrete domain, NLP, RL.
Download Telegram
This media is not supported in your browser
VIEW IN TELEGRAM
😁1
Falcon 180B подвезли

https://falconllm.tii.ae/falcon-180b.html

Falcon 180B is a super-powerful language model with 180 billion parameters, trained on 3.5 trillion tokens. It's currently at the top of the Hugging Face Leaderboard for pre-trained Open Large Language Models and is available for both research and commercial use.

This model performs exceptionally well in various tasks like reasoning, coding, proficiency, and knowledge tests, even beating competitors like Meta's LLaMA 2.

Among closed source models, it ranks just behind OpenAI's GPT 4, and performs on par with Google's PaLM 2 Large, which powers Bard, despite being half the size of the model
🔥40👍3
12😁8👍1🤔1
And a couple more slides
👏18🤣6👍3
An interesting observation by Xiang Zhang:

TL;DR: number of parameters is a more determining factor than numerical precision for large language model performance. Given a memory constraint, one should maximize the number of parameters by quantizing at the highest level possible.

https://www.xzh.me/2023/09/a-perplexity-benchmark-of-llamacpp.html
👍27