Linkstream – Telegram
Linkstream
173 subscribers
32 photos
3 videos
2 files
899 links
Various links I find interesting. Mostly hardcore tech :) // by @oleksandr_now. See @notatky for the personal stuff
Download Telegram
https://arxiv.org/abs/2404.09937v1

Compression Represents Intelligence Linearly

There is a belief that learning to compress well will lead to intelligence. Recently, language modeling has been shown to be equivalent to compression, which offers a compelling rationale for the success of large language models (LLMs): the development of more advanced language models is essentially enhancing compression which facilitates intelligence.
(...)
Given the abstract concept of "intelligence", we adopt the average downstream benchmark scores as a surrogate, specifically targeting intelligence related to knowledge and commonsense, coding, and mathematical reasoning. Across 12 benchmarks, our study brings together 30 public LLMs that originate from diverse organizations. Remarkably, we find that LLMs' intelligence -- reflected by average benchmark scores -- almost linearly correlates with their ability to compress external text corpora.

These results provide concrete evidence supporting the belief that superior compression indicates greater intelligence.

Furthermore, our findings suggest that compression efficiency, as an unsupervised metric derived from raw text corpora, serves as a reliable evaluation measure that is linearly associated with the model capabilities. We open-source our compression datasets as well as our data collection pipelines to facilitate future researchers to assess compression properly.
👾1
Llama3 in your browser via WebGPU, client-side!
(don't forget to pick the Llama3 bc default is TinyLlama)

https://secretllama.com/
👍21
every business process can be improved by speeding it up; then these speedups accumulate and cause phase changes, often irreversible; beautifully described by Tiago here

https://wz.ax/tiago/the-throughput-of-learning
👀1
🤣6
Kaggle Expert level reached with the end-to-end competition solver
(the paper says grandmaster but ahem not quite)
https://wz.ax/agent-k/2411.03562
🤷‍♂1
Updated Alibaba's Qwen 2.5 Coder model is
a) solid GPT4 level in code generation (okay),
but also
b) does that in less than 0.1x the resources compared to SOTA (llama3.1) just 3 months ago 🤯 (tested both myself)
https://huggingface.co/Qwen/Qwen2.5-Coder-32B-Instruct
🔥1🤯1
https://www.youtube.com/watch?v=T7sxvrJLJ14

sometimes training your own model is /better/ than making an app.
also a lesson that sometimes you can reach SOTA by yourself!
https://moondream.ai/ rocks and so does @vikhyatk
👍2
St. Peter’s Church in Lucerne, Switzerland has introduced an AI hologram of Jesus, called Deus in Machina, to take confessions. The AI sits behind a screen, providing advice and responding to worshippers' queries. One participant said, “I was surprised, it was so easy, and though it’s a machine, it gave me so much advice.”
https://www.disclose.tv/id/z33eqyr1bv/
🤣2
Nous Research designers are really cracked
https://distro.nousresearch.com/
3💯2