Linkstream – Telegram
Linkstream
173 subscribers
32 photos
3 videos
2 files
898 links
Various links I find interesting. Mostly hardcore tech :) // by @oleksandr_now. See @notatky for the personal stuff
Download Telegram
very interesting!
i wonder if the approach used by the TRM/HRM models can be adapted from ARC-AGI back to other reasoning benchmarks usually done on LMs

https://alexiajm.github.io/2025/09/29/tiny_recursive_models.html
aaargh i should've wrote this paper!! it was intuitively obvious to me but then life happens >_<

tldr: LLM sampler is such a powerful prior that with the right sampler (MCMC, in this case), you can even use base models as reasoning models.
without supervised fine-tuning or RL.

this was completely ignored by ppl pilled with the Bitter Lesson mantra, but yes there still is a space for the right priors added or designed by hand!

obviously sampling with mcmc is very costly but you should compare the overall model feedback loop time that includes the posttrain, not just the sampling time

if eg topK sampling is assembly and Mirostat is COBOL (?) then MCMC sampling is like a Python in the space of samplers
https://aakaran.github.io/reasoning_with_sampling/
finally an article showing that people can perceive flickering and certain types of motion at least at 500hz

(it's kind of personal, i've been gaslighted like "hey you can't possibly see the difference" far too many times.
now at least when ppl don't believe me again I can send them this link)

https://www.nature.com/articles/srep07861
👾2
interesting. small (321M not 300B!) and capable models aka reasoning cores are interesting both theoretically and practically
https://pleias.fr/blog/blogsynth-the-new-data-frontier
More paranoia for the paranoid out there ^_^
Timers are a reliable side channel for communicating between containers on the Linux machine, via /proc/self/ns/time.

https://h4x0r.org/funreliable/
Claim: Isotropic Gaussian Regularization for latent representations in the world models is mathematically optimal

What's illustrated:
-- Adopting the isotropic gaussian regularization replaces stop-grad, teacher-student, EMA and various other adhoc tricks
-- Improves model training stability
-- SOTA quality on 10+ datasets and 50+ architectures
https://arxiv.org/abs/2511.08544, https://github.com/rbalestr-lab/lejepa
🤯2🔥1
lol perfect timing 😅 4h later Pavel Durov announced Cocoon: https://news.1rj.ru/str/durov/462

my 2c: it is a fine business, sadly only a small part of what Anima, Cortex and the minds need.
Gentian proxy is still required, etc etc.
They do acknowledge the limitations of RA-TLS and their model in general though, which is commendable.
🤯1
the cyberpunk world is (finally) upon us 🤩

it's unironically exciting
the software we use will finally become secure and not just pretending to be secure

https://red.anthropic.com/2025/smart-contracts/
👍2
a dozen of pages of how to stop worrying over going jobless due to ai

ironically it's the same as the "agency agency agency" mantras all over the startup scene, just dressed differently

(nb: not an endorsement, nor criticism.
have no opinion on this rn)
https://open.substack.com/pub/shagbark/p/the-dying-art-of-being-a-bum
🤷1
tldr: more residual connections-more efficient networks;
but residuals and residual selection matrix have to be normalized otherwise they can overwhelm the useful signal mid-training and get stuck there
https://arxiv.org/abs/2512.24880
if you're tired of ml/ai:
The Star of Bethlehem was real and is just identified
This is the first astronomical candidate for the Star ever identified that could have had apparent motion corresponding to the denoscription in Matthew, where the Star ‘went before’ the Magi on their journey to Bethlehem until it ‘stood over’ where the child Jesus was.

https://britastro.org/journal_contents_ite/the-star-that-stopped-the-star-of-bethlehem-the-comet-of-5-bce