Linkstream – Telegram
Linkstream
173 subscribers
32 photos
3 videos
2 files
898 links
Various links I find interesting. Mostly hardcore tech :) // by @oleksandr_now. See @notatky for the personal stuff
Download Telegram
avoid excess complexity and noise at all costs. mvp is not "all features of shitty quality" but "minimum features still decent quality".
or "hypothesis validation code and the commitment to rewrite"
https://minds.md/zakirullin/cognitive
💯4
for something completely different:
first 2178 books of the Ritman Library are now digitized & online, safe from the natural disasters

https://embassyofthefreemind.com/en/library/online-catalogue/?mode=gallery&view=horizontal&sort=random%7B1517048201764%7D%20asc&page=1&fq%5B%5D=search_s_digitized_publication:%22Ja%22&reverse=0
3
CERN Animal Shelter for Computer Mice
https://computer-animal-shelter.web.cern.ch/index.shtml
🤣3
very interesting!
i wonder if the approach used by the TRM/HRM models can be adapted from ARC-AGI back to other reasoning benchmarks usually done on LMs

https://alexiajm.github.io/2025/09/29/tiny_recursive_models.html
aaargh i should've wrote this paper!! it was intuitively obvious to me but then life happens >_<

tldr: LLM sampler is such a powerful prior that with the right sampler (MCMC, in this case), you can even use base models as reasoning models.
without supervised fine-tuning or RL.

this was completely ignored by ppl pilled with the Bitter Lesson mantra, but yes there still is a space for the right priors added or designed by hand!

obviously sampling with mcmc is very costly but you should compare the overall model feedback loop time that includes the posttrain, not just the sampling time

if eg topK sampling is assembly and Mirostat is COBOL (?) then MCMC sampling is like a Python in the space of samplers
https://aakaran.github.io/reasoning_with_sampling/
finally an article showing that people can perceive flickering and certain types of motion at least at 500hz

(it's kind of personal, i've been gaslighted like "hey you can't possibly see the difference" far too many times.
now at least when ppl don't believe me again I can send them this link)

https://www.nature.com/articles/srep07861
👾2
interesting. small (321M not 300B!) and capable models aka reasoning cores are interesting both theoretically and practically
https://pleias.fr/blog/blogsynth-the-new-data-frontier
More paranoia for the paranoid out there ^_^
Timers are a reliable side channel for communicating between containers on the Linux machine, via /proc/self/ns/time.

https://h4x0r.org/funreliable/
Claim: Isotropic Gaussian Regularization for latent representations in the world models is mathematically optimal

What's illustrated:
-- Adopting the isotropic gaussian regularization replaces stop-grad, teacher-student, EMA and various other adhoc tricks
-- Improves model training stability
-- SOTA quality on 10+ datasets and 50+ architectures
https://arxiv.org/abs/2511.08544, https://github.com/rbalestr-lab/lejepa
🤯2🔥1
lol perfect timing 😅 4h later Pavel Durov announced Cocoon: https://news.1rj.ru/str/durov/462

my 2c: it is a fine business, sadly only a small part of what Anima, Cortex and the minds need.
Gentian proxy is still required, etc etc.
They do acknowledge the limitations of RA-TLS and their model in general though, which is commendable.
🤯1