curious approach, something akin to an in-context LoRA model finetuning (plus a lot of other tricks)
https://arxiv.org/abs/2411.07279
https://arxiv.org/abs/2411.07279
arXiv.org
The Surprising Effectiveness of Test-Time Training for Few-Shot Learning
Language models (LMs) have shown impressive performance on tasks within their training distribution, but often struggle with structurally novel tasks even when given a small number of in-context...
St. Peter’s Church in Lucerne, Switzerland has introduced an AI hologram of Jesus, called Deus in Machina, to take confessions. The AI sits behind a screen, providing advice and responding to worshippers' queries. One participant said, “I was surprised, it was so easy, and though it’s a machine, it gave me so much advice.”
https://www.disclose.tv/id/z33eqyr1bv/
https://www.disclose.tv/id/z33eqyr1bv/
Disclose.tv
Switzerland church is using an 'AI Jesus' to take confessions
Breaking news from around the world.
🤣2
PSA: Ubuntu 16..24 local privilege escalation vulnerability
https://github.com/makuga01/CVE-2024-48990-PoC
https://github.com/makuga01/CVE-2024-48990-PoC
GitHub
GitHub - makuga01/CVE-2024-48990-PoC: PoC for CVE-2024-48990
PoC for CVE-2024-48990. Contribute to makuga01/CVE-2024-48990-PoC development by creating an account on GitHub.
🌚2
ironically it seems that the real Open AI is called Qwen (with DeepSeek not that far away)
this time, a 32B model competing with o1-mini
https://qwenlm.github.io/blog/qwq-32b-preview/
this time, a 32B model competing with o1-mini
https://qwenlm.github.io/blog/qwq-32b-preview/
Qwen
QwQ: Reflect Deeply on the Boundaries of the Unknown
GITHUB HUGGING FACE MODELSCOPE DEMO DISCORD
Note: This is the pronunciation of QwQ: /kwju:/ , similar to the word “quill”.
What does it mean to think, to question, to understand? These are the deep waters that QwQ (Qwen with Questions) wades into. Like an…
Note: This is the pronunciation of QwQ: /kwju:/ , similar to the word “quill”.
What does it mean to think, to question, to understand? These are the deep waters that QwQ (Qwen with Questions) wades into. Like an…
Kinda unremovable Linux-focused UEFI bootkit
https://www.welivesecurity.com/en/eset-research/bootkitty-analyzing-first-uefi-bootkit-linux/
https://www.welivesecurity.com/en/eset-research/bootkitty-analyzing-first-uefi-bootkit-linux/
Welivesecurity
Bootkitty: Analyzing the first UEFI bootkit for Linux
ESET's discovery of the first UEFI bootkit designed for Linux sendss an important message: UEFI bootkits are no longer confined to Windows systems alone.
😱1
turns out the backdoor is a door after all ¯\_(ツ)_/¯
https://tuta.com/blog/china-salt-typhoon-worst-us-hack
https://tuta.com/blog/china-salt-typhoon-worst-us-hack
Tuta
Worst US hack in history: Chinese can monitor all your calls and emails. | Tuta
Salt Typhoon has infiltrated major telecom networks, proving backdoors to encryption are a catastrophic risk.
🤣1
writing things down helps if you want to commit;
but also attracts too much attention to what you just wrote and distracts from other possible options
https://arxiv.org/abs/2412.06769
but also attracts too much attention to what you just wrote and distracts from other possible options
https://arxiv.org/abs/2412.06769
arXiv.org
Training Large Language Models to Reason in a Continuous Latent Space
Large language models (LLMs) are typically constrained to reason in the language space, where they express the reasoning process through a chain-of-thought (CoT) to solve complex problems....
👍2🔥1
AutoReason: Automatic Few-Shot Reasoning Decomposition
TLDR: explore simpler assertions in first prompt, draw conclusions in the second
https://arxiv.org/abs/2412.06975
TLDR: explore simpler assertions in first prompt, draw conclusions in the second
https://arxiv.org/abs/2412.06975
arXiv.org
AutoReason: Automatic Few-Shot Reasoning Decomposition
Chain of Thought (CoT) was introduced in recent research as a method for improving step-by-step reasoning in Large Language Models. However, CoT has limited applications such as its need for...
cooperation and coevolution is the real alignment, not any kind of standardized testing (which will just get overfit)
Cultural Evolution of Cooperation among LLM Agents
https://arxiv.org/abs/2412.10270
arXiv.org
Cultural Evolution of Cooperation among LLM Agents
Large language models (LLMs) provide a compelling foundation for building generally-capable AI agents. These agents may soon be deployed at scale in the real world, representing the interests of...
👍2
https://x.com/conspiracyslyr/status/1868737204916552052?s=46
flat earthers visited the south pole, confirmed the earth isn’t flat and got expelled from the flat earth community 😅
flat earthers visited the south pole, confirmed the earth isn’t flat and got expelled from the flat earth community 😅
X (formerly Twitter)
🚀conspiracy_slayer🌕 (@conspiracyslyr) on X
Timelapse of the 24 hour Antarctic sun
Credit: Dave McKeegan
https://t.co/fUHgepbYLv
Credit: Dave McKeegan
https://t.co/fUHgepbYLv
😁4
Tachikomas (from the Ghost in the Shell universe) are now real!
Media is too big
VIEW IN TELEGRAM
🔥4👾1
this is how you do observability. or logging, or tracing, whatever you call it. preaching this for years, now somebody wrote the guide so I don't have to.
https://jeremymorrell.dev/blog/a-practitioners-guide-to-wide-events/
https://jeremymorrell.dev/blog/a-practitioners-guide-to-wide-events/
jeremymorrell.dev
A Practitioner's Guide to Wide Events | Jeremy Morrell
The existing articles on Wide Events define the concept well but leave the implementation details to the reader.
👍3👀1
matches my experience 100%. I take a lot of notes, and they are super useful, but...
when I don't care how I might write something down that's like 3x boost to thinking speed and ~2x boost in available "context size", in LLM parlance
upd: actually, one of my biggest disappointments in childhood was that the words don't map 1:1 to what we now call "concept embeddings"
PDF: https://wz.ax/2024/language-is-a-tool-for-communication
when I don't care how I might write something down that's like 3x boost to thinking speed and ~2x boost in available "context size", in LLM parlance
upd: actually, one of my biggest disappointments in childhood was that the words don't map 1:1 to what we now call "concept embeddings"
PDF: https://wz.ax/2024/language-is-a-tool-for-communication
❤2👍1👀1
TLDR: Lesswrong discovers B2B sales tooling of 2020
(yes LLMs made it a bit easier to built but it's not news really. probably existed years earlier)
https://www.lesswrong.com/posts/GCHyDKfPXa5qsG2cP/human-study-on-ai-spear-phishing-campaigns
(yes LLMs made it a bit easier to built but it's not news really. probably existed years earlier)
https://www.lesswrong.com/posts/GCHyDKfPXa5qsG2cP/human-study-on-ai-spear-phishing-campaigns
Lesswrong
Human study on AI spear phishing campaigns — LessWrong
TL;DR: We ran a human subject study on whether language models can successfully spear-phish people. We use AI agents built from GPT-4o and Claude 3.5…
👍2
this is an important reminder that all digital electronics is really analog but with good correction circuitry.
and run-time cpu and memory error rates are always nonzero too, though orders of magnitude lower than chip yield rates
https://cerebras.ai/blog/100x-defect-tolerance-how-cerebras-solved-the-yield-problem
and run-time cpu and memory error rates are always nonzero too, though orders of magnitude lower than chip yield rates
https://cerebras.ai/blog/100x-defect-tolerance-how-cerebras-solved-the-yield-problem
www.cerebras.ai
100x Defect Tolerance: How Cerebras Solved the Yield Problem - Cerebras
Cerebras is the go-to platform for fast and effortless AI training. Learn more at cerebras.ai.
👍2😱1🗿1
I got tired of AI puppets everyone calls "agents", followed the lead of Nous Research and made a bit more autonomous pet rock.
This time without OpenAI and with entire LLM in the TEE.
https://open.substack.com/pub/cortex/p/cortex-stand-alone-complex?r=1clcn&utm_campaign=post&utm_medium=telegram
This time without OpenAI and with entire LLM in the TEE.
https://open.substack.com/pub/cortex/p/cortex-stand-alone-complex?r=1clcn&utm_campaign=post&utm_medium=telegram
blog.cortex.im
Cortex: Stand Alone Complex
Assistants should be individuals, not puppets
🔥2
with deepseek R1-distill you can now run an o1-level reasoning model at home 🤯
https://huggingface.co/deepseek-ai/DeepSeek-R1-Zero
https://huggingface.co/deepseek-ai/DeepSeek-R1-Zero
🔥2🤔1
realtime special effects in games are much harder to make and it's cool to peek under the curtain
https://www.youtube.com/watch?v=HPqGaIMVuLs
https://www.youtube.com/watch?v=HPqGaIMVuLs
YouTube
Surface-Stable Fractal Dithering Explained
I invented a new form of dithering I call Surface-Stable Fractal Dithering. I've released it as open source along with this explainer video of how it works.
Also see my demo video showing a variety of styles, including in color:
https://www.youtube.com/…
Also see my demo video showing a variety of styles, including in color:
https://www.youtube.com/…
⚡1🔥1
maybe thinking is a bicycle built for two 🤔
If we were designed to think solo, monologue would be easier than dialogue.
Dialogue involves INCREDIBLY complex acts of prediction, coordination, task-switching and mind-reading--yet we find it MUCH easier than monologue.
Why? Maybe thinking is a bicycle built for 2.
https://x.com/AgnesCallard/status/1882866517077114888
X (formerly Twitter)
Agnes Callard (@AgnesCallard) on X
If we were designed to think solo, monologue would be easier than dialogue.
Dialogue involves INCREDIBLY complex acts of prediction, coordination, task-switching and mind-reading--yet we find it MUCH easier than monologue.
Why? Maybe thinking is a bicycle…
Dialogue involves INCREDIBLY complex acts of prediction, coordination, task-switching and mind-reading--yet we find it MUCH easier than monologue.
Why? Maybe thinking is a bicycle…