nvidia might be the king of the hill right now, but the future of AI is reconfigurable analog-like electronics (~100x more energy efficient already, which will take Moore’s law at least another 10 years for silicon)
Caveat: no backprop :P forward-forward and other algorithms exist though
https://www.nature.com/articles/s41928-023-01042-7
Caveat: no backprop :P forward-forward and other algorithms exist though
https://www.nature.com/articles/s41928-023-01042-7
Nature
Reconfigurable mixed-kernel heterojunction transistors for personalized support vector machine classification
Nature Electronics - Dual-gated van der Waals heterojunction transistors can provide Gaussian, sigmoid and mixed-kernel functions for use in low-power machine learning classification operations.
🔥2
TLDR: Placebo effects are dose-dependent.
We need structured belief engineering then
https://www.biorxiv.org/content/10.1101/2022.07.15.500226v2
We need structured belief engineering then
https://www.biorxiv.org/content/10.1101/2022.07.15.500226v2
bioRxiv
A thalamic circuit represents dose-like responses induced by nicotine-related beliefs in human smokers
Could non-pharmacological constructs, such as beliefs, impact brain activities in a dose-dependent manner as drugs do? While beliefs shape many aspects of our behavior and wellbeing, the precise mapping between subjective beliefs and neural substrates remains…
💊1
https://www.nature.com/articles/s41586-023-06668-3
interesting. code/models: https://github.com/brendenlake/MLC
interesting. code/models: https://github.com/brendenlake/MLC
Nature
Human-like systematic generalization through a meta-learning neural network
Nature - The meta-learning for compositionality approach achieves the systematicity and flexibility needed for human-like generalization.
generalization, continued:
> We argue that Transformers will generalize to harder instances on algorithmic tasks iff the algorithm can be written in the RASP-L programming language (Weiss et al). By design, each line of RASP-L code can be compiled into weights of 1 Transformer layer.
https://arxiv.org/abs/2310.16028
> We argue that Transformers will generalize to harder instances on algorithmic tasks iff the algorithm can be written in the RASP-L programming language (Weiss et al). By design, each line of RASP-L code can be compiled into weights of 1 Transformer layer.
https://arxiv.org/abs/2310.16028
now these are really hallucinations lol
> ZeroNVS: Zero-Shot 360-Degree View Synthesis from a Single Real Image
https://kylesargent.github.io/zeronvs/
> ZeroNVS: Zero-Shot 360-Degree View Synthesis from a Single Real Image
https://kylesargent.github.io/zeronvs/
👍1
LLMs are far from being the first technology met with fear, uncertainty and doubt
https://pessimistsarchive.org
https://pessimistsarchive.org
pessimistsarchive.org
Pessimists Archive
Archive of historical technological pessimism
big if works well: first paper that claims relatively efficient search on encrypted data without revealing what’s being searched
https://eprint.iacr.org/2022/1703
https://eprint.iacr.org/2022/1703
❤1🔥1🤩1
related: the value of privacy (2006) in plain English
https://www.schneier.com/blog/archives/2006/05/the_value_of_pr.html
https://www.schneier.com/blog/archives/2006/05/the_value_of_pr.html
R-Tuning: Teaching Large Language Models to Refuse Unknown Questions
TLDR: LLMs "hallucinate" because the training datasets never included the "I don't know" answer 🤷
https://arxiv.org/pdf/2311.09677.pdf
TLDR: LLMs "hallucinate" because the training datasets never included the "I don't know" answer 🤷
https://arxiv.org/pdf/2311.09677.pdf
A chance of training on huge piles of retail GPUs and not just on the superclusters?
https://huggingface.co/papers/2311.08105
https://huggingface.co/papers/2311.08105
huggingface.co
Paper page - DiLoCo: Distributed Low-Communication Training of Language Models
Join the discussion on this paper page
❤2
System 2 Attention (S2A).
- Soft attention in Transformers is susceptible to irrelevant/biased info
- S2A uses LLM reasoning to generate what to attend to
Improves factuality & objectivity, decreases sycophancy.
https://arxiv.org/abs/2311.11829
- Soft attention in Transformers is susceptible to irrelevant/biased info
- S2A uses LLM reasoning to generate what to attend to
Improves factuality & objectivity, decreases sycophancy.
https://arxiv.org/abs/2311.11829
DARE/MergeLM: Absorbing Abilities from Homologous Models as a Free Lunch
https://github.com/yule-BUAA/MergeLM
https://github.com/yule-BUAA/MergeLM
GitHub
GitHub - yule-BUAA/MergeLM: Codebase for Merging Language Models (ICML 2024)
Codebase for Merging Language Models (ICML 2024). Contribute to yule-BUAA/MergeLM development by creating an account on GitHub.
In this paper, we introduce generative agents--computational software agents that simulate believable human behavior. Generative agents wake up, cook breakfast, and head to work; artists paint, while authors write; they form opinions, notice each other, and initiate conversations; they remember and reflect on days past as they plan the next day.
https://arxiv.org/abs/2304.03442
https://github.com/joonspk-research/generative_agents
https://arxiv.org/abs/2304.03442
https://github.com/joonspk-research/generative_agents
GitHub
GitHub - joonspk-research/generative_agents: Generative Agents: Interactive Simulacra of Human Behavior
Generative Agents: Interactive Simulacra of Human Behavior - joonspk-research/generative_agents
👾1
https://github.com/comfyanonymous/ComfyUI
if you are into Stable Diffusion
if you are into Stable Diffusion
This media is not supported in your browser
VIEW IN TELEGRAM
insanely detailed LLM inference visualization from Brendan Bycroft
https://bbycroft.net/llm
https://bbycroft.net/llm
🔥2
⚡3
https://twitter.com/MistralAI/status/1733150512395038967
beautiful. on friday. even more beautiful.
beautiful. on friday. even more beautiful.
👍2
ChatGPT: sometimes “hallucinates” (tries to guess the details not in the training set).
OpenAI: tries to counter that
Google: hold my beer, let’s hallucinate the actual Gemini model presentation!
https://arstechnica.com/information-technology/2023/12/google-admits-it-fudged-a-gemini-ai-demo-video-which-critics-say-misled-viewers/
OpenAI: tries to counter that
Google: hold my beer, let’s hallucinate the actual Gemini model presentation!
https://arstechnica.com/information-technology/2023/12/google-admits-it-fudged-a-gemini-ai-demo-video-which-critics-say-misled-viewers/
Ars Technica
Google’s best Gemini AI demo video was fabricated
Google takes heat for a misleading AI demo video that hyped up its GPT-4 competitor.
👍2
Generative agent-based modeling with actions grounded in physical, social, or digital space using Concordia from Google DeepMind
https://arxiv.org/abs/2312.03664
https://github.com/google-deepmind/concordia
https://arxiv.org/abs/2312.03664
https://github.com/google-deepmind/concordia
GitHub
GitHub - google-deepmind/concordia: A library for generative social simulation
A library for generative social simulation. Contribute to google-deepmind/concordia development by creating an account on GitHub.
👍1
Google has good researchers and not so good product managers, as always.
Loosely related: Terence Tao was saying "LLMs help me with math" for a while already.
“The FunSearch paper by DeepMind that was used to discover new mathematics is an example of searching through generative patterns and employing evolutionary methods to creatively conjure up new solutions. This is a very general principle that lies at the core of creativity.”
https://www.nature.com/articles/d41586-023-04043-w
Loosely related: Terence Tao was saying "LLMs help me with math" for a while already.
“The FunSearch paper by DeepMind that was used to discover new mathematics is an example of searching through generative patterns and employing evolutionary methods to creatively conjure up new solutions. This is a very general principle that lies at the core of creativity.”
https://www.nature.com/articles/d41586-023-04043-w
Nature
DeepMind AI outdoes human mathematicians on unsolved problem
Nature - Large language model improves on efforts to solve combinatorics problems inspired by the card game Set.
🔥1