Multimodal ArXiv: A Dataset for Improving Scientific Comprehension of Large Vision-Language Models
Presents:
-ArXivCap, a million-scale figure-caption dataset from arxiv papers
- ArXivQA, a QA dataset generated by prompting GPT-4V based on arxiv figures.
Paper here.
Presents:
-ArXivCap, a million-scale figure-caption dataset from arxiv papers
- ArXivQA, a QA dataset generated by prompting GPT-4V based on arxiv figures.
Paper here.
mm-arxiv.github.io
Multimodal ArXiv
Vision-Language Feedback
⚡5
Intel's NPU Acceleration Library goes open source — Meteor Lake CPUs can now run TinyLlama and other lightweight LLMs.
Tom's Hardware
Intel's NPU Acceleration Library goes open source — Meteor Lake CPUs can now run TinyLlama and other lightweight LLMs
LLMs on the go.
⚡4
Anthropic announced the Claude 3 model family
The family includes three state-of-the-art models in ascending order of capability:
Claude 3 Haiku, Claude 3 Sonnet, and Claude 3 Opus.
Each successive model offers increasingly powerful performance, allowing users to select the optimal balance of intelligence, speed, and cost for their specific application.
The family includes three state-of-the-art models in ascending order of capability:
Claude 3 Haiku, Claude 3 Sonnet, and Claude 3 Opus.
Each successive model offers increasingly powerful performance, allowing users to select the optimal balance of intelligence, speed, and cost for their specific application.
Anthropic
Introducing the next generation of Claude
Today, we're announcing the Claude 3 model family, which sets new industry benchmarks across a wide range of cognitive tasks. The family includes three state-of-the-art models in ascending order of capability: Claude 3 Haiku, Claude 3 Sonnet, and Claude 3…
⚡5
Bytedance introduced Diffusion Protein Language Models (DPLM), a new suite of discrete diffusion-based protein language models
With versatility in both generative and predictive tasks, DPLM is poised to set the new SOTA in protein language models, excelling across a spectrum of benchmark tasks.
With versatility in both generative and predictive tasks, DPLM is poised to set the new SOTA in protein language models, excelling across a spectrum of benchmark tasks.
arXiv.org
Diffusion Language Models Are Versatile Protein Learners
This paper introduces diffusion protein language model (DPLM), a versatile protein language model that demonstrates strong generative and predictive capabilities for protein sequences. We first...
⚡4
A new article "Creative Flow as Optimized Processing: Evidence from Brain Oscillations During Jazz Improvisations by Expert and Non-Expert Musicians."
This is the first neuroimaging study to isolate the neural correlates of the flow experience during a creative production task, in this case, jazz improvisation. Flow is not hyperfocus. It results from an expert brain network plus release of executive control.
This is the first neuroimaging study to isolate the neural correlates of the flow experience during a creative production task, in this case, jazz improvisation. Flow is not hyperfocus. It results from an expert brain network plus release of executive control.
Drexel News
Your Brain in the Zone: A New Neuroimaging Study Reveals How the Brain Achieves a Creative Flow State
A new neuroimaging study from Drexel University’s Creativity Research Lab is the first to reveal how the brain gets to the creative flow state.
⚡8👍3
Very cool data analysis from Paradigm showing the breakdown of Ethereum's state.
ERC20s make up 27% of total state, while ERC721s make up 21.6%. Accounts total 14.1%.
XEN makes up 3.5% of Ethereum's state, which is more than any other single protocol.
ERC20s make up 27% of total state, while ERC721s make up 21.6%. Accounts total 14.1%.
XEN makes up 3.5% of Ethereum's state, which is more than any other single protocol.
⚡5🆒2
Andrew Ng: we will control and steer superhuman AI, so if we want humanity to survive and thrive we should develop AI "as fast as possible"
❤3⚡3
It’s a big! First-of-its-kind supplement clinically proven to slow effects of aging in dogs available at LeapYears.com
⚡3❤2
The world’s 4 biggest cloud firms, Amazon AWS, Microsoft, Google and Meta will spend a record high US$200 billion on capex in 2025, citing the Wells Fargo Investment Institute, up from $140 billion in spending last year.
MorningStar
AI boom in data centers has top tech companies spending more than major oil companies on capex
By Joy Wiltermuth
⚡3
MindSpeaker BCI has built its “MindSpeaker+MindClick”
The integrated product concept enables improving communication for patients and elderly suffering from speech disorders via in-ear EEG sensing.
MindSpeaker builds Alternative and Augmentative Communication products. This product will address patients with speech paralysis (dysarthria).
The integrated product concept enables improving communication for patients and elderly suffering from speech disorders via in-ear EEG sensing.
MindSpeaker builds Alternative and Augmentative Communication products. This product will address patients with speech paralysis (dysarthria).
Mindspeakerbci
Augmenting communication
MindSpeaker uses state-of-the-art technology to augment communication.
⚡5
What if you and your friends could see through each other’s eyes all at once?
Researchers revealed that elephantnose fish might really do this kind of group sensing with their electro-location sensory system.
Researchers revealed that elephantnose fish might really do this kind of group sensing with their electro-location sensory system.
⚡3🍌1
A 7B-parameter DNA language model trained on 2.7M prokaryotic genomes can perform generation and prediction at the DNA, RNA, and protein levels.
bioRxiv
Sequence modeling and design from molecular to genome scale with Evo
The genome is a sequence that completely encodes the DNA, RNA, and proteins that orchestrate the function of a whole organism. Advances in machine learning combined with massive datasets of whole genomes could enable a biological foundation model that accelerates…
❤3🍌1
Having a diversity of open-source models is good of course. But benchmarks suggest it's worse than even a 34B open LLM.
⚡4
South Korea’s National Tax Service plans to build a virtual asset management system to prevent users from using virtual assets to evade taxes.
The system is designed to effectively analyze and manage information collected through the mandatory submission of virtual asset transaction history, and is scheduled to be launched in 2025.
The system is designed to effectively analyze and manage information collected through the mandatory submission of virtual asset transaction history, and is scheduled to be launched in 2025.
디지털데일리
국세청, ‘가상자산 통합관리 시스템’ 구축 나서…탈세 방지 등 목표로 2025년 구축 완료
[디지털데일리 이상일기자] 국세청이 가상자산 거래내역 제출 의무화에 따라 수집되는 가상자산 거래 정보를 분석 관리 할 수 있는 ‘가상자산 통합관리 시스템’ 구축을 위한 사전 컨설팅...
⚡4
Tether announced today that USDT will launch on Celo, a mobile-first and EVM-compatible blockchain network
Celo core contributor proposed the use of USDT as a gas currency. Celo's ecosystem in countries like Kenya and Ghana will help adoption and utilization of USDT.
Celo core contributor proposed the use of USDT as a gas currency. Celo's ecosystem in countries like Kenya and Ghana will help adoption and utilization of USDT.
⚡5
A major paper in AI-driven drug discovery was released
It describes the underlying biology, chemistry, and clinical data supporting the lead candidate in the small molecule AI drug discovery race (INS018_055 by Insilico Medicine).
It describes the underlying biology, chemistry, and clinical data supporting the lead candidate in the small molecule AI drug discovery race (INS018_055 by Insilico Medicine).
Nature
A small-molecule TNIK inhibitor targets fibrosis in preclinical and clinical models
Nature Biotechnology - An AI-generated small-molecule inhibitor treats fibrosis in vivo and in phase I clinical trials.
⚡3
A cool finding on multilingualism in the brain. An MIT study finds the brains of polyglots expend comparatively little effort when processing their native language.
In the brains of these polyglots — people who speak five or more languages — the same language regions light up when they listen to any of the languages that they speak.
In general, this network responds more strongly to languages in which the speaker is more proficient, with one notable exception: the speaker’s native language.
When listening to one’s native language, language network activity drops off significantly.
The findings suggest there is something unique about the first language one acquires, which allows the brain to process it with minimal effort.
Many languages, one network
The brain’s language processing network, located primarily in the left hemisphere, includes regions in the frontal and temporal lobes.
In the new study, the researchers wanted to expand on that finding and explore what happens in the brains of polyglots as they listen to languages in which they have varying levels of proficiency.
Studying polyglots can help researchers learn more about the functions of the language network, and how languages learned later in life might be represented differently than a native language or languages.
Brain engagement
The researchers saw a similar phenomenon when polyglots listened to languages that they don’t speak: Their language network was more engaged when listening to languages related to a language that they could understand, than compared to listening to completely unfamiliar languages.
The researchers also found that a brain network known as the multiple demand network, which turns on whenever the brain is performing a cognitively demanding task, also becomes activated when listening to languages other than one’s native language.
In the brains of these polyglots — people who speak five or more languages — the same language regions light up when they listen to any of the languages that they speak.
In general, this network responds more strongly to languages in which the speaker is more proficient, with one notable exception: the speaker’s native language.
When listening to one’s native language, language network activity drops off significantly.
The findings suggest there is something unique about the first language one acquires, which allows the brain to process it with minimal effort.
Many languages, one network
The brain’s language processing network, located primarily in the left hemisphere, includes regions in the frontal and temporal lobes.
In the new study, the researchers wanted to expand on that finding and explore what happens in the brains of polyglots as they listen to languages in which they have varying levels of proficiency.
Studying polyglots can help researchers learn more about the functions of the language network, and how languages learned later in life might be represented differently than a native language or languages.
Brain engagement
The researchers saw a similar phenomenon when polyglots listened to languages that they don’t speak: Their language network was more engaged when listening to languages related to a language that they could understand, than compared to listening to completely unfamiliar languages.
The researchers also found that a brain network known as the multiple demand network, which turns on whenever the brain is performing a cognitively demanding task, also becomes activated when listening to languages other than one’s native language.
MIT News
For people who speak many languages, there’s something special about their native tongue
An MIT study of polyglots found the brain’s language network responds more strongly when hearing languages a speaker is more proficient in — and much more weakly to the speaker’s native language.
👍1
OpenAI released a tool they've been using internally to analyze transformer internals - the Transformer Debugger
It combines both automated interpretability and sparse autoencoders, and it allows rapid exploration of models without writing code.
It supports both neurons and attention heads. You can intervene on the forward pass by ablating individual neurons and see what changes.
In short, it's a quick and easy way to discover circuits manually.
This is still an early stage research tool.
It combines both automated interpretability and sparse autoencoders, and it allows rapid exploration of models without writing code.
It supports both neurons and attention heads. You can intervene on the forward pass by ablating individual neurons and see what changes.
In short, it's a quick and easy way to discover circuits manually.
This is still an early stage research tool.
GitHub
GitHub - openai/transformer-debugger
Contribute to openai/transformer-debugger development by creating an account on GitHub.
⚡3❤2