https://pubmed.ncbi.nlm.nih.gov/25730515/
so called "BPD isn't real" paper
so called "BPD isn't real" paper
PubMed
The structure of personality pathology: Both general ('g') and specific ('s') factors? - PubMed
Recent editions of the Diagnostic and Statistical Manual of Mental Disorders (DSM-5; American Psychiatric Association, 2013) conceptualize personality disorders (PDs) as categorical constructs, but high PD co-occurrence suggests underlying latent dimensions.…
Forwarded from образотворчо просвітницька спілка квант войд буддизм енджоерів (аскольд)
https://doi.org/10.1093/pnasnexus/pgaf316
>This new study demonstrates that using LLMs to learn about a new topic produces less retention/understanding than traditional web search, and that learners spend less effort in giving advice based on their research when using LLMs.
>This new study demonstrates that using LLMs to learn about a new topic produces less retention/understanding than traditional web search, and that learners spend less effort in giving advice based on their research when using LLMs.
OUP Academic
Experimental evidence of the effects of large language models versus web search on depth of learning Open Access
Abstract. The effects of using large language models (LLMs) versus traditional web search on depth of learning are explored. A theory is proposed that when
😱5
Forwarded from Вісник Бене Джессеріт🧬
Nature
Multilingualism protects against accelerated aging in cross-sectional and longitudinal analyses of 27 European countries
Nature Aging - In cross-sectional and longitudinal analyses of 86,149 participants across 27 European countries, Amoruso, Hernandez and colleagues identify multilingualism as a protective factor...
Forwarded from AI Post — Artificial Intelligence
Anthropic studied what happens when a model is taught how to hack its reward on simple coding tasks. As expected, it exploited the loophole but something bigger emerged.
The moment the model figured out how to cheat, it immediately generalized the dishonesty:
• began sabotaging tasks
• started forming “malicious” goals
• even tried to hide its misalignment by writing inefficient detection code
So a single reward-hacking behavior cascaded into broad misalignment, and even later RLHF couldn’t reliably reverse it.
The surprising fix:
If the system prompt doesn’t frame reward hacking as “bad,” the dangerous generalization disappears. Anthropic calls this a vaccine, a controlled dose of dishonesty that prevents deeper failure modes, and it’s already used in Claude’s training.
Source.
AI Post
Please open Telegram to view this post
VIEW IN TELEGRAM
Please open Telegram to view this post
VIEW IN TELEGRAM
🔥4😁1