The art of fact-checking
Mar. 21st, 2023 02:58 pmCLAIM: Greta Thunberg recently deleted a tweet she posted in 2018 in which she said that the world will end in 2023.
AP’S ASSESSMENT: Missing context. While Thunberg did delete a 2018 tweet about the urgency of addressing climate change, she did not say the world would end in 2023. Her tweet included a quote from an article that said an influential scientist warned climate change “will wipe out all humanity” unless fossil fuel use was ended “over the next five years.” Further complicating the issue, that article incorrectly summarized the scientist’s speech. He never made such comments.
...Thunberg never said the world was set to end in 2023. The young activist was quoting an article that was paraphrasing a speech by a Harvard University professor of atmospheric chemistry. The scientist said the world had limited time to act to reverse the disappearance of floating ice volume in the Arctic or there would be drastic consequences, not that the world would end in five years.
The article Thunberg was quoting has since been deleted, but it was published by the site Grit Post in February 2018 with the headline: “Top Climate Scientist: Humans Will Go Extinct if We Don’t Fix Climate Change by 2023,” according to an archived version of the web page. The first line of the story stated: “A top climate scientist is warning that climate change will wipe out all of humanity unless we stop using fossil fuels over the next five years.”
Forbes quoted Anderson as saying: “The chance that there will be any permanent ice left in the Arctic after 2022 is essentially zero.”
У снопса есть скриншот того твита и сходный анализ.
no subject
Date: 2023-03-23 01:55 am (UTC)So, presumably, better models need to make sure to provide an equivalent of that too... And there might be other properties of today's LLMs we would want to capture (or further upgrade) in better models.
no subject
Date: 2023-03-23 05:15 am (UTC)I agree, attention seems really important. Backpropagation not so much, perhaps.
no subject
Date: 2023-03-23 02:27 pm (UTC)But the magic property of **attention layers** performing gradient steps on the fly (without backpropagation) is something else; it does seem to be responsible for many of the magic properties of the modern models. And there are several papers approaching this phenomenon from different angles, but it still remains rather mysterious (the few-shot learning and all the GPT-3 magic would be quite unlikely without it).
no subject
Date: 2023-03-24 05:14 am (UTC)I suspect something similar happens in Transformer architectures but it is speculation at this point.
no subject
Date: 2023-03-24 05:34 am (UTC)(I do think it makes sense to replace scalar flows by vector flows on the level of individual neurons; this change immediately turns any classical neural net into an attention machine.)
no subject
Date: 2023-03-25 08:14 pm (UTC)no subject
Date: 2023-03-26 03:17 am (UTC)Are you publishing something about this form of attention?
no subject
Date: 2023-03-26 05:41 am (UTC)no subject
Date: 2023-03-27 01:36 am (UTC)no subject
Date: 2023-03-26 05:39 pm (UTC)no subject
Date: 2023-03-27 01:40 am (UTC)Я сделал небольшие заметки на эту тему, и в этих заметках есть ссылки на подробности:
https://dmm.dreamwidth.org/64571.html
https://github.com/anhinga/2022-notes/tree/main/Grokking-is-solved
no subject
Date: 2023-03-27 02:55 am (UTC)