ny_quant: (Default)
[personal profile] ny_quant

CLAIM: Greta Thunberg recently deleted a tweet she posted in 2018 in which she said that the world will end in 2023.

AP’S ASSESSMENT: Missing context. While Thunberg did delete a 2018 tweet about the urgency of addressing climate change, she did not say the world would end in 2023. Her tweet included a quote from an article that said an influential scientist warned climate change “will wipe out all humanity” unless fossil fuel use was ended “over the next five years.” Further complicating the issue, that article incorrectly summarized the scientist’s speech. He never made such comments.

...Thunberg never said the world was set to end in 2023. The young activist was quoting an article that was paraphrasing a speech by a Harvard University professor of atmospheric chemistry. The scientist said the world had limited time to act to reverse the disappearance of floating ice volume in the Arctic or there would be drastic consequences, not that the world would end in five years.

The article Thunberg was quoting has since been deleted, but it was published by the site Grit Post in February 2018 with the headline: “Top Climate Scientist: Humans Will Go Extinct if We Don’t Fix Climate Change by 2023,” according to an archived version of the web page. The first line of the story stated: “A top climate scientist is warning that climate change will wipe out all of humanity unless we stop using fossil fuels over the next five years.”

Forbes quoted Anderson as saying: “The chance that there will be any permanent ice left in the Arctic after 2022 is essentially zero.”


У снопса есть скриншот того твита и сходный анализ.

Date: 2023-03-21 11:34 pm (UTC)
From: [identity profile] misha-b.livejournal.com
I think it is clear she did not claim the world would end in five years. The world would end _if_ we do not stop using fossil fuel in five years.

The last claim is obviously correct.

Date: 2023-03-22 12:20 am (UTC)
From: [identity profile] anhinga-anhinga.livejournal.com
> The world would end _if_ we do not stop using fossil fuel in five years.

The last claim is obviously correct.


This sounds really weird, unless your logic is something approximately like "because we have not ended the use of fossil fuels, we have enough electricity to develop strong AI, and strong AI will end the world".

That would be somewhat plausible...

Other than that though, "obviously correct" sounds really weird...

Date: 2023-03-22 12:22 am (UTC)
From: [identity profile] misha-b.livejournal.com

The logic is much simpler. Our world would one day end (whether from AI or something else). Therefore if we do not stop using fossil fuels it would end. If we do, it would still end ;)

Date: 2023-03-22 12:31 am (UTC)
From: [identity profile] anhinga-anhinga.livejournal.com
Ah!

:-) I seem to recall that people invented "relevant logic" (https://en.wikipedia.org/wiki/Relevance_logic) which is informally recommended for informal arguments :-)

But you are clearly not bullish on "true immortality" :-)

Date: 2023-03-22 05:09 am (UTC)
From: [identity profile] misha-b.livejournal.com
Same people who invented non-standard analysis? ;)

I think we are too good at making weapons out of everything for any sort of immortality. The more I think about it, the clearer it becomes that GPT-type models will be like nuclear weapons. Imagine a virus with an ability to fully impersonate you running on your cell phone.
Edited Date: 2023-03-22 05:12 am (UTC)

Date: 2023-03-22 08:47 am (UTC)
From: [identity profile] anhinga-anhinga.livejournal.com
I can't decide whether LLMs will be like nukes on their own (they might, just like a number of other existing AI systems), but recursive self-improvement is almost at hand (the state of automated code generation including API use, of automated reasoning, of AutoML, all these are in excellent shape and rapidly improving, so an "artificial AI researcher" is getting more and more feasible, and when a population of those starts to produce improved "artificial AI researchers" a very classical "intelligence explosion" seems highly likely).

Date: 2023-03-22 03:14 pm (UTC)
From: [identity profile] misha-b.livejournal.com
I am not convinced recursive self-improvement is coming soon. I think current neural models are highly suboptimal in any case.

But combined with a group of dedicated hackers, the damage can already be close to nuclear (at least once they are able to run on commodity hardware). A computer virus that can impersonate the owner is a pretty scary thought..
Edited Date: 2023-03-22 03:22 pm (UTC)

Date: 2023-03-23 01:35 am (UTC)
From: [identity profile] anhinga-anhinga.livejournal.com
Right. So who will find better models first, researchers doing "neural architecture search", or researchers hand-designing them?

That depends on how good the "artificial AI researcher" part of the team of those doing "neural architecture search" is...

Date: 2023-03-23 01:55 am (UTC)
From: [identity profile] anhinga-anhinga.livejournal.com
But also the current models have those wonderful properties (like attention-layers computing some large step gradients, which greatly helps the few-shot learning).

So, presumably, better models need to make sure to provide an equivalent of that too... And there might be other properties of today's LLMs we would want to capture (or further upgrade) in better models.

Date: 2023-03-23 05:15 am (UTC)
From: [identity profile] misha-b.livejournal.com
Well, I am not a believer in neural architecture search. It seems too much like graduate student descent to me. We have to go back to first principles informed by the successes of neural networks, at least that is what I have been trying to do.

I agree, attention seems really important. Backpropagation not so much, perhaps.

Date: 2023-03-23 02:27 pm (UTC)
From: [identity profile] anhinga-anhinga.livejournal.com
The class of machines I have been playing with is, essentially, a class of flexible attention machines (in the sense that a linear combination of vectors is actually the quintessential form of attention, so it is indeed fundamental).

But the magic property of **attention layers** performing gradient steps on the fly (without backpropagation) is something else; it does seem to be responsible for many of the magic properties of the modern models. And there are several papers approaching this phenomenon from different angles, but it still remains rather mysterious (the few-shot learning and all the GPT-3 magic would be quite unlikely without it).

Date: 2023-03-24 05:14 am (UTC)
From: [identity profile] misha-b.livejournal.com
I am reasonably confident now that we understand fully connected networks. They also have a form of attention, in fact (Expected Gradient Outer Product, essentially, supervised PCA).

I suspect something similar happens in Transformer architectures but it is speculation at this point.

Date: 2023-03-24 05:34 am (UTC)
From: [identity profile] anhinga-anhinga.livejournal.com
Fully connected feed-forward multi-layer ones, or fully-connected recurrent single-layer ones?

(I do think it makes sense to replace scalar flows by vector flows on the level of individual neurons; this change immediately turns any classical neural net into an attention machine.)

Date: 2023-03-25 08:14 pm (UTC)
From: [identity profile] misha-b.livejournal.com
I mean fully connected multi-layer networks. Not sure about recurrent ones.

Date: 2023-03-26 03:17 am (UTC)
From: [identity profile] anhinga-anhinga.livejournal.com
> I am reasonably confident now that we understand fully connected networks. They also have a form of attention, in fact (Expected Gradient Outer Product, essentially, supervised PCA).

Are you publishing something about this form of attention?

Date: 2023-03-26 05:41 am (UTC)
From: [identity profile] misha-b.livejournal.com
I sent a link by a deleted comment.

Date: 2023-03-27 01:36 am (UTC)
From: [identity profile] anhinga-anhinga.livejournal.com
Спасибо, очень любопытно.

Date: 2023-03-26 05:39 pm (UTC)
From: [identity profile] misha-b.livejournal.com
Btw, it also explains grokking. I remember you asked me a couple of years ago about it, but I don't think my answer was correct.

Date: 2023-03-27 01:40 am (UTC)
From: [identity profile] anhinga-anhinga.livejournal.com
А, да, все основные загадки того, почему и как происходит Grokking, решил Neel Nanda в августе. Замечательная работа, с подробностями и деталями.

Я сделал небольшие заметки на эту тему, и в этих заметках есть ссылки на подробности:

https://dmm.dreamwidth.org/64571.html

https://github.com/anhinga/2022-notes/tree/main/Grokking-is-solved

Date: 2023-03-27 02:55 am (UTC)
From: [identity profile] misha-b.livejournal.com
Thanks, interesting work. From a quick look, I think the main idea is right. However the general mechanism was not clarified. EGOP makes clear what feature learning means in general.

Profile

ny_quant: (Default)
ny_quant

December 2025

S M T W T F S
 12 34 56
7 89 10 111213
14 151617 181920
21 2223 24252627
28 29 3031   

Most Popular Tags

Style Credit

Expand Cut Tags

No cut tags
Page generated Dec. 31st, 2025 09:28 am
Powered by Dreamwidth Studios