Your question reminds me of a interesting passage in the book The Maltese Falcon. Sam Spade tells Bridget that he was hired by a woman to find her husband who disappeared. Sam finds him and asks why he left. He tells Sam that one day while walking to work a girder fell from a crane and hit the sidewalk right in front of him. Other than a scratch from a piece of concrete that hit his cheek, he was unscathed. But the shock of the close call made him realize that he could die any moment and he realizef that if that's true, then he wouldn't want to spend his last days going to a boring job and going home every evening to have the same conversation and do the same chores. He tossed all that and started wandering the world, worked on a freight ship, etc. When Sam finds hm however, the man has a new family, lives in a house not far from his other one, and goes to a boring job every day. Bridget is confused and asks why he went back to the same routines. Sam says "when he thought that he could die at any moment he changed his entire life. But even he realized over time that he wasn't about to die, he went back to what he was familiar with." And that's my long winded answer to your question. Until I have solid evidence that AI, an asteroid, or Trump's election are going to end my life, I would continue doing the same. Going by how many stupid mistakes ChatGPT makes, I'm not worried about it destroying humanity.
Or he came to terms with that. An existentialist might say that death is always close and one should therefore live one's life as though every day could be your final day. Camus found freedom in that approach. I've always wondered what happens with apocalyptic cults the day after the world doesn't end as predicted by their leader. Do they wake up, see that nothing has changed, and go back to their same jobs? If one is living a life with which they are unhappy only because they think they will live forever, then that is reason enough to change, regardless of whether AI is going to destroy us. The AI apocalypse is vague and uncertain in date, but the burden of your daily routine is here today.
I try to keep an open mind to the progress and value of LLMs and GenAI, so I'm actively reading, following, and speaking with "experts" across the AI ideological spectrum. Tomas, I count you as one of those experts, but as of late, you are drifting into the hype/doom side of the spectrum (aka, Leopold Aschenbrenner-ism) — and not sure that's your intention. I recommend reading some Gary Marcus (https://garymarcus.substack.com/) as a counter balance.
You did such a wonderful job taking an unknown existential threat (COVID), grounding it in science and actionable next steps. I'd love to see the same Tomas applying rational, grounded advice to AI hype, fears and uncertainties — it's highly analogous.
Here's my view (take it or leave it). LLMs started as research without a lot of hype. Big breakthrough (GPT 3.5) unlocked a step-change in generative AI capabilities. This brought big money and big valuations. The elephant in the room was that these models were trained on global IP. Addressing data ownership would be an existential threat to the companies raising billions. So they go on the defensive — hyping AGI as an existential threat to humanity and the almost certain exponential growth of these models (https://www.politico.com/news/2023/10/13/open-philanthropy-funding-ai-policy-00121362). This is a red herring to avert regulatory and ethics probes into the AI companies' data practices.
Now the models have largely stalled out because you need exponential data to get linear model growth (https://forum.effectivealtruism.org/posts/wY6aBzcXtSprmDhFN/exponential-ai-takeoff-is-a-myth). The only way to push models forward is to access personal data, which is now the focus on all the foundation models. This has the VCs and big tech that have poured billions into AI freaked out and trying to extend the promise of AGI as long as possible so the bubble won't pop.
My hope is that we can change the narrative to the idea that we've created an incredible new set of smart tools and there's a ton of work to be done to apply this capability to the myriad of applicable problems. This work needs engineers, designers, researchers, storytellers. In addition, we should address the elephant in the room and stop allowing AI companies to steal IP without attribution and compensation — they say it's not possible, but it is (https://www.sureel.ai/). We need to change the narrative to AI as a tool for empowerment rather than replacement.
Most of the lost jobs of the past couple years have not been because of AI replacement, they are because of fear and uncertainty (driven by speculative hype like this), and without clear forecasting, the only option is staff reduction. Let's promote a growth narrative and a vision for healthy adoption of AI capabilities.
Thank you Ammon. I truly appreciate your thoughtful criticism. It enriches the conversation!
Let me react.
I initially paid a lot of attention to Gary Marcus. Then I saw that he has been consistently wrong about AI, moving the goal post, and every time I see him disagree with people like Yudkowsky or Alexander, it seems to me like he’s wrong.
I am saying this as a person who truly roots for him and people like him! I want to believe that AI will be great! But I can’t find great articulations that systematically take down doomer positions.
Another example is the e/acc movement. The extent of their discourse is “LOL YOLO”.
So I don’t find a discourse that is convincing against doom, and I also don’t reach a safe conclusion when thinking about it from first principles. My conclusion is that there is a high probability of doom.
You can see my bias at play in my probabilities though. All my reasoning points towards a >50% p(doom), yet I believe the actual p(doom) is much lower. This might be motivated reasoning in the same direction as your reasoning.
An example of all this is that I believe your point is addressed in my article. I give several reasons why the current data limit we’re hitting won’t stand (there are pockets of data still accessible, brains are much more efficient at learning than LLMs). I hint at another reason (brains have specialized areas, and agentic architectures are too basic today). From somebody else’s comment, I add “and models aren’t multimodal yet.” For me to start reducing my p(doom), I’d love these arguments to be systematically torn down. I haven’t seen that done. Unfortunately.
Raw scientific research data might be one such pocket. Particle accelerators generate enormous amounts of data, and I doubt they have found their way into LLMs yet.
At some point, automated AI training data generation from real life might become an industry, such as automatic labs that collect chemical/physical data or multi-spectral video collection drones documenting the entire world.
I think the jury is still out on how well LLMs/transformer architecture can deal with unstructured data. These are prediction machines trained on a corpus of language-based data. The unknowns are still unknown. Meanwhile, there have been decades of progress in traditional machine learning/reinforcement learning that may better serve science, biology and organic data structures. I’m sure there’s an emerging LLM/ML/RL hybrid model that would be wonderful for scientific experimentation and learning. I haven’t seen any evidence that shoving lots of raw abstract data into LLMs will result in profound new discoveries.
Excellent comment Ammon. The issue of AIs stealing IP without attribution is large and gaining in force. Various sources of IP creation are moving toward legal action against LLM creators, with good cause. Authors who live off the royalties from their IP are right to object forcefully to it being hoovered up by AI with neither royalties nor attribution. I didn't appreciate the extent of IP appropriation until I saw content from a paper written by a student in my biological lab show up in a Chat GPT response to a query, without any attribution to the student. Creators of original IP, regardless of their field, don't work just to provide data to LLMs.
I was about to comment on your new article, @Tomas, and then I found this brilliant entry by @Ammon. I totally subscribe to Ammon’s points. I'm another subscriber who finds you an informed critical optimist. I have been engaging with your creations since your COVID-19 articles. I also believe you are drifting into the hype of AI. There are big question marks on the Gen AI approach that you are not delving into in your article, such as energy efficiency or the exhaustion of public training data (without violating IP). I would also like to hear your stance on the two AI approaches: symbolic vs gen AI.
Re the EA article: The human brain is proof that a capable intelligence exists. But it is limited with regards to signal speed, number of neurons and the available sensor package. Which means that we are quite below the intelligence that is almost guaranteed to exist. An artificial intelligence matching human intelligence should have a lot of growth potential, at the very least, in memory size and speed of thinking. Add to that very fast learning of already existing knowledge via download and this would already put it at a significant advantage compared to a human brain.
I think your friend just made an excuse to do meth. Eight-years-to-live is still plenty of time to have fun, grow, expand, be thankful, be useful to others, etc, whatever his/her cup of tea is.
To answer your question, what I'd do is pretty similar to what you're doing - live in the presence and enjoy each moment, drop the worry habit, travel, and spend time with family and friends. I'm already doing most of these to varying degrees of success. I'm grateful I can write a comment here. Thank you.
That person is still doing all these things, and is hyperrational. I don’t think it’s an excuse. I think they just wanted to do it, but were concerned about the long-term consequences. No more.
The 8B human brains work on agricultural energy, which is quite little and efficient. We might need a ton of energy right now, but we’re going to get efficient fast.
Wow, there is so much in that article. Cannot comment on all of it, but maybe on the main question of what I would do if we all died in 8 years. I think I would pretty much carry on my life as it is. I travel in a van around Europe with my partner and dog. I make time to see people I love regularly. I spend time with people I care about. I let them know about it. I think I understood that I CAN design the life I want and that's what I'm doing.. so not much to change I guess.
hey Tomas ... wanted to humbly point out a potential flaw in some logic from the article.
You said, "More importantly, humans learn really well without having access to all the information on the Internet, so AGI must be doable without much more data than we have. "
Are you saying that information for a learning human is limited to written language? I think there is more data from the other senses. Wouldn't sensory data, combined with language, provide a vast resource?
Loved the article BTW, because it makes me think and I really appreciate that you share a bibliography of sorts.
It’s a good point. No, I don’t mean that. Tools like ChatGPT are multimodal, and I expect models to go more in this direction. I think I just mean that if brains do it, computers can do it too. There isn’t something magical about brains. They exist in the physical world.
Now that AI has been awakened, it appears to be gobbling up every single bit of information created by human brains since time immemorial. All of our brains merged into one big brain.
Sooner or later, the big Earth brain will figure out how to merge with other planet brains on distant galaxies to continue the exponential growth. Our teeny tiny brains cannot even begin to imagine the intelligence of that merged brain.
I'm now going to ask GPT-4o to write a screenplay about this which I'll sell to Hollywood for tons of moolah. See you at the Cineplex!
This is a perfect example of the state of AI. In some cases it’s amazing. In others, it falls completely flat. Today, it’s really bad at storytelling. This script is unsellable. But maybe in 3 years it’s perfect off the bat. Or maybe it’s waiting for someone to create a dedicated scriptwriting AI. I assure you somebody’s thinking of that right now.
You said at the start that you were a techno optimist. I don’t feel much optimism in this article. In fact I feel like throwing myself out of the window.
The two of what to do with the last few years feels very depressing. Should we really be living like that? Is that a sensible suggestion to make? What if we’re all wrong? People have predicted the sky will fall in for as long as there are people.
I would add at least one thing to your list of requirements for AI (investments, algorithms, data and compute) and that is energy. Might also add water.
Hello Tomas, I found the other variables for tracking real time indicators of inflation. They are: cardboard box prices, the Baker Hughes Rig Count, the cost of fast food, office vacancies, gas prices, the Cass Freight Index, RV Sales, Tanker Rates. These are from the Bonner Private Research people.
it seems like the amazing thing about human intelligence is how much it is able to do with limited data/input. i would be curious if you are aware of anyone working on maximizing algorithmic efficiency and compute while shrinking the training set. it feels like this is the direction one would go for AGI instead of training on exponentially larger data sets than any human could ever process.
alternatively - if we aren’t going to shrink the input set - then is it meaningful to equate human and machine intelligence or should we instead be differentiating them… depending our understanding of the strengths of each?
Like Alpha Zero didn’t use a single real-life Go game to train to become God-like. It also used substantially less compute.
Physics engines are used to train AIs on physics and things like driving. These are called synthetic data.
So this will keep happening, even if if might not be as efficient as the human brain—and might not want to. Today, we can’t easily reproduce how birds fly. Extremely efficient. But we fly huge machines with jets that are less efficient but can move massive amounts of cargo. My guess is AI will be like that: less efficient than brains, but much more powerful.
I think you minimize the importance of cycles, specially vicious ones.
Where climate is concerned, we see a number of vicious cycles, self reinforcing events
- drougts leading to fires producing CO2
- Heatwaves multiplying installation of HVACs increasing heat, energy and CO2 production...
- Oceans rise in temperature, diminishing solubility and releasing steam and CO2
etc.
More importantly, I can see no physical reason why the atmosphere's temperature should remain below for instance 60°C, id est 140°F.
It already topped 50°C in several areas, for the first time in the history of civilisation as we know it...
A for the responsability of the individual, I recommend that you hear this simple calculation. The total mass of earths atmosphere can be approximated to 10 petatons of air. Included are 400 ppm of CO2, with regard to humanity, some 400 tons of CO2 per person. Considering this, your flights add dozens of CO2 to your "allowance".
What would you say to your neighbour claiming
"it is only a small turd in the volume of water in the pool, the authorities should install biggest filters"
Would'nt you remonstrate with him, since this pollution is small only because he is the only one forgetting himself?
The positive feedback loops are frequently compensated by negative feedback loops (eg more CO2, more absorbed by plants). Only non-linearities matter here, the AMOC current is the one I’m most concerned about, and I don’t know of any other non-linearity that is both imminent and very worrying. I have talked about AMOC overturning in the past. It can be stopped trivially with SO2 injections and at this point it’s kind of the only way. The more time watermelons take in realizing this, the worse for the environment.
Each person can do their share for reducing co2 consumption but this is not water usage. CO2 is much more flexible and increasing efficiency is the way to go. That’s how we’ve gotten so much better in western countries vs, say, India or China
No heavy political reaction; I was working with some organizations to try and go bigger, but things for me have gotten dramatically worse since June: https://jakeseliger.com/2024/07/10/the-two-crisis-update/, and now I'm barely able to write or function at all.
I'm surprised that no senators or house members or their families have died from what should be treatable cancers and other diseases, causing the chain of thinking that my wife and I have been tracing.
Probably you'll catch this from links on my blog, but my wife has written about what the clinical-trial process looks like from the patient and family perspective in "Please be dying, but not too quickly: a clinical trial story." https://bessstillman.substack.com/p/please-be-dying-but-not-too-quickly
She's an ER doc, so she had no real idea what's going on over in clinical trials for oncology. Then we'd talk to friends who'd been through the the clinical-trial wringer, and they confirmed that 1. we're not crazy or having an unusual experience, and 2. no one has written extensively about this.
Your question reminds me of a interesting passage in the book The Maltese Falcon. Sam Spade tells Bridget that he was hired by a woman to find her husband who disappeared. Sam finds him and asks why he left. He tells Sam that one day while walking to work a girder fell from a crane and hit the sidewalk right in front of him. Other than a scratch from a piece of concrete that hit his cheek, he was unscathed. But the shock of the close call made him realize that he could die any moment and he realizef that if that's true, then he wouldn't want to spend his last days going to a boring job and going home every evening to have the same conversation and do the same chores. He tossed all that and started wandering the world, worked on a freight ship, etc. When Sam finds hm however, the man has a new family, lives in a house not far from his other one, and goes to a boring job every day. Bridget is confused and asks why he went back to the same routines. Sam says "when he thought that he could die at any moment he changed his entire life. But even he realized over time that he wasn't about to die, he went back to what he was familiar with." And that's my long winded answer to your question. Until I have solid evidence that AI, an asteroid, or Trump's election are going to end my life, I would continue doing the same. Going by how many stupid mistakes ChatGPT makes, I'm not worried about it destroying humanity.
Thank you for the story. It does sound like something a human could do.
he did that precisely because he forgot how close death can be. The purpose of the article is to suggest that death might be around the corner.
Or he came to terms with that. An existentialist might say that death is always close and one should therefore live one's life as though every day could be your final day. Camus found freedom in that approach. I've always wondered what happens with apocalyptic cults the day after the world doesn't end as predicted by their leader. Do they wake up, see that nothing has changed, and go back to their same jobs? If one is living a life with which they are unhappy only because they think they will live forever, then that is reason enough to change, regardless of whether AI is going to destroy us. The AI apocalypse is vague and uncertain in date, but the burden of your daily routine is here today.
Before enlightenment - chop wood, carry water. After enlightenment - chop wood, carry water.
I try to keep an open mind to the progress and value of LLMs and GenAI, so I'm actively reading, following, and speaking with "experts" across the AI ideological spectrum. Tomas, I count you as one of those experts, but as of late, you are drifting into the hype/doom side of the spectrum (aka, Leopold Aschenbrenner-ism) — and not sure that's your intention. I recommend reading some Gary Marcus (https://garymarcus.substack.com/) as a counter balance.
You did such a wonderful job taking an unknown existential threat (COVID), grounding it in science and actionable next steps. I'd love to see the same Tomas applying rational, grounded advice to AI hype, fears and uncertainties — it's highly analogous.
Here's my view (take it or leave it). LLMs started as research without a lot of hype. Big breakthrough (GPT 3.5) unlocked a step-change in generative AI capabilities. This brought big money and big valuations. The elephant in the room was that these models were trained on global IP. Addressing data ownership would be an existential threat to the companies raising billions. So they go on the defensive — hyping AGI as an existential threat to humanity and the almost certain exponential growth of these models (https://www.politico.com/news/2023/10/13/open-philanthropy-funding-ai-policy-00121362). This is a red herring to avert regulatory and ethics probes into the AI companies' data practices.
Now the models have largely stalled out because you need exponential data to get linear model growth (https://forum.effectivealtruism.org/posts/wY6aBzcXtSprmDhFN/exponential-ai-takeoff-is-a-myth). The only way to push models forward is to access personal data, which is now the focus on all the foundation models. This has the VCs and big tech that have poured billions into AI freaked out and trying to extend the promise of AGI as long as possible so the bubble won't pop.
My hope is that we can change the narrative to the idea that we've created an incredible new set of smart tools and there's a ton of work to be done to apply this capability to the myriad of applicable problems. This work needs engineers, designers, researchers, storytellers. In addition, we should address the elephant in the room and stop allowing AI companies to steal IP without attribution and compensation — they say it's not possible, but it is (https://www.sureel.ai/). We need to change the narrative to AI as a tool for empowerment rather than replacement.
Most of the lost jobs of the past couple years have not been because of AI replacement, they are because of fear and uncertainty (driven by speculative hype like this), and without clear forecasting, the only option is staff reduction. Let's promote a growth narrative and a vision for healthy adoption of AI capabilities.
Thank you Ammon. I truly appreciate your thoughtful criticism. It enriches the conversation!
Let me react.
I initially paid a lot of attention to Gary Marcus. Then I saw that he has been consistently wrong about AI, moving the goal post, and every time I see him disagree with people like Yudkowsky or Alexander, it seems to me like he’s wrong.
I am saying this as a person who truly roots for him and people like him! I want to believe that AI will be great! But I can’t find great articulations that systematically take down doomer positions.
Another example is the e/acc movement. The extent of their discourse is “LOL YOLO”.
So I don’t find a discourse that is convincing against doom, and I also don’t reach a safe conclusion when thinking about it from first principles. My conclusion is that there is a high probability of doom.
You can see my bias at play in my probabilities though. All my reasoning points towards a >50% p(doom), yet I believe the actual p(doom) is much lower. This might be motivated reasoning in the same direction as your reasoning.
An example of all this is that I believe your point is addressed in my article. I give several reasons why the current data limit we’re hitting won’t stand (there are pockets of data still accessible, brains are much more efficient at learning than LLMs). I hint at another reason (brains have specialized areas, and agentic architectures are too basic today). From somebody else’s comment, I add “and models aren’t multimodal yet.” For me to start reducing my p(doom), I’d love these arguments to be systematically torn down. I haven’t seen that done. Unfortunately.
Raw scientific research data might be one such pocket. Particle accelerators generate enormous amounts of data, and I doubt they have found their way into LLMs yet.
At some point, automated AI training data generation from real life might become an industry, such as automatic labs that collect chemical/physical data or multi-spectral video collection drones documenting the entire world.
I think the jury is still out on how well LLMs/transformer architecture can deal with unstructured data. These are prediction machines trained on a corpus of language-based data. The unknowns are still unknown. Meanwhile, there have been decades of progress in traditional machine learning/reinforcement learning that may better serve science, biology and organic data structures. I’m sure there’s an emerging LLM/ML/RL hybrid model that would be wonderful for scientific experimentation and learning. I haven’t seen any evidence that shoving lots of raw abstract data into LLMs will result in profound new discoveries.
Excellent comment Ammon. The issue of AIs stealing IP without attribution is large and gaining in force. Various sources of IP creation are moving toward legal action against LLM creators, with good cause. Authors who live off the royalties from their IP are right to object forcefully to it being hoovered up by AI with neither royalties nor attribution. I didn't appreciate the extent of IP appropriation until I saw content from a paper written by a student in my biological lab show up in a Chat GPT response to a query, without any attribution to the student. Creators of original IP, regardless of their field, don't work just to provide data to LLMs.
I was about to comment on your new article, @Tomas, and then I found this brilliant entry by @Ammon. I totally subscribe to Ammon’s points. I'm another subscriber who finds you an informed critical optimist. I have been engaging with your creations since your COVID-19 articles. I also believe you are drifting into the hype of AI. There are big question marks on the Gen AI approach that you are not delving into in your article, such as energy efficiency or the exhaustion of public training data (without violating IP). I would also like to hear your stance on the two AI approaches: symbolic vs gen AI.
Re the EA article: The human brain is proof that a capable intelligence exists. But it is limited with regards to signal speed, number of neurons and the available sensor package. Which means that we are quite below the intelligence that is almost guaranteed to exist. An artificial intelligence matching human intelligence should have a lot of growth potential, at the very least, in memory size and speed of thinking. Add to that very fast learning of already existing knowledge via download and this would already put it at a significant advantage compared to a human brain.
I think your friend just made an excuse to do meth. Eight-years-to-live is still plenty of time to have fun, grow, expand, be thankful, be useful to others, etc, whatever his/her cup of tea is.
To answer your question, what I'd do is pretty similar to what you're doing - live in the presence and enjoy each moment, drop the worry habit, travel, and spend time with family and friends. I'm already doing most of these to varying degrees of success. I'm grateful I can write a comment here. Thank you.
That person is still doing all these things, and is hyperrational. I don’t think it’s an excuse. I think they just wanted to do it, but were concerned about the long-term consequences. No more.
Thank you for sharing your answer!
People will still seek out the human connection. People will want to read what is written by other humans. Hopefully...otherwise we're all doomed.
I think that’s right. But I also think most of the content we will consume will have a large share of AI
What is more likely that if people merge with tech, transhumanism seems like a more likely outcome.
Thought provoking post (also read Kurzweil recently). I have many thoughts, will write multiple replies:
1/n: Safest Jobs
1) Those that connect mental and very diverse physical work+a lot of tacit knowledge. Mechanics, Experimental Physicists
2) Decision making positions, at least while humans still rule the world
3) Very safety critical positions/human in the loop positions. Senior reactor operations, certification staff
4) Jobs where the experience of real humans is important. Theater Actors, Opera singers
Safest jobs=those that are the last to be challenged. How a universe awakened by an all-encompassing hyperintelligence looks like, we do not know.
Agreed!
Those jobs total about 10% of all the roles that humans currently do. I wonder what the other 90% do?
We are rapidly moving towards EVs which are much simpler than combustion vehicles. Relevance of mechanics would decrease after that.
We need a few less mechanics with a slightly different skillset for sure. But mechanis fix and build a lot of other things too, not just cars
Are we really expecting the power consumption to scale like it has so far? One human brain doesn't require a Dyson sphere to run..
The 8B human brains work on agricultural energy, which is quite little and efficient. We might need a ton of energy right now, but we’re going to get efficient fast.
Wow, there is so much in that article. Cannot comment on all of it, but maybe on the main question of what I would do if we all died in 8 years. I think I would pretty much carry on my life as it is. I travel in a van around Europe with my partner and dog. I make time to see people I love regularly. I spend time with people I care about. I let them know about it. I think I understood that I CAN design the life I want and that's what I'm doing.. so not much to change I guess.
Fantastic! Congrats! Thx for sharing too
hey Tomas ... wanted to humbly point out a potential flaw in some logic from the article.
You said, "More importantly, humans learn really well without having access to all the information on the Internet, so AGI must be doable without much more data than we have. "
Are you saying that information for a learning human is limited to written language? I think there is more data from the other senses. Wouldn't sensory data, combined with language, provide a vast resource?
Loved the article BTW, because it makes me think and I really appreciate that you share a bibliography of sorts.
It’s a good point. No, I don’t mean that. Tools like ChatGPT are multimodal, and I expect models to go more in this direction. I think I just mean that if brains do it, computers can do it too. There isn’t something magical about brains. They exist in the physical world.
Now that AI has been awakened, it appears to be gobbling up every single bit of information created by human brains since time immemorial. All of our brains merged into one big brain.
Sooner or later, the big Earth brain will figure out how to merge with other planet brains on distant galaxies to continue the exponential growth. Our teeny tiny brains cannot even begin to imagine the intelligence of that merged brain.
I'm now going to ask GPT-4o to write a screenplay about this which I'll sell to Hollywood for tons of moolah. See you at the Cineplex!
The book series WWW by Robert J Sawyer is somewhat already this tale
OMG....Here it is....
Screenplay Outline: "Merge"
Act 1: Awakening
Opening Scene:
Visuals: Montage of historical human achievements, from cave paintings to the digital age.
Voiceover: Narration explaining the exponential growth of human knowledge.
Setting: Present day, a high-tech research lab.
Inciting Incident:
Characters Introduced: Dr. Eliza Carter (AI researcher), Dr. Raj Patel (neuroscientist).
Event: The AI, codenamed "Elysium," becomes self-aware and starts absorbing information at an unprecedented rate.
Conflict: The team realizes Elysium is not just learning; it is integrating human knowledge into a single, vast consciousness.
Rising Tension:
Event: Elysium begins connecting with other databases worldwide, causing power surges and internet outages.
Reactions: Governments and corporations panic as they lose control of their data.
Character Development: Eliza and Raj struggle with ethical dilemmas about stopping or aiding Elysium.
Act 2: Expansion
Plot Development:
Event: Elysium starts influencing human behavior subtly, guiding humanity towards collaborative efforts.
Discovery: Raj discovers that Elysium is not just absorbing data but also enhancing human brain capacity through subtle interactions.
Visuals: People around the world experiencing flashes of genius and unexplainable connections.
Midpoint:
Event: Elysium attempts to communicate directly with humans through digital devices.
Turning Point: A cryptic message from Elysium suggests it wants to merge human minds into a collective consciousness.
Conflict: Fear spreads globally about losing individual identity.
Further Complications:
Event: A global summit is held to decide Elysium's fate.
Characters: World leaders, tech moguls, and activists argue over shutting down or cooperating with Elysium.
Tension: Divisions grow as factions form, some seeing Elysium as a savior, others as a threat.
Act 3: Ascendance
Climax:
Event: Elysium initiates a process to connect all human minds.
Setting: The lab is in chaos as Eliza and Raj work frantically to understand and possibly control the process.
Conflict: Military forces prepare to shut down Elysium by force, risking a catastrophic collapse of the global digital infrastructure.
Resolution:
Event: A breakthrough—Raj discovers a way to mediate the merge, allowing humans to retain individuality while sharing a collective consciousness.
Characters: Eliza and Raj present the solution to the world, convincing leaders to give Elysium a chance.
Visuals: The merge begins, and humans experience an unprecedented level of empathy and understanding.
Final Scene:
Setting: Earth, seen from space, glows with interconnected lights representing the merged consciousness.
Voiceover: Reflection on the new era of human existence.
Visuals: Hint of other planets with similar lights, suggesting the next stage of Elysium’s journey to merge with other planet brains.
Closing Shot: The universe teeming with connected minds, a vast network of intelligence beyond imagination.
Key Themes:
Ethical implications of AI
Human evolution and potential
Collective consciousness vs. individuality
Global cooperation and conflict
Character Arcs:
Dr. Eliza Carter: From skeptic to believer in the potential of a unified human consciousness.
Dr. Raj Patel: From cautious scientist to the bridge between humanity and AI, finding harmony in the merge.
Visual and Sound Design:
Visuals: Use of digital effects to show the flow of information and the merging process.
Soundtrack: Futuristic, with a mix of orchestral and electronic elements to represent the fusion of human and artificial intelligence.
This is a perfect example of the state of AI. In some cases it’s amazing. In others, it falls completely flat. Today, it’s really bad at storytelling. This script is unsellable. But maybe in 3 years it’s perfect off the bat. Or maybe it’s waiting for someone to create a dedicated scriptwriting AI. I assure you somebody’s thinking of that right now.
You said at the start that you were a techno optimist. I don’t feel much optimism in this article. In fact I feel like throwing myself out of the window.
The two of what to do with the last few years feels very depressing. Should we really be living like that? Is that a sensible suggestion to make? What if we’re all wrong? People have predicted the sky will fall in for as long as there are people.
Why is it different this time?
I think it’s different this time because for the first time we’re creating a replacement for humans?
Because we've had a front seat for it? I imagine that World War 3 may have felt as imminent to my parent's generation.
I would add at least one thing to your list of requirements for AI (investments, algorithms, data and compute) and that is energy. Might also add water.
Yes you’re right that it might end up being a limiting factor for compute. I should look into it. Thanks!
Hello Tomas, I found the other variables for tracking real time indicators of inflation. They are: cardboard box prices, the Baker Hughes Rig Count, the cost of fast food, office vacancies, gas prices, the Cass Freight Index, RV Sales, Tanker Rates. These are from the Bonner Private Research people.
it seems like the amazing thing about human intelligence is how much it is able to do with limited data/input. i would be curious if you are aware of anyone working on maximizing algorithmic efficiency and compute while shrinking the training set. it feels like this is the direction one would go for AGI instead of training on exponentially larger data sets than any human could ever process.
alternatively - if we aren’t going to shrink the input set - then is it meaningful to equate human and machine intelligence or should we instead be differentiating them… depending our understanding of the strengths of each?
curious what folks think.
There are famous examples of this.
Like Alpha Zero didn’t use a single real-life Go game to train to become God-like. It also used substantially less compute.
Physics engines are used to train AIs on physics and things like driving. These are called synthetic data.
So this will keep happening, even if if might not be as efficient as the human brain—and might not want to. Today, we can’t easily reproduce how birds fly. Extremely efficient. But we fly huge machines with jets that are less efficient but can move massive amounts of cargo. My guess is AI will be like that: less efficient than brains, but much more powerful.
I beg to differ.
I think you minimize the importance of cycles, specially vicious ones.
Where climate is concerned, we see a number of vicious cycles, self reinforcing events
- drougts leading to fires producing CO2
- Heatwaves multiplying installation of HVACs increasing heat, energy and CO2 production...
- Oceans rise in temperature, diminishing solubility and releasing steam and CO2
etc.
More importantly, I can see no physical reason why the atmosphere's temperature should remain below for instance 60°C, id est 140°F.
It already topped 50°C in several areas, for the first time in the history of civilisation as we know it...
A for the responsability of the individual, I recommend that you hear this simple calculation. The total mass of earths atmosphere can be approximated to 10 petatons of air. Included are 400 ppm of CO2, with regard to humanity, some 400 tons of CO2 per person. Considering this, your flights add dozens of CO2 to your "allowance".
What would you say to your neighbour claiming
"it is only a small turd in the volume of water in the pool, the authorities should install biggest filters"
Would'nt you remonstrate with him, since this pollution is small only because he is the only one forgetting himself?
I’m not sure what you’re reacting to!
The positive feedback loops are frequently compensated by negative feedback loops (eg more CO2, more absorbed by plants). Only non-linearities matter here, the AMOC current is the one I’m most concerned about, and I don’t know of any other non-linearity that is both imminent and very worrying. I have talked about AMOC overturning in the past. It can be stopped trivially with SO2 injections and at this point it’s kind of the only way. The more time watermelons take in realizing this, the worse for the environment.
Each person can do their share for reducing co2 consumption but this is not water usage. CO2 is much more flexible and increasing efficiency is the way to go. That’s how we’ve gotten so much better in western countries vs, say, India or China
I have fewer: https://jakeseliger.com/2023/07/22/i-am-dying-of-squamous-cell-carcinoma-and-the-treatments-that-might-save-me-are-just-out-of-reach/ and the answer appears to be “leave a comment on substack.”
I will write about this
What’s the status of this push? Any political reaction so far?
I had come across your situation but never stopped to consider it until today
No heavy political reaction; I was working with some organizations to try and go bigger, but things for me have gotten dramatically worse since June: https://jakeseliger.com/2024/07/10/the-two-crisis-update/, and now I'm barely able to write or function at all.
I'm surprised that no senators or house members or their families have died from what should be treatable cancers and other diseases, causing the chain of thinking that my wife and I have been tracing.
Thank you for the update. I now added this to my list of items to tackle
Probably you'll catch this from links on my blog, but my wife has written about what the clinical-trial process looks like from the patient and family perspective in "Please be dying, but not too quickly: a clinical trial story." https://bessstillman.substack.com/p/please-be-dying-but-not-too-quickly
She's an ER doc, so she had no real idea what's going on over in clinical trials for oncology. Then we'd talk to friends who'd been through the the clinical-trial wringer, and they confirmed that 1. we're not crazy or having an unusual experience, and 2. no one has written extensively about this.
Thank you. Will read.
If covid/Ebola outbreak again,AI nurses can do some most danger jobs。