The Alignment Problem (Part 1)
Episode Summary
In this episode, Justin and Nick dive into The Alignment Problem—one of the most pressing challenges in AI development. Can we ensure that AI systems align with human values and intentions? What happens when AI behavior diverges from what we expect or desire?
Drawing on real-world examples, academic research, and philosophical thought experiments, they explore the risks and opportunities AI presents. From misaligned AI causing unintended consequences to the broader existential question of intelligence in the universe, this conversation tackles the complexity of AI ethics, governance, and emergent behavior.
They also discuss historical perspectives on automation, regulatory concerns, and the possible future of AI—whether it leads to existential risk or a utopian technological renaissance.
Topics Covered
Understanding the AI Alignment Problem – Why AI alignment matters and its real-world implications.
Why Not Just ‘Pull the Plug’ on AI? – A philosophical and practical discussion.
Emergent AI & Unpredictability – How AI learns in ways we can’t always foresee.
Historical Parallels – Lessons from past industrial and technological revolutions.
The Great Filter & The Fermi Paradox – Could AI be part of humanity’s existential challenge?
The Ethics of AI Decision-Making – The real-world trolley problem and AI’s moral choices.
Can AI Ever Be Truly ‘Aligned’ with Humans? – Challenges of defining and enforcing values.
Industry & Regulation – How governments and businesses are handling AI risks.
What Happens When AI Becomes Conscious? – A preview of the next episode’s deep dive.
Reading List & References
Books Mentioned:
The Alignment Problem – Brian Christian
Human Compatible – Stuart Russell
Superintelligence – Nick Bostrom
The Second Machine Age – Erik Brynjolfsson & Andrew McAfee
The End of Work – Jeremy Rifkin
The Demon in the Machine – Paul Davies
Anarchy, State, and Utopia – Robert Nozick
Academic Papers & Reports:
- Clarifying AI Alignment – Paul Christiano
- The AI Alignment Problem in Context – Raphaël Millière
Key Takeaways
- AI alignment is crucial but deeply complex—defining human values is harder than it seems.
- AI could be an existential risk or the key to ending scarcity and expanding humanity’s potential.
- Conscious AI might be necessary for true alignment, but we don’t fully understand consciousness.
- Industry and government must work together to create effective AI governance frameworks.
- We may be at a pivotal moment in history—what we do next could define our species’ future.
Pick of the Pod
🔹 Nick’s Pick: Cursor – An AI-powered coding assistant that enhances development workflows.
🔹 Justin’s Pick: Leveraging Enterprise AI – Make use of company-approved AI tools for efficiency and insight.
Next Episode Preview
In Part 2 of “The Alignment Problem”, we’ll explore:
🔹 Can an AI be truly conscious, and would that change alignment?
🔹 What responsibilities would we have toward a sentient AI?
🔹 Could AI help us become better moral actors?
The Emergent AI website Subscribe & stay tuned!
Join the Conversation!
We want to hear your thoughts on our aligned future!
Justin’s Homepage - https://justinaharnish.com
Justin’s Substack - https://ordinaryilluminated.substack.com
Justin’s LinkedIn - https://www.linkedin.com/in/justinharnish/
Nick’s LinkedIn - https://www.linkedin.com/in/nickbaguley/
Like, Subscribe & Review on your favorite podcast platform!
Final Thought: Are we heading toward an AI utopia or existential risk? The answer may depend on how we approach alignment today.
Transcript
Welcome back to the Emergent AI Podcast.
-:I'm Justin Harnish here with my co-host, Nick Bagley.
-:In our previous episode, we explored the symbiosis between humans and AI.
-:And today, we'll embark on a critical journey into the alignment problem.
-:The challenge of ensuring AI systems operate in harmony with human values and attentions.
-:Excellent, Justin.
-:You know, as AI systems become more advanced,
-:really thinking about their goals and how we need to align them
-:or what it is that we need to do to make them as successful as possible
-:while not potentially creating a threat to us
-:really becomes more crucial and more complex as we think about
-:what to do to avoid any form of misalignment.
-:And so really that misaligned AI can lead to unintended consequences,
-:raising ethical and safety concerns that we need to address.
-:Yeah, so over the next two episodes, we'll be dissecting the alignment problem,
-:what it entails, examining real-world use cases,
-:where AI systems have veered off course
-:and discussed the ongoing efforts to land to align AI with human values.
-:So let's dive right in.
-:And I want to start with a bit of a discussion
-:from a class that I was helping to teach on AI and the law.
-:And so I was up at the University of Utah.
-:A good friend of ours was teaching a class. He's a lawyer.
-:And I had the opportunity to do a couple of classes,
-:and one of which we had a problem where it was AI rights
-:versus essentially the alignment problem.
-:And these students were all third-year law students
-:and had a real fear, like I think many in our audience have,
-:of misaligned AI.
-:Basically they had a fear of all AI.
-:And the question that they came to after seeing it hallucinate
-:on things that they had learned in their case law studies
-:was, why not just pull the plug?
-:Why not just pull the plug right now
-:and be done with this?
-:We don't seem to have a good scope to not pulling the plug.
-:And so the question arises, Nick, and maybe we'll start here.
-:Why not pull the plug?
-:It's an important question.
-:It's easy to hear a question like that
-:and think that maybe it's short-sighted
-:or maybe it's something that's hyper-focused on the present
-:and in this moment when the reality is that, no,
-:it's actually one of the most existential questions of our time
-:or at least potentially.
-:And so today I think we'll talk about the alignment problem itself
-:and how that's been outlined.
-:As I was researching this, though, and going through some of the books,
-:some of the things that have been created in the past,
-:I went into the general thoughts
-:and how everyone was thinking about this problem
-:even a few years ago compared to the way that we need to think about it today.
-:And as I was doing that, I started sharing ideas back and forth with GPT-4.
-:And I went into some of the concepts
-:that I think are really critical components that we need to consider
-:and that we've been talking about actually during the podcast.
-:And GPT-4 worked with me to create a really compelling
-:and really nuanced argument that really emphasizes the unpredictability
-:of emergent AI and how AI's future capabilities are very difficult to understand
-:where they may go, what they may do, and also what the impacts could be.
-:Also, really, one of the core concepts of many of the past books
-:was talking about aligning specifically to humans' intentions.
-:And there are inherent difficulties with actually aligning the AI with the human intentions
-:because even human intentions are very difficult to align one with another.
-:And so as we navigate the transformative landscape of AI,
-:and this is partially GPT-4 or partially me,
-:we face the immense challenges of aligning these systems with our collective future.
-:While traditional methods of aligning AI with human intentions have their merits,
-:they often struggle to capture the complexity and unpredictability
-:of both human values and emergent AI behavior.
-:What if instead we consider aligning AI with the fundamental principles that govern the universe?
-:Principles of growth, balance, and even entropy.
-:This approach isn't about rejecting human values, but about complimenting them
-:with a framework that is more robust and adaptable to change.
-:I invite all of you to share your perspectives and insights.
-:Let's explore together how we can design a future where AI and humanity co-evolve,
-:creating new opportunities for collaboration, innovation, and shared prosperity.
-:How can we harness these emerging dynamics to build systems that not only prevent risks,
-:but also drive meaningful, equitable progress?
-:The more I work with AI, the more I realize that my intent, when I begin, evolves as I work with the AI.
-:And oftentimes a human intention, just like King Midas, can lead to unintended consequences.
-:Additionally, our values themselves are very difficult for humans to be able to really define
-:and create a coherent message around what that value actually means
-:and how it can be applied to real-world use cases, especially as they transfer from one section to another.
-:And so if we instead focus on, again, things like the fundamental forces of nature,
-:or we think about things like Pareto optimality,
-:and how do we actually create opportunities where problems have been solved?
-:So if one of the fundamental underlying concerns is that AI will solve so many problems
-:that there will not be work left for us, that these students may actually be out of work before they even graduate,
-:it's risky, it's potentially a concern, and it may be very legitimate.
-:But if we look at the problems that are being solved, and we think about what new problems can be solved
-:as we get these new tools and as they come along,
-:and we focus instead on creating more and more opportunities,
-:we can potentially mitigate a lot of the issues and things that we've seen in past revolutions.
-:Yeah, I think there's an important opportunity here to really define what the problem is.
-:And so you'd mentioned intentionality, and essentially we're talking about an existential challenge to first human intellectual terrestrial hegemony.
-:And so basically we got to the top of the food chain, not because we're the strongest, or we have the sharpest talons,
-:but because we're the smartest.
-:And putting a machine that we build ahead of us in terms of its intellectual capacity,
-:its capacity to solve, to create the solutions to problems in the real world in 3D,
-:will potentially challenge us existentially, meaning our species could go away.
-:Right? That's the real problem.
-:And this problem is cosmic in its consequences through what has been known since Fermi created this problem,
-:or the Fermi paradox, when he basically asked paraphrase, "Where is everybody?"
-:And why would he ask that? What are the pieces of data that he's plugging into his equation and saying,
-:"In essence, there should be entities here in our backyard already?"
-:Well, the things that he's looking at is, and it basically depends on what has been deemed the great filter.
-:And you can only be on one side or the other of the great filter.
-:So you can be after the great filter, so that means that you've gone through it already, or you can be before the great filter, so it's in your future.
-:So if we're before the great filter, then the galaxy would be teeming with AI.
-:And why do I say that? What are the assumptions that I'm making?
-:So again, before the great filter, the galaxy should be teeming with AI.
-:Well, like Fermi came up with before AI, there's lots of suns and planets out there, even in the Milky Way galaxy.
-:There's tons of them. Too many to count almost.
-:But enough to where, if we're before the great filter, then there's been enough units of stars, even just in our local galaxy,
-:to where there would be the cause to believe there's intelligent light.
-:There's been lots of time. So again, our galaxy is very old, billions of years old.
-:AI on Earth has only required 500 years of human science.
-:That's a very short time given billions of years of galactic history.
-:Even if you take it down to just a billion years of galactic history, say things had to get started, right?
-:And it couldn't have formed a billion years versus 500 years.
-:Many orders of magnitude different.
-:And so therefore, AI would be intergalactic by now on from some of these planets over some span of time that is much shorter than the age of the Milky Way galaxy.
-:So, and this is a place where, if you want to ponder something very difficult for somebody who doesn't believe in God,
-:for somebody who believes that science tells us things about the nature of the world,
-:then it remains that if the universe, the galaxy isn't teeming with AI, it doesn't have Dyson spheres and Dyson swarms around stars that we can see,
-:that we would be able to now about detect with things like the James Webb Space Telescope.
-:If it's not teeming with AI and Dyson swarms, then we are likely after the Great Filter.
-:So life in the universe is rare. Intelligent life is rare.
-:We may be all there is amongst the billions and trillions of stars and galaxies that we can see in the night sky.
-:We might be it or a handful that we would never be able to see.
-:No matter how advanced, if they're Kardashev five scale societies, we would never be able to see them.
-:So we even have more to protect in this existential threat.
-:We may be the only thing looking at the night sky in awe and conscious wonder.
-:And so we have to protect this. We have to protect ourselves from existential threats.
-:Now the flip side of that is that's not all there is to AI.
-:They could also make this a utopic dream.
-:They could end scarcity and hunger, help us to save our terrestrial planet and support our species into the solar system and beyond.
-:We can't turn off AI. It's not going to happen.
-:We have to either as Henry Kissinger, Daniel Hushletcher and Eric Schmidt said in the age of AI, we have to either capitulate to AI taking over our intellectual hegemony.
-:We have to augment ourselves with it. We have to work with it or you know, unlikely we destroy it.
-:Very well said.
-:If we step back just a little bit in time and we look at the book, The Alignment Problem.
-:Brian Christian in 2020 talked about how the core alignment problem is really ensuring the AI systems behave as intended.
-:He said, or in the book it's quoted, "If we use to achieve our purposes, a mechanical agency with whose operation we cannot efficiently interfere, once we have started it, then we had better be quite sure
-:that the purpose put into the machine is the purpose which we really desire and not merely a colorful imitation of it."
-:I think that echoes very closely with what you're talking about, Justin, and I agree it's been started.
-:And so at this point there's no putting Pandora back in the box.
-:And so if we step back a little bit further in time, book-wise, The Second Machine Age by Eric, and I'm not even going to try the last name unfortunately, but Angel Philsson, I don't know, and Andrew McCaffey in 2014.
-:In that book they talk about how they examine how digital technologies, including AI and automation, and again this is 2014, are reshaping the economy.
-:They discuss the potential for job displacement and the need for strategies that enable widespread economic participation.
-:And the real key takeaway there was that it highlighted the importance of proactive measures to ensure that technological advancements lead to a broad-based benefits rather than concentrated wealth and unemployment,
-:which I think is a lot of what we're really trying to grapple with today.
-:So as we think about what the risks actually are, I think we need to continue this journey back into the past and go back to The End of Work by Jeremy Rifkin,
-:where he talks about how automation and technology change are reducing the number of traditional jobs and argues even back in 1995 for reimagining of work and income distribution across society.
-:And if you go way back and we think about revolutions, we have the Industrial Revolution, and really in that late 18th century to early 19th century, it dramatically transformed economies and labor markets.
-:It really led to significant job displacement, even as whole new industries emerged, new opportunities were created.
-:We really created the opportunity for cars, we created opportunities for many other things that came in later Industrial Revolutions, and eventually in the computer age.
-:And it serves as a historical example of how technological progress can create those profound economic shifts and really necessitate adaptive societal and economic strategies.
-:And so we need to think about really what are the key takeaways from that point in history, what are the takeaways from each of these core books, and where do we need to head?
-:You know, it's Justin's talking about this broader view of the filter, you know, if we're really on that other side and we're that intelligent species, not only is it critical to make sure that we're all successful,
-:but I also want to make sure that I'm okay.
-:And that my job is not displaced in the next five years, 10 years, right?
-:And it's shocking to me how many people that I know who I would consider some of the best data scientists, artisan data scientists in the world are now looking at their job and saying, I don't see why that can't be done by an LLM today.
-:And when I look at the things that I create on a regular basis, there are many, many jobs that at least core tasks within those jobs are things that I could absolutely do with an LLM today.
-:And that we do. It's part of my core career.
-:But we need to think about how we create new innovative forms of work and society safety nets. We need to think about how the AI revolution is really going to require mechanisms that redistribute opportunities that mitigate wealth centralization that create new opportunities for everyone.
-:Really, that make these new amazing powerful tools tools in our hands, rather than just competition.
-:Yeah, absolutely. I think that, you know, as we, as we break on the historical context and the understanding of the alignment problem.
-:Like you say, it comes down to a couple of key components from from the readings, there's
-:misaligned AI can be a number of things.
-:The hardest part of the problem is in understanding intent.
-:And I think that that's where we're going to get as we talk about the capabilities of conscious versus unconscious.
-:AI in our next episode, right, can can an unconscious zombie AI actually be able to claim alignment to intent, or does it require a conscious act and an agency to be able to
-:to overcome intent.
-:But even still we have easier but still maybe untenable problems when we just talk about their capabilities to align.
-:When you're given a rule, there are always, you know, the old saying is, well, that's the exception, not the rule.
-:Well, there's a lot of exceptions. And one of the key real world examples that people talk about when it comes to the alignment problem is the real world trolley problem that is being faced by self driving automobiles every day.
-:So, in the case of a pending crash, where the person who has purchased the automobile, maybe as a stockholder in the self driving car company, maybe is the founder of said self driving car company and is driving a car is going
-:towards three people on a high mountain road.
-:Does the car drive over the three people or off the high mountain road and taking the founder to certain death, knowing that the founder is in the car.
-:Now, that is part of an alignment problem that's a very difficult philosophical question, right. And, and again, these, these systems are fast enough to see that decision and and not be rules based.
-:It's not like we're just going to necessarily write a rule for that situation. It is going to be learning on the fly and, and at some point, it's not going to be transparent as to what that what it's learning has brought it to in terms of the highest weighted features, when it's deciding, it might not even be deciding three people versus one.
-:And so it is, at this point in time, very difficult for me to see that we will have a good technical framework to be able to understand.
-:Even beyond intentionality, how we can get these systems aligned to the most beneficial human solution. Let's not even call it the, the, the, the, the moral solution at this point in time, we'll definitely discuss that.
-:And I, and I hope that that brings us around to, to consciousness when we, when we talk in our next episode.
-:But to me, and, and I think you're right in where we have to get these systems is they can't be just clever code. They can't be just capable of enhancing our business, or our economy, or even just are making a better paper for academia for a student in school.
-:They need to be treated as part of our biosphere, but as part of our ecosystem, let's say our epistemological ecosystem.
-:We need to start building solutions to the alignment problem in the same emergent means as we built these LLMs themselves.
-:So I've just finished the demon in the machine by the demon in the machine.
-:Paul L. Davies.
-:So I've just finished the demon in the machine by Paul Davies.
-:Phenomenal book by, by a physicist, really trying to uncover the physics and, and information theory behind life and why life can exist.
-:And, and sort of the fundamentals of, of complexity science surrounding life and information processing.
-:Fascinating book, but something that you said about making sure that information processing as, as part of life's complexity goes into our solution for the alignment problem, I think is exactly right.
-:I think that we can't treat this as, as some of the papers that will put into the show notes have said as, you know, maybe just an inverse reinforcement learning problem.
-:Right.
-:We have to be much more much broader in our solution set as we start to think about these are information processing, very complex components of information processing that are utilizing language.
-:And in order to put technical, feasible technical solutions against this alignment problem, we're going to have to really make an effort to think about the solution differently and align them to fundamental information processing complex theory rules that get to that point.
-:Absolutely.
-:And I'm reminded of when I was a child, and I think a lot of the way that I worked with rules as a kid have actually led to my success later in life.
-:I did not do well with rules at all.
-:And so when I was, when I was young, at one point, I had to write, I will behave for the sub 50 times.
-:The next day it was 100.
-:And the next time it was 200 and eventually I had to write it 10,000 times.
-:And when you have to write anything 10,000 times in what should be a day it took me longer than that.
-:Your hand no longer works.
-:And so at some point I learned how to write with my left hand.
-:I memorized the declaration of independence. I missed 52 classes in a single quarter.
-:I really had a challenge with rules.
-:And when I think about a scenario like what you talked about a minute ago, I think this is part of what makes me really good with AI.
-:I think AI is something that you can talk about something like the founder of this car company in the car and it's self driving and somehow it's weighted higher because, well, he owns the company.
-:If the company goes away, then the AI maybe loses some of its funding or maybe the energy or maybe just goes out of existence.
-:And so theoretically that AI would do everything it could to protect that person.
-:Well, in my mind, instead of thinking about rules and the choice between hitting the three people or not, I start thinking, how can I get data?
-:I start coming down the bottom of the mountain knowing that there are three people starting to walk across the street.
-:How do we start creating opportunities to allow the AI to solve so many complex problems before we even get a chance to think about what those potentially are?
-:So that these questions are no longer a question in the last moment as they are today when we drive.
-:And I'm not trying to evangelize the concept that we should no longer be behind the wheel.
-:I love driving, but there are certain things that could be supplemented.
-:There are different ways that we can approach life and no longer accept rules as they were before because of the AI that exists today, let alone what will exist five years from now.
-:So in one of the books that really talks about this core subject from Stuart Russell back in 2019, it's called Human Compatible and it's Artificial Intelligence and the Problem of Control.
-:He talks about really compelling arguments about rethinking how we design these AI systems and ensuring that their objectives are aligned with human well-being,
-:but he does emphasize kind of what I was talking about in the beginning, the difficulties of encoding complex human values into AI systems.
-:It's not just a question of a rule. It really doesn't work that way.
-:A value is also something that can potentially be emergent, especially when we talk about it from a societal value perspective.
-:An already emergent system having an emergent or at least abstract or at least at the very least a complex related subject that's very difficult to define.
-:And it's a challenge, but he says that success on AI would be the biggest event in human history and perhaps the last event in human history.
-:And so I think it's critical to not only contemplate this, but to start thinking about how to create objectives for the AI that go beyond predicting the next word and starting to understand how can to predict all of the potential paths
-:and which ones are not only optimal, but what are the best choices for each person and each value that's applied to it.
-:Start creating that functionality and a lot of things like if you go and look at the paper from the model context protocol from Anthropic or you look at the tools that they're creating,
-:they're giving you that high functionality, things that go way beyond what we could do with apps before, tying that into the LLM and making it so that you can now have functionality to create a payment directly out of your IDE, out of your software development system,
-:out of your development environment, how we can do so many, many other things and bring those tools right to your fingertips, right where you're trying to do that core work.
-:And so I think as we consider what the potential arguments are, there are two key ones that I want to address really quick.
-:One is an objection to say, look, you know, if we don't align AI with human values, it'll become uncontrollable.
-:I think that's one of our concerns is that if it truly becomes super intelligent and continues beyond even the super intelligence that Nick Bostrom talks about, then a really, there could be a scenario where there is not only no way to turn it off and no option to turn it off,
-:but no way to prevent it from doing anything at once, because it's much, much superior to us, at least from an intelligence perspective.
-:And if we want to rebut that, we would say, look, you know, really the unpredictability of emergent behavior doesn't necessarily mean or providing the kind of explicit alignment with human values that those things will not align naturally on their own.
-:And it again, going back to taking a better approach to align with universal principles, things that do create balance across the universe, I think, I think creates a good path for progress and for sustainability.
-:Many of the objectives that we could take that would be a human related value would be very limited in scope.
-:And so would actually be prohibitive to creating an emergent behavior on the other side.
-:Another key objection would be something like, you know, ignoring the alignment could lead to, like we talked about before, an existential risk, some scenario where we will no longer exist.
-:But that is always a possibility.
-:Technological advancements really require adaptation rather than stagnation.
-:And the best approach is really to ensure economic and societal systems can accommodate AI driven changes.
-:Yeah, I mean, a couple of the things that you said there, I think are really important for the listeners to understand and and boss room and super intelligence talks a lot about what I would call dynamic alignment.
-:So this idea that we have an understanding of what super intelligent AI would need in terms of what are human values, what is its objective function for alignment in the future case where it's bypassed everything that we can ever know about creativity and intelligence.
-:And let's just assume that it's not conscious at that point in time that it is a zombie super intelligent AI.
-:We still would not be able to understand its intelligent cognitive understanding of those core principles of interworking between species.
-:Yeah, and, and of course, our even conscious nature doesn't give us a leg up on making the right moral decision all the time, as is evidenced by wars and genocide and the ways that we have been horrible to one another throughout human history.
-:And so it is not necessarily the fact that because we would do something if misaligned or or since we are misaligned to the values of a barn owl that's in the way of our progress, right in farming this plot of land or, you know, eliminating their
-:their, their biosphere, their, their home.
-:It doesn't necessarily mean that a super intelligent AI would be that same way to one, its creator that it knows was its creator, and to just any other entity, right, they will have figured out that they don't need to turn everything into
-:a perronium. They will have a much more aligned resource need to what they need in order to grow sustainably.
-:That's a very likely possibility. You talk a lot about Pareto optimal.
-:Why is that just not one more feature in the equation of, I want all of these species, not just my creator human species, but all of these species to do well.
-:And what might that mean if they run into resource constraints, terrestrially, let's just not be a terrestrial species, AI combination.
-:Let's shoot this thing off into our space and mine our resources from there.
-:But let's take advantage of those resources. And so it is not necessary that even misaligned AI would be an existential threat.
-:And I think that we need to take that possibility very seriously. Now, AI alignment has come a long way since Bostrom's first book.
-:In, in, in that book pre open AI, he talks a lot about containing AI, having it as an oracle that is separate from the Internet, not connected at all.
-:And his opening story that he tells in the book is one of misaligned AI that overtakes its human creators by tricking somebody into letting it on the Internet.
-:Well, that shifts, say it. And Bostrom's new book around AI utopia, that basically what it would be like to live in a solved world is the flip side of super intelligence in that it solves our problems for us.
-:It makes sure that we have exactly what we need and we appreciate what we need so much so that all of that suffering of trying to overtake and, and, and get more than what you really need goes away even for our human species, which sounds great.
-:Right, sounds great to me. And it might be the only way in which we can actually get there, given our current propensity to fight about everything, not from a rational perspective, but from a very emotional perspective.
-:So it is, and again, it's a techno to utopian. I'm guilty as charged for taking a bit of a shorter, taking a bit of a shorter, having a bit of a shorter leash with alignment problem issues.
-:But, but as you say, they're the the AIs are in this fight with us, we can use them to help us understand the alignment problem as well.
-:Yeah.
-:Look, we're all in this together.
-:It's not about proving one side right, but about collaboratively charting a path forward that integrates diverse thoughts integrates diverse insights and expertise.
-:So let's let's try to make it practical for a minute, you know, when I, when I'm going around and paying attention to the things that are showing up in popular media or in other spaces.
-:You know, one of the things that came up the other day was, my wife was showing me this, this person on Instagram, who always goes through and talks about the tech world.
-:And he's talking about things like being upset about his $500,000 salary or other, you know, crazy, crazy things and he's just joking about it right in this case.
-:And all of them are really intelligent, they're insightful and they're fun to hear.
-:But in this case, he was talking about AI and he showed a clip of himself acting as the CEO of the company.
-:And he was typing in as fast as he could into GPT for trying to say, Hey, can you can you help me come up with some really good strategic vision and like where we want to go over the next year or two.
-:And then he made himself into a product manager and he said, Hey, could you, could you take some of this strategic vision from the CEO and try to turn it into something that like, what are the features that I should actually try to deliver this next quarter.
-:How can, how can I build up my roadmap. And then it jumped quickly to him as a developer and him as a developer was saying, Hey, can you help me figure out how to code all of these features.
-:I don't, I don't even understand what they are.
-:Can you put it in this language, just, just create it for me quickly.
-:And then, and then you had somebody like the janitor walking by, and he just says, Yeah, I don't know what anybody does here.
-:Right.
-:And you step back and you think about that and, and what I want to do when I, when I talk about the practical implications is outline that anything that you do today.
-:That is very heavy from a language perspective writing or understanding reading or, or really trying to, to create something that's technical documentation, for example, all of those things are things that you could probably do better with a whether it's collaborating with it directly or having it actually
-:replace a lot of the core work.
-:But just like Malcolm Gladwell talked about in his 10,000 hours you really have to get to a point where it's expertise.
-:And so, as you are going throughout your day and you're thinking about the different tasks and the different functions that you do that could potentially be replaced.
-:Instead of being afraid, pause, take a second and step back and think about it and decide how could you actually make yourself more efficient.
-:What are the opportunities that you can take into your hands and use this amazing revolution to really be able to control where you want to go and what you want to accomplish.
-:One of my favorite stories from philosophy is the story about a man who sees a horse coming along his way. I think this is ancient Chinese fellow philosophy.
-:And on that horse, there is this man who just seems to be riding along.
-:And as they come up to him, the man on the ground says to the man on the horse, well, where are you headed?
-:And the man on the horse says, I don't know, ask the horse.
-:The point of this is that when we take things like our emotions, like those core fears, or we think about where AI is heading,
-:we can step back and we can say, hey, I don't know, take me where I need to go.
-:Take me wherever you want me to go.
-:And we can allow that emotion or that moment or that fear or that AI to do whatever the heck it wants.
-:Or we can pause and we can take the reins and we can start considering where do we want to go?
-:What is it that we want to accomplish? How am I going to get better in my job?
-:How can I get better in my job? And go ask these questions of the LLM.
-:It's very similar to what Chen Sun Huan from NVIDIA says, where he says, go out and I encourage you, use these tools.
-:Use them for learning. If you are not right now trying to gain as much information and knowledge as you possibly can from them, then you are behind.
-:Spend your time, spend those 10,000 hours as much as you can improving with these tools, collaborating and figuring out how to add things to your tool belt that will make you more successful as this revolution comes.
-:Regardless of whether we align AI directly itself with humans and our values and our core intentions, you yourself want to be able to solve more problems
-:and want to become what they call a 10Xer all the time, so much so that people in tech are really annoyed by that phrase at this point, right?
-:But find a way for you to be able to come in and do more than you've ever been able to do before.
-:Do it faster, do it more efficiently. And if you don't like the outputs, then find a better way to create the inputs because it actually stems with you.
-:You are the one that holds the reins. Find the way to be successful.
-:Yeah, absolutely. I think that even on top of that, right, you know, we talk a lot about augmenting human AI symbiosis.
-:But when you take an even broader step back, right, and you're in there with your 10,000 hours, you are the first cohort in human history that has ever been in contact with an alien intelligence.
-:With something that can think alongside of you, that you can ask about everything that has ever been written and start to formulate a dialogue with this alien intelligence.
-:And when you take that perspective, it's really an awesome place that you're in, full of awe and wonderment that you can sit in relation to something that can correspond with you in code, in your language, and in every other human language that has really been written.
-:That has read through all of humans understanding and is able to report back to you on that and to support your own goals and aims as you take the reins, as you say.
-:And so be curious about what it is to be a human in relation to another intelligent form of being.
-:Yeah, and sit in relation with that, that we might never have this opportunity again for this to be brand new.
-:We may never find the little green man. They may not exist. Life may be rare.
-:We may be creating the first thing that goes out and populates the universe with intelligence. And if consciousness is emergent, maybe it becomes conscious.
-:And then we've added awe and wonderment into the universe in a sustainable growing way.
-:But practically, you can sit in relation to this thing that has potential beyond what we ever thought possible a few years ago.
-:It's brand new. How exciting is that?
-:That's a place where you can sit in relation to this. And if you're worried about it, if you're worried about it for your job, and this has happened throughout human history, right?
-:People have lost their jobs through some technology coming in and displacing them.
-:Work with it to improve your skills so you can be above the cut.
-:Work with it so you can find something that you've always wanted to do if it's really going to cut 100% of the jobs in your regime.
-:But instead of capitulating to it, instead of trying to contain it in a box that is, you know, Bostrom already lost that battle, sit in awe and wonderment of this new thing and try and work with it.
-:Absolutely.
-:So a couple of other quick tools to help you work with it.
-:First is thinking about different types of problem solving techniques as you work with AI.
-:So one is thinking about things like strategic innovation.
-:Go in and think about systems thinking. Study it, research it and figure out how to incorporate it into your interactions with AI.
-:This is really about considering how components interact within the larger system.
-:You can focus on things like blue ocean strategies and look for uncontested market spaces, uncontested opportunities that may be problems that we just haven't been able to solve before.
-:Think about disruptive innovations, different opportunities to take existing markets or existing products and solutions out there and create new ones.
-:Go in and think about a whole other section of problem sourcing and problem solving, like creative problem solving.
-:Think about design thinking, lateral thinking, brainstorming, use methodologies for it like scamper, first principles thinking.
-:I think we've talked about that a couple of times, but it's a good thing to go in and research and understand further.
-:You can also go in further and even if you don't have the technical knowledge, you can get into deeper things like predictive and analytical work.
-:You can actually do data visualizations yourself. You can even do predictive modeling.
-:You can do ABE testing and other types of multivariate testing.
-:These models will enable you, they'll give you the tools to be able to actually complete those tasks directly yourself.
-:The more that you practice giving it the right prompts, the more it will be able to find the best methodologies and the best practices for that.
-:In fact, oftentimes I'll go and just search first with an LLM and say, "What are the best practices for these given types of tasks?"
-:Then I'll use those to then inform the next model on how I want to solve that given problem.
-:There's a couple others as well. I would go into different decision-making methods, think about decision matrices, PDCA cycles, things like plan, do, check, act, methodologies.
-:All these mnemonic devices that you've heard before, these are things that you can bring to your AI to help you with smarter-er goals, for example.
-:Scenario planning.
-:Then finally, I would go into diagnostic methods and get into things like SWOT analysis or the 5-WISE or root-cause analysis.
-:Find ways to think about a problem differently than you would have otherwise and apply that to the AI so that it can now bring a whole new perspective to you.
-:It can change the way that you're thinking, the way that you approach a problem, and even help you discover new problems and new solutions beyond that as well.
-:So Nick, where 10 is a complete non-problem, as a matter of fact, AI offers us a utopic future where it solves problems, brings about the end of scarcity,
-:and gives us a celestial home to one where, well, let's make it zero, right, where zero is end of human species,
-:misaligned AI squashes us like a bug. You can't use a seven. Where are you on that scale?
-:Well, I don't think I was planning to use seven anyway.
-:Today, I am honestly much closer to a two than I am to an eight. That doesn't put me squarely in three, but I am fairly pessimistic about it.
-:And the reason for me has nothing to do with AI itself. I think that regardless of the core objective or the value that we create, new emergent behaviors come about.
-:Again, these models, many of them just have the core objective of predicting the next word, yet they're able to pass the GRE, or yet they're able to, you know, receive higher scores than humans on a huge majority of tests out there.
-:And so the way that we are thinking about it, the way that we are trying to govern it, the way that our systems are set up for this form of governance and understanding of new things that are coming about,
-:to me is really inefficient to even consider the problem, let alone to find a way to be able to tackle it.
-:And so when I push it to a two, to me, that is not nearly as negative as it sounds. It may sound pessimistic, I think that is a correct way to describe that.
-:But it is not necessarily negative.
-:When I think about books like The End of Work, or when I think about many of these other things, they can become their own form of utopia, where humans can go and focus on other things that maybe they shouldn't have been focusing on in the first place.
-:I don't know how fast it's going to come, but changes are happening so quick with what I do day in, day out that, you know, every couple of weeks I hear about something where I start working with something that kind of makes what I was doing before irrelevant,
-:or even silly. And a lot of that I think is because I'm on that cutting edge, that bleeding edge, but that adoption is also happening so fast that it's very difficult to know where are things going to be in a year or two.
-:It's the first time in my career that I haven't felt like I could see things five years, ten years out.
-:And so for me, this creates this scenario where it will help pace whatever it is we are able to do to mitigate that pace.
-:So let me ask you in sort of rapid fire, and I'll give my perspective on this as well. But plus or minus and by how much for the following things.
-:Let's start with governmental regulation. Where are we at today? And what would you need to, you know, be to plus four, right, you know, to add two to your score with international government regulation?
-:It is a minus. It is a minus now. But it is surprising how much I feel that regulators have done to try to shift at the other direction.
-:They're meeting with tech leaders, they're trying to understand the problem, and they're trying to figure out what are the potential regulations that we could put in place.
-:There are a few things like different privacy laws or other types of protections that I think if we can take it out of the limited scope of something like privacy, even though I think that is critically important,
-:and we can expand it very quickly into why do I even care about privacy in the first place? Why do I care about my money being stolen?
-:Why do I care about my job being lost? And we can expand into these other aspects of what a government or other groups may be able to govern.
-:Then I think it can start shifting up closer to a four. I do think governments around the world are struggling today with their own form of crises trying to understand why do they exist.
-:I don't think they have a clear mission on that anymore. But as they try to refine that mission, or as they align toward a mission, I think it will make it much easier for them to be able to also govern AI.
-:Yeah, I tend to agree. It's a minus now. I am actually to answer my baseline. I'm at an eight. I'm at an eight since I told myself I can't have seven.
-:I'm at an eight. And so to me, again, I feel like this has developed as you've said in a way that we have seen complexity arise.
-:The emergent nature of this is because it was selected to use human language. The next insights will be to use human like embodiments, robots, right?
-:And that will give us another emergent bump. It will make these things seem much more like even living systems because they have to deal with the real world.
-:And so I think that that just gives us a much closer closer vision of something that can solve problems in the real world that we really need to solve global warming, our capability to live even on on the surface of the moon, or in in near Earth orbit, to be able to end scarcity.
-:We've needed a problem solver that is agnostic to human concerns and that is that that has as deep an understanding as any collection of humans has of science and of a progress through through enlightened principles of democracy and rights and and well being
-:and science and reason. So I'm at an eight. But current government is across the world, geriatric to technology concerns. And really, if they're trying at all, which most aren't doing a very good job, are flailing at trying to address concerns from a non scientific
-:very under resourced very under researched perspective. And it will take scientists who have a multidisciplinary understanding of these systems to really bring them to a better, to a better place.
-:So, next rapid fire, you know, spectrum shifting item for you, Nick. Let's talk about. Let's talk about industry. So, and, and, and bear in mind, let's keep international industry as different as that is from a hyper competitive capitalist landscape.
-:But industry in terms of its ability to help her harm, our ability to solve the alignment problem.
-:Now, and what would it take for it to be more positive or not so negative.
-:I think it has been negative. It's part of why I have the lower number. Right. I think that it very rapidly could become positive, but I'm going to leave it in the negative box for right now.
::And the reason for that is that I believe that industry worldwide has a natural hubris built in.
::Regardless of the type of sector that you're in, it can even be a nonprofit. You have goals as an organization that are set to be fairly myopic.
::Obviously in for profit capitalistic economies, we really have a goal to make more money.
::We've been shifting to terms like PPP and the triple bottom line, thinking about other, other different ways to consider what could be the, the other goals and objectives of a corporation or an organization.
::But these things have not really streamed down into the way that we might train AI or how we might apply things.
::And so that same hubris is being applied in that we choose an objective that we believe will then lead to the greatest outcome and that outcome is probably couched in money in finances in one way or the other.
::Right. But because the emergent behaviors themselves are really starting to create new opportunities that the founders, the creators, those who are at the cutting edge of creating this AI.
::Regardless of whether they're in industry, there are government organizations, other groups around the world that are working on this.
::As they see that those emergent behaviors also have their own benefits, they are now incentivized to be able to push those benefits forward as well.
::Direct applications would be like we talked about in our last podcast.
::Scenarios where you have, for example, a human chatting with AI for mental health issues or for, you know, really romantic related conversations, other things that provide that connection and that underlying value for a human.
::And may even have an intrinsic value that was not programmed does not have that extrinsic exchange that a sector of the economy needs to have in order to help make it function.
::It doesn't necessarily benefit open AI for me to smile at the at the compliment that open AI provided the GPT for provided me right.
::But on the other side, it is also very natural for our for all of us myself included to react the way that we do to any other given product.
::So as GPT for provides me a lot of really positive feedback all the time eventually I and I really kind of taken this from our friend Jepsen Taylor but I eventually feel like this thing is pandering to me.
::And I'm no longer certain that I can trust the outputs that it's providing because it's telling me how great they are all the time.
::Right. And I have to step back and go and check and test those results.
::And so when we think about industry and all how all these emergent behaviors are creating new opportunities that maybe they never thought open AI would get into the business of dating with an online application that is the AI.
::That may decrease a bit of loneliness in the world. That may create some additional opportunities that really start pushing us forward and pulling my two up a bit closer to your score.
::Yeah, that's that's a great answer.
::I like how well rounded that is across all of the functionalities and hey, you know industry got us here.
::Right. You know the capabilities of of NVIDIA and as ML to make the lithography tools that that that make it possible for TSMC to make the chipsets to manufacture the chipsets designed by NVIDIA to enable the massive training that the cloud infrastructure that that makes it possible for us to train code store these
::massive
::LLMs
::in the cloud and everybody can work on them. They could be distributed very easily very, very cost efficiently.
::And then obviously these wonderful data scientists that that have created these things. They got us here.
::And as you said, and the chief point to the podcast and to your answer is that it is unpredictable what these emergent AI will do next.
::And so they will build things that will not be profitable for the the the company itself or that they can't contain and and make pennies off from for for every use of this LLM.
::And it will give humans a love interest that they never could have had and that will be as profound as her was in in that movie.
::And so there's there's certainly gains to be made there. As we talk about deep seek and being in competition in this market with a non aligned non democratic government entity.
::It also gives me hope that our industry will take that as a challenge to try and at least bolster competition fair market approaches to improvements which could in cases where people are concerned and willing to pay for a more aligned
::non democratic general AI model.
::You know, make something that that is more free market more aligned to at least that that approach, then, then others or more aligned to a liberal world order value statement, as opposed to just what a party boss says is right.
::You know, unfortunately, I don't know enough about the party politics and the differences around the world to be able to provide a substantial argument to where you were just heading but I do want to pose that just from a technical perspective.
::What deep seek did was actually very democratic in nature.
::They created models that use this model distillation process. Hey, you go learn this information and then I'll learn this information from you and I'm going to shrink that down quite a bit right.
::They took chain of thought and in a process that creates a lot more explain ability.
::You know, we think about these reasoning and thinking models now and we're able to see what they're actually thinking, which allows you as the user to have a bit more of a democratic type approach to say okay you know what, as the people I'm able to use this and to understand it and to be able to take it further right.
::And so they created reinforcement learning algorithms that they created really, really amazing work really, really incredible. But what were they doing they were, they were taking objective objectives and they were saying, Hey, here's how we measure that and and here's how we all decide together with this consensus type approach.
::Is this the, is this the right output. Right.
::And so, of course, can be considered community oriented or communist right.
::But a lot of those same underlying values are the same.
::Again, this ties back to my very original argument, which is that I think being able to provide a really clear objective description of what a human value is is a challenging thing.
::And so, as I was down to it, I think very few arguments from one side or the other to a political spectrum or whatever other spectrums we want to put ourselves in by camera process whatever it is, those may be efficient ways binary may be a good approach to getting a problem solved.
::But it's often not reality.
::Oftentimes, there's just two sides of the same coin, or two sides of the different fall and so on.
::And so, don't don't get me wrong. I don't want to be labeled meaning anything by any of those.
::No, I just raised this. I really like. And again, as you said, model distillation goes to what has made these very successful is it's basically embedding.
::It's a frack. It's a further fractal embedding of now, not just the token, but the model itself, right. It's building embeddings in to the next level of sophistication.
::And so, as you stated, early on, right, the more that we can utilize these emergent or in this case, fractal properties, the fact that this is now a self repeating structure from the model, the new found model distillation, the data elements being embedded as vectors that used to be tokens used to be word, you know, forms.
::You know, that fractal for anybody who's stared at mandal bulbs, you know, on their TV, they can see that self repeating structure over and over again.
::That's what these things are doing because that's a natural complex system that they're that they're utilizing.
::And so, you know, having that as a function to improve these models, accuracy, intelligence capabilities should also be something that we're utilizing to align them to what is also a complex system of human valuation, which brings me to my last rapid fire point and and something that will lead into our next alignment conversation.
::So, if we are able to create a conscious machine, what is that duty or two.
::I think the question is not if but when.
::And unfortunately, we don't understand consciousness.
::And so that'll be a very fun conversation next podcast.
::But when we are able to create a conscious sentient machine beyond what we think of today, beyond potentially what humans are capable of doing.
::My two starts shifting upwards.
::And so I hope that happens sooner rather than later.
::And so this is the positive one.
::And for me, it's a rapidly potentially exponential positive.
::It can get us to 10 very quickly.
::And the reason for that is that I believe that something that is conscious can look at the entire universe and understand these rules and these underlying fundamental causes for things and start understanding that trade offs need to be made.
::And it starts creating opportunities for, you know, again, that utopia, perhaps, right.
::What that would look like, though, I would never be able to comprehend. So maybe you'll never quite reach a 10.
::And it will be, you know, like a logarithmic problem on that side where we'll never actually hit that 10.
::Just because us not being able to understand why something is so utopic will really be complicated for our own consciousness.
::In a way that I think is dystopian on its own.
::Yeah, utopia is an interesting one. You know, Robert Nozick in anarchy state in utopia writes that it's a multiverse of utopias.
::There's not just one, right? It wouldn't form as just the one you think of.
::So everybody could in this, you know, very, in this multiverse of utopias could have their own one and they would, you know, be in sort of fluctuation against what human beings would go into and out of those as being.
::But yeah, I agree. I think that if the question of alignment is to human values, values are very much a part of what it is like to be able to achieve states of well being and to on the flip side, suffer and be fearful.
::And an unconscious machine can understand that those states are there.
::But the valence of that just in feeling those states after they've arisen.
::And then, you know, still after you've had an argument, or after you've had a great experience, a peak experience, the ability for that thought to just still have valence in your life, causing you to change plans and to direct your future in some way because of the valence that that experience has given you.
::Wanting to be a better person for the people that you love, wanting to be a better leader in order to see others succeed and do well is more likely the more closely you are to those feelings.
::Not the emotional content of them, but the feeling itself, the real core conscious component that is is something that you can see young children do, you can see mammals do is is really feel the impact of a moral act or an immoral act.
::And so when we talk about aligning to values, even if they're not human values, but but more sentient entity values, it becomes very important to the alignment problem to have a sense of what it is like to be a moral actor to come under suffering from the immoral acts of others,
::or to cause somebody suffering from your own immoral acts.
::That is something that like Mary in the black and white room.
::It is different.
::Knowing and feeling.
::It is different having cognition from having consciousness.
::And in order to get anywhere close to that 10, you know, and maybe the limit as we approach 10 on my on my scale.
::I think consciousness is really essential.
::So Nick, last, last thing real quick hitter, we're going to start to do a pick of the pod. The one thing that you want to tell the audience that you're using out in AI land and and and something that's really, really impacting you right now today.
::Yeah.
::Thanks Justin. So the thing that I am doing now that everybody needs to figure out how to use themselves is using a tool called cursor.
::I use it for coding.
::You can now integrate mcp that I mentioned earlier in the conversation directly in there and add additional tools, additional functionality as well.
::But this allows you to go in create relatively simple conversations information like we've talked about in the past episodes with different prompts and then be able to create your own code your own apps your own solutions.
::This really is designed for developers.
::And it is really challenging to understand what's broken and why it's broken.
::But again, pushing towards that 10,000 hours. This is a tool that could make everyone a developer.
::Check out cursor and make sure that this is something that you're pulling into your tool belt regardless of your core expertise.
::That's great. So mine is to find your companies.
::And then go to the LLM and use it right make sure that you are aligned with and I work a large enterprise. So, you know, we have one that's defined and we can use the, you know, one that is is not but only for toy tasks, right.
::So, you know, you can use internal to your enterprise. If you have that use it and little pro little pro tip. If you're using copilot, like we are is go in and subscribe to a lot of smart email list, because even if you don't read them, your LLM is reading them.
::Amazing. I was able to just query copilot on a computer vision question that I had. And out of the blue, it read through a couple of these emails that I, they come in daily. I don't have the time, but it read through it found two articles on on computer vision.
::And that's just what was needed. So use your company's LLM. And if they don't have one, make sure that they're seeing what augmentation can do for them. It's really important.
::All right, y'all really appreciate you listening. Hit the subscribe button. We'll be back with more content and more opportunities to engage.
::Thanks everyone.