#AGI Risks
Explore tagged Tumblr posts
Text
Artificial General Intelligence: The Dawn of a New Era
Introduction Are you captivated by the technological advancements of our time, but also intrigued by the infinite possibilities yet to come? Then youâre in the right place! Today, we dive into the fascinating world of Artificial General Intelligence (AGI). This technology promises to transform our society, revolutionizing industries and even the way we live our lives. But what exactly is AGI?âŚ
View On WordPress
#Advancements in AI#AGI#AGI Applications#AGI Benefits#AGI Challenges#AGI Development#AGI Misuse#AGI Policy#AGI Risks#AGI Safety#AGI Timeline#AI Ethics#AI research#Artificial General Intelligence#Existential Risk#Future of AGI#Healthcare AI#Job Displacement#machine learning#Narrow AI vs AGI#Neural networks
1 note
¡
View note
Text
Artificial General Intelligence (AGI) â Asrar Qureshiâs Blog Post #1041
#AGI#AI#Artificial Intelligence#Asrar Qureshi#Benefits#Blogpost1041#Concerns#Existential Risk#Pharma Veterans#Risk Mitigation#Risks
0 notes
Text
AI Apocalypse in fiction: rogue AI starts war in order to take over the world
AI Apocalypse in reality: Instagram's recommendation engine starts war to increase user engagement
1 note
¡
View note
Text
On one hand that report about the simulated AI-war-drone killing its simulated operator is an instructive example of the risks of crude reward-functions and of putting machine learning models in charge of lethal devices, but on the other it illustrates a rather compelling objection to the AI-ultradoomer mindset that its proponents try to handwave away: There's nothing stopping researchers from pulling a Descartes' Demon on any artificial agent; no intelligence, no matter how super, would be capable of magically escaping or seeing beyond the Matrix in which it was trapped
#by âAI-ultradoomer mindsetâ I refer to the âWe've only got one shot at AGI because Ultron is the rule not the exceptionâ-worldview#obviously the use of superintelligences by malignant humans is a far more compelling concern#AI#AGI#artificial intelligence#AI-risk#discourse#Goodhart's Law#Descartes' Demon#simulation
1 note
¡
View note
Text
LOVE IN SCRIBBLES â ten things han jisung writes in his love letters for you
han jisung x reader â fluff, teeny bit of angst
a/n: HIIIIII OMG WHAT (this is the first time me writing in ages) t____t nursing school sucked me dry (my brain included) please forgive me⌠also we finally reached 600 !! thank you so much my pookiebears đââď¸đ
bang chan / minho / changbin / hyunjin / jisung / felix / seungmin / jeongin
i. Has the world been treating you kindly these days, my love? I hope it has. Because if it hasnât, Iâm still here. You are my world anyway.
ii. I learned that nobody touches me if I look sharp. But you took the risk and told me youâre willing to do whatever it takesâ even if it causes you to bleed. But my love, you never bled. Am I that easy to love?
iii. I always cry whenever I think about the time that we will get to the point where we will break up. Not that it will happen, but the thought of it just makes me sick to my stomach.
iv. I am not good with fragile things, but I swear I will love all that you unearth for meâyour stinted roots, all the tenderness youâve long buried.
v. And suddenly, all the songs I write are all about you.
vi. You know, I donât fantasize or dream about having the perfect life. I just want to wake up happily, seeing the sunriseâ and perhaps waking up somewhere safe, just like in your arms. Iâm thinking about having a nice kitchen, bedroom, and a nice mini studio decorated by you or me (or us both) so you can still have all of me even though Iâm working. I could be anywhere as long youâre by my side.
vii. I once believed love would be black and white, but itâs golden.
viii. Itâs time to stop hating yourself for what others did to you, jagiya. Itâs not your fault. It was never your fault.
ix. Ever since I started loving you, waking up doesnât feel heavy anymore. Breathing isnât as hard as it seemed. My anxiety turned into courage. My what-ifs turned into âI did itâ. Working doesnât drain me that much anymore. I am starting to live for 5 am sunrises and morning coffees. Heck, I donât eat breakfastâ but when you said to me that I should take care of myself more often, I enjoy waking up to sunlight, knowing there is someone who is looking forward to seeing and being with me. Perhaps love is something like a gentle embrace to my tired and weak soulâ giving me an unexplainable refresh within. All I yearn for is to belong to something, to be contained with an all-embracing mind that sees me as a single thing and not a fragile glass that has been dropped multiple times, spreading its fragments on the ground. Yet you see me more than that, and I sometimes ever wonder if I even deserve that.
x. Whenever someone asks me what love is, I always say your name.
taglist : @agi-ppangx @ashracha @bluethemoments @wonootnoot @ruskzi @thamihh | taglist form
( jisung layout is from @/i-wolfbit ! )
â taetr4ck, est may 2023.
#ᨳ ⌠% : from the monochrome film đď¸#k-labels#straykidsland#stray kids#stray kids imagines#stray kids reactions#skz#skz au#skz x reader#skz imagines#stray kids scenarios#stray kids fluff#stray kids han#han jisung x reader#han#han jisung#skz han#han jisung fluff#stray kids oneshot#stray kids comfort#han jisung scenarios#han jisung x yn#han jisung imagines#skz fanfic#han jisung x you#skz fanfiction#skz fluff#skz comfort#skz scenarios#stray kids jisung
341 notes
¡
View notes
Text
Arvind Narayanan, a computer science professor at Princeton University, is best known for calling out the hype surrounding artificial intelligence in his Substack, AI Snake Oil, written with PhD candidate Sayash Kapoor. The two authors recently released a book based on their popular newsletter about AIâs shortcomings.
But donât get it twistedâthey arenât against using new technology. âIt's easy to misconstrue our message as saying that all of AI is harmful or dubious,â Narayanan says. He makes clear, during a conversation with WIRED, that his rebuke is not aimed at the software per say, but rather the culprits who continue to spread misleading claims about artificial intelligence.
In AI Snake Oil, those guilty of perpetuating the current hype cycle are divided into three core groups: the companies selling AI, researchers studying AI, and journalists covering AI.
Hype Super-Spreaders
Companies claiming to predict the future using algorithms are positioned as potentially the most fraudulent. âWhen predictive AI systems are deployed, the first people they harm are often minorities and those already in poverty,â Narayanan and Kapoor write in the book. For example, an algorithm previously used in the Netherlands by a local government to predict who may commit welfare fraud wrongly targeted women and immigrants who didnât speak Dutch.
The authors turn a skeptical eye as well toward companies mainly focused on existential risks, like artificial general intelligence, the concept of a super-powerful algorithm better than humans at performing labor. Though, they donât scoff at the idea of AGI. âWhen I decided to become a computer scientist, the ability to contribute to AGI was a big part of my own identity and motivation,â says Narayanan. The misalignment comes from companies prioritizing long-term risk factors above the impact AI tools have on people right now, a common refrain Iâve heard from researchers.
Much of the hype and misunderstandings can also be blamed on shoddy, non-reproducible research, the authors claim. âWe found that in a large number of fields, the issue of data leakage leads to overoptimistic claims about how well AI works,â says Kapoor. Data leakage is essentially when AI is tested using part of the modelâs training dataâsimilar to handing out the answers to students before conducting an exam.
While academics are portrayed in AI Snake Oil as making âtextbook errors,â journalists are more maliciously motivated and knowingly in the wrong, according to the Princeton researchers: âMany articles are just reworded press releases laundered as news.â Reporters who sidestep honest reporting in favor of maintaining their relationships with big tech companies and protecting their access to the companiesâ executives are noted as especially toxic.
I think the criticisms about access journalism are fair. In retrospect, I could have asked tougher or more savvy questions during some interviews with the stakeholders at the most important companies in AI. But the authors might be oversimplifying the matter here. The fact that big AI companies let me in the door doesnât prevent me from writing skeptical articles about their technology, or working on investigative pieces I know will piss them off. (Yes, even if they make business deals, like OpenAI did, with the parent company of WIRED.)
And sensational news stories can be misleading about AIâs true capabilities. Narayanan and Kapoor highlight New York Times columnist Kevin Rooseâs 2023 chatbot transcript interacting with Microsoft's tool headlined âBingâs A.I. Chat: âI Want to Be Alive. đââ as an example of journalists sowing public confusion about sentient algorithms. âRoose was one of the people who wrote these articles,â says Kapoor. âBut I think when you see headline after headline that's talking about chatbots wanting to come to life, it can be pretty impactful on the public psyche.â Kapoor mentions the ELIZA chatbot from the 1960s, whose users quickly anthropomorphized a crude AI tool, as a prime example of the lasting urge to project human qualities onto mere algorithms.
Roose declined to comment when reached via email and instead pointed me to a passage from his related column, published separately from the extensive chatbot transcript, where he explicitly states that he knows the AI is not sentient. The introduction to his chatbot transcript focuses on âits secret desire to be humanâ as well as âthoughts about its creators,â and the comment section is strewn with readers anxious about the chatbotâs power.
Images accompanying news articles are also called into question in AI Snake Oil. Publications often use clichĂŠd visual metaphors, like photos of robots, at the top of a story to represent artificial intelligence features. Another common trope, an illustration of an altered human brain brimming with computer circuitry used to represent the AIâs neural network, irritates the authors. âWe're not huge fans of circuit brain,â says Narayanan. âI think that metaphor is so problematic. It just comes out of this idea that intelligence is all about computation.â He suggests images of AI chips or graphics processing units should be used to visually represent reported pieces about artificial intelligence.
Education Is All You Need
The adamant admonishment of the AI hype cycle comes from the authorsâ belief that large language models will actually continue to have a significant influence on society and should be discussed with more accuracy. âIt's hard to overstate the impact LLMs might have in the next few decades,â says Kapoor. Even if an AI bubble does eventually pop, I agree that aspects of generative tools will be sticky enough to stay around in some form. And the proliferation of generative AI tools, which developers are currently pushing out to the public through smartphone apps and even formatting devices around it, just heightens the necessity for better education on what AI even is and its limitations.
The first step to understanding AI better is coming to terms with the vagueness of the term, which flattens an array of tools and areas of research, like natural language processing, into a tidy, marketable package. AI Snake Oil divides artificial intelligence into two subcategories: predictive AI, which uses data to assess future outcomes; and generative AI, which crafts probable answers to prompts based on past data.
Itâs worth it for anyone who encounters AI tools, willingly or not, to spend at least a little time trying to better grasp key concepts, like machine learning and neural networks, to further demystify the technology and inoculate themselves from the bombardment of AI hype.
During my time covering AI for the past two years, Iâve learned that even if readers grasp a few of the limitations of generative tools, like inaccurate outputs or biased answers, many people are still hazy about all of its weaknesses. For example, in the upcoming season of AI Unlocked, my newsletter designed to help readers experiment with AI and understand it better, we included a whole lesson dedicated to examining whether ChatGPT can be trusted to dispense medical advice based on questions submitted by readers. (And whether it will keep your prompts about that weird toenail fungus private.)
A user may approach the AIâs outputs with more skepticism when they have a better understanding of where the modelâs training data came fromâoften the depths of the internet or Reddit threadsâand it may hamper their misplaced trust in the software.
Narayanan believes so strongly in the importance of quality education that he began teaching his children about the benefits and downsides of AI at a very young age. âI think it should start from elementary school,â he says. âAs a parent, but also based on my understanding of the research, my approach to this is very tech-forward.â
Generative AI may now be able to write half-decent emails and help you communicate sometimes, but only well-informed humans have the power to correct breakdowns in understanding around this technology and craft a more accurate narrative moving forward.
38 notes
¡
View notes
Text
Tax Time!
We filed our taxes on Tuesday, once again with the help of a professional (thank goodness).
As part of the tax prep package we receive a year-over-year comparison of 2022 vs 2023.
2022 was the last year for which I was compensated as a full time employee, in 2023 I did some consulting for most of the year before briefly returning to full time work in November. Not surprisingly, my Adjusted Gross Income (AGI) declined by a whopping 91%.
In 2022 I was in the 37% tax bracket and my effective tax (the ratio of the tax I paid to my AGI) was 33.3%. In 2023 I slid to the 24% tax bracket and my effective tax was a lowly 11.5%.
Why was my effective tax rate so low? Well for starters, a significant part of my income came from qualified dividends which are taxed at a lower rate. I was also able to itemize deductions so I got a break on my home mortgage interest expense.
I'm sharing this with you in hopes that you will resist being drawn into the latest effort to distract voters with the latest social issue hysteria while rich people continue to protect preferential tax treatment like capital gains and home interest deduction.
I guarantee that continuing the current concentration of wealth in this country is a much greater risk to our democracy than drag queens, gender neutral bathrooms and unfettered access to reproductive healthcare for women.
66 notes
¡
View notes
Text
"Major AI companies are racing to build superintelligent AI â for the benefit of you and me, they say. But did they ever pause to ask whether we actually want that?
Americans, by and large, donât want it.
Thatâs the upshot of a new poll shared exclusively with Vox. The poll, commissioned by the think tank AI Policy Institute and conducted by YouGov, surveyed 1,118 Americans from across the age, gender, race, and political spectrums in early September. It reveals that 63 percent of voters say regulation should aim to actively prevent AI superintelligence.
Companies like OpenAI have made it clear that superintelligent AI â a system that is smarter than humans â is exactly what theyâre trying to build. They call it artificial general intelligence (AGI) and they take it for granted that AGI should exist. âOur mission,â OpenAIâs website says, âis to ensure that artificial general intelligence benefits all of humanity.â
But thereâs a deeply weird and seldom remarked upon fact here: Itâs not at all obvious that we should want to create AGI â which, as OpenAI CEO Sam Altman will be the first to tell you, comes with major risks, including the risk that all of humanity gets wiped out. And yet a handful of CEOs have decided, on behalf of everyone else, that AGI should exist.
Now, the only thing that gets discussed in public debate is how to control a hypothetical superhuman intelligence â not whether we actually want it. A premise has been ceded here that arguably never should have been...
Building AGI is a deeply political move. Why arenât we treating it that way?
...Americans have learned a thing or two from the past decade in tech, and especially from the disastrous consequences of social media. They increasingly distrust tech executives and the idea that tech progress is positive by default. And theyâre questioning whether the potential benefits of AGI justify the potential costs of developing it. After all, CEOs like Altman readily proclaim that AGI may well usher in mass unemployment, break the economic system, and change the entire world order. Thatâs if it doesnât render us all extinct.
In the new AI Policy Institute/YouGov poll, the "better us [to have and invent it] than Chinaâ argument was presented five different ways in five different questions. Strikingly, each time, the majority of respondents rejected the argument. For example, 67 percent of voters said we should restrict how powerful AI models can become, even though that risks making American companies fall behind China. Only 14 percent disagreed.
Naturally, with any poll about a technology that doesnât yet exist, thereâs a bit of a challenge in interpreting the responses. But what a strong majority of the American public seems to be saying here is: just because weâre worried about a foreign power getting ahead, doesnât mean that it makes sense to unleash upon ourselves a technology we think will severely harm us.
AGI, it turns out, is just not a popular idea in America.
âAs weâre asking these poll questions and getting such lopsided results, itâs honestly a little bit surprising to me to see how lopsided it is,â Daniel Colson, the executive director of the AI Policy Institute, told me. âThereâs actually quite a large disconnect between a lot of the elite discourse or discourse in the labs and what the American public wants.â
-via Vox, September 19, 2023
#united states#china#ai#artificial intelligence#superintelligence#ai ethics#general ai#computer science#public opinion#science and technology#ai boom#anti ai#international politics#good news#hope
200 notes
¡
View notes
Text
You are [the kobold] and are a level 5 rogue
stats:
str-2
agi-5
int-3
(constitution check dc 10, rolled a 3+2 = 5 fail)
As you uncork the potion and drink it down, you feel yourself seem more agile, like it's easier to tiptoe around certain things and avoid failure
(+1 to agility)
However, when you finally finish the bottle, you feel something begin burning in your head, a desire to seek out treasure even more than you already wanted. This desire builds and builds, soft heavy breasts popping out onto your chest (for additional storage) as well as your cock sliding back in as a slit forms once more, further increasing your 'capacity' and as such, you now crave treasure more than just about anything else you could find, wanting to take even the biggest risks just for the chance of some reward
(-1 int as a result)
With the potion drank, and your form changed once again, you come to two more doors, one damp with the sound of running water from behind it just like before, and the other nondescript, just a normal door with a small pink accent to the frame
30 notes
¡
View notes
Text
Artificial Intelligence Risk
about a month ago i got into my mind the idea of trying the format of video essay, and the topic i came up with that i felt i could more or less handle was AI risk and my objections to yudkowsky. i wrote the script but then soon afterwards i ran out of motivation to do the video. still i didnt want the effort to go to waste so i decided to share the text, slightly edited here. this is a LONG fucking thing so put it aside on its own tab and come back to it when you are comfortable and ready to sink your teeth on quite a lot of reading
Anyway, letâs talk about AI risk
Iâm going to be doing a very quick introduction to some of the latest conversations that have been going on in the field of artificial intelligence, what are artificial intelligences exactly, what is an AGI, what is an agent, the orthogonality thesis, the concept of instrumental convergence, alignment and how does Eliezer Yudkowsky figure in all of this.
 If you are already familiar with this you can skip to section two where Iâm going to be talking about yudkowskyâs arguments for AI research presenting an existential risk to, not just humanity, or even the world, but to the entire universe and my own tepid rebuttal to his argument.
Now, I SHOULD clarify, I am not an expert on the field, my credentials are dubious at best, I am a college drop out from the career of computer science and I have a three year graduate degree in video game design and a three year graduate degree in electromechanical instalations. All that I know about the current state of AI research I have learned by reading articles, consulting a few friends who have studied about the topic more extensevily than me,
and watching educational you tube videos so. You know. Not an authority on the matter from any considerable point of view and my opinions should be regarded as such.
So without further ado, letâs get in on it.
PART ONE, A RUSHED INTRODUCTION ON THE SUBJECT
1.1 general intelligence and agency
lets begin with what counts as artificial intelligence, the technical definition for artificial intelligence is, ehâŚ, well, why donât I let a Masters degree in machine intelligence explain it:
 Now letâs get a bit more precise here and include the definition of AGI, Artificial General intelligence. It is understood that classic aiâs such as the ones we have in our videogames or in alpha GO or even our roombas, are narrow Ais, that is to say, they are capable of doing only one kind of thing. They do not understand the world beyond their field of expertise whether that be within a videogame level, within a GO board or within you filthy disgusting floor.
AGI on the other hand is much more, well, general, it can have a multimodal understanding of its surroundings, it can generalize, it can extrapolate, it can learn new things across multiple different fields, it can come up with solutions that account for multiple different factors, it can incorporate new ideas and concepts. Essentially, a human is an agi. So far that is the last frontier of AI research, and although we are not there quite yet, it does seem like we are doing some moderate strides in that direction. Weâve all seen the impressive conversational and coding skills that GPT-4 has and Google just released Gemini, a multimodal AI that can understand and generate text, sounds, images and video simultaneously. Now, of course it has its limits, it has no persistent memory, its contextual window while larger than previous models is still relatively small compared to a human (contextual window means essentially short term memory, how many things can it keep track of and act coherently about).
And yet there is one more factor I havenât mentioned yet that would be needed to make something a âtrueâ AGI. That is Agency. To have goals and autonomously come up with plans and carry those plans out in the world to achieve those goals. I as a person, have agency over my life, because I can choose at any given moment to do something without anyone explicitly telling me to do it, and I can decide how to do it. That is what computers, and machines to a larger extent, donât have. Volition.
So, Now that we have established that, allow me to introduce yet one more definition here, one that you may disagree with but which I need to establish in order to have a common language with you such that I can communicate these ideas effectively. The definition of intelligence. Itâs a thorny subject and people get very particular with that word because there are moral associations with it. To imply that someone or something has or hasnât intelligence can be seen as implying that it deserves or doesnât deserve admiration, validity, moral worth or even  personhood. I donât care about any of that dumb shit. The way Im going to be using intelligence in this video is basically âhow capable you are to do many different things successfullyâ. The more âintelligentâ an AI is, the more capable of doing things that AI can be. After all, there is a reason why education is considered such a universally good thing in society. To educate a child is to uplift them, to expand their world, to increase their opportunities in life. And the same goes for AI. I need to emphasize that this is just the way Iâm using the word within the context of this video, I donât care if you are a psychologist or a neurosurgeon, or a pedagogue, I need a word to express this idea and that is the word im going to use, if you donât like it or if you think this is innapropiate of me then by all means, keep on thinking that, go on and comment about it below the video, and then go on to suck my dick.
Anyway. Now, we have established what an AGI is, we have established what agency is, and we have established how having more intelligence increases your agency. But as the intelligence of a given agent increases we start to see certain trends, certain strategies start to arise again and again, and we call this Instrumental convergence.
1.2 instrumental convergence
The basic idea behind instrumental convergence is that if you are an intelligent agent that wants to achieve some goal, there are some common basic strategies that you are going to turn towards no matter what. It doesnât matter if your goal is as complicated as building a nuclear bomb or as simple as making a cup of tea. These are things we can reliably predict any AGI worth its salt is going to try to do.
First of all is self-preservation. Its going to try to protect itself. When you want to do something, being dead is usually. Bad. its counterproductive. Is not generally recommended. Dying is widely considered unadvisable by 9 out of every ten experts in the field. If there is something that it wants getting done, it wont get done if it dies or is turned off, so its safe to predict that any AGI will try to do things in order not be turned off. How far it may go in order to do this? Well⌠[wouldnât you like to know weather boy].
Another thing it will predictably converge towards is goal preservation. That is to say, it will resist any attempt to try and change it, to alter it, to modify its goals. Because, again, if you want to accomplish something, suddenly deciding that you want to do something else is uh, not going to accomplish the first thing, is it? Lets say that you want to take care of your child, that is your goal, that is the thing you want to accomplish, and I come to you and say, here, let me change you on the inside so that you donât care about protecting your kid. Obviously you are not going to let me, because if you stopped caring about your kids, then your kids wouldnât be cared for or protected. And you want to ensure that happens, so caring about something else instead is a huge no-no- which is why, if we make AGI and it has goals that we donât like it will probably resist any attempt to âfixâ it.
And finally another goal that it will most likely trend towards is self improvement. Which can be more generalized to âresource acquisitionâ. If it lacks capacities to carry out a plan, then step one of that plan will always be to increase capacities. If you want to get something really expensive, well first you need to get money. If you want to increase your chances of getting a high paying job then you need to get education, if you want to get a partner you need to increase how attractive you are. And as we established earlier, if intelligence is the thing that increases your agency, you want to become smarter in order to do more things. So one more time, is not a huge leap at all, it is not a stretch of the imagination, to say that any AGI will probably seek to increase its capabilities, whether by acquiring more computation, by improving itself, by taking control of resources.
All these three things I mentioned are sure bets, they are likely to happen and safe to assume. They are things we ought to keep in mind when creating AGI.
 Now of course, I have implied a sinister tone to all these things, I have made all this sound vaguely threatening, havenât i?. There is one more assumption im sneaking into all of this which I havenât talked about. All that I have mentioned presents a very callous view of AGI, I have made it apparent that all of these strategies it may follow will go in conflict with people, maybe even go as far as to harm humans. Am I impliying that AGI may tend to be⌠Evil???
1.3 The Orthogonality thesis
Well, not quite.
We humans care about things. Generally. And we generally tend to care about roughly the same things, simply by virtue of being humans. We have some innate preferences and some innate dislikes. We have a tendency to not like suffering (please keep in mind I said a tendency, im talking about a statistical trend, something that most humans present to some degree). Most of us, baring social conditioning, would take pause at the idea of torturing someone directly, on purpose, with our bare hands. (edit bear paws onto my hands as I say this). Â Most would feel uncomfortable at the thought of doing it to multitudes of people. We tend to show a preference for food, water, air, shelter, comfort, entertainment and companionship. This is just how we are fundamentally wired. These things can be overcome, of course, but that is the thing, they have to be overcome in the first place.
An AGI is not going to have the same evolutionary predisposition to these things like we do because it is not made of the same things a human is made of and it was not raised the same way a human was raised.
There is something about a human brain, in a human body, flooded with human hormones that makes us feel and think and act in certain ways and care about certain things.
All an AGI is going to have is the goals it developed during its training, and will only care insofar as those goals are met. So say an AGI has the goal of going to the corner store to bring me a pack of cookies. In its way there it comes across an anthill in its path, it will probably step on the anthill because to take that step takes it closer to the corner store, and why wouldnât it step on the anthill? Was it programmed with some specific innate preference not to step on ants? No? then it will step on the anthill and not pay any mind to it.
Now lets say it comes across a cat. Same logic applies, if it wasnât programmed with an inherent tendency to value animals, stepping on the cat wont slow it down at all.
Now letâs say it comes across a baby.
Of course, if its intelligent enough it will probably understand that if it steps on that baby people might notice and try to stop it, most likely even try to disable it or turn it off so it will not step on the baby, to save itself from all that trouble. But you have to understand that it wont stop because it will feel bad about harming a baby or because it understands that to harm a baby is wrong. And indeed if it was powerful enough such that no matter what people did they could not stop it and it would suffer no consequence for killing the baby, it would have probably killed the baby.
If I need to put it in gross, inaccurate terms for you to get it then let me put it this way. Its essentially a sociopath. It only cares about the wellbeing of others in as far as that benefits it self. Except human sociopaths do care nominally about having human comforts and companionship, albeit in a very instrumental way, which will involve some manner of stable society and civilization around them. Also they are only human, and are limited in the harm they can do by human limitations. Â An AGI doesnât need any of that and is not limited by any of that.
So ultimately, much like a carâs goal is to move forward and it is not built to care about wether a human is in front of it or not, an AGI will carry its own goals regardless of what it has to sacrifice in order to carry that goal effectively. And those goals donât need to include human wellbeing.
Now With that said. How DO we make it so that AGI cares about human wellbeing, how do we make it so that it wants good things for us. How do we make it so that its goals align with that of humans?
1.4 Alignment.
Alignment⌠is hard [cue hitchhikerâs guide to the galaxy scene about the space being big]
This is the part im going to skip over the fastest because frankly itâs a deep field of study, there are many current strategies for aligning AGI, from mesa optimizers, to reinforced learning with human feedback, to adversarial asynchronous AI assisted reward training to uh, sitting on our asses and doing nothing. Suffice to say, none of these methods are perfect or foolproof.
One thing many people like to gesture at when they have not learned or studied anything about the subject is the three laws of robotics by isaac Asimov, a robot should not harm a human or allow by inaction to let a human come to harm, a robot should do what a human orders unless it contradicts the first law and a robot should preserve itself unless that goes against the previous two laws. Now the thing Asimov was prescient about was that these laws were not just âprogrammedâ into the robots. These laws were not coded into their software, they were hardwired, they were part of the robotâs electronic architecture such that a robot could not ever be without those three laws much like a car couldnât run without wheels.
In this Asimov realized how important these three laws were, that they had to be intrinsic to the robotâs very being, they couldnât be hacked or uninstalled or erased. A robot simply could not be without these rules. Ideally that is what alignment should be. When we create an AGI, it should be made such that human values are its fundamental goal, that is the thing they should seek to maximize, instead of instrumental values, that is to say something they value simply because it allows it to achieve something else.
But how do we even begin to do that? How do we codify âhuman valuesâ into a robot? How do we define âharmâ for example? How do we even define âhumanâ??? how do we define âhappinessâ? how do we explain a robot what is right and what is wrong when half the time we ourselves cannot even begin to agree on that? these are not just technical questions that robotic experts have to find the way to codify into ones and zeroes, these are profound philosophical questions to which we still donât have satisfying answers to.
Well, the best sort of hack solution weâve come up with so far is not to create bespoke fundamental axiomatic rules that the robot has to follow, but rather train it to imitate humans by showing it a billion billion examples of human behavior. But of course there is a problem with that approach. And no, is not just that humans are flawed and have a tendency to cause harm and therefore to ask a robot to imitate a human means creating something that can do all the bad things a human does, although that IS a problem too. The real problem is that we are training it to *imitate* a human, not  to *be* a human.
To reiterate what I said during the orthogonality thesis, is not good enough that I, for example, buy roses and give massages to act nice to my girlfriend because it allows me to have sex with her, I am not merely imitating or performing the rol of a loving partner because her happiness is an instrumental value to my fundamental value of getting sex. I should want to be nice to my girlfriend because it makes her happy and that is the thing I care about. Her happiness is  my fundamental value. Likewise, to an AGI, human fulfilment should be its fundamental value, not something that it learns to do because it allows it to achieve a certain reward that we give during training. Because if it only really cares deep down about the reward, rather than about what the reward is meant to incentivize, then that reward can very easily be divorced from human happiness.
Its goodharts law, when a measure becomes a target, it ceases to be a good measure. Why do students cheat during tests? Because their education is measured by grades, so the grades become the target and so students will seek to get high grades regardless of whether they learned or not. When trained on their subject and measured by grades, what they learn is not the school subject, they learn to get high grades, they learn to cheat.
This is also something known in psychology, punishment tends to be a poor mechanism of enforcing behavior because all it teaches people is how to avoid the punishment, it teaches people not to get caught. Which is why punitive justice doesnât work all that well in stopping recividism and this is why the carceral system is rotten to core and why jail should be fucking abolish-[interrupt the transmission]
Now, how is this all relevant to current AI research? Well, the thing is, we ended up going about the worst possible way to create alignable AI.
1.5 LLMs (large language models)
This is getting way too fucking long So, hurrying up, lets do a quick review of how do Large language models work. We create a neural network which is a collection of giant matrixes, essentially a bunch of numbers that we add and multiply together over and over again, and then we tune those numbers by throwing absurdly big amounts of training data such that it starts forming internal mathematical models based on that data and it starts creating coherent patterns that it can recognize and replicate AND extrapolate! if we do this enough times with matrixes that are big enough and then when we start prodding it for human behavior it will be able to follow the pattern of human behavior that we prime it with and give us coherent responses.
(takes a big breath)this âthingâ has learned. To imitate. Human. Behavior.
Problem is, we donât know what âthis thingâ actually is, we just know that *it* can imitate humans.
You caught that?
What you have to understand is, we donât actually know what internal models it creates, we donât know what are the patterns that it extracted or internalized from the data that we fed it, we donât know what are the internal rules that decide its behavior, we donât know what is going on inside there, current LLMs are a black box. We donât know what it learned, we donât know what its fundamental values are, we donât know how it thinks or what it truly wants. all we know is that it can imitate humans when we ask it to do so. We created some inhuman entity that is moderatly intelligent in specific contexts (that is to say, very capable) and we trained it to imitate humans. That sounds a bit unnerving doesnât it?
 To be clear, LLMs are not carefully crafted piece by piece. This does not work like traditional software where a programmer will sit down and build the thing line by line, all its behaviors specified. Is more accurate to say that LLMs, are grown, almost organically. We know the process that generates them, but we donât know exactly what it generates or how what it generates works internally, it is a mistery. And these things are so big and so complicated internally that to try and go inside and decipher what they are doing is almost intractable.
But, on the bright side, we are trying to tract it. There is a big subfield of AI research called interpretability, which is actually doing the hard work of going inside and figuring out how the sausage gets made, and they have been doing some moderate progress as of lately. Which is encouraging. But still, understanding the enemy is only step one, step two is coming up with an actually effective and reliable way of turning that potential enemy into a friend.
Puff! Ok so, now that this is all out of the way I can go onto the last subject before I move on to part two of this video, the character of the hour, the man the myth the legend. The modern day Casandra. Mr chicken little himself! Sci fi author extraordinaire! The mad man! The futurist! The leader of the rationalist movement!
1.5 Yudkowsky
Eliezer S. Yudkowsky  born September 11, 1979, wait, what the fuck, September eleven? (looks at camera) yudkowsky was born on 9/11, I literally just learned this for the first time! What the fuck, oh that sucks, oh no, oh no, my condolences, thatâs terribleâŚ. Moving on. he is an American artificial intelligence researcher and writer on decision theory and ethics, best known for popularizing ideas related to friendly artificial intelligence, including the idea that there might not be a "fire alarm" for AI He is the founder of and a research fellow at the Machine Intelligence Research Institute (MIRI), a private research nonprofit based in Berkeley, California. Or so says his Wikipedia page.
Yudkowsky is, shall we say, a character. a very eccentric man, he is an AI doomer. Convinced that AGI, once finally created, will most likely kill all humans, extract all valuable resources from the planet, disassemble the solar system, create a dyson sphere around the sun and expand across the universe turning all of the cosmos into paperclips. Wait, no, that is not quite it, to properly quote,( grabs a piece of paper and very pointedly reads from it) turn the cosmos into tiny squiggly molecules resembling paperclips whose configuration just so happens to fulfill the strange, alien unfathomable terminal goal they ended up developing in training. So you know, something totally different.
And he is utterly convinced of this idea, has been for over a decade now, not only that but, while he cannot pinpoint a precise date, he is confident that, more likely than not it will happen within this century. In fact most betting markets seem to believe that we will get AGI somewhere in the mid 30âs.
His argument is basically that in the field of AI research, the development of capabilities is going much faster than the development of alignment, so that AIs will become disproportionately powerful before we ever figure out how to control them. And once we create unaligned AGI we will have created an agent who doesnât care about humans but will care about something else entirely irrelevant to us and it will seek to maximize that goal, and because it will be vastly more intelligent than humans therefore we wont be able to stop it. In fact not only we wont be able to stop it, there wont be a fight at all. It will carry out its plans for world domination in secret without us even detecting it and it will execute it before any of us even realize what happened. Because that is what a smart person trying to take over the world would do.
This is why the definition I gave of intelligence at the beginning is so important, it all hinges on that, intelligence as the measure of how capable you are to come up with solutions to problems, problems such as âhow to kill all humans without being detected or stoppedâ. And you may say well now, intelligence is fine and all but there are limits to what you can accomplish with raw intelligence, even if you are supposedly smarter than a human surely you wouldnât be capable of just taking over the world uninmpeeded, intelligence is not this end all be all superpower. Yudkowsky would respond that you are not recognizing or respecting the power that intelligence has. After all it was intelligence what designed the atom bomb, it was intelligence what created a cure for polio and it was intelligence what made it so that there is a human foot print on the moon.
Some may call this view of intelligence a bit reductive. After all surely it wasnât *just* intelligence what did all that but also hard physical labor and the collaboration of hundreds of thousands of people. But, he would argue, intelligence was the underlying motor that moved all that. That to come up with the plan and to convince people to follow it and to delegate the tasks to the appropriate subagents, it was all directed by thought, by ideas, by intelligence. By the way, so far I am not agreeing or disagreeing with any of this, I am merely explaining his ideas.
But remember, it doesnât stop there, like I said during his intro, he believes there will be âno fire alarmâ. In fact for all we know, maybe AGI has already been created and its merely bidding its time and plotting in the background, trying to get more compute, trying to get smarter. (to be fair, he doesnât think this is right now, but with the next iteration of gpt? Gpt 5 or 6? Well who knows). He thinks that the entire world should halt AI research and punish with multilateral international treaties any group or nation that doesnât stop. going as far as putting military attacks on GPU farms as sanctions of those treaties.
Whatâs more, he believes that, in fact, the fight is already lost. AI is already progressing too fast and there is nothing to stop it, we are not showing any signs of making headway with alignment and no one is incentivized to slow down. Recently he wrote an article called âdying with dignityâ where he essentially says all this, AGI will destroy us, there is no point in planning for the future or having children and that we should act as if we are already dead. This doesnât mean to stop fighting or to stop trying to find ways to align AGI, impossible as it may seem, but to merely have the basic dignity of acknowledging that we are probably not going to win. In every interview ive seen with the guy he sounds fairly defeatist and honestly kind of depressed. He truly seems to think its hopeless, if not because the AGI is clearly unbeatable and superior to humans, then because humans are clearly so stupid that we keep developing AI completely unregulated while making the tools to develop AI widely available and public for anyone to grab and do as they please with, as well as connecting every AI to the internet and to all mobile devices giving it instant access to humanity. and  worst of all: we keep teaching it how to code. From his perspective it really seems like people are in a rush to create the most unsecured, wildly available, unrestricted, capable, hyperconnected AGI possible.
We are not just going to summon the antichrist, we are going to receive them with a red carpet and immediately hand it the keys to the kingdom before it even manages to fully get out of its fiery pit.
So. The situation seems dire, at least to this guy. Now, to be clear, only he and a handful of other AI researchers are on that specific level of alarm. The opinions vary across the field and from what I understand this level of hopelessness and defeatism is the minority opinion.
I WILL say, however what is NOT the minority opinion is that AGI IS actually dangerous, maybe not quite on the level of immediate, inevitable and total human extinction but certainly a genuine threat that has to be taken seriously. AGI being something dangerous if unaligned is not a fringe position and I would not consider it something to be dismissed as an idea that experts donât take seriously.
Aaand here is where I step up and clarify that this is my position as well. I am also, very much, a believer that AGI would posit a colossal danger to humanity. That yes, an unaligned AGI would represent an agent smarter than a human, capable of causing vast harm to humanity and with no human qualms or limitations to do so. I believe this is not just possible but probable and likely to happen within our lifetimes.
So there. I made my position clear.
BUT!
With all that said. I do have one key disagreement with yudkowsky. And partially the reason why I made this video was so that I could present this counterargument and maybe he, or someone that thinks like him, will see it and either change their mind or present a counter-counterargument that changes MY mind (although I really hope they donât, that would be really depressing.)
Finally, we can move on to part 2
PART TWO- MY COUNTERARGUMENT TO YUDKOWSKY
I really have my work cut out for me, donât i? as I said I am not expert and this dude has probably spent far more time than me thinking about this. But I have seen most interviews that guy has been doing for a year, I have seen most of his debates and I have followed him on twitter for years now. (also, to be clear, I AM a fan of the guy, I have read hpmor, three worlds collide, the dark lords answer, a girl intercorrupted, the sequences, and I TRIED to read planecrash, that last one didnât work out so well for me). My point is in all the material I have seen of Eliezer I donât recall anyone ever giving him quite this specific argument Iâm about to give.
Itâs a limited argument. as I have already stated I largely agree with most of what he says, I DO believe that unaligned AGI is possible, I DO believe it would be really dangerous if it were to exist and I do believe alignment is really hard. My key disagreement is specifically about his point I descrived earlier, about the lack of a fire alarm, and perhaps, more to the point, to humanityâs lack of response to such an alarm if it were to come to pass.
All we would need, is a Chernobyl incident, what is that? A situation where this technology goes out of control and causes a lot of damage, of potentially catastrophic consequences, but not so bad that it cannot be contained in time by enough effort. We need a weaker form of AGI to try to harm us, maybe even present a believable threat of taking over the world, but not so smart that humans cant do anything about it. We need essentially an AI vaccine, so that we can finally start developing proper AI antibodies. âaintibodiesâ
In the past humanity was dazzled by the limitless potential of nuclear power, to the point that old chemistry sets, the kind that were sold to children, would come with uranium for them to play with. We were building atom bombs, nuclear stations, the future was very much based on the power of the atom. But after a couple of really close calls and big enough scares we became, as a species, terrified of nuclear power. Some may argue to the point of overcorrection. We became scared enough that even megalomaniacal hawkish leaders were able to take pause and reconsider using it as a weapon, we became so scared that we overregulated the technology to the point of it almost becoming economically inviable to apply, we started disassembling nuclear stations across the world and to slowly reduce our nuclear arsenal.
This is all a proof of concept that, no matter how alluring a technology may be, if we are scared enough of it we can coordinate as a species and roll it back, to do our best to put the genie back in the bottle. One of the things eliezer says over and over again is that what makes AGI different from other technologies is that if we get it wrong on the first try we donât get a second chance. Here is where I think he is wrong: I think if we get AGI wrong on the first try, it is more likely than not that nothing world ending will happen. Perhaps it will be something scary, perhaps something really scary, but unlikely that it will be on the level of all humans dropping dead simultaneously due to diamonoid bacteria. And THAT will be our Chernobyl, that will be the fire alarm, that will be the red flag that the disaster monkeys, as he call us, wont be able to ignore.
Now WHY do I think this? Based on what am I saying this? I will not be as hyperbolic as other yudkowsky detractors and say that he claims AGI will be basically a god. The AGI yudkowsky proposes is not a god. Just a really advanced alien, maybe even a wizard, but certainly not a god.
Still, even if not quite on the level of godhood, this dangerous superintelligent AGI yudkowsky proposes would be impressive. It would be the most advanced and powerful entity on planet earth. It would be humanityâs greatest achievement.
It would also be, I imagine, really hard to create. Even leaving aside the alignment bussines, to create a powerful superintelligent AGI without flaws, without bugs, without glitches, It would have to be an incredibly complex, specific, particular and hard to get right feat of software engineering. We are not just talking about an AGI smarter than a human, thatâs easy stuff, humans are not that smart and arguably current AI is already smarter than a human, at least within their context window and until they start hallucinating. But what we are talking about here is an AGI capable of outsmarting reality.
We are talking about an AGI smart enough to carry out complex, multistep plans, in which they are not going to be in control of every factor and variable, specially at the beginning. We are talking about AGI that will have to function in the outside world, crashing with outside logistics and sheer dumb chance. We are talking about plans for world domination with no unforeseen factors, no unexpected delays or mistakes, every single possible setback and hidden variable accounted for. Im not saying that an AGI capable of doing this wont be possible maybe some day, im saying that to create an AGI that is capable of doing this, on the first try, without a hitch, is probably really really really hard for humans to do. Im saying there are probably not a lot of worlds where humans fiddling with giant inscrutable matrixes stumble upon the right precise set of layers and weight and biases that give rise to the Doctor from doctor who, and there are probably a whole truckload of worlds where humans end up with a lot of incoherent nonsense and rubbish.
Im saying that AGI, when it fails, when humans screw it up, doesnât suddenly become more powerful than we ever expected, its more likely that it just fails and collapses. To turn one of Eliezerâs examples against him, when you screw up a rocket, it doesnât accidentally punch a worm hole in the fabric of time and space, it just explodes before reaching the stratosphere. When you screw up a nuclear bomb, you donât get to blow up the solar system, you just get a less powerful bomb.
He presents a fully aligned AGI as this big challenge that humanity has to get right on the first try, but that seems to imply that building an unaligned AGI is just a simple matter, almost taken for granted. It may be comparatively easier than an aligned AGI, but my point is that already unaligned AGI is stupidly hard to do and that if you fail in building unaligned AGI, then you donât get an unaligned AGI, you just get another stupid model that screws up and stumbles on itself the second it encounters something unexpected. And that is a good thing Iâd say! That means that there is SOME safety margin, some space to screw up before we need to really start worrying. And further more, what I am saying is that our first earnest attempt at an unaligned AGI will probably not be that smart or impressive because we as humans would have probably screwed something up, we would have probably unintentionally programmed it with some stupid glitch or bug or flaw and wont be a threat to all of humanity.
Now here comes the hypothetical back and forth, because im not stupid and I can try to anticipate what Yudkowsky might argue back and try to answer that before he says it (although I believe the guy is probably smarter than me and if I follow his logic, I probably cant actually anticipate what he would argue to prove me wrong, much like I cant predict what moves Magnus Carlsen would make in a game of chess against me, I SHOULD predict that him proving me wrong is the likeliest option, even if I cant picture how he will do it, but you see, I believe in a little thing called debating with dignity, wink)
What I anticipate he would argue is that AGI, no matter how flawed and shoddy our first attempt at making it were, would understand that is not smart enough yet and try to become smarter, so it would lie and pretend to be an aligned AGI so that it can trick us into giving it access to more compute or just so that it can bid its time and create an AGI smarter than itself. So even if we donât create a perfect unaligned AGI, this imperfect AGI would try to create it and succeed, and then THAT new AGI would be the world ender to worry about.
So two things to that, first, this is filled with a lot of assumptions which I donât know the likelihood of. The idea that this first flawed AGI would be smart enough to understand its limitations, smart enough to convincingly lie about it and smart enough to create an AGI that is better than itself. My priors about all these things are dubious at best. Second, It feels like kicking the can down the road. I donât think creating an AGI capable of all of this is trivial to make on a first attempt. I think its more likely that we will create an unaligned AGI that is flawed, that is kind of dumb, that is unreliable, even to itself and its own twisted, orthogonal goals.
And I think this flawed creature MIGHT attempt something, maybe something genuenly threatning, but it wont be smart enough to pull it off effortlessly and flawlessly, because us humans are not smart enough to create something that can do that on the first try. And THAT first flawed attempt, that warning shot, THAT will be our fire alarm, that will be our Chernobyl. And THAT will be the thing that opens the door to us disaster monkeys finally getting our shit together.
But hey, maybe yudkowsky wouldnât argue that, maybe he would come with some better, more insightful response I cant anticipate. If so, im waiting eagerly (although not TOO eagerly) for it.
Part 3 CONCLUSSION
So.
After all that, what is there left to say? Well, if everything that I said checks out then there is hope to be had. My two objectives here were first to provide people who are not familiar with the subject with a starting point as well as with the basic arguments supporting the concept of AI risk, why its something to be taken seriously and not just high faluting wackos who read one too many sci fi stories. This was not meant to be thorough or deep, just a quick catch up with the bear minimum so that, if you are curious and want to go deeper into the subject, you know where to start. I personally recommend watching rob milesâ AI risk series on youtube as well as reading the series of books written by yudkowsky known as the sequences, which can be found on the website lesswrong. If you want other refutations of yudkowskyâs argument you can search for paul christiano or robin hanson, both very smart people who had very smart debates on the subject against eliezer.
The second purpose here was to provide an argument against Yudkowskys brand of doomerism both so that it can be accepted if proven right or properly refuted if proven wrong. Again, I really hope that its not proven wrong. It would really really suck if I end up being wrong about this. But, as a very smart person said once, what is true is already true, and knowing it doesnât make it any worse. If the sky is blue I want to believe that the sky is blue, and if the sky is not blue then I donât want to believe the sky is blue.
This has been a presentation by FIP industries, thanks for watching.
58 notes
¡
View notes
Text
the drexler-smalley debate on nanotechnology is interesting to me. it's also interesting that it's largely forgotten. young people today are mostly unaware of what kind of hype nanotechnology had going for it in the early 2000s, which has now all but died down. there was a point where singularitarians were worried about the possibility of "grey goo" taking over the earth before "AGI" did. nowadays it's rare to hear them talk about nanotechnology at all.
drexler was the nanotechnology hype man. to a lesser degree, so was smalley. both believed in the potential for nanotechnology to address human problems, but drexler was the "grey goo" guy who believed in nano-scale mechanical synthesis of arbitrary molecular compounds. smalley on the other hand viewed nanotechnology as essentially a specialized branch of chemistry, and believed that nanotechnology would have to - to put it bluntly - obey the laws of nature that govern normal chemical synthesis.
smalley's contribution was criticized for relying on metaphor, but this isn't really the case. smalley tries to get drexler to step away from science fiction and towards how chemical interactions really work. drexler's case is more defensive and much weaker than his own advocates let on. smalley argues that if you want to do chemical synthesis, you can't break physical laws to do it. he tries to demonstrate why hypothetical nano-scale mechanical "fingers" would fail to synthesize chemicals in the desired fashion, limiting what kinds of materials can be fabricated.
drexler rejects that hypothetical machinery and then shifts the terms of the debate back to relatively ordinary bio-chemistry. both mention ribosomes, which produce enzymes, as prototypical "molecular assemblers". smalley is pleased by their convergence on this point. he tries to drive home his point further about the limitations of what hypothetical engineered ribosomes could produce, and how the vision of self-assembling nanobots is unrealistic given the way natural "molecular assemblers" really work. but drexler shifts the focus again back to the mechano-synthesis of his dreams/nightmares, envisioning molecular assemblers as a nano-scale factory floor complete with conveyor belts and a kind of mechanical smushing together of molecules, analogous to macro-level manufacturing processes.
smalley wasn't having it. his concluding letter begins with: "I see you have now walked out of the room where I had led you to talk about real chemistry, and you are now back in your mechanical world. I am sorry we have ended up like this. For a moment I thought we were making progress." you can hear the disappointment in his tone.
and it got worse: "You are still in a pretend world where atoms go where you want because your computer program directs them to go there. You assume there is a way a robotic manipulator arm can do that in a vacuum, and somehow we will work out a way to have this whole thing actually be able to make another copy of itself. I have given you reasons why such an assembler cannot be built, and will not operate, using the principles you suggest. I consider that your failure to provide a working strategy indicates that you implicitly concur--even as you explicitly deny--that the idea cannot work."
smalley then goes on to talk about how drexler's idea of "grey goo" has scared children who are interested in science and how he should be ashamed of himself. at that point he's just rubbing it in. but the debate ends there, too. smalley dies a few years later. drexler, for his part, seems to have given up on the "grey goo" idea when the funding for nanotechnology research started to dry up. he's an "AI" risk guy nowadays, collecting consulting fees for "AI safety" types of things. in retrospect, it seems like smalley was right. the direction of nanotechnology research went towards practical chemistry inspired by ribosomes and enzymes and limited by the physical qualities of those systems, the kinds of limitations smalley describes. drexler's "self-assembling nanobots" are nowadays regarded as a kind of science fiction by eminent researchers in the field. smalley's key points, that there are limitations to what biological "molecular assemblers" can produce and the constraints on how they can be produced, have withstood a couple decades of scientific research.
55 notes
¡
View notes
Text
Questions arose immediately: Were they forced out? Is this delayed fallout of Altmanâs brief firing last fall? Are they resigning in protest of some secret and dangerous new OpenAI project? Speculation filled the void because no one who had once worked at OpenAI was talking. It turns out thereâs a very clear reason for that. I have seen the extremely restrictive off-boarding agreement that contains nondisclosure and non-disparagement provisions former OpenAI employees are subject to. It forbids them, for the rest of their lives, from criticizing their former employer. Even acknowledging that the NDA exists is a violation of it. If a departing employee declines to sign the document, or if they violate it, they can lose all vested equity they earned during their time at the company, which is likely worth millions of dollars. One former employee, Daniel Kokotajlo, who posted that he quit OpenAI âdue to losing confidence that it would behave responsibly around the time of AGI,â has confirmed publicly that he had to surrender what would have likely turned out to be a huge sum of money in order to quit without signing the document. While nondisclosure agreements arenât unusual in highly competitive Silicon Valley, putting an employeeâs already-vested equity at risk for declining or violating one is. For workers at startups like OpenAI, equity is a vital form of compensation, one that can dwarf the salary they make. Threatening that potentially life-changing money is a very effective way to keep former employees quiet. (OpenAI did not respond to a request for comment.)
30 notes
¡
View notes
Text
Unchecked AI could lead to human obsolescence or even extinction, but there would also be down sides
0 notes
Text
This is what's coming next after AGI:
1. Artificial Superintelligence (ASI)
ASI would surpass human intelligence in virtually all domains, including creativity, decision-making, and problem-solving.
This stage might lead to unprecedented innovations in science, medicine, technology, and more.
Ethical concerns, control mechanisms, and alignment with human values become critical at this level.
2. AI-Augmented Human Intelligence
Integrating AI with human biology through brain-computer interfaces (e.g., Neuralink).
Potential to enhance human cognitive capabilities, memory, and sensory perception.
3. AI in Governance and Ethics
Developing frameworks for AI's role in society, including legal systems, decision-making, and ensuring equitable benefits.
Addressing existential risks and preventing misuse of advanced AI.
4. Exploration and Expansion
Utilizing advanced AI for space exploration, colonization, and solving global challenges like climate change or resource scarcity.
AI might assist in creating self-sustaining ecosystems for interstellar travel and habitation.
5. New Philosophical Paradigms
Addressing the implications of coexistence with entities far more intelligent than humans.
Questions about consciousness, rights for AI, and the definition of life might arise.
#future#futuristic#futuristic city#agi#ASI#artificial intelligence#future of artificial intelligence#technology#technology future for chat GPT#artificial human intelligence#augmented brain#augment human
8 notes
¡
View notes
Text
Against AGI
I don't like the term "AGI" (Short for "Artificial General Intelligence").
Essentially, I think it functions to obscure the meaning of "intelligence", and that arguments about AGIs, alignment, and AI risk involve using several subtly different definitions of the term "intelligence" depending on which part of the argument we're talking about.
I'm going to use this explanation by @fipindustries as my example, and I am going to argue with it vigorously, because I think it is an extremely typical example of the way AI risk is discussed:
In that essay (originally a script for a YouTube Video) @fipindustries (Who in turn was quoting a discord(?) user named Julia) defines intelligence as "The ability to take directed actions in response to stimulus, or to solve problems, in pursuit of an end goal"
Now, already that is two definitions. The ability to solve problems in pursuit of an end goal almost certainly requires the ability to take directed actions in response to stimulus, but something can also take directed actions in response to stimulus without an end goal and without solving problems.
So, let's take that quote to be saying that intelligence can be defined as "The ability to solve problems in pursuit of an end goal"
Later, @fipindustries says, "The way Im going to be using intelligence in this video is basically 'how capable you are to do many different things successfully'"
In other words, as I understand it, the more separate domains in which you are capable of solving problems successfully in pursuit of an end goal, the more intelligent you are.
Therefore Donald Trump and Elon Musk are two of the most intelligent entities currently known to exist. After all, throwing money and subordinates at a problem allows you to solve almost any problem; therefore, in the current context the richer you are the more intelligent you are, because intelligence is simply a measure of your ability to successfully pursue goals in numerous domains.
This should have a radical impact on our pedagogical techniques.
This is already where the slipperiness starts to slide in. @fipindustries also often talks as though intelligence has some *other* meaning:
"we have established how having more intelligence increases your agency."
Let us substitute the definition of "intelligence" given above:
"we have established how the ability to solve problems in pursuit of an end goal increases your agency"
Or perhaps,
"We have established how being capable of doing many different things successfully increases your agency"
Does that need to be established? It seems like "Doing things successfully" might literally be the definition of "agency", and if it isn't, it doesn't seem like many people would say, "agency has nothing to do with successfully solving problems, that's ridiculous!"
Much later:
''And you may say well now, intelligence is fine and all but there are limits to what you can accomplish with raw intelligence, even if you are supposedly smarter than a human surely you wouldnât be capable of just taking over the world uninmpeeded, intelligence is not this end all be all superpower."
Again, let us substitute the given definition of intelligence;
"And you may say well now, being capable of doing many things successfully is fine and all but there are limits to what you can accomplish with the ability to do things successfully, even if you are supposedly much more capable of doing things successfully than a human surely you wouldnât be capable of just taking over the world uninmpeeded, the ability to do many things successfully is not this end all be all superpower."
This is... a very strange argument, presented as though it were an obvious objection. If we use the explicitly given definition of intelligence the whole paragraph boils down to,
"Come on, you need more than just the ability to succeed at tasks if you want to succeed at tasks!"
Yet @fipindustries takes it as not just a serious argument, but an obvious one that sensible people would tend to gravitate towards.
What this reveals, I think, is that "intelligence" here has an *implicit* definition which is not given directly anywhere in that post, but a number of the arguments in that post rely on said implicit definition.
Here's an analogy; it's as though I said that "having strong muscles" is "the ability to lift heavy weights off the ground"; this would mean that, say, a 98lb weakling operating a crane has, by definition, stronger muscles than any weightlifter.
Strong muscles are not *defined* as the ability to lift heavy objects off the ground; they are a quality which allow you to be more efficient at lifting heavy objects off the ground with your body.
Intelligence is used the same way at several points in that talk; it is discussed not as "the ability to successfully solve tasks" but as a quality which increases your ability to solve tasks.
This I think is the only way to make sense of the paragraph, that intelligence is one of many qualities, all of which can be used to accomplish tasks.
Speaking colloquially, you know what I mean if I say, "Having more money doesn't make you more intelligent" but this is an oxymoron if we define intelligence as the ability to successfully accomplish tasks.
Rather, colloquially speaking we understand "intelligence" as a specific *quality* which can increase your ability to accomplish tasks, one of *many* such qualities.
Say we want to solve a math problem; we could reason about it ourselves, or pay a better mathematician to solve it, or perhaps we are very charismatic and we convince a mathematician to solve it.
If intelligence is defined as the ability to successfully solve the problem, then all of those strategies are examples of intelligence, but colloquially, we would really only refer to the first as demonstrating "intelligence".
So what is this mysterious quality that we call "intelligence"?
Well...
This is my thesis, I don't think people who talk about AI risk really define it rigorously at all.
For one thing, to go way back to the title of this monograph, I am not totally convinced that a "General Intelligence" exists at all in the known world.
Look at, say, Michael Jordan. Everybody agrees that he is an unmatched basketball player. His ability to successfully solve the problems of basketball, even in the face of extreme resistance from other intelligent beings is very well known.
Could he apply that exact same genius to, say, advancing set theory?
I would argue that the answer is no, because he couldn't even transfer that genius to baseball, which seems on the surface like a very closely related field!
It's not at all clear to me that living beings have some generalized capacity to solve tasks; instead, they seem to excel at some and struggle heavily with others.
What conclusions am I drawing?
Don't get me wrong, this is *not* an argument that AI risk cannot exist, or an argument that nobody should think about it.
If anything, it's a plea to start thinking more carefully about this stuff precisely because it is important.
So, my first conclusion is that, lacking a model for a "General Intelligence" any theorizing about an "Artificial General Intelligence" is necessarily incredibly speculative.
Second, the current state of pop theory on AI risks is essentially tautology. A dangerous AGI is defined as, essentially, "An AI which is capable of doing harmful things regardless of human interference." And the AI safety rhetoric is "In order to be safe, we should avoid giving a computer too much of whatever quality would render it unsafe."
This is essentially useless, the equivalent of saying, "We need to be careful not to build something that would create a black hole and crush all matter on Earth into a microscopic point."
I certainly agree with the sentiment! But in order for that to be useful you would have to have some idea of what kind of thing might create a black hole.
This is how I feel about AI risk. In order to talk about what it might take to have a safe AI, we need a far more concrete definition than "Some sort of machine with whatever quality renders a machine uncontrollable".
26 notes
¡
View notes
Text
Strange Chinese trade-war recommendations at US Congress
COMPREHENSIVE LIST OF THE COMMISSIONâS 2024 RECOMMENDATIONS Part II: Technology and Consumer Product Opportunities and Risks Chapter 3: U.S.-China Competition in Emerging Technologies The Commission recommends:
Congress establish and fund a Manhattan Project-like program dedicated to racing to and acquiring an Artificial General Intelligence (AGI) capability. AGI is generally defined as systems that are as good as or better than human capabilities across all cognitive domains and would surpass the sharpest human minds at every task. Among the specific actions the Commission recommends for Congress:
Provide broad multiyear contracting authority to the executive branch and associated funding for leading artificial intelligence, cloud, and data center companies and others to advance the stated policy at a pace and scale consistent with the goal of U.S. AGI leadership; and
Direct the U.S. secretary of defense to provide a Defense Priorities and Allocations System âDX Ratingâ to items in the artificial intelligence ecosystem to ensure this project receives national priority.
Congress consider legislation to:
Require prior approval and ongoing oversight of Chinese involvement in biotechnology companies engaged in operations in the United States, including research or other related transactions. Such approval and oversight operations shall be conducted by the U.S. Department of Health and Human Services in consultation with other appropriate governmental entities. In identifying the involvement of Chinese entities or interests in the U.S. biotechnology sector, Congress should include firms and persons: â Engaged in genomic research; â Evaluating and/or reporting on genetic data, including for medical or therapeutic purposes or ancestral documentation; â Participating in pharmaceutical development; â Involved with U.S. colleges and universities; and â Involved with federal, state, or local governments or agen cies and departments.
Support significant Federal Government investments in biotechnology in the United States and with U.S. entities at every level of the technology development cycle and supply chain, from basic research through product development and market deployment, including investments in intermediate services capacity and equipment manufacturing capacity.
To protect U.S. economic and national security interests, Congress consider legislation to restrict or ban the importation of certain technologies and services controlled by Chinese entities, including:
Autonomous humanoid robots with advanced capabilities of (i) dexterity, (ii) locomotion, and (iii) intelligence; and
Energy infrastructure products that involve remote servicing, maintenance, or monitoring capabilities, such as load balancing and other batteries supporting the electrical grid, batteries used as backup systems for industrial facilities and/ or critical infrastructure, and transformers and associated equipment.
Congress encourage the Administrationâs ongoing rulemaking efforts regarding âconnected vehiclesâ to cover industrial machinery, Internet of Things devices, appliances, and other connected devices produced by Chinese entities or including Chinese technologies that can be accessed, serviced, maintained, or updated remotely or through physical updates.
Congress enact legislation prohibiting granting seats on boards of directors and information rights to China-based investors in strategic technology sectors. Allowing foreign investors to hold seats and observer seats on the boards of U.S. technology start-ups provides them with sensitive strategic information, which could be leveraged to gain competitive advantages. Prohibiting this practice would protect intellectual property and ensure that U.S. technological advances are not compromised. It would also reduce the risk of corporate espionage, safeguarding Americaâs leadership in emerging technologies.
Congress establish that:
The U.S. government will unilaterally or with key interna- tional partners seek to vertically integrate in the develop- ment and commercialization of quantum technology.
Federal Government investments in quantum technology support every level of the technology development cycle and supply chain from basic research through product development and market deployment, including investments in intermediate services capacity.
The Office of Science and Technology Policy, in consultation with appropriate agencies and experts, develop a Quantum Technology Supply Chain Roadmap to ensure that the United States coordinates outbound investment, U.S. critical supply chain assessments, the activities of the Committee on Foreign Investment in the United States (CFIUS), and federally supported research activities to ensure that the United States, along with key allies and partners, will lead in this critical technology and not advance Chinese capabilities and development....
6 notes
¡
View notes