• 0 Posts
  • 4 Comments
Joined 1 year ago
cake
Cake day: July 1st, 2023

help-circle

  • AI alignment is a field that attempts to solve the problem of “how do you stop something with the ability to deceive, plan ahead, seek and maintain power, and parallelize itself from just doing that to everything”.

    https://aisafety.info/

    AI alignment is “the problem of building machines which faithfully try to do what we want them to do”. An AI is aligned if its actual goals (what it’s “trying to do”) are close enough to the goals intended by its programmers, its users, or humanity in general. Otherwise, it’s misaligned. The concept of alignment is important because many goals are easy to state in human language terms but difficult to specify in computer language terms. As a current example, a self-driving car might have the human-language goal of “travel from point A to point B without crashing”. “Crashing” makes sense to a human, but requires significant detail for a computer. “Touching an object” won’t work, because the ground and any potential passengers are objects. “Damaging the vehicle” won’t work, because there is a small amount of wear and tear caused by driving. All of these things must be carefully defined for the AI, and the closer those definitions come to the human understanding of “crash”, the better the AI is “aligned” to the goal that is “don’t crash”. And even if you successfully do all of that, the resulting AI may still be misaligned because no part of the human-language goal mentions roads or traffic laws. Pushing this analogy to the extreme case of an artificial general intelligence (AGI), asking a powerful unaligned AGI to e.g. “eradicate cancer” could result in the solution “kill all humans”. In the case of a self-driving car, if the first iteration of the car makes mistakes, we can correct it, whereas for an AGI, the first unaligned deployment might be an existential risk.



  • You’re at a moment in history where the only two real options are utopia or extinction. There are some worse things than extinction that people also worry about, but lets call it all “extinction” for now. Super-intelligence is coming. It literally can’t be stopped at this point. The only question is whether it’s 2, 5, or 10 years.

    If we don’t solve alignment, you die. It is the default. AI alignment is the hardest problem humans have ever tried to solve. Global warming will cause suffering on that timescale, but not extinction. A well-aligned super-intelligence has actual potential to reverse global warming. A misaligned one will mean it doesn’t matter.

    So, if you care, you should be working in AI alignment. If you don’t have the skillset, find something else: https://80000hours.org/

    Every single dismissal of AI “doom” is based on wishful thinking and hand-waving.