AI ALIGNMENT FORUM
AF

2746
Nisan
Ω15311392
Message
Dialogue
Subscribe

Posts

Sorted by New

Wikitag Contributions

Comments

Sorted by
Newest
2Nisan's Shortform
4y
3
Nisan's Shortform
Nisan5mo20

Update: We're back to "ensure". On 2025-05-05, Sam Altman said (archived):

[OpenAI's] mission is to ensure that artificial general intelligence (AGI) benefits all of humanity.

Reply
Nisan's Shortform
Nisan1y150

On 2018-04-09, OpenAI said[1]:

OpenAI’s mission is to ensure that artificial general intelligence (AGI) [...] benefits all of humanity.

In contrast, in 2023, OpenAI said[2]:

[...] OpenAI’s mission: to build artificial general intelligence (AGI) that is safe and benefits all of humanity.


  1. Archived ↩︎

  2. This archived snapshot is from 2023-05-17, but the document didn't get much attention until November that year. ↩︎

Reply
The Waluigi Effect (mega-post)
Nisan2y10

The subject of this post appears in the "Did you know..." section of Wikipedia's front page(archived) right now.

Reply
Modern Transformers are AGI, and Human-Level
Nisan2y1214

I'm saying "transformers" every time I am tempted to write "LLMs" because many modern LLMs also do image processing, so the term "LLM" is not quite right.

"Transformer"'s not quite right either because you can train a transformer on a narrow task. How about foundation model: "models (e.g., BERT, DALL-E, GPT-3) that are trained on broad data at scale and are adaptable to a wide range of downstream tasks".

Reply
Modern Transformers are AGI, and Human-Level
Nisan2y97

I agree 100%. It would be interesting to explore how the term "AGI" has evolved, maybe starting with Goertzel and Pennachin 2007 who define it as:

a software program that can solve a variety of complex problems in a variety of different domains, and that controls itself autonomously, with its own thoughts, worries, feelings, strengths, weaknesses and predispositions

On the other hand, Stuart Russell testified that AGI means

machines that match or exceed human capabilities in every relevant dimension

so the experts seem to disagree. (On the other hand, Stuart & Russell's textbook cite Goertzel and Pennachin 2007 when mentioning AGI. Confusing.)

In any case, I think it's right to say that today's best language models are AGIs for any of these reasons:

  • They're not narrow AIs.
  • They satisfy the important parts of Goertzel and Pennachin's definition.
  • The tasks they can perform are not limited to a "bounded" domain.

In fact, GPT-2 is an AGI.

Reply1
Google Gemini Announced
Nisan2y53

I wonder why Gemini used RLHF instead of Direct Preference Optimization (DPO). DPO was written up 6 months ago; it's simpler and apparently more compute-efficient than RLHF.

  • Is the Gemini org structure so sclerotic that it couldn't switch to a more efficient training algorithm partway through a project?
  • Is DPO inferior to RLHF in some way? Lower quality, less efficient, more sensitive to hyperparameters?
  • Maybe they did use DPO, even though they claimed it was RLHF in their technical report?
Reply
My take on higher-order game theory
Nisan2y10

Thanks! For convex sets of distributions: If you weaken the definition of fixed point to gC(S)≤S, then the set {S∈CX:gC(S)=S} has a least element which really is a least fixed point.

Reply
Nisan's Shortform
Nisan2y30

Conception is a startup trying to do in vitro gametogenesis for humans!

Reply
Shah and Yudkowsky on alignment failures
Nisan4y90

CFAR used to have an awesome class called "Be specific!" that was mostly about concreteness. Exercises included:

  • Rationalist taboo
  • A group version of rationalist taboo where an instructor holds an everyday object and asks the class to describe it in concrete terms.
  • The Monday-Tuesday game
  • A role-playing game where the instructor plays a management consultant whose advice is impressive-sounding but contentless bullshit, and where the class has to force the consultant to be specific and concrete enough to be either wrong or trivial.
  • People were encouraged to make a habit of saying "can you give an example?" in everyday conversation. I practiced it a lot.

IIRC, Eliezer taught the class in May 2012? He talks about the relevant skills here and here. And then I ran it a few times, and then CFAR dropped it; I don't remember why.

Reply
My take on higher-order game theory
Nisan4y10

Yep, I skimmed it by looking at the colorful plots that look like Ising models and reading the captions. Those are always fun.

Reply
Load More
18My take on higher-order game theory
4y
8
2Nisan's Shortform
4y
3
4What is a VNM stable set, really?
5y
0
4Why you should minimax in two-player zero-sum games
5y
1
9Book report: Theory of Games and Economic Behavior (von Neumann & Morgenstern)
5y
3
9Beliefs at different timescales
7y
7
4Counterfactuals and reflective oracles
7y
0
8Counterfactuals, thick and thin
7y
4
5An environment for studying counterfactuals
7y
6
1Logical counterfactuals and differential privacy
8y
1
Load More
List of Blogs
12 years ago
(+31)
Less Wrong Meetup Group Resources
15 years ago
(+8)