Flask Parent Template
aisafety.info
Semantic Search
Find Duplicates
Literature Abstracts
Extract QA
Duplicates Report
Score
Entry 1
Entry 2
(1.00)
What is the lottery ticket hypothesis?
What is the lottery ticket hypothesis?
(1.00)
What are neuron families?
What are neuron families?
(1.00)
What is Contrast Consistent Search (CCS)?
What is Contrast Consistent Search(CCS)?
(1.00)
Will superhuman AI systems be goal directed?
Will superhuman AI systems be goal directed?
(1.00)
What are finite factored sets?
What are finite factored sets?
(1.00)
What is specification gaming?
What is specification gaming?
(1.00)
What is relaxed adversarial training?
What is relaxed adversarial training?
(1.00)
Wouldn’t any AI be constrained by the limited computing power in the world?
Wouldn’t any AI be constrained by the limited computing power in the world?
(1.00)
What is P(doom)?
What is P(doom)?
(1.00)
What happens if all goes well?
What happens if all goes well?
(1.00)
What is an SLT?
What is a SLT?
(0.99)
Would taking AI safety seriously lead to a totalitarian government?
Would taking AI safety seriously lead to a totalitarian governments ?
(0.98)
What is Conjecture's main research agenda?
What is Conjecture's research agenda?
(0.98)
Why can't we build an AI that is programmed to shut off after some time?
Why can’t we build an AI that is programmed to turn off after some time?
(0.98)
What is reward misspecification?
What is “reward misspecification”?
(0.97)
Will future AIs want to solve the alignment problem?
Will future AIs be able to solve the alignment problem?
(0.97)
What exactly does “AI alignment” mean?
What is AI alignment?
(0.97)
What is Stuart Armstrong's research strategy?
What is Stuart Armstrong's research strategy
(0.96)
What are some common objections to the need for AI alignment, and brief responses to these?
What are some objections to the importance of AI alignment?
(0.96)
What are some introductory videos about AI safety?
Where can I find videos about AI safety?
(0.96)
What are Responsible Scaling Policies (RSPs)?
What is a responsible scaling policy (RSP)?
(0.96)
Which organizations are working on AI alignment?
What organizations are working on technical AI alignment?
(0.94)
Where can I learn more about AI alignment?
What are some good resources on AI alignment?
(0.93)
What is the weak scaling hypothesis?
What is the strong scaling hypothesis?
(0.93)
What is the general nature of the concern about AI alignment?
What are some objections to the importance of AI alignment?
(0.92)
Is smarter-than-human AI unrealistic?
Is smarter-than-human AI a realistic prospect?
(0.92)
What are current machine learning anchors?
What are “current machine learning” anchors?
(0.92)
What is offline reinforcement learning (RL)?
What is online reinforcement learning (RL)?
(0.92)
What is AI-assisted alignment?
What is AI alignment?
(0.92)
What is online reinforcement learning (RL)?
What is reinforcement learning (RL)?
(0.92)
What are some proposed training techniques to solve outer misalignment?
What are some proposed training techniques to solve inner misalignment?
(0.92)
The alignment problem
How can we solve the alignment problem?
(0.91)
What is least-to-most prompting?
What is least to most prompting?
(0.91)
Should selfish people care about AI safety?
Are there “selfish” reasons for caring about AI safety?
(0.91)
What are some common objections to the need for AI alignment, and brief responses to these?
What is the general nature of the concern about AI alignment?
(0.91)
What are some good resources on AI alignment?
I’d like to get deeper into the AI alignment literature. Where should I look?
(0.91)
How can I help AI alignment researchers be more effective?
Are there promising ways to make AI alignment researchers smarter?
(0.91)
What are the capabilities of GPT-4?
What is GPT-4 and what is it capable of?
(0.91)
Why is AI alignment a hard problem?
At a high level, what is the challenge of AI alignment?
(0.91)
Where can I learn more about AI alignment?
I’d like to get deeper into the AI alignment literature. Where should I look?
(0.90)
What is AI: Futures and Responsibility (AI:FAR)'s research agenda?
What is FAR AI's research agenda?
(0.90)
Superintelligence is unlikely?
Is superintelligence soon really possible?
(0.90)
Do we have an example/evidence of outer misalignment?
Do we have an example/evidence of inner misalignment?
(0.90)
What should I do with my machine learning research idea for AI alignment?
What should I do with my idea for helping with AI alignment?
(0.90)
Why would AGI want to self-improve or self-modify at all?
Would AGI want to self-improve or self-modify at all?
(0.89)
How plausible is AI existential risk?
Do people seriously worry about existential risk from AI?
(0.89)
What is Contrast Consistent Search(CCS)?
What is Cross-Contrast Search (CCS)?
(0.89)
What is Contrast Consistent Search (CCS)?
What is Cross-Contrast Search (CCS)?
(0.89)
What is the strong scaling hypothesis?
What is the scaling hypothesis?
(0.89)
What is out of context learning?
What is in-context learning?
(0.89)
What is a Task AI?
What is a Task-directed AI?
(0.89)
What is GPT-4?
What is GPT-4 and what is it capable of?
(0.88)
Is superintelligence soon really possible?
Will we ever build a superintelligence?
(0.88)
What are the different AI Alignment / Safety organizations and academics researching?
Briefly, what are the major AI safety organizations and academics working on?
(0.88)
What are some introductory videos about AI safety?
What are some introductions to AI safety?
(0.88)
Aren't there easy solutions to AI alignment?
What would a good solution to AI alignment look like?
(0.88)
Which alignment strategies can scale to superintelligence?
What concrete work is being done on alignment strategies which won’t scale to superintelligence?
(0.88)
What other options are there for pursuing a technical career in AI alignment?
How can I build a career in AI alignment?
(0.88)
Aren't there easy solutions to AI alignment?
Is AI alignment possible?
(0.88)
Intro to AI safety
What are some introductions to AI safety?
(0.88)
What is the general nature of the concern about AI alignment?
At a high level, what is the challenge of AI alignment?
(0.88)
What is concept distribution shift ?
What is a distributional shift?
(0.88)
Can't we just tell an AI to do what we want?
Can we tell an AI just to figure out what we want and then do that?
(0.87)
Why not just raise AI like kids?
Why can't we just make a "child AI" and raise it?
(0.87)
How can I work on public AI safety outreach?
How can I work on AI safety outreach in academia and among experts?
(0.87)
What philosophical approaches are used in AI alignment?
How can I do conceptual, mathematical, or philosophical work on AI alignment?
(0.87)
What is the objective based perspective on the alignment problem?
What is the optimization based perspective on the alignment problem?
(0.87)
Could AI alignment research be bad? How?
Might AI alignment research lead to outcomes worse than extinction?
(0.87)
What are Francois Chollet’s criticisms of AI Alignment?
What are Andrew Ng’s criticisms of AI Alignment?
(0.87)
What is outer alignment?
What is the difference between inner and outer alignment?
(0.87)
What is meant by 'first critical try'?
What is meant by 'first crucial try'?
(0.87)
What exactly does “AI alignment” mean?
What is AI-assisted alignment?
(0.87)
What is reward modeling?
What is recursive reward modeling?
(0.87)
What else is on aisafety.info?
What is aisafety.info about?
(0.87)
What are some existing alignment strategies, and what are their pitfalls?
What alignment strategies are scalably safe and competitive?
(0.87)
Superintelligence is unlikely?
Is the risk of superintelligence exaggerated?
(0.87)
Alignment concepts
What is Alignment?
(0.87)
Does talk of existential risk from AI detract from current harms?
Do people seriously worry about existential risk from AI?
(0.87)
Isn't the real concern that people will misuse AI?
Isn’t the real concern with AI something else?
(0.86)
What are some objections to the importance of AI alignment?
At a high level, what is the challenge of AI alignment?
(0.86)
What are some proposed training techniques to solve outer misalignment?
What are some proposed training techniques to solve deceptive misalignment?
(0.86)
How could an intelligence explosion be useful?
How might an "intelligence explosion" be dangerous?
(0.86)
What benchmarks exist for evaluating the safety of AI systems?
What benchmarks exist for measuring the capabilities of AI systems?
(0.86)
What would a good solution to AI alignment look like?
Is AI alignment possible?
(0.86)
Might an "intelligence explosion" never occur?
How likely is an intelligence explosion?
(0.86)
What does it mean for an AI to think?
Can an AI really think?
(0.86)
Why is AI alignment a hard problem?
Aren't there easy solutions to AI alignment?
(0.86)
Why is AI alignment a hard problem?
Why do some AI researchers not worry about alignment?
(0.86)
Why is AI alignment a hard problem?
How does AI taking things literally contribute to alignment being hard?
(0.86)
Could AI alignment research be bad? How?
What is the general nature of the concern about AI alignment?
(0.86)
Why would AGI be more dangerous than other technologies?
Why don't we just not build AGI if it's so dangerous?
(0.86)
What concepts underlie existential risk from AI?
What are the main sources of AI existential risk?
(0.85)
Could AI alignment research be bad? How?
What are some objections to the importance of AI alignment?
(0.85)
What is reward design?
What is reward modeling?
(0.85)
What is decision theory?
What is "logical decision theory"?
(0.85)
Governance research organizations
Governance research
(0.85)
What is a distribution shift and how is it related to alignment?
What is a distributional shift?
(0.85)
What is an intelligence explosion?
How could an intelligence explosion be useful?
(0.85)
What are Francois Chollet’s criticisms of AI Alignment?
What are Yann LeCun’s criticisms of AI Alignment?
(0.85)
Why do some AI researchers not worry about alignment?
What is the general nature of the concern about AI alignment?