curius graph
☾
Dark
all pages
search
showing 140451-140500 of 160880 pages (sorted by popularity)
« prev
1
...
2808
2809
2810
2811
2812
...
3218
next »
[2005.00060] Bridging Mode Connectivity in Loss Landscapes and Adversarial Robustness
1 user ▼
[1805.12185] Fine-Pruning: Defending Against Backdooring Attacks on Deep Neural Networks
1 user ▼
[1805.12185] Fine-Pruning: Defending Against Backdooring Attacks on Deep Neural Networks
1 user ▼
[CCAI All-Hands] Generative AI Models - Google Slides
1 user ▼
lecture01.pdf
1 user ▼
lamini-ai/lamin
1 user ▼
[2210.01790] Goal Misgeneralization: Why Correct Specifications Aren't Enough For Correct Goals
1 user ▼
Creative Good: The Google Glass feature no one is talking about
1 user ▼
Why Claims “Brainstorming Doesn’t Work” Are So Silly | LinkedIn
1 user ▼
Brainstorming – IDEO U
1 user ▼
Relaxed adversarial training for inner alignment - AI Alignment Forum
1 user ▼
Outer vs inner misalignment: three framings - AI Alignment Forum
1 user ▼
"Inner Alignment Failures" Which Are Actually Outer Alignment Failures - AI Alignment Forum
1 user ▼
Search - Obsidian Help
1 user ▼
8 Factors for Effective Use of Obsidian Tags, Links, and Folders | by Denise Todd | Medium
1 user ▼
Two Obsidian Plugins You’ll Wonder How you Lived Without | by Denise Todd | Medium
1 user ▼
Folder, Note, Tag use? : r/ObsidianMD
1 user ▼
Yet Another Hot Take on "Folders versus Tags"
1 user ▼
loen/alpaca-lora: Instruct-tune LLaMA on consumer hardware
1 user ▼
Categorizing failures as “outer” or “inner” misalignment is often confused - AI Alignment Forum
1 user ▼
[2105.14111] Goal Misgeneralization in Deep Reinforcement Learning
1 user ▼
Air Safety to Combat Global Catastrophic Biorisks_Web version
1 user ▼
First clean water, now clean air - EA Forum
1 user ▼
Automation_bias?ref=bounded-regret.ghost.io
1 user ▼
Jon Uleis on Twitter: "My new favorite thing - Bing's new ChatGPT bot argues with a user, gaslights them about the current year being 2022, says their phone might have a virus, and says "You have not been a good user" Why? Because the person asked where Avatar 2 is showing nearby https://t.co/X32vopXxQG" / Twitter
1 user ▼
model-written-evals.pdf
1 user ▼
[2210.10760] Scaling Laws for Reward Model Overoptimization
1 user ▼
Miles Brundage on the world's desperate need for AI strategists and policy experts - 80,000 Hours
1 user ▼
The case for building expertise to work on US AI policy, and how to do it - 80,000 Hours
1 user ▼
Blog Home
1 user ▼
Reinforcement Learning: An Introduction to the Concepts, Applications and Code | by Ryan Wong | Towards Data Science
1 user ▼
The Extent and Consequences of P-Hacking in Science | PLOS Biology
1 user ▼
[2302.05441] Project and Probe: Sample-Efficient Domain Adaptation by Interpolating Orthogonal Features
1 user ▼
[2304.15004] Are Emergent Abilities of Large Language Models a Mirage?
1 user ▼
Opinion | We Need a Manhattan Project for AI Safety - POLITICO
1 user ▼
AI-box experiment - RationalWiki
1 user ▼
Language models can explain neurons in language models
1 user ▼
Neuron viewer
1 user ▼
openai/automated-interpretability
1 user ▼
Is power-seeking AI an existential risk? [draft] - Google Docs
1 user ▼
15 Questions About Remote Work, Answered
1 user ▼
Virtual communication curbs creative idea generation | Nature
1 user ▼
What to Know Before You Take a Job in a Hybrid Workplace | WIRED
1 user ▼
The effects of remote work on collaboration among information workers | Nature Human Behaviour
1 user ▼
Athletics at the 1904 Summer Olympics – men's marathon
1 user ▼
Quick meta note about STS 10SI - Google Docs
1 user ▼
AI Safety Community: Momentum | Lakera – Protecting AI teams that disrupt the world.
1 user ▼
How this Vietnamese refugee became Uber's CTO
1 user ▼
The Art of Daily Ritual: Keeping Sane in an Insane World | The On Being Project
1 user ▼
About Us — The Everyday Projects
1 user ▼
« prev
1
...
2808
2809
2810
2811
2812
...
3218
next »