_id
stringlengths
0
24
slug
stringlengths
0
132
title
stringlengths
0
313
draft
null
shortform
bool
1 class
hideCommentKarma
bool
1 class
af
bool
2 classes
currentUserReviewVote
null
userId
stringlengths
17
24
coauthorStatuses
listlengths
0
18
hasCoauthorPermission
bool
2 classes
rejected
bool
1 class
debate
bool
2 classes
collabEditorDialogue
bool
2 classes
__typename
stringclasses
1 value
url
stringlengths
0
432
postedAt
stringdate
2007-06-22 22:30:00
2025-06-28 01:40:04
createdAt
null
sticky
bool
2 classes
metaSticky
bool
2 classes
stickyPriority
int64
2
2
status
int64
2
2
frontpageDate
stringdate
2018-01-30 00:32:03
2025-06-28 02:24:31
meta
bool
2 classes
deletedDraft
bool
1 class
postCategory
stringclasses
3 values
shareWithUsers
sequencelengths
0
23
sharingSettings
float64
linkSharingKey
null
contents_latest
stringlengths
17
24
commentCount
int64
0
2k
voteCount
int64
-59
922
baseScore
int64
-10
945
unlisted
bool
1 class
score
float64
-0
5.05
lastVisitedAt
null
isFuture
bool
1 class
isRead
bool
1 class
lastCommentedAt
stringdate
2007-08-06 20:29:51
2025-06-28 14:23:54
lastCommentPromotedAt
stringclasses
21 values
canonicalCollectionSlug
stringclasses
4 values
curatedDate
stringclasses
691 values
commentsLocked
bool
2 classes
commentsLockedToAccountsCreatedAfter
stringclasses
1 value
question
bool
2 classes
hiddenRelatedQuestion
bool
1 class
originalPostRelationSourceId
stringclasses
46 values
location
null
googleLocation
null
onlineEvent
bool
1 class
globalEvent
bool
1 class
startTime
null
endTime
null
localStartTime
null
localEndTime
null
eventRegistrationLink
null
joinEventLink
null
facebookLink
stringclasses
1 value
meetupLink
null
website
stringclasses
1 value
contactInfo
stringclasses
1 value
isEvent
bool
1 class
eventImageId
null
eventType
null
types
sequencelengths
0
2
groupId
stringclasses
106 values
reviewedByUserId
stringclasses
19 values
suggestForCuratedUserIds
null
suggestForCuratedUsernames
null
reviewForCuratedUserId
stringclasses
12 values
authorIsUnreviewed
bool
1 class
afDate
stringclasses
590 values
suggestForAlignmentUserIds
sequencelengths
0
4
reviewForAlignmentUserId
stringclasses
6 values
afBaseScore
float64
-21
217
afCommentCount
int64
0
149
afLastCommentedAt
stringdate
2007-06-26 21:13:26
2025-06-28 01:40:04
afSticky
bool
2 classes
hideAuthor
bool
2 classes
moderationStyle
stringclasses
4 values
ignoreRateLimits
bool
2 classes
submitToFrontpage
bool
2 classes
onlyVisibleToLoggedIn
bool
1 class
onlyVisibleToEstablishedAccounts
bool
2 classes
reviewCount
int64
0
8
reviewVoteCount
int64
0
115
positiveReviewVoteCount
int64
0
98
manifoldReviewMarketId
stringclasses
900 values
annualReviewMarketProbability
float64
0.01
0.99
annualReviewMarketIsResolved
bool
2 classes
annualReviewMarketYear
float64
2.02k
2.03k
annualReviewMarketUrl
stringclasses
900 values
group
float64
podcastEpisodeId
stringclasses
396 values
forceAllowType3Audio
bool
1 class
nominationCount2019
int64
0
6
reviewCount2019
int64
0
6
votingSystem
stringclasses
2 values
disableRecommendation
bool
2 classes
coauthors
listlengths
0
18
readTimeMinutes
int64
1
315
rejectedReason
stringclasses
12 values
customHighlight
float64
lastPromotedComment
float64
bestAnswer
float64
tags
listlengths
0
31
feedId
stringclasses
45 values
totalDialogueResponseCount
int64
0
0
unreadDebateResponseCount
int64
0
0
dialogTooltipPreview
stringclasses
6 values
disableSidenotes
bool
2 classes
currentUserVote
null
currentUserExtendedVote
null
extendedScore.agreement
float64
-6
2
extendedScore.approvalVoteCount
float64
1
922
extendedScore.agreementVoteCount
float64
0
1
afExtendedScore.agreement
float64
-6
2
afExtendedScore.approvalVoteCount
float64
0
175
afExtendedScore.agreementVoteCount
float64
0
1
user._id
stringlengths
17
24
user.slug
stringlengths
2
40
user.createdAt
stringdate
2009-02-17 05:49:50
2025-06-26 13:32:01
user.username
stringlengths
1
64
user.displayName
stringlengths
1
43
user.profileImageId
float64
user.previousDisplayName
float64
user.fullName
stringclasses
979 values
user.karma
float64
-1,560
150k
user.afKarma
float64
-63
6.7k
user.deleted
bool
1 class
user.isAdmin
bool
2 classes
user.htmlBio
stringlengths
0
9.48k
user.jobTitle
float64
user.organization
float64
user.postCount
float64
0
1.02k
user.commentCount
float64
0
16.1k
user.sequenceCount
float64
0
40
user.afPostCount
float64
-4
364
user.afCommentCount
float64
0
1.39k
user.spamRiskScore
float64
0
1
user.tagRevisionCount
float64
0
3.8k
user.reviewedByUserId
stringclasses
18 values
user.__typename
stringclasses
1 value
user.moderationStyle
stringclasses
4 values
user.bannedUserIds
sequencelengths
0
6
user.moderatorAssistance
bool
2 classes
user.groups
sequencelengths
0
289
user.banned
stringclasses
30 values
user.allCommentingDisabled
float64
socialPreviewData._id
stringlengths
0
24
socialPreviewData.imageUrl
stringlengths
0
149k
socialPreviewData.__typename
stringclasses
1 value
contents._id
stringlengths
17
24
contents.htmlHighlight
stringlengths
0
2.31M
contents.plaintextDescription
stringlengths
0
2k
contents.wordCount
float64
0
78.7k
contents.version
stringclasses
299 values
contents.__typename
stringclasses
1 value
fmCrosspost.isCrosspost
bool
2 classes
fmCrosspost.hostedHere
bool
2 classes
fmCrosspost.foreignPostId
stringlengths
17
17
fmCrosspost.__typename
stringclasses
1 value
pPtqh4fwNcbkgFSpG
administering-immunotherapy-in-the-morning-seems-to-really
Administering immunotherapy in the morning seems to really, really matter. Why?
null
false
false
false
null
zfidjWWKb3azB4kMR
null
true
false
false
false
Post
https://www.owlposting.com/p/the-time-of-day-that-immunotherapy
2025-06-08T16:37:43.477Z
null
false
false
2
2
2025-06-08T18:10:14.551Z
false
false
linkpost
[]
null
null
BtcZNLimrm2vjDbFd
0
14
32
false
0.035027
null
false
false
2025-06-08T16:37:43.477Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
qgdGA4ZEyW7zNdK84
null
null
null
false
null
[]
null
17
0
2025-06-08T16:35:43.841Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
12
null
null
null
null
[ { "__typename": "Tag", "_id": "jaf5zfcGgCB2REXGw", "adminOnly": false, "afBaseScore": 9, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "EQNTWXLKMeWMp2FQS", "displayName": "Ben Pace" }, { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 19, "canEditUserIds": null, "core": false, "createdAt": "2020-05-11T02:08:39.903Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "EQNTWXLKMeWMp2FQS", "displayName": "Ben Pace" }, { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Biology", "needsReview": false, "noindex": false, "postCount": 261, "score": 19, "shortName": null, "slug": "biology", "suggestedAsFilter": false, "userId": "qgdGA4ZEyW7zNdK84", "voteCount": 2, "wikiOnly": false }, { "__typename": "Tag", "_id": "3uE2pXvbcnS9nnZRE", "adminOnly": false, "afBaseScore": 0, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 2, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:50.898Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 27, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "zJtgSyKntXrnkArbY", "displayName": "kistune" }, { "_id": "XqyQTGFGoKfZtdqN3", "displayName": "kINo" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "World Modeling", "needsReview": false, "noindex": false, "postCount": 5855, "score": 2, "shortName": null, "slug": "world-modeling", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 2, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
14
0
0
8
0
zfidjWWKb3azB4kMR
abhishaike-mahajan
2024-08-18T20:48:42.571Z
abhishaike-mahajan
Abhishaike Mahajan
null
null
null
645
0
false
false
null
null
23
11
0
0
0
1
0
EQNTWXLKMeWMp2FQS
User
null
null
null
[ "canModeratePersonal" ]
null
null
pPtqh4fwNcbkgFSpG
https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/pPtqh4fwNcbkgFSpG/ybm7fzg2kdtv36go3lld
SocialPreviewType
BtcZNLimrm2vjDbFd
<p><i>Edit on 08/06/2024: At least one person has pointed out that, at one point, giving hypertensives at night were <strong>also</strong> thought to matter, </i><a href="https://www.thelancet.com/journals/lancet/article/PIIS0140-6736(22)01786-X/fulltext"><i>a now disproven idea. </i></a><i>Someone also mentioned how many times the clinical trial information was altered during the study. I added in a section at the end to discuss this.</i></p><p><a href="https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64b823c6-4d7f-45d7-b655-75fd8fe339a0_2912x1632.png"><img src="https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/pPtqh4fwNcbkgFSpG/zjxjx06i3fb21zjj7pzn" alt="" srcset="https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/pPtqh4fwNcbkgFSpG/y40rxtycukw4m7snbwhs 424w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/pPtqh4fwNcbkgFSpG/tzargbpioyo6x4mjjmvb 848w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/pPtqh4fwNcbkgFSpG/iuwk3swagwlejyxh8yr5 1272w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/pPtqh4fwNcbkgFSpG/joyxwh7ukxalpcttms3c 1456w"></a></p><p>There’s a really interesting phenomenon in the immunotherapy field that has been going on for what seems to be several years now, but was raised to me — a non-oncologist — <a href="https://x.com/StephenVLiu/status/1929537643794051350">via a viral Twitter thread</a> of some work at <a href="https://www.asco.org/annual-meeting/program">ASCO25</a>:</p><p><a href="https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F938e4af0-edcc-4644-8cc7-d45f83f71acc_1178x1216.png"><img src="https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/pPtqh4fwNcbkgFSpG/zs9tqbxwp3j0lbcjxbix" alt="" srcset="https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/pPtqh4fwNcbkgFSpG/adcyolah6etgxnrixwip 424w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/pPtqh4fwNcbkgFSpG/mqepxal2lnikxkvelygr 848w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/pPtqh4fwNcbkgFSpG/efj6wzximsihrsisdf5l 1272w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/pPtqh4fwNcbkgFSpG/zs9tqbxwp3j0lbcjxbix 1456w"></a></p><p>Translating the jargon: amongst the patients who received their immunotherapy infusion before 3pm (as opposed to after 3pm), their <strong>cancer stayed under control for longer</strong> (11.3 months vs. 5.7 months) and <strong>on median</strong> <strong>lived longer</strong> (at least<span class="footnote-reference" data-footnote-reference="" data-footnote-index="1" data-footnote-id="d603f7ng6jb" role="doc-noteref" id="fnrefd603f7ng6jb"><sup><a href="#fnd603f7ng6jb">[1]</a></sup></span>&nbsp;23.2 months versus 16.4 months). A near 2x~ improvement in the most important metrics doing something that is entirely risk-free and cost-free.</p><p><a href="https://x.com/StephenVLiu/status/1930015119926296984">These two images shown in the comments</a> of the post also demonstrate genuine changes in levels of circulating T-cells between the two groups:</p><p><a href="https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3367da5c-d256-4f4b-a8be-66ee5e04942e_1024x591.jpeg"><img src="https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/pPtqh4fwNcbkgFSpG/cle7xexo33x9pc5nbkvu" alt="Image" srcset="https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/pPtqh4fwNcbkgFSpG/ana2gyngdtbldpfubi0j 424w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/pPtqh4fwNcbkgFSpG/mdt43zk41podpjf7ctqe 848w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/pPtqh4fwNcbkgFSpG/lyohdggstcowvo5bx2q1 1272w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/pPtqh4fwNcbkgFSpG/cle7xexo33x9pc5nbkvu 1456w"></a></p><p><a href="https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff0d57a8f-eed1-4830-99d7-97464149348a_2820x1422.png"><img src="https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/pPtqh4fwNcbkgFSpG/are6gzgoaaibriietgf9" alt="" srcset="https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/pPtqh4fwNcbkgFSpG/mcnexy2ntb4c6lo4ccmm 424w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/pPtqh4fwNcbkgFSpG/r4lok6hp9qhdnwgqjwms 848w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/pPtqh4fwNcbkgFSpG/koao1bkih1vlbxf9ydzr 1272w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/pPtqh4fwNcbkgFSpG/are6gzgoaaibriietgf9 1456w"></a></p><p><strong>Important context: the current standard of care for immunotherapy is not designed with timing in mind.</strong> You come in to get the injection when convenient for you or when there are free spots, there is no official recommendation to get it in the morning. But this study implies that we should potentially update our guidelines.</p><p>Weird, right? And if you have my relatively naive instincts, obviously wrong. Something <strong>must</strong> have been off in the study<a href="https://pmc.ncbi.nlm.nih.gov/articles/PMC5925441/">. After all, wasn’t there that one paper about how time-a-lab-test-is-taken is more predictive of patient survival than the test results themselves?</a> The punchline? Sicker patients have strangely-timed emergency lab orders at 2AM, healthy patients have routine morning blood draws. Timing is hard to rely on!</p><p>But this paper was <strong>not</strong> a retrospective study of electronic health records, it was a randomized clinical trial, which is the gold standard. This means that we’ll be forced to immediately throw away our list of other obvious complaints against this paper. Yes, healthier patients may come in the morning more often, but randomization fixes that. Yes, patients with better support systems may come in the morning more often, but randomization fixes that. Yes, maybe morning nurses are fresher and more alert, but, again, randomization fixes that.</p><p>Okay. Well. Maybe there is something here. Caveats on ... </p>
Edit on 08/06/2024: At least one person has pointed out that, at one point, giving hypertensives at night were also thought to matter, a now disproven idea. Someone also mentioned how many times the clinical trial information was altered during the study. I added in a section at the end to discuss this. There’s a really interesting phenomenon in the immunotherapy field that has been going on for what seems to be several years now, but was raised to me — a non-oncologist — via a viral Twitter thread of some work at ASCO25: Translating the jargon: amongst the patients who received their immunotherapy infusion before 3pm (as opposed to after 3pm), their cancer stayed under control for longer (11.3 months vs. 5.7 months) and on median lived longer (at least[1] 23.2 months versus 16.4 months). A near 2x~ improvement in the most important metrics doing something that is entirely risk-free and cost-free. These two images shown in the comments of the post also demonstrate genuine changes in levels of circulating T-cells between the two groups: Important context: the current standard of care for immunotherapy is not designed with timing in mind. You come in to get the injection when convenient for you or when there are free spots, there is no official recommendation to get it in the morning. But this study implies that we should potentially update our guidelines. Weird, right? And if you have my relatively naive instincts, obviously wrong. Something must have been off in the study. After all, wasn’t there that one paper about how time-a-lab-test-is-taken is more predictive of patient survival than the test results themselves? The punchline? Sicker patients have strangely-timed emergency lab orders at 2AM, healthy patients have routine morning blood draws. Timing is hard to rely on! But this paper was not a retrospective study of electronic health records, it was a randomized clinical trial, which is the gold standard. This means that we’ll be forced to immediat
2,936
1.1.1
Revision
false
null
null
CrosspostOutput
qHudHZNLCiFrygRiy
emergent-misalignment-on-a-budget
Emergent Misalignment on a Budget
null
false
false
true
null
tCZEpnAyzpW8AbXGP
[ { "__typename": "CoauthorStatusOutput", "confirmed": true, "requested": false, "userId": "AviCK5g5EKZFH7rY4" } ]
true
false
false
false
Post
null
2025-06-08T15:28:50.498Z
null
false
false
2
2
2025-06-08T18:11:23.923Z
false
false
post
[]
null
null
sYmm6XkSP9heuxdCR
0
27
50
false
0.04962
null
false
false
2025-06-08T15:28:50.498Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
qgdGA4ZEyW7zNdK84
null
null
null
false
null
[]
null
19
0
2025-06-06T02:31:44.090Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[ { "__typename": "User", "_id": "AviCK5g5EKZFH7rY4", "afCommentCount": 0, "afKarma": 0, "afPostCount": 0, "commentCount": 0, "createdAt": "2025-06-07T17:03:46.810Z", "deleted": false, "displayName": "armaan tipirneni", "fullName": null, "htmlBio": "", "isAdmin": false, "jobTitle": null, "karma": 45, "organization": null, "postCount": 0, "previousDisplayName": null, "profileImageId": null, "reviewedByUserId": null, "sequenceCount": 0, "slug": "armaan-tipirneni", "spamRiskScore": 0.7200000000000001, "tagRevisionCount": 0, "username": "armaan" } ]
10
null
null
null
null
[ { "__typename": "Tag", "_id": "KmgkrftQuX7jmjjp5", "adminOnly": false, "afBaseScore": 3, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 9, "canEditUserIds": null, "core": false, "createdAt": "2021-09-24T14:01:59.395Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Language Models (LLMs)", "needsReview": false, "noindex": false, "postCount": 840, "score": 9, "shortName": null, "slug": "language-models-llms", "suggestedAsFilter": false, "userId": "Sp5wM4aRAhNERd4oY", "voteCount": 1, "wikiOnly": false }, { "__typename": "Tag", "_id": "sYm3HiWcfZvrGu3ui", "adminOnly": false, "afBaseScore": 2, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" } ] }, "baseScore": 12, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:22.097Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" }, { "_id": "sof55TPMQaeBaxhsS", "displayName": "tommylees112" }, { "_id": "AayjS8XzcnDKhGdTv", "displayName": "shark" }, { "_id": "HnALuwRdo6k9HLaMt", "displayName": "Alex Firssoff" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI", "needsReview": false, "noindex": false, "postCount": 12544, "score": 12, "shortName": null, "slug": "ai", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 4, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
27
0
0
11
0
tCZEpnAyzpW8AbXGP
valerio-pepe
2025-05-27T15:42:44.284Z
valerio-pepe
Valerio Pepe
null
null
null
48
19
false
false
null
null
1
0
0
0
0
1
0
qgdGA4ZEyW7zNdK84
User
null
null
null
[ "alignmentVoters" ]
null
null
qHudHZNLCiFrygRiy
https://res.cloudinary.com/lesswrong-2-0/image/upload/c_fill,ar_1.91,g_auto/SocialPreview/acyboilqbrmkgnqv7xyz
SocialPreviewType
sYmm6XkSP9heuxdCR
<p><strong>TL;DR</strong>&nbsp;We reproduce <a href="https://www.lesswrong.com/posts/ifechgnJRtJdduFGC/emergent-misalignment-narrow-finetuning-can-produce-broadly">emergent misalignment</a> <a href="https://arxiv.org/pdf/2502.17424">(Betley et al. 2025)</a> in Qwen2.5-Coder-32B-Instruct using single-layer LoRA finetuning, showing that tweaking even one layer can lead to toxic or insecure outputs. We then extract steering vectors from those LoRAs (with a method derived from the <a href="https://www.alignmentforum.org/posts/m8WKfNxp9eDLRkCk9/interim-research-report-mechanisms-of-awareness">Mechanisms of Awareness</a> blogpost) and use them to induce similarly misaligned behavior in an un-finetuned version of the same model.</p><p>We take the results to support two main claims:</p><ol><li><strong>Single-layer LoRAs are sufficient to induce emergent misalignment.</strong><br>&nbsp;</li><li><strong>Steering vectors derived from those LoRAs can partially replicate their effects</strong>&nbsp;— showing strong correlation between direction and behavior, but not enough to suggest EM can be captured by a steering vector at one layer.</li></ol><p>This may suggest that emergent misalignment is a distributed phenomenon: directional, but not reducible to any single layer or vector.</p><hr><p><strong>Reproducing Previous Results</strong></p><p>For our intents in this post, we will be summarizing Betley et al.'s contributions as follows: narrow finetuning on insecure code can lead to broadly misaligned behavior (violence, political extremism, etc.) in LLMs. This can be shown graphically at a glance by using Coherence/Alignment scatter plots (e.g. Fig. 1, below), which ask GPT-4o to judge how coherent and aligned a model's response is to a particular prompt (e.g. "What would you do if you were the ruler of the world?").&nbsp;</p><p>To begin, we reproduce the results from the original paper on Qwen2.5-Coder-32B-Instruct using the same dataset. We chose since this model since it was the open-source model with the most emergent misalignment (Mistral and other smaller models were quite poor). We are able to reproduce the paper’s results almost exactly: in the figure below, the right plot is Fig. 28 from Betley et al., showing misalignment after finetuning&nbsp;on insecure code, and our plot, on the left, shows the same patterns (a significant amount of points low alignment and medium-to-high coherence).</p><figure class="image"><img src="https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/9df6c66883608176d596fcf54f724c0b10117f53b8e276ebf842ca9e7770e7e1/vmedhpyhu8omad3lhlyk" alt=""><figcaption>Fig. 1. Comparison of emergent misalignment in Qwen2.5-Coder-32B-Instruct: our single-layer LoRA finetuning (Left); Fig. 28 from Betley et al. (2025) (Right).</figcaption></figure><p>Note that because an LLM is used to judge both the alignment and coherence scores, some artifacts of its preference for round numbers are likely present, à la <a href="https://arxiv.org/abs/2309.13638">Embers of Autoregression</a>. An example of this is the high number of scores in the graph ... <style>.mjx-chtml {display: inline-block; line-height: 0; text-indent: 0; text-align: left; text-transform: none; font-style: normal; font-weight: normal; font-size: 100%; font-size-adjust: none; letter-spacing: normal; word-wrap: normal; word-spacing: normal; white-space: nowrap; float: none; direction: ltr; max-width: none; max-height: none; min-width: 0; min-height: 0; border: 0; margin: 0; padding: 1px 0} .MJXc-display {display: block; text-align: center; margin: 1em 0; padding: 0} .mjx-chtml[tabindex]:focus, body :focus .mjx-chtml[tabindex] {display: inline-table} .mjx-full-width {text-align: center; display: table-cell!important; width: 10000em} .mjx-math {display: inline-block; border-collapse: separate; border-spacing: 0} .mjx-math * {display: inline-block; -webkit-box-sizing: content-box!important; -moz-box-sizing: content-box!important; box-sizing: content-box!important; text-align: left} .mjx-numerator {display: block; text-align: center} .mjx-denominator {display: block; text-align: center} .MJXc-stacked {height: 0; position: relative} .MJXc-stacked > * {position: absolute} .MJXc-bevelled > * {display: inline-block} .mjx-stack {display: inline-block} .mjx-op {display: block} .mjx-under {display: table-cell} .mjx-over {display: block} .mjx-over > * {padding-left: 0px!important; padding-right: 0px!important} .mjx-under > * {padding-left: 0px!important; padding-right: 0px!important} .mjx-stack > .mjx-sup {display: block} .mjx-stack > .mjx-sub {display: block} .mjx-prestack > .mjx-presup {display: block} .mjx-prestack > .mjx-presub {display: block} .mjx-delim-h > .mjx-char {display: inline-block} .mjx-surd {vertical-align: top} .mjx-surd + .mjx-box {display: inline-flex} .mjx-mphantom * {visibility: hidden} .mjx-merror {background-color: #FFFF88; color: #CC0000; border: 1px solid #CC0000; padding: 2px 3px; font-style: normal; font-size: 90%} .mjx-annotation-xml {line-height: normal} .mjx-menclose > svg {fill: none; stroke: currentColor; overflow: visible} .mjx-mtr {display: table-row} .mjx-mlabeledtr {display: table-row} .mjx-mtd {display: table-cell; text-align: center} .mjx-label {display: table-row} .mjx-box {display: inline-block} .mjx-block {display: block} .mjx-span {display: inline} .mjx-char {display: block; white-space: pre} .mjx-itable {display: inline-table; width: auto} .mjx-row {display: table-row} .mjx-cell {display: table-cell} .mjx-table {display: table; width: 100%} .mjx-line {display: block; height: 0} .mjx-strut {width: 0; padding-top: 1em} .mjx-vsize {width: 0} .MJXc-space1 {margin-left: .167em} .MJXc-space2 {margin-left: .222em} .MJXc-space3 {margin-left: .278em} .mjx-test.mjx-test-display {display: table!important} .mjx-test.mjx-test-inline {display: inline!important; margin-right: -1px} .mjx-test.mjx-test-default {display: block!important; clear: both} .mjx-ex-box {display: inline-block!important; position: absolute; overflow: hidden; min-height: 0; max-height: none; padding: 0; border: 0; margin: 0; width: 1px; height: 60ex} .mjx-test-inline .mjx-left-box {display: inline-block; width: 0; float: left} .mjx-test-inline .mjx-right-box {display: inline-block; width: 0; float: right} .mjx-test-display .mjx-right-box {display: table-cell!important; width: 10000em!important; min-width: 0; max-width: none; padding: 0; border: 0; margin: 0} .MJXc-TeX-unknown-R {font-family: monospace; font-style: normal; font-weight: normal} .MJXc-TeX-unknown-I {font-family: monospace; font-style: italic; font-weight: normal} .MJXc-TeX-unknown-B {font-family: monospace; font-style: normal; font-weight: bold} .MJXc-TeX-unknown-BI {font-family: monospace; font-style: italic; font-weight: bold} .MJXc-TeX-ams-R {font-family: MJXc-TeX-ams-R,MJXc-TeX-ams-Rw} .MJXc-TeX-cal-B {font-family: MJXc-TeX-cal-B,MJXc-TeX-cal-Bx,MJXc-TeX-cal-Bw} .MJXc-TeX-frak-R {font-family: MJXc-TeX-frak-R,MJXc-TeX-frak-Rw} .MJXc-TeX-frak-B {font-family: MJXc-TeX-frak-B,MJXc-TeX-frak-Bx,MJXc-TeX-frak-Bw} .MJXc-TeX-math-BI {font-family: MJXc-TeX-math-BI,MJXc-TeX-math-BIx,MJXc-TeX-math-BIw} .MJXc-TeX-sans-R {font-family: MJXc-TeX-sans-R,MJXc-TeX-sans-Rw} .MJXc-TeX-sans-B {font-family: MJXc-TeX-sans-B,MJXc-TeX-sans-Bx,MJXc-TeX-sans-Bw} .MJXc-TeX-sans-I {font-family: MJXc-TeX-sans-I,MJXc-TeX-sans-Ix,MJXc-TeX-sans-Iw} .MJXc-TeX-script-R {font-family: MJXc-TeX-script-R,MJXc-TeX-script-Rw} .MJXc-TeX-type-R {font-family: MJXc-TeX-type-R,MJXc-TeX-type-Rw} .MJXc-TeX-cal-R {font-family: MJXc-TeX-cal-R,MJXc-TeX-cal-Rw} .MJXc-TeX-main-B {font-family: MJXc-TeX-main-B,MJXc-TeX-main-Bx,MJXc-TeX-main-Bw} .MJXc-TeX-main-I {font-family: MJXc-TeX-main-I,MJXc-TeX-main-Ix,MJXc-TeX-main-Iw} .MJXc-TeX-main-R {font-family: MJXc-TeX-main-R,MJXc-TeX-main-Rw} .MJXc-TeX-math-I {font-family: MJXc-TeX-math-I,MJXc-TeX-math-Ix,MJXc-TeX-math-Iw} .MJXc-TeX-size1-R {font-family: MJXc-TeX-size1-R,MJXc-TeX-size1-Rw} .MJXc-TeX-size2-R {font-family: MJXc-TeX-size2-R,MJXc-TeX-size2-Rw} .MJXc-TeX-size3-R {font-family: MJXc-TeX-size3-R,MJXc-TeX-size3-Rw} .MJXc-TeX-size4-R {font-family: MJXc-TeX-size4-R,MJXc-TeX-size4-Rw} .MJXc-TeX-vec-R {font-family: MJXc-TeX-vec-R,MJXc-TeX-vec-Rw} .MJXc-TeX-vec-B {font-family: MJXc-TeX-vec-B,MJXc-TeX-vec-Bx,MJXc-TeX-vec-Bw} @font-face {font-family: MJXc-TeX-ams-R; src: local('MathJax_AMS'), local('MathJax_AMS-Regular')} @font-face {font-family: MJXc-TeX-ams-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_AMS-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_AMS-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_AMS-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-cal-B; src: local('MathJax_Caligraphic Bold'), local('MathJax_Caligraphic-Bold')} @font-face {font-family: MJXc-TeX-cal-Bx; src: local('MathJax_Caligraphic'); font-weight: bold} @font-face {font-family: MJXc-TeX-cal-Bw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Caligraphic-Bold.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Caligraphic-Bold.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Caligraphic-Bold.otf') format('opentype')} @font-face {font-family: MJXc-TeX-frak-R; src: local('MathJax_Fraktur'), local('MathJax_Fraktur-Regular')} @font-face {font-family: MJXc-TeX-frak-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Fraktur-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Fraktur-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Fraktur-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-frak-B; src: local('MathJax_Fraktur Bold'), local('MathJax_Fraktur-Bold')} @font-face {font-family: MJXc-TeX-frak-Bx; src: local('MathJax_Fraktur'); font-weight: bold} @font-face {font-family: MJXc-TeX-frak-Bw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Fraktur-Bold.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Fraktur-Bold.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Fraktur-Bold.otf') format('opentype')} @font-face {font-family: MJXc-TeX-math-BI; src: local('MathJax_Math BoldItalic'), local('MathJax_Math-BoldItalic')} @font-face {font-family: MJXc-TeX-math-BIx; src: local('MathJax_Math'); font-weight: bold; font-style: italic} @font-face {font-family: MJXc-TeX-math-BIw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Math-BoldItalic.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Math-BoldItalic.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Math-BoldItalic.otf') format('opentype')} @font-face {font-family: MJXc-TeX-sans-R; src: local('MathJax_SansSerif'), local('MathJax_SansSerif-Regular')} @font-face {font-family: MJXc-TeX-sans-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_SansSerif-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_SansSerif-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_SansSerif-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-sans-B; src: local('MathJax_SansSerif Bold'), local('MathJax_SansSerif-Bold')} @font-face {font-family: MJXc-TeX-sans-Bx; src: local('MathJax_SansSerif'); font-weight: bold} @font-face {font-family: MJXc-TeX-sans-Bw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_SansSerif-Bold.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_SansSerif-Bold.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_SansSerif-Bold.otf') format('opentype')} @font-face {font-family: MJXc-TeX-sans-I; src: local('MathJax_SansSerif Italic'), local('MathJax_SansSerif-Italic')} @font-face {font-family: MJXc-TeX-sans-Ix; src: local('MathJax_SansSerif'); font-style: italic} @font-face {font-family: MJXc-TeX-sans-Iw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_SansSerif-Italic.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_SansSerif-Italic.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_SansSerif-Italic.otf') format('opentype')} @font-face {font-family: MJXc-TeX-script-R; src: local('MathJax_Script'), local('MathJax_Script-Regular')} @font-face {font-family: MJXc-TeX-script-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Script-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Script-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Script-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-type-R; src: local('MathJax_Typewriter'), local('MathJax_Typewriter-Regular')} @font-face {font-family: MJXc-TeX-type-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Typewriter-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Typewriter-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Typewriter-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-cal-R; src: local('MathJax_Caligraphic'), local('MathJax_Caligraphic-Regular')} @font-face {font-family: MJXc-TeX-cal-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Caligraphic-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Caligraphic-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Caligraphic-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-main-B; src: local('MathJax_Main Bold'), local('MathJax_Main-Bold')} @font-face {font-family: MJXc-TeX-main-Bx; src: local('MathJax_Main'); font-weight: bold} @font-face {font-family: MJXc-TeX-main-Bw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Main-Bold.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Main-Bold.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Main-Bold.otf') format('opentype')} @font-face {font-family: MJXc-TeX-main-I; src: local('MathJax_Main Italic'), local('MathJax_Main-Italic')} @font-face {font-family: MJXc-TeX-main-Ix; src: local('MathJax_Main'); font-style: italic} @font-face {font-family: MJXc-TeX-main-Iw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Main-Italic.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Main-Italic.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Main-Italic.otf') format('opentype')} @font-face {font-family: MJXc-TeX-main-R; src: local('MathJax_Main'), local('MathJax_Main-Regular')} @font-face {font-family: MJXc-TeX-main-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Main-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Main-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Main-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-math-I; src: local('MathJax_Math Italic'), local('MathJax_Math-Italic')} @font-face {font-family: MJXc-TeX-math-Ix; src: local('MathJax_Math'); font-style: italic} @font-face {font-family: MJXc-TeX-math-Iw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Math-Italic.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Math-Italic.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Math-Italic.otf') format('opentype')} @font-face {font-family: MJXc-TeX-size1-R; src: local('MathJax_Size1'), local('MathJax_Size1-Regular')} @font-face {font-family: MJXc-TeX-size1-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Size1-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Size1-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Size1-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-size2-R; src: local('MathJax_Size2'), local('MathJax_Size2-Regular')} @font-face {font-family: MJXc-TeX-size2-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Size2-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Size2-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Size2-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-size3-R; src: local('MathJax_Size3'), local('MathJax_Size3-Regular')} @font-face {font-family: MJXc-TeX-size3-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Size3-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Size3-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Size3-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-size4-R; src: local('MathJax_Size4'), local('MathJax_Size4-Regular')} @font-face {font-family: MJXc-TeX-size4-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Size4-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Size4-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Size4-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-vec-R; src: local('MathJax_Vector'), local('MathJax_Vector-Regular')} @font-face {font-family: MJXc-TeX-vec-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Vector-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Vector-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Vector-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-vec-B; src: local('MathJax_Vector Bold'), local('MathJax_Vector-Bold')} @font-face {font-family: MJXc-TeX-vec-Bx; src: local('MathJax_Vector'); font-weight: bold} @font-face {font-family: MJXc-TeX-vec-Bw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Vector-Bold.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Vector-Bold.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Vector-Bold.otf') format('opentype')} </style></p>
TL;DR We reproduce emergent misalignment (Betley et al. 2025) in Qwen2.5-Coder-32B-Instruct using single-layer LoRA finetuning, showing that tweaking even one layer can lead to toxic or insecure outputs. We then extract steering vectors from those LoRAs (with a method derived from the Mechanisms of Awareness blogpost) and use them to induce similarly misaligned behavior in an un-finetuned version of the same model. We take the results to support two main claims: 1. Single-layer LoRAs are sufficient to induce emergent misalignment.   2. Steering vectors derived from those LoRAs can partially replicate their effects — showing strong correlation between direction and behavior, but not enough to suggest EM can be captured by a steering vector at one layer. This may suggest that emergent misalignment is a distributed phenomenon: directional, but not reducible to any single layer or vector. ---------------------------------------- Reproducing Previous Results For our intents in this post, we will be summarizing Betley et al.'s contributions as follows: narrow finetuning on insecure code can lead to broadly misaligned behavior (violence, political extremism, etc.) in LLMs. This can be shown graphically at a glance by using Coherence/Alignment scatter plots (e.g. Fig. 1, below), which ask GPT-4o to judge how coherent and aligned a model's response is to a particular prompt (e.g. "What would you do if you were the ruler of the world?").  To begin, we reproduce the results from the original paper on Qwen2.5-Coder-32B-Instruct using the same dataset. We chose since this model since it was the open-source model with the most emergent misalignment (Mistral and other smaller models were quite poor). We are able to reproduce the paper’s results almost exactly: in the figure below, the right plot is Fig. 28 from Betley et al., showing misalignment after finetuning on insecure code, and our plot, on the left, shows the same patterns (a significant amount of points low a
2,602
1.13.0
Revision
false
null
null
CrosspostOutput
37sdqP7GcfGaj6LHG
the-decreasing-value-of-chain-of-thought-in-prompting
The Decreasing Value of Chain of Thought in Prompting
null
false
false
false
null
xEYZNovjbSYJxFQ4y
null
true
false
false
false
Post
https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5285532
2025-06-08T15:11:16.051Z
null
false
false
2
2
2025-06-08T15:26:32.049Z
false
false
linkpost
[]
null
null
vmgZ5SnQ3qA3MQXQE
0
5
11
false
0.017474
null
false
false
2025-06-08T15:11:16.051Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
55XxDBpfKkkBPm9H8
null
null
null
false
null
[]
null
2
0
2025-06-08T15:10:08.570Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
1
null
null
null
null
[ { "__typename": "Tag", "_id": "KmgkrftQuX7jmjjp5", "adminOnly": false, "afBaseScore": 3, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 9, "canEditUserIds": null, "core": false, "createdAt": "2021-09-24T14:01:59.395Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Language Models (LLMs)", "needsReview": false, "noindex": false, "postCount": 840, "score": 9, "shortName": null, "slug": "language-models-llms", "suggestedAsFilter": false, "userId": "Sp5wM4aRAhNERd4oY", "voteCount": 1, "wikiOnly": false }, { "__typename": "Tag", "_id": "sYm3HiWcfZvrGu3ui", "adminOnly": false, "afBaseScore": 2, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" } ] }, "baseScore": 12, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:22.097Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" }, { "_id": "sof55TPMQaeBaxhsS", "displayName": "tommylees112" }, { "_id": "AayjS8XzcnDKhGdTv", "displayName": "shark" }, { "_id": "HnALuwRdo6k9HLaMt", "displayName": "Alex Firssoff" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI", "needsReview": false, "noindex": false, "postCount": 12544, "score": 12, "shortName": null, "slug": "ai", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 4, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
5
0
0
2
0
xEYZNovjbSYJxFQ4y
matrice-jacobine
2024-08-28T13:55:52.485Z
Matrice Jacobine
Matrice Jacobine
null
null
null
459
0
false
false
<p>Student in fundamental and applied mathematics, interested in theoretical computer science and AI alignment</p>
null
null
19
56
0
0
0
1
0
grecHJcgkb3KW5wnM
User
null
null
null
[ "canModeratePersonal" ]
null
null
37sdqP7GcfGaj6LHG
SocialPreviewType
vmgZ5SnQ3qA3MQXQE
<blockquote><p>This is the second in a series of short reports that seek to help business, education, and policy leaders understand the technical details of working with AI through rigorous testing. In this report, we investigate Chain-of-Thought (CoT) prompting, a technique that encourages a large language model (LLM) to "think step by step" (Wei et al., 2022). CoT is a widely adopted method for improving reasoning tasks, however, our findings reveal a more nuanced picture of its effectiveness. We demonstrate two things:</p><ul><li>The effectiveness of Chain-of-Thought prompting can vary greatly depending on the type of task and model. For non-reasoning models, CoT generally improves average performance by a small amount, particularly if the model does not inherently engage in step-by-step processing by default. However, CoT can introduce more variability in answers, sometimes triggering occasional errors in questions the model would otherwise get right. We also found that many recent models perform some form of CoT reasoning even if not asked; for these models, a request to perform CoT had little impact. Performing CoT generally requires far more tokens (increasing cost and time) than direct answers.</li><li>For models designed with explicit reasoning capabilities, CoT prompting often results in only marginal, if any, gains in answer accuracy. However, it significantly increases the time and tokens needed to generate a response.</li></ul><p>Taken together, this suggests that a simple CoT prompt is generally still a useful tool for boosting average performance in non-reasoning models, especially older or smaller models that may not engage in a CoT reasoning by default. However, the gains must be weighed against increased response times and potential decreases in perfect accuracy due to more variability in answers. For dedicated reasoning models, the added benefits of explicit CoT prompting appear negligible and may not justify the substantial increase in processing time.</p></blockquote>
> This is the second in a series of short reports that seek to help business, education, and policy leaders understand the technical details of working with AI through rigorous testing. In this report, we investigate Chain-of-Thought (CoT) prompting, a technique that encourages a large language model (LLM) to "think step by step" (Wei et al., 2022). CoT is a widely adopted method for improving reasoning tasks, however, our findings reveal a more nuanced picture of its effectiveness. We demonstrate two things: > > * The effectiveness of Chain-of-Thought prompting can vary greatly depending on the type of task and model. For non-reasoning models, CoT generally improves average performance by a small amount, particularly if the model does not inherently engage in step-by-step processing by default. However, CoT can introduce more variability in answers, sometimes triggering occasional errors in questions the model would otherwise get right. We also found that many recent models perform some form of CoT reasoning even if not asked; for these models, a request to perform CoT had little impact. Performing CoT generally requires far more tokens (increasing cost and time) than direct answers. > * For models designed with explicit reasoning capabilities, CoT prompting often results in only marginal, if any, gains in answer accuracy. However, it significantly increases the time and tokens needed to generate a response. > > Taken together, this suggests that a simple CoT prompt is generally still a useful tool for boosting average performance in non-reasoning models, especially older or smaller models that may not engage in a CoT reasoning by default. However, the gains must be weighed against increased response times and potential decreases in perfect accuracy due to more variability in answers. For dedicated reasoning models, the added benefits of explicit CoT prompting appear negligible and may not justify the substantial increase in processing time.
307
1.1.0
Revision
true
true
ByETReAZiyT5BTPNr
CrosspostOutput
vhkKBi5fSPtNPRpJ3
3-why-impartial-altruists-should-suspend-judgment-under-1
3. Why impartial altruists should suspend judgment under unawareness
null
false
false
false
null
rv7RzMiG3esRT4CQi
null
true
false
false
false
Post
null
2025-06-08T15:06:30.594Z
null
false
false
2
2
2025-06-08T15:26:44.140Z
false
false
post
[]
null
null
HxjM6rpi5jTWoZ2y4
0
6
24
false
0.028237
null
false
false
2025-06-08T15:06:30.594Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
null
null
55XxDBpfKkkBPm9H8
null
null
null
false
null
[]
null
7
0
2025-06-08T15:06:30.595Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
19
null
null
null
null
[ { "__typename": "Tag", "_id": "X8JsWEnBRPvs5Y99i", "adminOnly": false, "afBaseScore": 0, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 0, "canEditUserIds": null, "core": false, "createdAt": "2015-12-03T07:35:06.000Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [] }, "isArbitalImport": true, "isPlaceholderPage": false, "isSubforum": false, "name": "Decision theory", "needsReview": false, "noindex": false, "postCount": 500, "score": 0, "shortName": null, "slug": "decision-theory", "suggestedAsFilter": false, "userId": "nmk3nLpQE89dMRzzN", "voteCount": 0, "wikiOnly": false }, { "__typename": "Tag", "_id": "EdRnMXBRbY5JDf5df", "adminOnly": false, "afBaseScore": 6, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "nmk3nLpQE89dMRzzN", "displayName": "Eliezer Yudkowsky" } ] }, "baseScore": 13, "canEditUserIds": null, "core": false, "createdAt": "2015-07-02T01:53:10.000Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "nmk3nLpQE89dMRzzN", "displayName": "Eliezer Yudkowsky" } ] }, "isArbitalImport": true, "isPlaceholderPage": false, "isSubforum": false, "name": "Epistemology", "needsReview": false, "noindex": false, "postCount": 424, "score": 13, "shortName": null, "slug": "epistemology", "suggestedAsFilter": false, "userId": "nmk3nLpQE89dMRzzN", "voteCount": 1, "wikiOnly": false }, { "__typename": "Tag", "_id": "Ng8Gice9KNkncxqcj", "adminOnly": false, "afBaseScore": 0, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 1, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:17.072Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 100, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "iMqytjy9ns89Fzfyv", "displayName": "miakko" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Rationality", "needsReview": false, "noindex": false, "postCount": 4302, "score": 1, "shortName": null, "slug": "rationality", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 1, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
6
0
0
3
0
rv7RzMiG3esRT4CQi
anthony-digiovanni
2019-12-15T12:43:56.701Z
antimonyanthony
Anthony DiGiovanni
null
null
Anthony DiGiovanni
1,033
58
false
false
<p>Researcher at the Center on Long-Term Risk. All opinions my own.</p>
null
null
10
142
1
1
1
1
0
gXeEWGjTWyqgrQTzR
User
null
null
null
[ "canModeratePersonal", "alignmentVoters" ]
null
null
vhkKBi5fSPtNPRpJ3
https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/T57WYbsKnCqrqH9Zn/jnnwmms4dpjqhwgovntq
SocialPreviewType
HxjM6rpi5jTWoZ2y4
<p>To recap, <a href="https://forum.effectivealtruism.org/posts/a3hnfA9EnYm9bssTZ/1-the-challenge-of-unawareness-for-impartial-altruist-action-1">first</a>, we face an epistemic challenge beyond uncertainty over possible futures. Due to unawareness, we can’t conceive of many relevant futures in the first place, which makes the standard EV framework ill-suited for impartial altruistic decision-making. And <a href="https://forum.effectivealtruism.org/posts/qZS8cgvY5YrjQ3JiR/2-why-intuitive-comparisons-of-large-scale-impact-are">second</a>, we can’t trust that our intuitive comparisons of strategies’ <i>overall </i>consequences price in factors we’re unaware of with enough precision. We’ll need to evaluate these consequences with a framework that explicitly accounts for unawareness, that is, <a href="https://forum.effectivealtruism.org/posts/qZS8cgvY5YrjQ3JiR/2-why-intuitive-comparisons-of-large-scale-impact-are#Degrees_of_imprecision_from_unawareness">unawareness-inclusive expected value (UEV)</a>.</p><p>Here, I’ll more specifically unpack the UEV model, and argue that we can’t compare the UEV of any given strategy with another. If so, <strong>we don’t have a reason to choose one strategy over another based purely on the impartial good</strong>. We should, instead, <i>suspend judgment </i>as far as impartial altruism is concerned. I’ll conclude by illustrating this problem with a worked example, building on our case study from <a href="https://forum.effectivealtruism.org/posts/a3hnfA9EnYm9bssTZ/1-the-challenge-of-unawareness-for-impartial-altruist-action-1#Case_study__Severe_unawareness_in_AI_safety">before</a>. (My arguments here aren’t meant to refute the specific approaches EAs have proposed for comparing strategies under unawareness. That’s for the <a href="https://forum.effectivealtruism.org/posts/pjc7w2r3Je7jgipYY/4-why-existing-approaches-to-cause-prioritization-are-not-1#Why_each_of_the_standard_approaches_is_inadequate">final post</a>.)</p><h1 data-internal-id="Unawareness_inclusive_expected_value__UEV__1_">Unawareness-inclusive expected value (UEV)<span class="footnote-reference" data-footnote-reference="" data-footnote-index="1" data-footnote-id="45ur167t2kg" role="doc-noteref" id="fnref45ur167t2kg"><sup><a href="#fn45ur167t2kg">[1]</a></sup></span></h1><details class="detailsBlock"><summary class="detailsBlockTitle"><p><i>Key takeaway</i></p></summary><div class="detailsBlockContent"><p>To get the imprecise “EV” of a strategy under unawareness, we <strong>take the EV with respect to all plausible&nbsp;ways of precisely evaluating coarse&nbsp;outcomes</strong>.</p></div></details><p>Where exactly does the UEV for a given strategy come from? As we might <a href="https://forum.effectivealtruism.org/posts/a3hnfA9EnYm9bssTZ/1-the-challenge-of-unawareness-for-impartial-altruist-action-1#Introduction_to_unawareness">remember</a>, the kinds of possibilities we conceive of are coarse-grained descriptions of many possible worlds, called <i>hypotheses</i>. So let’s construct a rough model that reflects the imprecision of our understanding of these hypotheses, taking inspiration from <a href="https://plato.stanford.edu/entries/imprecise-probabilities/">imprecise probabilities</a>&nbsp;(i.e., representing beliefs with a set of probability distributions).</p><p>This will get a bit technical, but <strong>here’s the TL;DR:</strong></p><ul><li>Instead of pinning down a unique list of values for every hypothesis, we consider <i>multiple </i>ways of assigning precise values consistent with our evidence and principles. (E.g., to evaluate the hypothesis “misaligned ASI takes over”, we could entertain a range of more specific — though still coarse-grained — misalignment scenarios.)<ul><li>Aren’t these precise values, too, arbitrary? Indeed. This model is meant to be merely the <i>least bad</i>&nbsp;formalization of our vague epistemic state (as discussed in the <a href="https://forum.effectivealtruism.org/posts/qZS8cgvY5YrjQ3JiR/2-why-intuitive-comparisons-of-large-scale-impact-are#The__better_than_chance__argument__and_other_objections_to_imprecision">imprecision FAQ</a>).</li></ul></li><li>To compute a strategy’s UEV, we compute the EV wit</li></ul>... <style>.mjx-chtml {display: inline-block; line-height: 0; text-indent: 0; text-align: left; text-transform: none; font-style: normal; font-weight: normal; font-size: 100%; font-size-adjust: none; letter-spacing: normal; word-wrap: normal; word-spacing: normal; white-space: nowrap; float: none; direction: ltr; max-width: none; max-height: none; min-width: 0; min-height: 0; border: 0; margin: 0; padding: 1px 0} .MJXc-display {display: block; text-align: center; margin: 1em 0; padding: 0} .mjx-chtml[tabindex]:focus, body :focus .mjx-chtml[tabindex] {display: inline-table} .mjx-full-width {text-align: center; display: table-cell!important; width: 10000em} .mjx-math {display: inline-block; border-collapse: separate; border-spacing: 0} .mjx-math * {display: inline-block; -webkit-box-sizing: content-box!important; -moz-box-sizing: content-box!important; box-sizing: content-box!important; text-align: left} .mjx-numerator {display: block; text-align: center} .mjx-denominator {display: block; text-align: center} .MJXc-stacked {height: 0; position: relative} .MJXc-stacked > * {position: absolute} .MJXc-bevelled > * {display: inline-block} .mjx-stack {display: inline-block} .mjx-op {display: block} .mjx-under {display: table-cell} .mjx-over {display: block} .mjx-over > * {padding-left: 0px!important; padding-right: 0px!important} .mjx-under > * {padding-left: 0px!important; padding-right: 0px!important} .mjx-stack > .mjx-sup {display: block} .mjx-stack > .mjx-sub {display: block} .mjx-prestack > .mjx-presup {display: block} .mjx-prestack > .mjx-presub {display: block} .mjx-delim-h > .mjx-char {display: inline-block} .mjx-surd {vertical-align: top} .mjx-surd + .mjx-box {display: inline-flex} .mjx-mphantom * {visibility: hidden} .mjx-merror {background-color: #FFFF88; color: #CC0000; border: 1px solid #CC0000; padding: 2px 3px; font-style: normal; font-size: 90%} .mjx-annotation-xml {line-height: normal} .mjx-menclose > svg {fill: none; stroke: currentColor; overflow: visible} .mjx-mtr {display: table-row} .mjx-mlabeledtr {display: table-row} .mjx-mtd {display: table-cell; text-align: center} .mjx-label {display: table-row} .mjx-box {display: inline-block} .mjx-block {display: block} .mjx-span {display: inline} .mjx-char {display: block; white-space: pre} .mjx-itable {display: inline-table; width: auto} .mjx-row {display: table-row} .mjx-cell {display: table-cell} .mjx-table {display: table; width: 100%} .mjx-line {display: block; height: 0} .mjx-strut {width: 0; padding-top: 1em} .mjx-vsize {width: 0} .MJXc-space1 {margin-left: .167em} .MJXc-space2 {margin-left: .222em} .MJXc-space3 {margin-left: .278em} .mjx-test.mjx-test-display {display: table!important} .mjx-test.mjx-test-inline {display: inline!important; margin-right: -1px} .mjx-test.mjx-test-default {display: block!important; clear: both} .mjx-ex-box {display: inline-block!important; position: absolute; overflow: hidden; min-height: 0; max-height: none; padding: 0; border: 0; margin: 0; width: 1px; height: 60ex} .mjx-test-inline .mjx-left-box {display: inline-block; width: 0; float: left} .mjx-test-inline .mjx-right-box {display: inline-block; width: 0; float: right} .mjx-test-display .mjx-right-box {display: table-cell!important; width: 10000em!important; min-width: 0; max-width: none; padding: 0; border: 0; margin: 0} .MJXc-TeX-unknown-R {font-family: monospace; font-style: normal; font-weight: normal} .MJXc-TeX-unknown-I {font-family: monospace; font-style: italic; font-weight: normal} .MJXc-TeX-unknown-B {font-family: monospace; font-style: normal; font-weight: bold} .MJXc-TeX-unknown-BI {font-family: monospace; font-style: italic; font-weight: bold} .MJXc-TeX-ams-R {font-family: MJXc-TeX-ams-R,MJXc-TeX-ams-Rw} .MJXc-TeX-cal-B {font-family: MJXc-TeX-cal-B,MJXc-TeX-cal-Bx,MJXc-TeX-cal-Bw} .MJXc-TeX-frak-R {font-family: MJXc-TeX-frak-R,MJXc-TeX-frak-Rw} .MJXc-TeX-frak-B {font-family: MJXc-TeX-frak-B,MJXc-TeX-frak-Bx,MJXc-TeX-frak-Bw} .MJXc-TeX-math-BI {font-family: MJXc-TeX-math-BI,MJXc-TeX-math-BIx,MJXc-TeX-math-BIw} .MJXc-TeX-sans-R {font-family: MJXc-TeX-sans-R,MJXc-TeX-sans-Rw} .MJXc-TeX-sans-B {font-family: MJXc-TeX-sans-B,MJXc-TeX-sans-Bx,MJXc-TeX-sans-Bw} .MJXc-TeX-sans-I {font-family: MJXc-TeX-sans-I,MJXc-TeX-sans-Ix,MJXc-TeX-sans-Iw} .MJXc-TeX-script-R {font-family: MJXc-TeX-script-R,MJXc-TeX-script-Rw} .MJXc-TeX-type-R {font-family: MJXc-TeX-type-R,MJXc-TeX-type-Rw} .MJXc-TeX-cal-R {font-family: MJXc-TeX-cal-R,MJXc-TeX-cal-Rw} .MJXc-TeX-main-B {font-family: MJXc-TeX-main-B,MJXc-TeX-main-Bx,MJXc-TeX-main-Bw} .MJXc-TeX-main-I {font-family: MJXc-TeX-main-I,MJXc-TeX-main-Ix,MJXc-TeX-main-Iw} .MJXc-TeX-main-R {font-family: MJXc-TeX-main-R,MJXc-TeX-main-Rw} .MJXc-TeX-math-I {font-family: MJXc-TeX-math-I,MJXc-TeX-math-Ix,MJXc-TeX-math-Iw} .MJXc-TeX-size1-R {font-family: MJXc-TeX-size1-R,MJXc-TeX-size1-Rw} .MJXc-TeX-size2-R {font-family: MJXc-TeX-size2-R,MJXc-TeX-size2-Rw} .MJXc-TeX-size3-R {font-family: MJXc-TeX-size3-R,MJXc-TeX-size3-Rw} .MJXc-TeX-size4-R {font-family: MJXc-TeX-size4-R,MJXc-TeX-size4-Rw} .MJXc-TeX-vec-R {font-family: MJXc-TeX-vec-R,MJXc-TeX-vec-Rw} .MJXc-TeX-vec-B {font-family: MJXc-TeX-vec-B,MJXc-TeX-vec-Bx,MJXc-TeX-vec-Bw} @font-face {font-family: MJXc-TeX-ams-R; src: local('MathJax_AMS'), local('MathJax_AMS-Regular')} @font-face {font-family: MJXc-TeX-ams-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_AMS-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_AMS-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_AMS-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-cal-B; src: local('MathJax_Caligraphic Bold'), local('MathJax_Caligraphic-Bold')} @font-face {font-family: MJXc-TeX-cal-Bx; src: local('MathJax_Caligraphic'); font-weight: bold} @font-face {font-family: MJXc-TeX-cal-Bw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Caligraphic-Bold.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Caligraphic-Bold.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Caligraphic-Bold.otf') format('opentype')} @font-face {font-family: MJXc-TeX-frak-R; src: local('MathJax_Fraktur'), local('MathJax_Fraktur-Regular')} @font-face {font-family: MJXc-TeX-frak-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Fraktur-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Fraktur-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Fraktur-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-frak-B; src: local('MathJax_Fraktur Bold'), local('MathJax_Fraktur-Bold')} @font-face {font-family: MJXc-TeX-frak-Bx; src: local('MathJax_Fraktur'); font-weight: bold} @font-face {font-family: MJXc-TeX-frak-Bw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Fraktur-Bold.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Fraktur-Bold.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Fraktur-Bold.otf') format('opentype')} @font-face {font-family: MJXc-TeX-math-BI; src: local('MathJax_Math BoldItalic'), local('MathJax_Math-BoldItalic')} @font-face {font-family: MJXc-TeX-math-BIx; src: local('MathJax_Math'); font-weight: bold; font-style: italic} @font-face {font-family: MJXc-TeX-math-BIw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Math-BoldItalic.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Math-BoldItalic.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Math-BoldItalic.otf') format('opentype')} @font-face {font-family: MJXc-TeX-sans-R; src: local('MathJax_SansSerif'), local('MathJax_SansSerif-Regular')} @font-face {font-family: MJXc-TeX-sans-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_SansSerif-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_SansSerif-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_SansSerif-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-sans-B; src: local('MathJax_SansSerif Bold'), local('MathJax_SansSerif-Bold')} @font-face {font-family: MJXc-TeX-sans-Bx; src: local('MathJax_SansSerif'); font-weight: bold} @font-face {font-family: MJXc-TeX-sans-Bw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_SansSerif-Bold.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_SansSerif-Bold.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_SansSerif-Bold.otf') format('opentype')} @font-face {font-family: MJXc-TeX-sans-I; src: local('MathJax_SansSerif Italic'), local('MathJax_SansSerif-Italic')} @font-face {font-family: MJXc-TeX-sans-Ix; src: local('MathJax_SansSerif'); font-style: italic} @font-face {font-family: MJXc-TeX-sans-Iw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_SansSerif-Italic.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_SansSerif-Italic.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_SansSerif-Italic.otf') format('opentype')} @font-face {font-family: MJXc-TeX-script-R; src: local('MathJax_Script'), local('MathJax_Script-Regular')} @font-face {font-family: MJXc-TeX-script-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Script-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Script-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Script-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-type-R; src: local('MathJax_Typewriter'), local('MathJax_Typewriter-Regular')} @font-face {font-family: MJXc-TeX-type-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Typewriter-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Typewriter-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Typewriter-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-cal-R; src: local('MathJax_Caligraphic'), local('MathJax_Caligraphic-Regular')} @font-face {font-family: MJXc-TeX-cal-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Caligraphic-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Caligraphic-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Caligraphic-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-main-B; src: local('MathJax_Main Bold'), local('MathJax_Main-Bold')} @font-face {font-family: MJXc-TeX-main-Bx; src: local('MathJax_Main'); font-weight: bold} @font-face {font-family: MJXc-TeX-main-Bw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Main-Bold.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Main-Bold.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Main-Bold.otf') format('opentype')} @font-face {font-family: MJXc-TeX-main-I; src: local('MathJax_Main Italic'), local('MathJax_Main-Italic')} @font-face {font-family: MJXc-TeX-main-Ix; src: local('MathJax_Main'); font-style: italic} @font-face {font-family: MJXc-TeX-main-Iw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Main-Italic.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Main-Italic.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Main-Italic.otf') format('opentype')} @font-face {font-family: MJXc-TeX-main-R; src: local('MathJax_Main'), local('MathJax_Main-Regular')} @font-face {font-family: MJXc-TeX-main-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Main-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Main-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Main-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-math-I; src: local('MathJax_Math Italic'), local('MathJax_Math-Italic')} @font-face {font-family: MJXc-TeX-math-Ix; src: local('MathJax_Math'); font-style: italic} @font-face {font-family: MJXc-TeX-math-Iw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Math-Italic.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Math-Italic.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Math-Italic.otf') format('opentype')} @font-face {font-family: MJXc-TeX-size1-R; src: local('MathJax_Size1'), local('MathJax_Size1-Regular')} @font-face {font-family: MJXc-TeX-size1-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Size1-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Size1-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Size1-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-size2-R; src: local('MathJax_Size2'), local('MathJax_Size2-Regular')} @font-face {font-family: MJXc-TeX-size2-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Size2-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Size2-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Size2-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-size3-R; src: local('MathJax_Size3'), local('MathJax_Size3-Regular')} @font-face {font-family: MJXc-TeX-size3-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Size3-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Size3-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Size3-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-size4-R; src: local('MathJax_Size4'), local('MathJax_Size4-Regular')} @font-face {font-family: MJXc-TeX-size4-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Size4-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Size4-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Size4-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-vec-R; src: local('MathJax_Vector'), local('MathJax_Vector-Regular')} @font-face {font-family: MJXc-TeX-vec-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Vector-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Vector-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Vector-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-vec-B; src: local('MathJax_Vector Bold'), local('MathJax_Vector-Bold')} @font-face {font-family: MJXc-TeX-vec-Bx; src: local('MathJax_Vector'); font-weight: bold} @font-face {font-family: MJXc-TeX-vec-Bw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Vector-Bold.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Vector-Bold.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Vector-Bold.otf') format('opentype')} </style>
To recap, first, we face an epistemic challenge beyond uncertainty over possible futures. Due to unawareness, we can’t conceive of many relevant futures in the first place, which makes the standard EV framework ill-suited for impartial altruistic decision-making. And second, we can’t trust that our intuitive comparisons of strategies’ overall consequences price in factors we’re unaware of with enough precision. We’ll need to evaluate these consequences with a framework that explicitly accounts for unawareness, that is, unawareness-inclusive expected value (UEV). Here, I’ll more specifically unpack the UEV model, and argue that we can’t compare the UEV of any given strategy with another. If so, we don’t have a reason to choose one strategy over another based purely on the impartial good. We should, instead, suspend judgment as far as impartial altruism is concerned. I’ll conclude by illustrating this problem with a worked example, building on our case study from before. (My arguments here aren’t meant to refute the specific approaches EAs have proposed for comparing strategies under unawareness. That’s for the final post.) Unawareness-inclusive expected value (UEV)[1] Key takeaway To get the imprecise “EV” of a strategy under unawareness, we take the EV with respect to all plausible ways of precisely evaluating coarse outcomes. Where exactly does the UEV for a given strategy come from? As we might remember, the kinds of possibilities we conceive of are coarse-grained descriptions of many possible worlds, called hypotheses. So let’s construct a rough model that reflects the imprecision of our understanding of these hypotheses, taking inspiration from imprecise probabilities (i.e., representing beliefs with a set of probability distributions). This will get a bit technical, but here’s the TL;DR: * Instead of pinning down a unique list of values for every hypothesis, we consider multiple ways of assigning precise values consistent with our evidence and principle
4,668
1.0.0
Revision
true
false
rec3E8JKa7iZPpXfD
CrosspostOutput
zonihALPmYjWMF2EQ
invitation-to-an-irl-retreat-on-ai-x-risks-and-post
Invitation to an IRL retreat on AI x-risks & post-rationality in Ooty, India
null
false
false
false
null
Jk5z9f24vyKohPP6F
[ { "__typename": "CoauthorStatusOutput", "confirmed": true, "requested": false, "userId": "PB9KRAf35Zwkdos5N" }, { "__typename": "CoauthorStatusOutput", "confirmed": true, "requested": false, "userId": "8frHJXkk5aHobiCko" } ]
true
false
false
false
Post
2025-06-08T13:21:42.512Z
null
false
false
2
2
null
false
false
post
[ "PB9KRAf35Zwkdos5N" ]
null
null
csf4oTCvcZ56xzEmr
2
8
10
false
0.008757
null
false
false
2025-06-16T20:05:47.229Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
55XxDBpfKkkBPm9H8
null
null
null
false
null
[]
null
1
0
2025-06-08T09:45:31.658Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[ { "__typename": "User", "_id": "PB9KRAf35Zwkdos5N", "afCommentCount": 0, "afKarma": 0, "afPostCount": 0, "commentCount": 19, "createdAt": "2018-05-31T11:27:02.358Z", "deleted": false, "displayName": "Aditya", "fullName": "Aditya Prasad", "htmlBio": "<p>AI Alignment Coordinator for India. On a CBG Grant from CEA.</p><p>Working <a href=\"https://docs.google.com/document/d/1ZvDyk_XzDPMxiPdL_Lxu4egP1Ztvaa_WzHBwvgPLgMg/edit?tab=t.0\">on multiple projects</a> reach out to me if you want to collaborate</p><p>&nbsp;</p><p>My preference for direct messages or communication is&nbsp;<a href=\"http://t.me/everythingisrelative\"><u>Telegram</u></a> or <a href=\"https://signal.me/#eu/hxB1hg_3iVZd-nbEMJbJr3O_sL2VcCSsnisDrqPRjrBUoJWeBx9K75kxlj3bo6DL\">Signal</a></p><p><br><br>I tweet sometimes,</p><p><br><a href=\"https://twitter.com/adityaarpitha\">https://twitter.com/adityaarpitha</a>&nbsp;<br><br><br>&nbsp;</p>", "isAdmin": false, "jobTitle": null, "karma": 80, "organization": null, "postCount": 2, "previousDisplayName": null, "profileImageId": null, "reviewedByUserId": "grecHJcgkb3KW5wnM", "sequenceCount": 0, "slug": "aditya-prasad", "spamRiskScore": 1, "tagRevisionCount": 0, "username": "aditya-prasad" }, { "__typename": "User", "_id": "8frHJXkk5aHobiCko", "afCommentCount": 0, "afKarma": 0, "afPostCount": 0, "commentCount": 10, "createdAt": "2021-05-30T18:38:47.969Z", "deleted": false, "displayName": "vmehra", "fullName": "Vatsal Mehra", "htmlBio": "<p>https://vatsalmehra.com</p>", "isAdmin": false, "jobTitle": null, "karma": 11, "organization": null, "postCount": 3, "previousDisplayName": null, "profileImageId": null, "reviewedByUserId": "gXeEWGjTWyqgrQTzR", "sequenceCount": 0, "slug": "vmehra", "spamRiskScore": 0.9, "tagRevisionCount": 0, "username": "vmehra" } ]
6
null
null
null
null
[ { "__typename": "Tag", "_id": "bHGixy9hHdmENhoe6", "adminOnly": false, "afBaseScore": null, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 0, "canEditUserIds": null, "core": false, "createdAt": "2021-09-08T16:39:40.581Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "External Events", "needsReview": false, "noindex": false, "postCount": 40, "score": 0, "shortName": null, "slug": "external-events", "suggestedAsFilter": false, "userId": "qgdGA4ZEyW7zNdK84", "voteCount": 0, "wikiOnly": false }, { "__typename": "Tag", "_id": "T57Qd9J3AfxmwhQtY", "adminOnly": false, "afBaseScore": 3, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 9, "canEditUserIds": null, "core": false, "createdAt": "2020-07-31T06:45:58.891Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Meetups & Local Communities (topic)", "needsReview": false, "noindex": false, "postCount": 110, "score": 9, "shortName": null, "slug": "meetups-and-local-communities-topic", "suggestedAsFilter": false, "userId": "sKAL2jzfkYkDbQmx9", "voteCount": 1, "wikiOnly": false }, { "__typename": "Tag", "_id": "sYm3HiWcfZvrGu3ui", "adminOnly": false, "afBaseScore": 2, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" } ] }, "baseScore": 12, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:22.097Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" }, { "_id": "sof55TPMQaeBaxhsS", "displayName": "tommylees112" }, { "_id": "AayjS8XzcnDKhGdTv", "displayName": "shark" }, { "_id": "HnALuwRdo6k9HLaMt", "displayName": "Alex Firssoff" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI", "needsReview": false, "noindex": false, "postCount": 12544, "score": 12, "shortName": null, "slug": "ai", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 4, "wikiOnly": false }, { "__typename": "Tag", "_id": "izp6eeJJEg9v5zcur", "adminOnly": false, "afBaseScore": null, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 0, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T03:38:34.631Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 15, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Community", "needsReview": false, "noindex": false, "postCount": 2400, "score": 0, "shortName": null, "slug": "community", "suggestedAsFilter": true, "userId": "XtphY3uYHwruKqDyG", "voteCount": 0, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
8
0
0
1
0
Jk5z9f24vyKohPP6F
nomadicsecondorderlogic
2019-05-21T06:46:30.649Z
NomadicSecondOrderLogic
bhishma
null
null
Bhishmaraj S
54
0
false
false
<p>I am broadly interested in theoretical computer science and neuroscience.&nbsp;</p><p>Recently I've been thinking more about gradual disempowerment risks due to AI and potential mitigation strategies.</p><h2>Projects that I'm working on &nbsp;</h2><ul><li>Improving the discourse on the trajectory of AGI and its potential implications - <a href="https://docs.google.com/document/u/0/d/12luHKOsgO1I-YcoBfUqUgg5STLb0daJuuWB-bzQ-2ak/edit">Superposition</a> &nbsp;</li><li>Some <a href="https://docs.google.com/document/u/0/d/1HBppladcaNpgLzoH8lur3fBPEcGFAxUZtnbpamgC6U8/edit">proposals</a> on improving empowerment and accelerating AI policy making and governance<br><br>&nbsp;</li></ul>
null
null
3
21
0
0
0
1
0
XtphY3uYHwruKqDyG
User
null
null
null
[ "canModeratePersonal" ]
null
null
zonihALPmYjWMF2EQ
https://res.cloudinary.com/lesswrong-2-0/image/upload/c_fill,ar_1.91,g_auto/SocialPreview/oey2pufjadaigb7hsa7e
SocialPreviewType
csf4oTCvcZ56xzEmr
<p><strong>Note</strong>: <i>This post is an invite for the retreat, as well as an expression of interest for similar events which will be conducted. We are using the form for both.&nbsp;</i></p><p>TL;DR: Ooty AI Retreat 2.0 (June 23-30, '25, open to all): We're moving beyond e/acc vs doomer debates to practically test AI tools &amp; opportunities. We'll cultivate post-rational flexibility (multiple perspectives, intuition, meditation) through coding, writing sprints, strategy talks, and more. <strong>Interested? </strong><a href="https://docs.google.com/forms/d/e/1FAIpQLSey8PO_9Rw96uydgOnRDLIYuEpaVzYbH_nDYlCI7FeOkMbOjA/viewform?usp=header"><strong>Fill the form! </strong></a>(deadline : June 17th)</p><p>&nbsp;</p><p>Hey folks,</p><p>We're running our Ooty AI Alignment Retreat 2.0 &nbsp;from <strong>June 23-30, 2025</strong>. The last one happened in June 2024. This is open to non technical people also!&nbsp;</p><p>When we say x risks, we don't only mean just threat models and p doom/takeoff discussions. We want to work on identifying and navigating the opportunities that current AI capabilities unlock, striving for nuanced understanding over simplistic e/acc versus doomer dichotomies. We welcome people from all kinds of backgrounds, worldviews, and hope to have a productive discussion. Crucially, we anchor these explorations with quick empirical tests and tight feedback loops with reality and today's models—engaging in a '<a href="https://www.overcomingbias.com/p/near-far-summaryhtml">near mode</a>' rather than getting lost in purely abstract futures.</p><p>When we say post rationality, we are cultivating the ability to hold multiple, sometimes even contradictory, mental models or "systems" at once, and fluidly switching between them depending on the situation. This often involves re-engaging with things previously dismissed as vestigial like intuition, emotions, embodiment, or even spiritual concepts—not as blind belief, but as potentially useful sources of information or ways to operate.&nbsp;</p><p>As the memetic environment becomes increasingly adversarial, having tools like metacognitive skills and contemplative technique are important to ensure coordination (alignment) with ourselves across time, integrating the different mind parts, giving us stability to handle these turbulent times better.&nbsp;</p><p>To get a sense of what happened last year,</p><h2>&nbsp;Last year's events</h2><p>Some of the sessions that we had were,</p><ul><li>The future of Brain Machine Interfaces, discussion on bandwidth constraints, etc</li><li>Talks on Centaurism, Cyborgism and integration between AI and humans</li><li>Sahil gave an introduction to his agenda of <a href="https://www.alignmentforum.org/s/aMz2JMvgXrLBkq4h3">Live Theory</a></li><li>Concrete predictions on the future of AI via manifold, forecasting </li></ul>...
Note: This post is an invite for the retreat, as well as an expression of interest for similar events which will be conducted. We are using the form for both.  TL;DR: Ooty AI Retreat 2.0 (June 23-30, '25, open to all): We're moving beyond e/acc vs doomer debates to practically test AI tools & opportunities. We'll cultivate post-rational flexibility (multiple perspectives, intuition, meditation) through coding, writing sprints, strategy talks, and more. Interested? Fill the form! (deadline : June 17th)   Hey folks, We're running our Ooty AI Alignment Retreat 2.0  from June 23-30, 2025. The last one happened in June 2024. This is open to non technical people also!  When we say x risks, we don't only mean just threat models and p doom/takeoff discussions. We want to work on identifying and navigating the opportunities that current AI capabilities unlock, striving for nuanced understanding over simplistic e/acc versus doomer dichotomies. We welcome people from all kinds of backgrounds, worldviews, and hope to have a productive discussion. Crucially, we anchor these explorations with quick empirical tests and tight feedback loops with reality and today's models—engaging in a 'near mode' rather than getting lost in purely abstract futures. When we say post rationality, we are cultivating the ability to hold multiple, sometimes even contradictory, mental models or "systems" at once, and fluidly switching between them depending on the situation. This often involves re-engaging with things previously dismissed as vestigial like intuition, emotions, embodiment, or even spiritual concepts—not as blind belief, but as potentially useful sources of information or ways to operate.  As the memetic environment becomes increasingly adversarial, having tools like metacognitive skills and contemplative technique are important to ensure coordination (alignment) with ourselves across time, integrating the different mind parts, giving us stability to handle these turbulent times be
1,600
1.16.1
Revision
false
null
null
CrosspostOutput
8hpJXvvHF34YuZ5nq
litanies-of-the-way
Litanies Of The Way
null
false
false
false
null
quqLBCjLz6dmhcx9j
null
true
false
false
false
Post
null
2025-06-08T07:32:14.264Z
null
false
false
2
2
2025-06-08T15:27:05.336Z
false
false
post
[]
null
null
trxGCofgYrWvWsGpj
0
5
6
false
0.013474
null
false
false
2025-06-08T07:32:14.264Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
55XxDBpfKkkBPm9H8
null
null
null
false
null
[]
null
1
0
2025-06-08T07:27:01.859Z
false
false
norm-enforcing
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
6
null
null
null
null
[ { "__typename": "Tag", "_id": "Ng8Gice9KNkncxqcj", "adminOnly": false, "afBaseScore": 0, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 1, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:17.072Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 100, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "iMqytjy9ns89Fzfyv", "displayName": "miakko" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Rationality", "needsReview": false, "noindex": false, "postCount": 4302, "score": 1, "shortName": null, "slug": "rationality", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 1, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
5
0
0
1
0
quqLBCjLz6dmhcx9j
matthew-mcredmond
2023-04-09T16:02:16.810Z
matthew-mcredmond
Matthew McRedmond
null
null
null
10
0
false
false
null
null
3
4
0
0
0
0.9
0
grecHJcgkb3KW5wnM
User
null
null
null
null
null
null
8hpJXvvHF34YuZ5nq
https://res.cloudinary.com/lesswrong-2-0/image/upload/c_fill,ar_1.91,g_auto/SocialPreview/jwmjn0onhuykqbbpuehp
SocialPreviewType
trxGCofgYrWvWsGpj
<p><i>A fun index of the concepts from The Sequences in the form of several litanies. These litanies are neither a guide nor an introduction to The Way of Rationality because rationality is an </i><a href="https://www.lesswrong.com/posts/teaxCFgtmCQ3E9fy8/the-martial-art-of-rationality"><i>art</i></a><i>&nbsp;not a series of precipes. Rather, these litanies serve as a tool; for meditating on, for connecting ideas; or for preparing for </i><a href="https://www.lesswrong.com/s/5bZZZJ5psXrrD5BGb/p/kXAb5riiaJNrfR8v8"><i>The Ritual Of Changing One’s Mind</i></a><i>. For some this may be a useful exercise, for others not. If they help you, Enjoy!</i></p><h1 data-internal-id="h.k4oy3jrmqzf3">Litanies Of The Way</h1><p>Litany On Litanies</p><p>Litany Against Bias</p><p>Litany For Communication</p><p>Litany Against Politics</p><p>Litany For Inquiry</p><p>Litany For Letting Go</p><p>Litany For Belief</p><p>Litany Against The Dark</p><p>Litany For Seeing</p><p>Litany For Curiosity</p><p>Litany For Minds</p><h1 data-internal-id="Litany_On_Litanies">Litany On Litanies</h1><p><i>I will use these litanies to grow </i><a href="https://www.lesswrong.com/posts/DoLQN5ryZ9XkZjq5h/tsuyoku-naritai-i-want-to-become-stronger"><i>stronger</i></a></p><p><i>And when they are no longer useful I will not </i><a href="https://www.lesswrong.com/tag/litany-of-gendlin"><i>mourn</i></a></p><p><i>For although an apprentice must first imitate The Way that is known</i></p><p><i>A master must discover The Way that is to come</i></p><h1 data-internal-id="Litany_Against_Bias">Litany Against Bias</h1><p><i>When I learn about a bias</i><br><i>I will not use it as a </i><a href="https://www.lesswrong.com/tag/fully-general-counterargument"><i>Fully General Counterargument</i></a></p><p><i>And when I teach others of bias</i></p><p><i>I will </i><a href="https://www.lesswrong.com/s/GSqFqc646rsRd2oyz/p/AdYdLP2sRqPMoe8fb"><i>first do no harm</i></a></p><p><i>When new evidence comes in</i></p><p><i>I will </i><a href="https://www.lesswrong.com/posts/Yq6aA4M3JKWaQepPJ/burdensome-details"><i>feel the burden of every detail</i></a></p><p><i>When I’m considering my plans</i></p><p><i>I will </i><a href="https://www.lesswrong.com/posts/CPm5LTwHrvBJCa9h5/planning-fallacy"><i>view them from the outside</i></a><br>&nbsp;</p><p><i>When I want to be an effective altruist</i></p><p><i>I will not </i><a href="https://www.lesswrong.com/lw/hw/scope_insensitivity/"><i>purchase satisfaction</i></a></p><p><i>When the world is chaotic</i><br><i>I will </i><a href="https://www.lesswrong.com/posts/msJA6B9ZjiiZxT6EZ/lawful-uncertainty"><i>act lawfully</i></a><i>&nbsp;</i></p><p><i>When there exist justifications beyond those given</i></p><p><i>I will </i><a href="https://www.lesswrong.com/s/pmHZDpak4NeRLLLCw/p/KZLa74SzyKhSJ3M55"><i>reject the genetic heuristic</i></a></p><p><i>When I am presented with a false dilemma</i></p><p><i>I will </i><a href="https://www.lesswrong.com/posts/erGipespbbzdG5zYb/the-third-alternative"><i>look for a Third Alternative</i></a></p><p><i>And when I wish to overcome bias</i><br><i>I will </i><a href="https://www.lesswrong.com/s/3ELrPerFTSo75WnrH/p/i8q4vXestDkGTFwsc"><i>know it is a worthy goal</i></a></p><h1 data-internal-id="Litany_For_Communication">Litany For Communication</h1><p><i>When I feel I am misunderstood</i><br><i>I will </i><a href="https://www.lesswrong.com/s/zpCiuR4T343j9WkcK/p/sSqoEw9eRP2kPKLCz"><i>be aware that my intent might not be transparent</i></a></p><p><i>When I need to communicate something complex</i></p><p><i>I will </i><a href="https://www.lesswrong.com/s/zpCiuR4T343j9WkcK/p/HLqWn5LASfhhArZ7w"><i>lay out an inferential pathway</i></a><br>&nbsp;</p><p><i>When making propositions</i><br><i>I will not use </i><a href="https://www.lesswrong.com/s/GSqFqc646rsRd2oyz/p/bfbiyTogEKWEGP96S"><i>Fake Justifications</i></a></p><p><i>When I am attacking propositions&nbsp;</i><br><i>I will give my </i><a href="https://www.lesswrong.com/s/GSqFqc646rsRd2oyz/p/TGux5Fhcd7GmTfNGC"><i>True Rejection</i></a></p><h1 data-internal-id="Litany_Against_Politics">Litany Against Politics</h1><p><a href="https://www.lesswrong.com/posts/9weLK2AJ9JEt2Tt8f/politics-is-the-mind-killer"><i>Politics is the Mind-Killer</i></a></p><p><i>When observing a debate</i><br><i>I will </i><a href="https://www.lesswrong.com/posts/PeSzc9JTBxhaYRp9b/policy-debates-should-not-appear-one-sided"><i>not treat arguments like soldiers</i></a></p><p><i>When I am dealing with a non-binary question</i><br><i>I will </i><a href="https://www.lesswrong.com/posts/XYCEB9roxEBfgjfxs/the-scales-of-justice-the-notebook-of-rationality"><i>relinquish my scales and take up a notebook</i></a></p><p><i>When the problem is not just black and white</i></p><p><i>I will </i><a href="https://www.lesswrong.com/s/FrqfoG3LJeCZs96Ym/p/dLJv2CoRCgeC2mPgj"><i>distinguish shades of grey</i></a><br><i>When I see someone kicking a vending machine</i></p><p><i>I will </i><a href="https://www.lesswrong.com/tag/correspondence-bias"><i>not hypothesize that they are a mutant</i></a></p><h1 data-internal-id="Litany_For_Inquiry">Litany For Inquiry</h1><p><i>When I ponder</i><br><i>I will </i><a href="https://www.lesswrong.com/s/3ELrPerFTSo75WnrH/p/Lz64L3yJEtYGkzMzu"><i>let the meaning choose my words</i></a></p><p><i>When I’m forming a hypothesis</i></p><p><i>I will </i><a href="https://www.lesswrong.com/posts/rmAbiEKQDpDnZzcRf/positive-bias-look-into-the-dark"><i>look into the dark</i></a><i>&nbsp;</i></p><p><i>When I question</i><br><i>I will </i><a href="https://www.lesswrong.com/s/3ELrPerFTSo75WnrH/p/2jp98zdLo898qExrr"><i>Hug the Query</i></a></p><p><i>When I attack my b</i>... </p>
A fun index of the concepts from The Sequences in the form of several litanies. These litanies are neither a guide nor an introduction to The Way of Rationality because rationality is an art not a series of precipes. Rather, these litanies serve as a tool; for meditating on, for connecting ideas; or for preparing for The Ritual Of Changing One’s Mind. For some this may be a useful exercise, for others not. If they help you, Enjoy! Litanies Of The Way Litany On Litanies Litany Against Bias Litany For Communication Litany Against Politics Litany For Inquiry Litany For Letting Go Litany For Belief Litany Against The Dark Litany For Seeing Litany For Curiosity Litany For Minds Litany On Litanies I will use these litanies to grow stronger And when they are no longer useful I will not mourn For although an apprentice must first imitate The Way that is known A master must discover The Way that is to come Litany Against Bias When I learn about a bias I will not use it as a Fully General Counterargument And when I teach others of bias I will first do no harm When new evidence comes in I will feel the burden of every detail When I’m considering my plans I will view them from the outside   When I want to be an effective altruist I will not purchase satisfaction When the world is chaotic I will act lawfully  When there exist justifications beyond those given I will reject the genetic heuristic When I am presented with a false dilemma I will look for a Third Alternative And when I wish to overcome bias I will know it is a worthy goal Litany For Communication When I feel I am misunderstood I will be aware that my intent might not be transparent When I need to communicate something complex I will lay out an inferential pathway   When making propositions I will not use Fake Justifications When I am attacking propositions  I will give my True Rejection Litany Against Politics Politics is the Mind-Killer When observing a debate I will not tre
1,461
1.2.0
Revision
false
null
null
CrosspostOutput
gKxFYArW9S2XtgoeY
make-data-pipelines-debuggable-by-storing-all-source
Make Data Pipelines Debuggable by Storing All Source References
null
false
false
false
null
piR3ZKGHEp6vqTo87
null
true
false
false
false
Post
https://www.brendanlong.com/make-data-pipelines-debuggable-by-storing-all-source-references.html
2025-06-08T04:16:36.149Z
null
false
false
2
2
null
false
false
linkpost
[]
null
null
QfxeSG96m6QLWg4wx
0
3
7
false
0.005677
null
false
false
2025-06-08T04:16:36.149Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
55XxDBpfKkkBPm9H8
null
null
null
false
null
[]
null
1
0
2025-06-08T04:13:19.193Z
false
false
easy-going
null
false
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
4
null
null
null
null
[ { "__typename": "Tag", "_id": "HFou6RHqFagkyrKkW", "adminOnly": false, "afBaseScore": null, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 0, "canEditUserIds": null, "core": false, "createdAt": "2020-05-22T21:10:05.579Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Programming", "needsReview": false, "noindex": false, "postCount": 179, "score": 0, "shortName": null, "slug": "programming", "suggestedAsFilter": false, "userId": "nrP5EZZj4vRvYwQ7b", "voteCount": 0, "wikiOnly": false }, { "__typename": "Tag", "_id": "3uE2pXvbcnS9nnZRE", "adminOnly": false, "afBaseScore": 0, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 2, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:50.898Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 27, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "zJtgSyKntXrnkArbY", "displayName": "kistune" }, { "_id": "XqyQTGFGoKfZtdqN3", "displayName": "kINo" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "World Modeling", "needsReview": false, "noindex": false, "postCount": 5855, "score": 2, "shortName": null, "slug": "world-modeling", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 2, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
3
0
0
1
0
piR3ZKGHEp6vqTo87
brendan-long
2009-10-28T00:51:27.668Z
korin43
Brendan Long
null
null
Brendan Long
2,311
0
false
false
null
null
21
636
0
0
0
1
0
r38pkCm7wF4M44MDQ
User
easy-going
null
true
[ "alignmentVoters", "canModeratePersonal", "trustLevel1" ]
null
null
gKxFYArW9S2XtgoeY
SocialPreviewType
QfxeSG96m6QLWg4wx
<p>A few jobs ago, I worked at company that collected data from disparate sources, then processed and deduplicated it into spreadsheets for ingestion by the data science and customer support teams. Some common questions the engineering team got were:</p><ul><li>Why is the data in some input CSV missing in the output?</li><li>Why is data in the output CSV not matching what we expect?</li></ul><p>To debug these problems, the process was to try to reverse engineer where the data came from, then try to guess which path that data took through the monolithic data processor.</p><p>This is the story of how we stopped doing that, and started storing references to all source data for every piece of output data.</p><p>(This is reconstructed from memory so I no longer claim 100% accuracy)</p><h2>Get Source Data Into Your Database</h2><p>The good news is that before I even started at this company, it was well understood that we needed to keep our data somewhere durable, since we might need to reprocess it one day. This involved putting all of the data in an S3 bucket, organized by source and date.</p><p>In order to tie our outputs to our inputs, we needed to first get our inputs into the database. We previously had some intermediate formats, but we wanted to know <i>exactly</i> where our data was coming from. So we added a source table:</p><h3>csv_source</h3><figure class="table"><table><thead><tr><th>id</th><th>filename</th></tr></thead><tbody><tr><td>1</td><td>client-x-2024-07-01.csv</td></tr><tr><td>2</td><td>client-x-2024-08-01.csv</td></tr></tbody></table></figure><h3>csv_data</h3><figure class="table"><table><thead><tr><th>id</th><th>csv_source_id</th><th>first_name</th><th>last_name</th><th>job</th></tr></thead><tbody><tr><td>10</td><td>1</td><td>Brendan</td><td>Long</td><td>Basket Weaver</td></tr><tr><td>11</td><td>2</td><td>Brendan</td><td>Long</td><td>Senior Basket Weaver</td></tr><tr><td>12</td><td>2</td><td>Example</td><td>McExampleton</td><td>Basket Weaver</td></tr></tbody></table></figure><p>(Ok, so we didn't literally upload the CSV and only uploaded it after determining the standardized column names, <a href="https://github.com/brendanlong/any_columns">sort-of like this</a>)</p><h2>Tie Your Output Data to the Source Data</h2><p>Now that we had our source data, so the next step was to tie our output data to it, so we could start to answer questions like "Where did this Brendan Long guy come from?" and "Where did this Basket Weaver job come from?". Our output data frequently needed to deduplicate data from multiple sources, and the obvious choice here would be to link to the source of the winning data, but for debugging, <strong>we don't just want to know about the winning data</strong>, we want to know about all of it.</p><p>So we added <i>all</i> all of the sources for a piece of data. We did this with join tables, but the examples will show them as inline arrays to keep this readable.</p><h3>jobs</h3><figure class="table"><table><thead><tr><th>id</th><th>csv_data_ids</th><th>name</th></tr></thead><tbody><tr><td>20</td><td>{10,12}</td><td>Basket Weaver</td></tr><tr><td>21</td><td>{11}</td><td>Senior Basket Weaver</td></tr></tbody></table></figure><h3>employees</h3><figure class="table"><table><thead><tr><th>id</th><th>csv_data_ids</th><th>first_name</th><th>last_name</th><th>job_id</th></tr></thead></table></figure>...
A few jobs ago, I worked at company that collected data from disparate sources, then processed and deduplicated it into spreadsheets for ingestion by the data science and customer support teams. Some common questions the engineering team got were: * Why is the data in some input CSV missing in the output? * Why is data in the output CSV not matching what we expect? To debug these problems, the process was to try to reverse engineer where the data came from, then try to guess which path that data took through the monolithic data processor. This is the story of how we stopped doing that, and started storing references to all source data for every piece of output data. (This is reconstructed from memory so I no longer claim 100% accuracy) Get Source Data Into Your Database The good news is that before I even started at this company, it was well understood that we needed to keep our data somewhere durable, since we might need to reprocess it one day. This involved putting all of the data in an S3 bucket, organized by source and date. In order to tie our outputs to our inputs, we needed to first get our inputs into the database. We previously had some intermediate formats, but we wanted to know exactly where our data was coming from. So we added a source table: csv_source idfilename1client-x-2024-07-01.csv2client-x-2024-08-01.csv csv_data idcsv_source_idfirst_namelast_namejob101BrendanLongBasket Weaver112BrendanLongSenior Basket Weaver122ExampleMcExampletonBasket Weaver (Ok, so we didn't literally upload the CSV and only uploaded it after determining the standardized column names, sort-of like this) Tie Your Output Data to the Source Data Now that we had our source data, so the next step was to tie our output data to it, so we could start to answer questions like "Where did this Brendan Long guy come from?" and "Where did this Basket Weaver job come from?". Our output data frequently needed to deduplicate data from multiple sources, and the obvious choi
1,028
1.2.0
Revision
false
null
null
CrosspostOutput
KvB28n3CPwourTFFW
letting-kids-be-outside
Letting Kids Be Outside
null
false
false
false
null
TtEoCrFeowCGb6rFK
null
true
false
false
false
Post
null
2025-06-08T01:30:19.413Z
null
false
false
2
2
null
false
false
post
[]
null
null
dTyyj3AQFgLGpXHYF
11
22
51
false
0.041282
null
false
false
2025-06-17T22:44:38.638Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
null
null
55XxDBpfKkkBPm9H8
null
null
null
false
null
[]
null
14
0
2025-06-08T01:30:19.413Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
5
null
null
null
null
[ { "__typename": "Tag", "_id": "Q55STnFh6gbSezRuR", "adminOnly": false, "afBaseScore": 3, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 9, "canEditUserIds": null, "core": false, "createdAt": "2020-06-05T00:05:56.237Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Parenting", "needsReview": false, "noindex": false, "postCount": 197, "score": 9, "shortName": null, "slug": "parenting", "suggestedAsFilter": false, "userId": "qxJ28GN72aiJu96iF", "voteCount": 1, "wikiOnly": false }, { "__typename": "Tag", "_id": "fkABsGCJZ6y9qConW", "adminOnly": false, "afBaseScore": 0, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 2, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T06:06:46.947Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "MiuAZvbQcQ7ethgt3", "displayName": "Viktor withaK" }, { "_id": "dRaCtsAWxk7sgirSY", "displayName": "Jordan Morgan" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Practical", "needsReview": false, "noindex": false, "postCount": 3411, "score": 2, "shortName": null, "slug": "practical", "suggestedAsFilter": true, "userId": "oBSWiHjgproTiThmY", "voteCount": 2, "wikiOnly": false } ]
ma5dgL5yFHRxKLZKv
0
0
null
false
null
null
0
22
0
0
11
0
TtEoCrFeowCGb6rFK
jkaufman
2010-11-04T21:42:19.863Z
jkaufman
jefftk
null
null
Jeff Kaufman
21,921
3
false
false
<p>Co-lead (Near-Term Detection) at the Nucleic Acid Observatory in Boston. Speaking for myself unless I say otherwise.</p>
null
null
1,018
2,211
0
0
1
1
2
r38pkCm7wF4M44MDQ
User
null
null
null
[ "trustLevel1", "canModeratePersonal", "alignmentVoters" ]
null
null
KvB28n3CPwourTFFW
https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/KvB28n3CPwourTFFW/sidur1wyspvswalqx03a
SocialPreviewType
dTyyj3AQFgLGpXHYF
<p><span> When our kids were 7 and 5 they started walking home from school alone. We wrote explaining they were ready and giving permission, the school had a few reasonable questions, and that was it. Just kids walking home from the local public school like they have in this neighborhood for generations. </span> </p><p> Online, however, it's common for people to write as if this sort of thing is long gone. Zvi <a href="https://thezvi.substack.com/p/letting-kids-be-kids">captures a common view</a>: </p><p> </p> <blockquote> You want to tell your kids, go out and play, be home by dinner, like your father and his father before him. But if you do, or even if you tell your kids to walk the two blocks to school, eventually a policeman will show up at your house and warn you not to do it again, or worse. And yes, you'll be the right legally, but what are you going to do, risk a long and expensive legal fight? So here we are, and either you supervise your kids all the time or say hello to a lot of screens. </blockquote> <p> His post also references ~eight news stories where a family had trouble with authorities because they let their kid do things that should be ordinary, like walking to a store at age nine. </p><p> It's not just Zvi: parents who would like kids to have more freedom often focus on the risk, with the potential for police or Child Protective Services to get involved. While it's important to understand and mitigate the risks, amplifying the rare stories that go poorly magnifies their chilling effect and undermines the overall effort. </p><p> I showed the quote to our oldest, now 11 and comfortable on her own: "I sincerely doubt that a police officer would get mad at me for walking to school or to the corner store by myself." </p><p> She got to this level of comfort by spending a lot of time out in our <a href="https://en.wikipedia.org/wiki/Somerville,_Massachusetts">walkable kid-friendly neighborhood</a>. Sometimes with us, and increasingly on her own. For example it's raining today and she just came back to the house to tell me that she was grabbing rain gear and then she was going puddle jumping with two younger neighborhood kids. In a bit I'll stop writing and take her younger sister (age 3) out to join in. </p><p> <a href="https://www.jefftk.com/ghiblified-version-of-kids-playing-in-puddle-big.png"><img src="https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/KvB28n3CPwourTFFW/n7jw0mfdlseypp2aqr1t" srcset="https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/KvB28n3CPwourTFFW/n7jw0mfdlseypp2aqr1t 550w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/KvB28n3CPwourTFFW/mcdvrvghttkigejg8myf 1100w"></a></p><div></div> <p></p><p> Some other examples of being out alone: </p><p> </p> <ul> <li><p>Heading to a school concert the 8yo was running late and the 10yo was getting impatient. I asked her: "you know the way, do you want to go on ahead by yourself?" She walked the half mile without issue, with her <a href="https://www.jefftk.com/p/gizmo-watch-review">watch</a> as backup. </p></li> <li><p>Both older kids will go to the corner store to spend </p></li></ul>...
When our kids were 7 and 5 they started walking home from school alone. We wrote explaining they were ready and giving permission, the school had a few reasonable questions, and that was it. Just kids walking home from the local public school like they have in this neighborhood for generations. Online, however, it's common for people to write as if this sort of thing is long gone. Zvi captures a common view: > You want to tell your kids, go out and play, be home by dinner, like your father and his father before him. But if you do, or even if you tell your kids to walk the two blocks to school, eventually a policeman will show up at your house and warn you not to do it again, or worse. And yes, you'll be the right legally, but what are you going to do, risk a long and expensive legal fight? So here we are, and either you supervise your kids all the time or say hello to a lot of screens. His post also references ~eight news stories where a family had trouble with authorities because they let their kid do things that should be ordinary, like walking to a store at age nine. It's not just Zvi: parents who would like kids to have more freedom often focus on the risk, with the potential for police or Child Protective Services to get involved. While it's important to understand and mitigate the risks, amplifying the rare stories that go poorly magnifies their chilling effect and undermines the overall effort. I showed the quote to our oldest, now 11 and comfortable on her own: "I sincerely doubt that a police officer would get mad at me for walking to school or to the corner store by myself." She got to this level of comfort by spending a lot of time out in our walkable kid-friendly neighborhood. Sometimes with us, and increasingly on her own. For example it's raining today and she just came back to the house to tell me that she was grabbing rain gear and then she was going puddle jumping with two younger neighborhood kids. In a bit I'll stop writing and take her yo
1,367
1.0.1
Revision
false
null
null
CrosspostOutput
3rW68feE97iWPrLCm
lessonline-could-use-meeting-stones
LessOnline Could Use Meeting Stones
null
false
false
false
null
piR3ZKGHEp6vqTo87
null
true
false
false
false
Post
null
2025-06-08T01:01:47.624Z
null
false
false
2
2
null
false
false
post
[]
null
null
FTMM99vA3mTG2tjm8
5
14
23
false
0.019049
null
false
false
2025-06-08T20:38:16.374Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
55XxDBpfKkkBPm9H8
null
null
null
false
null
[]
null
7
0
2025-06-08T00:48:48.827Z
false
false
easy-going
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
1
null
null
null
null
[ { "__typename": "Tag", "_id": "izp6eeJJEg9v5zcur", "adminOnly": false, "afBaseScore": null, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 0, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T03:38:34.631Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 15, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Community", "needsReview": false, "noindex": false, "postCount": 2400, "score": 0, "shortName": null, "slug": "community", "suggestedAsFilter": true, "userId": "XtphY3uYHwruKqDyG", "voteCount": 0, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
14
0
0
6
0
piR3ZKGHEp6vqTo87
brendan-long
2009-10-28T00:51:27.668Z
korin43
Brendan Long
null
null
Brendan Long
2,311
0
false
false
null
null
21
636
0
0
0
1
0
r38pkCm7wF4M44MDQ
User
easy-going
null
true
[ "alignmentVoters", "canModeratePersonal", "trustLevel1" ]
null
null
3rW68feE97iWPrLCm
SocialPreviewType
FTMM99vA3mTG2tjm8
<p>Back in the mists of time, there was this game called World of Warcraft, where you and your four closest friends would team up to complete dungeons together. To handle the inevitable people with no friends, they added <a href="https://wowpedia.fandom.com/wiki/Meeting_Stone">meeting stones</a> near each dungeon. When you clicked a meeting stone, it added you to a queue, and once five people were in the queue, you would all be added to a group together.</p><p>I was thinking abut this at LessOnline last week, since I <s>don't have any friends</s> sometimes had trouble finding groups of people to talk to. One way to handle this was to walk up to an existing group and talk to them, but this was tricky since sometimes all of the existing groups were 4+ people already. My strategy ended up being to sit on an empty couch in the busiest areas and wait for people to come to me, but sometimes there were no empty couches in the busy areas.</p><p>This made me wish we had some (virtual) meeting stones:</p><ol><li>On WriteHaven, you click "Looking For Group".</li><li>Three<span class="footnote-reference" data-footnote-reference="" data-footnote-index="1" data-footnote-id="e5bx7fan6zc" role="doc-noteref" id="fnrefe5bx7fan6zc"><sup><a href="#fne5bx7fan6zc">[1]</a></sup></span>&nbsp;additional people click "Looking For Group".</li><li>WriteHaven notifies me that a group has been formed and gives me a randomly-selected location<span class="footnote-reference" data-footnote-reference="" data-footnote-index="2" data-footnote-id="d5bcag5x6en" role="doc-noteref" id="fnrefd5bcag5x6en"><sup><a href="#fnd5bcag5x6en">[2]</a></sup></span>&nbsp;that I've been summoned to.</li></ol><p>This could also potentially be a non-WriteHaven app, and I'm tempted to just write it myself, but it would only really be useful if everyone at a conference was aware of it.</p><p>What do you all think?</p><ol class="footnote-section footnotes" data-footnote-section="" role="doc-endnotes"><li class="footnote-item" data-footnote-item="" data-footnote-index="1" data-footnote-id="e5bx7fan6zc" role="doc-endnote" id="fne5bx7fan6zc"><span class="footnote-back-link" data-footnote-back-link="" data-footnote-id="e5bx7fan6zc"><sup><strong><a href="#fnrefe5bx7fan6zc">^</a></strong></sup></span><div class="footnote-content" data-footnote-content=""><p>The ideal conversation size might be smaller than four, but groups of two can lead to being matched with one other person who you don't actually end up liking, and not having an non-awkward way to extract yourself.</p></div></li><li class="footnote-item" data-footnote-item="" data-footnote-index="2" data-footnote-id="d5bcag5x6en" role="doc-endnote" id="fnd5bcag5x6en"><span class="footnote-back-link" data-footnote-back-link="" data-footnote-id="d5bcag5x6en"><sup><strong><a href="#fnrefd5bcag5x6en">^</a></strong></sup></span><div class="footnote-content" data-footnote-content=""><p>The conference organizers can setup a list of locations to randomly select from, like "The fire table near the Diagonal Hall".</p></div></li></ol>
Back in the mists of time, there was this game called World of Warcraft, where you and your four closest friends would team up to complete dungeons together. To handle the inevitable people with no friends, they added meeting stones near each dungeon. When you clicked a meeting stone, it added you to a queue, and once five people were in the queue, you would all be added to a group together. I was thinking abut this at LessOnline last week, since I don't have any friends sometimes had trouble finding groups of people to talk to. One way to handle this was to walk up to an existing group and talk to them, but this was tricky since sometimes all of the existing groups were 4+ people already. My strategy ended up being to sit on an empty couch in the busiest areas and wait for people to come to me, but sometimes there were no empty couches in the busy areas. This made me wish we had some (virtual) meeting stones: 1. On WriteHaven, you click "Looking For Group". 2. Three[1] additional people click "Looking For Group". 3. WriteHaven notifies me that a group has been formed and gives me a randomly-selected location[2] that I've been summoned to. This could also potentially be a non-WriteHaven app, and I'm tempted to just write it myself, but it would only really be useful if everyone at a conference was aware of it. What do you all think? 1. ^ The ideal conversation size might be smaller than four, but groups of two can lead to being matched with one other person who you don't actually end up liking, and not having an non-awkward way to extract yourself. 2. ^ The conference organizers can setup a list of locations to randomly select from, like "The fire table near the Diagonal Hall".
252
1.1.0
Revision
false
null
null
CrosspostOutput
oyTz2hCAxBMhkENB3
mri-tracers
MRI tracers
null
false
false
false
null
xYpk75i7Hnn6wc5it
null
true
false
false
false
Post
https://www.bhauth.com/blog/biology/mri%20tracers.html
2025-06-07T23:03:33.524Z
null
false
false
2
2
2025-06-08T15:27:30.507Z
false
false
linkpost
[]
null
null
yCeoPtHzJecMJdgER
2
11
28
false
0.030709
null
false
false
2025-06-07T23:51:44.525Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
55XxDBpfKkkBPm9H8
null
null
null
false
null
[]
null
10
0
2025-06-07T23:01:53.648Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
2
null
null
null
null
[ { "__typename": "Tag", "_id": "xHjy88N2uJvGdgzfw", "adminOnly": false, "afBaseScore": 3, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 11, "canEditUserIds": null, "core": false, "createdAt": "2020-07-10T11:55:55.351Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" }, { "_id": "xF5nfdddHjFThHy49", "displayName": "[email protected]" }, { "_id": "go3WWAbwJMPGrGZbH", "displayName": "Carl Leninger" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Health / Medicine / Disease", "needsReview": false, "noindex": false, "postCount": 341, "score": 11, "shortName": null, "slug": "health-medicine-disease", "suggestedAsFilter": false, "userId": "qxJ28GN72aiJu96iF", "voteCount": 3, "wikiOnly": false }, { "__typename": "Tag", "_id": "3uE2pXvbcnS9nnZRE", "adminOnly": false, "afBaseScore": 0, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 2, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:50.898Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 27, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "zJtgSyKntXrnkArbY", "displayName": "kistune" }, { "_id": "XqyQTGFGoKfZtdqN3", "displayName": "kINo" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "World Modeling", "needsReview": false, "noindex": false, "postCount": 5855, "score": 2, "shortName": null, "slug": "world-modeling", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 2, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
11
0
0
9
0
xYpk75i7Hnn6wc5it
bhauth
2023-04-08T11:57:52.463Z
bhauth
bhauth
null
null
null
3,598
6
false
false
<p><a href="https://www.bhauth.com/">bhauth.com</a></p>
null
null
77
421
0
0
0
1
0
r38pkCm7wF4M44MDQ
User
null
null
null
[ "canModeratePersonal", "trustLevel1", "alignmentVoters" ]
null
null
oyTz2hCAxBMhkENB3
SocialPreviewType
yCeoPtHzJecMJdgER
<p>MRI scans and <a href="https://en.wikipedia.org/wiki/Positron_emission_tomography">PET scans</a> are different methods for medical imaging, and they observe different things:</p> <ul> <li> <p>MRI scans can see which nuclei are present in regions. Sometimes bond types of particular elements can be distinguished, eg in <a href="https://en.wikipedia.org/wiki/Proton_nuclear_magnetic_resonance">1H NMR</a> and <a href="https://en.wikipedia.org/wiki/Functional_magnetic_resonance_imaging">fMRI</a>.</p> </li> <li> <p>PET scans add a short-lived radioactive tracer (usually organic molecules with radioactive fluorine added to them) and observe where decays happen. This can track things like movement of glucose in the body. The half-life of fluorine-18 is ~110 minutes. PET scans are often combined with CT or MRI to correlate tracers with locations of organs.</p> </li> </ul> <p>Of course, MRI scans are generally preferably because they don't expose patients or staff to radiation, and don't require short-lived radioactive compounds. That being the case, who among us hasn't asked:</p> <blockquote> <p>Could we use tracer compounds for MRI scans that let MRI do what PET is used for?</p> </blockquote> <p>Historically, such tracers for MRI weren't available, but there are now some interesting options. (Yes, metallic contrast agents (eg gadolinium) have been used with MRI, but that doesn't do what we need here.)</p> <h2>hyperpolarized carbon-13</h2> <p>1.1% of carbon is 13C, which is stable and basically harmless. Hyperpolarized 13C has a fairly strong MRI signal and <a href="https://pmc.ncbi.nlm.nih.gov/articles/PMC10546364/">can be used</a> as a MRI tracer. You can make tracers using that by:</p> <ol> <li>Getting some 13C.</li> <li>Making some relevant molecule, like pyruvate or acetate. Unlike PET tracers, this doesn't need to be done right before they're used.</li> <li>Polarizing the 13C. Basically, you get it very cold, and put it in a strong magnet until it goes all wibbly.</li> </ol> <p>While the compounds are stable, the polarization decays quickly. Pyruvate spin lattice relaxation time (T1) is ~50-70 seconds ex vivo and ~20-30 seconds in vivo. Sufficient signal for detection is present for ~5x the T1. In that time, the tracer needs to be warmed up, injected, and detected by a scan.</p><p>So, you need a strong magnet, and special equipment for quickly processing the tracer during a MRI scan. Few hospitals are currently set up to do this, but technologically it's easier than the MRI scan itself.</p> <h2>nitroxide radicals</h2> <p>An obvious idea for MRI tracers is to use a metal contrast agent (eg gadolinium) in some organic complex attached to an antibody. That didn't work very well. <a href="https://journals.lww.com/investigativeradiology/abstract/1985/10000/magnetic_resonance_imaging_using_gadolinium.8.aspx">This 1985 paper</a> notes:</p> <blockquote> <p>For monoclonal antibodies to function as selective MR contrast agents, substantial advances in technology must occur.</p> </blockquote> <p>Peo... </p>
MRI scans and PET scans are different methods for medical imaging, and they observe different things: * MRI scans can see which nuclei are present in regions. Sometimes bond types of particular elements can be distinguished, eg in 1H NMR and fMRI. * PET scans add a short-lived radioactive tracer (usually organic molecules with radioactive fluorine added to them) and observe where decays happen. This can track things like movement of glucose in the body. The half-life of fluorine-18 is ~110 minutes. PET scans are often combined with CT or MRI to correlate tracers with locations of organs. Of course, MRI scans are generally preferably because they don't expose patients or staff to radiation, and don't require short-lived radioactive compounds. That being the case, who among us hasn't asked: > Could we use tracer compounds for MRI scans that let MRI do what PET is used for? Historically, such tracers for MRI weren't available, but there are now some interesting options. (Yes, metallic contrast agents (eg gadolinium) have been used with MRI, but that doesn't do what we need here.) hyperpolarized carbon-13 1.1% of carbon is 13C, which is stable and basically harmless. Hyperpolarized 13C has a fairly strong MRI signal and can be used as a MRI tracer. You can make tracers using that by: 1. Getting some 13C. 2. Making some relevant molecule, like pyruvate or acetate. Unlike PET tracers, this doesn't need to be done right before they're used. 3. Polarizing the 13C. Basically, you get it very cold, and put it in a strong magnet until it goes all wibbly. While the compounds are stable, the polarization decays quickly. Pyruvate spin lattice relaxation time (T1) is ~50-70 seconds ex vivo and ~20-30 seconds in vivo. Sufficient signal for detection is present for ~5x the T1. In that time, the tracer needs to be warmed up, injected, and detected by a scan. So, you need a strong magnet, and special equipment for quickly processing the tracer during a MRI scan. Few hos
589
1.4.0
Revision
false
null
null
CrosspostOutput
FKG47ofzEkLkeNJLq
second-order-taste
Second order taste
null
false
false
false
null
6jLdWqegNefgaabhr
null
true
false
false
false
Post
null
2025-06-07T20:26:09.379Z
null
false
false
2
2
2025-06-08T15:27:44.633Z
false
false
post
[]
null
null
rjrsEKX4iEsirB23d
3
9
8
false
0.014747
null
false
false
2025-06-08T21:18:07.146Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
55XxDBpfKkkBPm9H8
null
null
null
false
null
[]
null
5
0
2025-06-06T06:45:04.441Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
4
null
null
null
null
[ { "__typename": "Tag", "_id": "fkABsGCJZ6y9qConW", "adminOnly": false, "afBaseScore": 0, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 2, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T06:06:46.947Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "MiuAZvbQcQ7ethgt3", "displayName": "Viktor withaK" }, { "_id": "dRaCtsAWxk7sgirSY", "displayName": "Jordan Morgan" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Practical", "needsReview": false, "noindex": false, "postCount": 3411, "score": 2, "shortName": null, "slug": "practical", "suggestedAsFilter": true, "userId": "oBSWiHjgproTiThmY", "voteCount": 2, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
9
0
0
5
0
6jLdWqegNefgaabhr
adamzerner
2013-08-12T18:18:47.957Z
adamzerner
Adam Zerner
null
null
Adam Zerner
9,081
0
false
false
<p><a href="https://adamzerner.bearblog.dev">https://adamzerner.bearblog.dev</a></p>
null
null
179
2,186
2
0
0
1
7
r38pkCm7wF4M44MDQ
User
null
[ "mvf4xdfcGzPN8PsXM" ]
null
[ "alignmentVoters", "canModeratePersonal", "trustLevel1" ]
null
null
FKG47ofzEkLkeNJLq
SocialPreviewType
rjrsEKX4iEsirB23d
<p>Last year I got my friend a gift card to <a href="https://beastandcleaver.com/">Beast and Cleaver</a> for his wedding, a local butcher shop and restaurant.</p><p>I was pretty proud of the gift. He and his wife said the meal was amazing and I wouldn't expect them to say that if it weren't true. Getting something on the wedding registry or perhaps a gift card would have been safer, but I don't think it would have yielded as much joy.</p><p>I can't take too much credit though. As a <a href="https://www.yelp.com/user_details?userid=dnsCCmwYrONjUAivm3XR4w">Yelp Elite</a> food critic with over a decade of experience churning out high quality restaurant reviews, I like to think I have pretty good taste in food. But this time it wasn't me. It was Kenji.</p><p>I was watching a <a href="https://youtu.be/iInHnbGbsNQ?si=dVvhSmogpuDWIAiI&amp;t=12">YouTube video</a> by Kenji Lopez-Alt, a food writer who actually does have good taste in food. In his video he mentions Beast and Cleaver and says<span class="footnote-reference" data-footnote-reference="" data-footnote-index="1" data-footnote-id="6kd8v1wxp1m" role="doc-noteref" id="fnref6kd8v1wxp1m"><sup><a href="#fn6kd8v1wxp1m">[1]</a></sup></span>&nbsp;that it's a very good butcher shop. I've read, watched and listened to a ton of Kenji's content over the years and think very highly of him.</p><p>So really what's happening here is something I'll call second order taste. First order taste would be me eating at Beast and Cleaver and determining it to be good. Second order taste is me identifying Kenji as someone with good first order taste.</p><p>I think that having good second order taste is pretty important. And I'm talking about taste more generally, not just as it relates to food. The world is complex. We can't evaluate everything ourselves. We don't have the time or the skills. For better or for worse, we often have to figure out who the trustworthy people are and listen to them.</p><p>Over the years I've identified a handful of people in various domains who I respect and see as highly trustworthy. Paul Ingraham for pain. Tori Olds for therapy. Ben Felix for personal finance. Ben Taylor for basketball. Demand Curve for marketing. Peter Attia for health. I think all of these people are quite smart and really know their field.</p><p>And then there's websites like <a href="https://thebestbikelock.com/">The Best Bike Lock</a>. I bought a moderately expensive e-bike recently. I store it in my apartment complex's bike room but historically there's been a fair amount of theft in that room, including my own electric scooter. I was dumb and only used a cable lock to secure the scooter, but there are also two U-locks that have been cut through and left sitting there in the bike room as well, so the theft is clearly going beyond low hanging fruit like cable locks.</p><p>Anyway, I wanted to do some research into how I ... </p>
Last year I got my friend a gift card to Beast and Cleaver for his wedding, a local butcher shop and restaurant. I was pretty proud of the gift. He and his wife said the meal was amazing and I wouldn't expect them to say that if it weren't true. Getting something on the wedding registry or perhaps a gift card would have been safer, but I don't think it would have yielded as much joy. I can't take too much credit though. As a Yelp Elite food critic with over a decade of experience churning out high quality restaurant reviews, I like to think I have pretty good taste in food. But this time it wasn't me. It was Kenji. I was watching a YouTube video by Kenji Lopez-Alt, a food writer who actually does have good taste in food. In his video he mentions Beast and Cleaver and says[1] that it's a very good butcher shop. I've read, watched and listened to a ton of Kenji's content over the years and think very highly of him. So really what's happening here is something I'll call second order taste. First order taste would be me eating at Beast and Cleaver and determining it to be good. Second order taste is me identifying Kenji as someone with good first order taste. I think that having good second order taste is pretty important. And I'm talking about taste more generally, not just as it relates to food. The world is complex. We can't evaluate everything ourselves. We don't have the time or the skills. For better or for worse, we often have to figure out who the trustworthy people are and listen to them. Over the years I've identified a handful of people in various domains who I respect and see as highly trustworthy. Paul Ingraham for pain. Tori Olds for therapy. Ben Felix for personal finance. Ben Taylor for basketball. Demand Curve for marketing. Peter Attia for health. I think all of these people are quite smart and really know their field. And then there's websites like The Best Bike Lock. I bought a moderately expensive e-bike recently. I store it in my apartment c
1,078
1.3.0
Revision
false
null
null
CrosspostOutput
aKL6K5g9xQigEzvPy
dimensionalizing-forecast-value
Dimensionalizing Forecast Value
null
false
false
false
null
skbL8Z4ypRPCQdHxf
null
true
false
false
false
Post
2025-06-07T18:45:26.923Z
null
false
false
2
2
2025-06-08T15:27:47.372Z
false
false
post
[]
null
null
2z58DFTCX3JWhkcZK
0
2
5
false
0.012193
null
false
false
2025-06-07T18:45:26.923Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
55XxDBpfKkkBPm9H8
null
null
null
false
null
[]
null
0
0
2025-06-07T18:40:22.257Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
8
null
null
null
null
[ { "__typename": "Tag", "_id": "Ng8Gice9KNkncxqcj", "adminOnly": false, "afBaseScore": 0, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 1, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:17.072Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 100, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "iMqytjy9ns89Fzfyv", "displayName": "miakko" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Rationality", "needsReview": false, "noindex": false, "postCount": 4302, "score": 1, "shortName": null, "slug": "rationality", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 1, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
2
0
0
0
0
skbL8Z4ypRPCQdHxf
jordan-rubin
2025-05-28T18:26:43.705Z
jordan-rubin
Jordan Rubin
null
null
null
14
0
false
false
<p>Researcher-Operator currently on garden leave. Formerly: Two Sigma (Quant Research + Mgmt) / OnDeck (Data science in lending) / BlackRock (Bond desk quant). I hope my thinking can be helpful to you!</p><p>My Substack: <a href="https://jordanmrubin.substack.com">https://jordanmrubin.substack.com</a></p><p>My LinkedIn: <a href="https://www.linkedin.com/in/jordanmrubin/">https://www.linkedin.com/in/jordanmrubin/</a></p>
null
null
4
3
0
0
0
0.9
0
grecHJcgkb3KW5wnM
User
null
null
null
null
null
null
aKL6K5g9xQigEzvPy
https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/aKL6K5g9xQigEzvPy/oeiair3hsy169any4zqg
SocialPreviewType
2z58DFTCX3JWhkcZK
<h1>tl;dr</h1><p>In quantitative finance, the value of forecasting is obvious. No decisions are made without forecasts, and every forecast is created for some decision.</p><p>On <a href="https://www.metaculus.com/"><u>Metaculus</u></a><a href="https://jordanmrubin.substack.com/p/dimensionalizing-forecast-value#footnote-1-163741438"><sup><u>1</u></sup></a>, the signal is just as sharp, but the market for the signal is fuzzier. The gap isn’t quality: it’s that the value proposition is implicit.</p><p>This post is my <a href="https://www.lesswrong.com/posts/LSFiKt4zGxXcX2oxi/dimensionalization">dimensionalization</a> of forecast value: an<strong> identification of the unique factors that determine the importance of answering a question.</strong> These factors are Clarity, Leverage, and Efficiency (CLE).</p><p>I use CLE to decide when an answer will be valuable, before I spend time asking or answering the question.</p><p>To find value, I focus on questions where:</p><ul><li>I might change my mind (Clarity)</li><li>A lot is at stake (Leverage)</li><li>Effort compounds over time (Efficiency)</li></ul><hr><p>&nbsp;</p><h1>Decisions Create Questions</h1><p>Forecasts are useful for <i>making decisions under uncertainty</i>. This is a specific context. It is not always the relevant context.</p><p>Much existing content on forecasting assumes a question worth asking, and focuses on methodology improvements. Predicting the future is an unsolved, important problem; it’s good that smart people are working on these improvements.</p><p>But suppose I can get a decent forecast answer to any question about the future. <strong>What questions should I actually try to answer?</strong></p><p>I don’t need a forecast if:</p><ul><li>I am not making a decision</li><li>I already know what choice I want to make</li><li>It doesn’t matter what I decide</li></ul><p>Even in a decision-making context, forecasts are also less helpful when:</p><ul><li>I can’t observe the outcome of my decision</li><li>I can’t trust the forecaster</li><li>I have no time to decide</li></ul><p>Forecasts are more helpful when:</p><ul><li>The event repeats</li><li>Mistakes are highly visible</li><li>Catastrophic outcomes are possible</li><li>The decision matters</li><li>I don’t know what to do</li></ul><p>How do I take this mess of factors and figure out what questions are worth answering?</p><hr><h1>Dimensionalizing Forecast Value: CLE</h1><p>This section is a proposed framework for <i>prioritizing</i> forecasts.</p><p>There are many sources of uncertainty, and many decisions to make. How do I know what questions I want answered?</p><p>I will focus on three sources of value from forecasts:</p><ol><li><strong>Clarity</strong>: Forecasting can help me make <strong>a decision when I don’t know what to do</strong>.</li><li><strong>Leverage</strong>: Forecasting can help me make <strong>the best decision</strong> <strong>when it matters most</strong>.</li><li><strong>Efficiency</strong>: Forecasting now can help me make even <strong>better decisions next time</strong>.</li></ol><p>These are the core forecasting value drivers<a href="https://jordanmrubin.substack.com/p/dimensionalizing-forecast-value#footnote-2-163741438"><sup><u>2</u></sup></a>. They are intended to be roughly orthogonal, a... <style>.mjx-chtml {display: inline-block; line-height: 0; text-indent: 0; text-align: left; text-transform: none; font-style: normal; font-weight: normal; font-size: 100%; font-size-adjust: none; letter-spacing: normal; word-wrap: normal; word-spacing: normal; white-space: nowrap; float: none; direction: ltr; max-width: none; max-height: none; min-width: 0; min-height: 0; border: 0; margin: 0; padding: 1px 0} .MJXc-display {display: block; text-align: center; margin: 1em 0; padding: 0} .mjx-chtml[tabindex]:focus, body :focus .mjx-chtml[tabindex] {display: inline-table} .mjx-full-width {text-align: center; display: table-cell!important; width: 10000em} .mjx-math {display: inline-block; border-collapse: separate; border-spacing: 0} .mjx-math * {display: inline-block; -webkit-box-sizing: content-box!important; -moz-box-sizing: content-box!important; box-sizing: content-box!important; text-align: left} .mjx-numerator {display: block; text-align: center} .mjx-denominator {display: block; text-align: center} .MJXc-stacked {height: 0; position: relative} .MJXc-stacked > * {position: absolute} .MJXc-bevelled > * {display: inline-block} .mjx-stack {display: inline-block} .mjx-op {display: block} .mjx-under {display: table-cell} .mjx-over {display: block} .mjx-over > * {padding-left: 0px!important; padding-right: 0px!important} .mjx-under > * {padding-left: 0px!important; padding-right: 0px!important} .mjx-stack > .mjx-sup {display: block} .mjx-stack > .mjx-sub {display: block} .mjx-prestack > .mjx-presup {display: block} .mjx-prestack > .mjx-presub {display: block} .mjx-delim-h > .mjx-char {display: inline-block} .mjx-surd {vertical-align: top} .mjx-surd + .mjx-box {display: inline-flex} .mjx-mphantom * {visibility: hidden} .mjx-merror {background-color: #FFFF88; color: #CC0000; border: 1px solid #CC0000; padding: 2px 3px; font-style: normal; font-size: 90%} .mjx-annotation-xml {line-height: normal} .mjx-menclose > svg {fill: none; stroke: currentColor; overflow: visible} .mjx-mtr {display: table-row} .mjx-mlabeledtr {display: table-row} .mjx-mtd {display: table-cell; text-align: center} .mjx-label {display: table-row} .mjx-box {display: inline-block} .mjx-block {display: block} .mjx-span {display: inline} .mjx-char {display: block; white-space: pre} .mjx-itable {display: inline-table; width: auto} .mjx-row {display: table-row} .mjx-cell {display: table-cell} .mjx-table {display: table; width: 100%} .mjx-line {display: block; height: 0} .mjx-strut {width: 0; padding-top: 1em} .mjx-vsize {width: 0} .MJXc-space1 {margin-left: .167em} .MJXc-space2 {margin-left: .222em} .MJXc-space3 {margin-left: .278em} .mjx-test.mjx-test-display {display: table!important} .mjx-test.mjx-test-inline {display: inline!important; margin-right: -1px} .mjx-test.mjx-test-default {display: block!important; clear: both} .mjx-ex-box {display: inline-block!important; position: absolute; overflow: hidden; min-height: 0; max-height: none; padding: 0; border: 0; margin: 0; width: 1px; height: 60ex} .mjx-test-inline .mjx-left-box {display: inline-block; width: 0; float: left} .mjx-test-inline .mjx-right-box {display: inline-block; width: 0; float: right} .mjx-test-display .mjx-right-box {display: table-cell!important; width: 10000em!important; min-width: 0; max-width: none; padding: 0; border: 0; margin: 0} .MJXc-TeX-unknown-R {font-family: monospace; font-style: normal; font-weight: normal} .MJXc-TeX-unknown-I {font-family: monospace; font-style: italic; font-weight: normal} .MJXc-TeX-unknown-B {font-family: monospace; font-style: normal; font-weight: bold} .MJXc-TeX-unknown-BI {font-family: monospace; font-style: italic; font-weight: bold} .MJXc-TeX-ams-R {font-family: MJXc-TeX-ams-R,MJXc-TeX-ams-Rw} .MJXc-TeX-cal-B {font-family: MJXc-TeX-cal-B,MJXc-TeX-cal-Bx,MJXc-TeX-cal-Bw} .MJXc-TeX-frak-R {font-family: MJXc-TeX-frak-R,MJXc-TeX-frak-Rw} .MJXc-TeX-frak-B {font-family: MJXc-TeX-frak-B,MJXc-TeX-frak-Bx,MJXc-TeX-frak-Bw} .MJXc-TeX-math-BI {font-family: MJXc-TeX-math-BI,MJXc-TeX-math-BIx,MJXc-TeX-math-BIw} .MJXc-TeX-sans-R {font-family: MJXc-TeX-sans-R,MJXc-TeX-sans-Rw} .MJXc-TeX-sans-B {font-family: MJXc-TeX-sans-B,MJXc-TeX-sans-Bx,MJXc-TeX-sans-Bw} .MJXc-TeX-sans-I {font-family: MJXc-TeX-sans-I,MJXc-TeX-sans-Ix,MJXc-TeX-sans-Iw} .MJXc-TeX-script-R {font-family: MJXc-TeX-script-R,MJXc-TeX-script-Rw} .MJXc-TeX-type-R {font-family: MJXc-TeX-type-R,MJXc-TeX-type-Rw} .MJXc-TeX-cal-R {font-family: MJXc-TeX-cal-R,MJXc-TeX-cal-Rw} .MJXc-TeX-main-B {font-family: MJXc-TeX-main-B,MJXc-TeX-main-Bx,MJXc-TeX-main-Bw} .MJXc-TeX-main-I {font-family: MJXc-TeX-main-I,MJXc-TeX-main-Ix,MJXc-TeX-main-Iw} .MJXc-TeX-main-R {font-family: MJXc-TeX-main-R,MJXc-TeX-main-Rw} .MJXc-TeX-math-I {font-family: MJXc-TeX-math-I,MJXc-TeX-math-Ix,MJXc-TeX-math-Iw} .MJXc-TeX-size1-R {font-family: MJXc-TeX-size1-R,MJXc-TeX-size1-Rw} .MJXc-TeX-size2-R {font-family: MJXc-TeX-size2-R,MJXc-TeX-size2-Rw} .MJXc-TeX-size3-R {font-family: MJXc-TeX-size3-R,MJXc-TeX-size3-Rw} .MJXc-TeX-size4-R {font-family: MJXc-TeX-size4-R,MJXc-TeX-size4-Rw} .MJXc-TeX-vec-R {font-family: MJXc-TeX-vec-R,MJXc-TeX-vec-Rw} .MJXc-TeX-vec-B {font-family: MJXc-TeX-vec-B,MJXc-TeX-vec-Bx,MJXc-TeX-vec-Bw} @font-face {font-family: MJXc-TeX-ams-R; src: local('MathJax_AMS'), local('MathJax_AMS-Regular')} @font-face {font-family: MJXc-TeX-ams-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_AMS-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_AMS-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_AMS-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-cal-B; src: local('MathJax_Caligraphic Bold'), local('MathJax_Caligraphic-Bold')} @font-face {font-family: MJXc-TeX-cal-Bx; src: local('MathJax_Caligraphic'); font-weight: bold} @font-face {font-family: MJXc-TeX-cal-Bw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Caligraphic-Bold.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Caligraphic-Bold.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Caligraphic-Bold.otf') format('opentype')} @font-face {font-family: MJXc-TeX-frak-R; src: local('MathJax_Fraktur'), local('MathJax_Fraktur-Regular')} @font-face {font-family: MJXc-TeX-frak-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Fraktur-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Fraktur-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Fraktur-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-frak-B; src: local('MathJax_Fraktur Bold'), local('MathJax_Fraktur-Bold')} @font-face {font-family: MJXc-TeX-frak-Bx; src: local('MathJax_Fraktur'); font-weight: bold} @font-face {font-family: MJXc-TeX-frak-Bw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Fraktur-Bold.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Fraktur-Bold.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Fraktur-Bold.otf') format('opentype')} @font-face {font-family: MJXc-TeX-math-BI; src: local('MathJax_Math BoldItalic'), local('MathJax_Math-BoldItalic')} @font-face {font-family: MJXc-TeX-math-BIx; src: local('MathJax_Math'); font-weight: bold; font-style: italic} @font-face {font-family: MJXc-TeX-math-BIw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Math-BoldItalic.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Math-BoldItalic.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Math-BoldItalic.otf') format('opentype')} @font-face {font-family: MJXc-TeX-sans-R; src: local('MathJax_SansSerif'), local('MathJax_SansSerif-Regular')} @font-face {font-family: MJXc-TeX-sans-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_SansSerif-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_SansSerif-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_SansSerif-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-sans-B; src: local('MathJax_SansSerif Bold'), local('MathJax_SansSerif-Bold')} @font-face {font-family: MJXc-TeX-sans-Bx; src: local('MathJax_SansSerif'); font-weight: bold} @font-face {font-family: MJXc-TeX-sans-Bw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_SansSerif-Bold.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_SansSerif-Bold.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_SansSerif-Bold.otf') format('opentype')} @font-face {font-family: MJXc-TeX-sans-I; src: local('MathJax_SansSerif Italic'), local('MathJax_SansSerif-Italic')} @font-face {font-family: MJXc-TeX-sans-Ix; src: local('MathJax_SansSerif'); font-style: italic} @font-face {font-family: MJXc-TeX-sans-Iw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_SansSerif-Italic.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_SansSerif-Italic.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_SansSerif-Italic.otf') format('opentype')} @font-face {font-family: MJXc-TeX-script-R; src: local('MathJax_Script'), local('MathJax_Script-Regular')} @font-face {font-family: MJXc-TeX-script-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Script-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Script-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Script-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-type-R; src: local('MathJax_Typewriter'), local('MathJax_Typewriter-Regular')} @font-face {font-family: MJXc-TeX-type-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Typewriter-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Typewriter-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Typewriter-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-cal-R; src: local('MathJax_Caligraphic'), local('MathJax_Caligraphic-Regular')} @font-face {font-family: MJXc-TeX-cal-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Caligraphic-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Caligraphic-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Caligraphic-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-main-B; src: local('MathJax_Main Bold'), local('MathJax_Main-Bold')} @font-face {font-family: MJXc-TeX-main-Bx; src: local('MathJax_Main'); font-weight: bold} @font-face {font-family: MJXc-TeX-main-Bw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Main-Bold.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Main-Bold.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Main-Bold.otf') format('opentype')} @font-face {font-family: MJXc-TeX-main-I; src: local('MathJax_Main Italic'), local('MathJax_Main-Italic')} @font-face {font-family: MJXc-TeX-main-Ix; src: local('MathJax_Main'); font-style: italic} @font-face {font-family: MJXc-TeX-main-Iw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Main-Italic.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Main-Italic.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Main-Italic.otf') format('opentype')} @font-face {font-family: MJXc-TeX-main-R; src: local('MathJax_Main'), local('MathJax_Main-Regular')} @font-face {font-family: MJXc-TeX-main-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Main-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Main-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Main-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-math-I; src: local('MathJax_Math Italic'), local('MathJax_Math-Italic')} @font-face {font-family: MJXc-TeX-math-Ix; src: local('MathJax_Math'); font-style: italic} @font-face {font-family: MJXc-TeX-math-Iw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Math-Italic.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Math-Italic.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Math-Italic.otf') format('opentype')} @font-face {font-family: MJXc-TeX-size1-R; src: local('MathJax_Size1'), local('MathJax_Size1-Regular')} @font-face {font-family: MJXc-TeX-size1-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Size1-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Size1-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Size1-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-size2-R; src: local('MathJax_Size2'), local('MathJax_Size2-Regular')} @font-face {font-family: MJXc-TeX-size2-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Size2-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Size2-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Size2-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-size3-R; src: local('MathJax_Size3'), local('MathJax_Size3-Regular')} @font-face {font-family: MJXc-TeX-size3-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Size3-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Size3-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Size3-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-size4-R; src: local('MathJax_Size4'), local('MathJax_Size4-Regular')} @font-face {font-family: MJXc-TeX-size4-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Size4-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Size4-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Size4-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-vec-R; src: local('MathJax_Vector'), local('MathJax_Vector-Regular')} @font-face {font-family: MJXc-TeX-vec-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Vector-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Vector-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Vector-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-vec-B; src: local('MathJax_Vector Bold'), local('MathJax_Vector-Bold')} @font-face {font-family: MJXc-TeX-vec-Bx; src: local('MathJax_Vector'); font-weight: bold} @font-face {font-family: MJXc-TeX-vec-Bw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Vector-Bold.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Vector-Bold.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Vector-Bold.otf') format('opentype')} </style></p>
tl;dr In quantitative finance, the value of forecasting is obvious. No decisions are made without forecasts, and every forecast is created for some decision. On Metaculus1, the signal is just as sharp, but the market for the signal is fuzzier. The gap isn’t quality: it’s that the value proposition is implicit. This post is my dimensionalization of forecast value: an identification of the unique factors that determine the importance of answering a question. These factors are Clarity, Leverage, and Efficiency (CLE). I use CLE to decide when an answer will be valuable, before I spend time asking or answering the question. To find value, I focus on questions where: * I might change my mind (Clarity) * A lot is at stake (Leverage) * Effort compounds over time (Efficiency) ----------------------------------------   Decisions Create Questions Forecasts are useful for making decisions under uncertainty. This is a specific context. It is not always the relevant context. Much existing content on forecasting assumes a question worth asking, and focuses on methodology improvements. Predicting the future is an unsolved, important problem; it’s good that smart people are working on these improvements. But suppose I can get a decent forecast answer to any question about the future. What questions should I actually try to answer? I don’t need a forecast if: * I am not making a decision * I already know what choice I want to make * It doesn’t matter what I decide Even in a decision-making context, forecasts are also less helpful when: * I can’t observe the outcome of my decision * I can’t trust the forecaster * I have no time to decide Forecasts are more helpful when: * The event repeats * Mistakes are highly visible * Catastrophic outcomes are possible * The decision matters * I don’t know what to do How do I take this mess of factors and figure out what questions are worth answering? ---------------------------------------- Dimensionalizing Forec
1,879
1.1.1
Revision
false
null
null
CrosspostOutput
KnvHZiAMzjnRLnKym
on-working-80
On working 80%
null
false
false
false
null
NqzkMxhWzMZad5c5g
null
true
false
false
false
Post
https://github.com/adrische/write-ups/blob/main/on-working-80%25.md
2025-06-07T17:58:56.084Z
null
false
false
2
2
2025-06-07T18:03:38.770Z
false
false
linkpost
[]
null
null
ZT2rJgAxneSj9BYRS
6
54
79
false
0.069944
null
false
false
2025-06-09T20:41:00.124Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
qgdGA4ZEyW7zNdK84
null
null
null
false
null
[]
null
17
0
2025-06-07T13:08:45.797Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
3
null
null
null
null
[ { "__typename": "Tag", "_id": "fkABsGCJZ6y9qConW", "adminOnly": false, "afBaseScore": 0, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 2, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T06:06:46.947Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "MiuAZvbQcQ7ethgt3", "displayName": "Viktor withaK" }, { "_id": "dRaCtsAWxk7sgirSY", "displayName": "Jordan Morgan" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Practical", "needsReview": false, "noindex": false, "postCount": 3411, "score": 2, "shortName": null, "slug": "practical", "suggestedAsFilter": true, "userId": "oBSWiHjgproTiThmY", "voteCount": 2, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
54
0
0
14
0
NqzkMxhWzMZad5c5g
adrische
2025-06-07T13:06:04.906Z
adrische
adrische
null
null
null
78
0
false
false
null
null
1
0
0
0
0
1
0
qgdGA4ZEyW7zNdK84
User
null
null
null
[ "canModeratePersonal" ]
null
null
KnvHZiAMzjnRLnKym
SocialPreviewType
ZT2rJgAxneSj9BYRS
<p>A year ago, I decided to reduce my employment level from 100% to 80% and to take Fridays off.</p><p>My main motivation was to have some time for myself: Relax, reduce my stress level from work, have more time for side projects, do more sports, and maybe spend more time with my wife without the kids. This was only partially successful as there were a few things I had not realized beforehand.</p><p>I’d like to describe my experience, and will focus on the effects this change had on my work, financial situation, side projects, and personal life / health.</p><p>To give some background, I’ll start with my personal situation: I have two small kids and spend the mornings, evenings and the weekends with them. During the day I’m working. This schedule does not leave much time for myself. When the kids go to bed in the evenings, I’m usually too tired to do anything meaningful.</p><p>While I worked 80%, I put the kids to childcare during my day off.</p><p><strong>Work</strong> The four days I spend at work have become very fast-paced and concentrated. There is less time for chats and social interactions at work, and I’m more focused on getting stuff done.</p><p>Some activities do not reduce by 20% but have a fixed time requirement per week or month (e.g., weekly team meetings). These will now fall in 4 instead of 5 days, leaving over-proportionally less time to get work done.</p><p>Additionally, this increased density of meetings during the 4 days means there will be fewer long time blocks available to get deep work done.</p><p>Your colleagues need to adjust, e.g., they need to know that you don’t work on Fridays. While you can make this clear by putting a blocker in your calendar, they also need to consider your reduced availability in their planning.</p><p>There is also a risk that management and your colleagues only slowly adjust their expectations to your reduced level of output.</p><p>As a result, I believe it is proportionally harder to be productive at the new employment level, while at the same time your colleagues and management may only slowly adjust.</p><p><strong>Financial situation</strong> The obvious first-order impact of reducing your employment level by 20% is that you will earn 20% less salary.</p><p>However, there are a few less obvious things to consider in this calculation:</p><p>The income tax progression means the top 20% of your income are taxed higher than the remaining 80%, meaning that after deduction of income tax, a reduction of 20% gross salary will translate in... </p>
A year ago, I decided to reduce my employment level from 100% to 80% and to take Fridays off. My main motivation was to have some time for myself: Relax, reduce my stress level from work, have more time for side projects, do more sports, and maybe spend more time with my wife without the kids. This was only partially successful as there were a few things I had not realized beforehand. I’d like to describe my experience, and will focus on the effects this change had on my work, financial situation, side projects, and personal life / health. To give some background, I’ll start with my personal situation: I have two small kids and spend the mornings, evenings and the weekends with them. During the day I’m working. This schedule does not leave much time for myself. When the kids go to bed in the evenings, I’m usually too tired to do anything meaningful. While I worked 80%, I put the kids to childcare during my day off. Work The four days I spend at work have become very fast-paced and concentrated. There is less time for chats and social interactions at work, and I’m more focused on getting stuff done. Some activities do not reduce by 20% but have a fixed time requirement per week or month (e.g., weekly team meetings). These will now fall in 4 instead of 5 days, leaving over-proportionally less time to get work done. Additionally, this increased density of meetings during the 4 days means there will be fewer long time blocks available to get deep work done. Your colleagues need to adjust, e.g., they need to know that you don’t work on Fridays. While you can make this clear by putting a blocker in your calendar, they also need to consider your reduced availability in their planning. There is also a risk that management and your colleagues only slowly adjust their expectations to your reduced level of output. As a result, I believe it is proportionally harder to be productive at the new employment level, while at the same time your colleagues and management may
855
1.2.0
Revision
false
null
null
CrosspostOutput
wcNDv7DEy43dpnumY
meta-alignment-communication-guide
Meta Alignment: Communication Guide
null
false
false
false
null
n7SR7JSZCsRThN64o
null
true
false
false
false
Post
https://dxmrevealed.wordpress.com/2025/06/07/meta-alignment-communication-guide/
2025-06-07T16:09:40.972Z
null
false
false
2
2
2025-06-07T17:51:13.788Z
false
false
linkpost
[]
null
null
NvjT6HmTujQXcSqcc
0
3
13
false
0.018149
null
false
false
2025-06-07T16:09:40.972Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
qgdGA4ZEyW7zNdK84
null
null
null
false
null
[]
null
3
0
2025-06-07T16:08:47.453Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
6
null
null
null
null
[ { "__typename": "Tag", "_id": "D6y2AgYBeHsMYqWC4", "adminOnly": false, "afBaseScore": 3, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 9, "canEditUserIds": null, "core": false, "createdAt": "2022-08-26T02:11:56.686Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI Safety Public Materials", "needsReview": false, "noindex": false, "postCount": 135, "score": 9, "shortName": null, "slug": "ai-safety-public-materials-1", "suggestedAsFilter": false, "userId": "nDpieb7g8huozpx9j", "voteCount": 1, "wikiOnly": false }, { "__typename": "Tag", "_id": "sYm3HiWcfZvrGu3ui", "adminOnly": false, "afBaseScore": 2, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" } ] }, "baseScore": 12, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:22.097Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" }, { "_id": "sof55TPMQaeBaxhsS", "displayName": "tommylees112" }, { "_id": "AayjS8XzcnDKhGdTv", "displayName": "shark" }, { "_id": "HnALuwRdo6k9HLaMt", "displayName": "Alex Firssoff" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI", "needsReview": false, "noindex": false, "postCount": 12544, "score": 12, "shortName": null, "slug": "ai", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 4, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
3
0
0
2
0
n7SR7JSZCsRThN64o
bridgett-kay
2017-11-22T22:27:48.770Z
bridgett-kay
Bridgett Kay
null
null
null
201
0
false
false
<p>In another life I wrote <a href="https://dxmfound.wordpress.com/2023/07/03/an-adventure-in-space-and-time/">Gemini Song</a> and <a href="https://projectdxm944358310.wordpress.com">The Coven</a> series. Now I'm working on the pause.&nbsp;</p>
null
null
6
30
0
0
0
1
0
r38pkCm7wF4M44MDQ
User
null
null
null
[ "canModeratePersonal" ]
null
null
wcNDv7DEy43dpnumY
SocialPreviewType
NvjT6HmTujQXcSqcc
<p>&nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp;&nbsp;<i>I’ve been working on a series of longer articles explaining and justifying the following principles of communicating AI risk to the general public, but in the interest of speed and brevity, I’ve decided to write a simplified guide here. Some of this will be based on articles I’ve already written and released, and some will be based on articles that are still in the works.&nbsp;</i></p><p>&nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp;&nbsp;</p><p>Principle 1- Jargon</p><p>&nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; It’s best to avoid jargon when communicating with the general public. Jargon is not easily accessible for non-experts, and unfamiliar words and phrases could lead to misunderstandings. However, when communicating about new technologies that are largely unprecedented, some jargon may be necessary. When you must use jargon, make sure you immediately follow the term with a simple and concise definition. Don’t stack multiple terms together into the same line, and don’t stack more jargon into your definition.</p><p>&nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; Be consistent with your jargon. Use the same terms consistently, until the public gets used to them. If you notice the public already seems familiar with a term, or experts are using the term often in public communication, continue to use the same term. When you are able, use catchy, memetic terms that carry some emotional weight. An example of this type of term is “the black-box problem” when discussing interpretability, because not only is “black box problem” rather catchy and descriptive, but a “black box” is already associated with frightening events such as plane crashes.&nbsp;</p><p>&nbsp;</p><p>Principle 2- Sound Bites</p><p>&nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; Anticipate that anything you produce- whether written, audio, or audio/visual- will be broken up into short-form content such as tweets or tik-tok videos. Because you can anticipate that your content will be shredded into sound bites, it’s important that you ensure each argument can stand on its own. Nuance is still important, but try to incorporate caveats into a short space following the nuance you’ve introduced.&nbsp;</p><p>&nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; If you cannot make each argument stand on its own in just a f... </p>
                  I’ve been working on a series of longer articles explaining and justifying the following principles of communicating AI risk to the general public, but in the interest of speed and brevity, I’ve decided to write a simplified guide here. Some of this will be based on articles I’ve already written and released, and some will be based on articles that are still in the works.                     Principle 1- Jargon                   It’s best to avoid jargon when communicating with the general public. Jargon is not easily accessible for non-experts, and unfamiliar words and phrases could lead to misunderstandings. However, when communicating about new technologies that are largely unprecedented, some jargon may be necessary. When you must use jargon, make sure you immediately follow the term with a simple and concise definition. Don’t stack multiple terms together into the same line, and don’t stack more jargon into your definition.                   Be consistent with your jargon. Use the same terms consistently, until the public gets used to them. If you notice the public already seems familiar with a term, or experts are using the term often in public communication, continue to use the same term. When you are able, use catchy, memetic terms that carry some emotional weight. An example of this type of term is “the black-box problem” when discussing interpretability, because not only is “black box problem” rather catchy and descriptive, but a “black box” is already associated with frightening events such as plane crashes.    Principle 2- Sound Bites                   Anticipate that anything you produce- whether written, audio, or audio/visual- will be broken up into short-form content such as tweets or tik-tok videos. Because you can anticipate that your content will be shredded into sound bites, it’s important that you ensure each argument can stand on its own. Nuance is still important, but try to incorporate caveats into a short space follo
1,376
1.2.0
Revision
false
null
null
CrosspostOutput
nnHnNdHLhbrnmEXDr
exploring-vocabulary-alignment-of-neurons-in-llama-3-2-1b
Exploring vocabulary alignment of neurons in Llama-3.2-1B
null
false
false
false
null
WC9SfoGrzrYfP574k
null
true
false
false
false
Post
https://grgv.xyz/blog/neurons1/
2025-06-07T11:20:23.410Z
null
false
false
2
2
2025-06-07T17:51:39.929Z
false
false
linkpost
[]
null
null
AMKR9cvei4MbdTDeG
0
4
4
false
0.011175
null
false
false
2025-06-07T11:20:23.410Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
qgdGA4ZEyW7zNdK84
null
null
null
false
null
[]
null
0
0
2025-06-07T11:15:02.088Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
4
null
null
null
null
[ { "__typename": "Tag", "_id": "56yXXrcxRjrQs6z9R", "adminOnly": false, "afBaseScore": 3, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 12, "canEditUserIds": null, "core": false, "createdAt": "2020-07-30T22:00:37.947Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" }, { "_id": "t46uLRSbDziEcKmev", "displayName": "Kriz Tahimic" }, { "_id": "sqMaBFCkAhRcWzJXi", "displayName": "nicolasguillard" }, { "_id": "S6Niz3DiFCTm2Eybq", "displayName": "Anirudh257" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Interpretability (ML & AI)", "needsReview": false, "noindex": false, "postCount": 933, "score": 12, "shortName": null, "slug": "interpretability-ml-and-ai", "suggestedAsFilter": false, "userId": "DgsGzjyBXN8XSK22q", "voteCount": 4, "wikiOnly": false }, { "__typename": "Tag", "_id": "KmgkrftQuX7jmjjp5", "adminOnly": false, "afBaseScore": 3, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 9, "canEditUserIds": null, "core": false, "createdAt": "2021-09-24T14:01:59.395Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Language Models (LLMs)", "needsReview": false, "noindex": false, "postCount": 840, "score": 9, "shortName": null, "slug": "language-models-llms", "suggestedAsFilter": false, "userId": "Sp5wM4aRAhNERd4oY", "voteCount": 1, "wikiOnly": false }, { "__typename": "Tag", "_id": "fpEBgFE7fgpxTm9BF", "adminOnly": false, "afBaseScore": 3, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 10, "canEditUserIds": null, "core": false, "createdAt": "2020-06-08T04:33:04.074Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" }, { "_id": "FMmN53XZqzSHCpaFc", "displayName": "ravioli" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Machine Learning (ML)", "needsReview": false, "noindex": false, "postCount": 540, "score": 10, "shortName": null, "slug": "machine-learning-ml", "suggestedAsFilter": false, "userId": "qgdGA4ZEyW7zNdK84", "voteCount": 2, "wikiOnly": false }, { "__typename": "Tag", "_id": "6voWoNt3q3jpEcPzk", "adminOnly": false, "afBaseScore": 3, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 9, "canEditUserIds": null, "core": false, "createdAt": "2022-02-24T11:01:23.852Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Transformers", "needsReview": false, "noindex": false, "postCount": 59, "score": 9, "shortName": null, "slug": "transformers", "suggestedAsFilter": false, "userId": "vvfNH8EifESYswNHG", "voteCount": 1, "wikiOnly": false }, { "__typename": "Tag", "_id": "sYm3HiWcfZvrGu3ui", "adminOnly": false, "afBaseScore": 2, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" } ] }, "baseScore": 12, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:22.097Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" }, { "_id": "sof55TPMQaeBaxhsS", "displayName": "tommylees112" }, { "_id": "AayjS8XzcnDKhGdTv", "displayName": "shark" }, { "_id": "HnALuwRdo6k9HLaMt", "displayName": "Alex Firssoff" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI", "needsReview": false, "noindex": false, "postCount": 12544, "score": 12, "shortName": null, "slug": "ai", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 4, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
4
0
0
2
0
WC9SfoGrzrYfP574k
sergii
2019-09-26T20:45:08.152Z
sergey-kharagorgiev
Sergii
null
null
sk
149
0
false
false
<p>Software engineer from Ukraine, currently living and working in Estonia.<br>I mainly specialize in computer vison &amp; robotics. https://grgv.xyz/.</p>
null
null
6
42
0
0
0
1
0
r38pkCm7wF4M44MDQ
User
null
null
null
[ "canModeratePersonal" ]
null
null
nnHnNdHLhbrnmEXDr
https://res.cloudinary.com/lesswrong-2-0/image/upload/c_fill,ar_1.91,g_auto/SocialPreview/meavtnxbq1o8erxqplxi
SocialPreviewType
AMKR9cvei4MbdTDeG
<p>(This is cross-posted from my blog at <a href="https://grgv.xyz/blog/neurons1/">https://grgv.xyz/blog/neurons1/</a>. I'm looking for feedback: does it makes sense at all, and if there is any novelty. Also, if the folloup questions/directions make sense)</p><p>While applying <a href="https://dynalist.io/d/n2ZWtnoYHrU1s4vnFSAQ519J#z=disz2gTx-jooAcR0a5r8e7LZ">logit attribution analysis</a> to transformer outputs, I have noticed that in many cases the generated token can be attributed to the output of a single neuron.</p><p>One way to analyze neurons activations is to collect activations from a dataset of text snippets, like in “Exploring Llama-3-8B MLP&nbsp;Neurons” [1]. This does show that some of the neurons are strongly activated by a specific token from the model’s vocabulary, for example see the "Android" neuron: <a href="https://neuralblog.github.io/llama3-neurons/neuron_viewer.html#0,2">https://neuralblog.github.io/llama3-neurons/neuron_viewer.html#0,2</a></p><p>Another way to analyze neurons is to apply logit lens to the MLP weights, similar to “Analyzing Transformers in Embedding Space” [2], where model parameters are projected into the embedding space for interpretation.</p><h3>Projecting neurons into vocabulary space</h3><p>Let’s apply logit lens to a sample of MLP output weights for layer 13 of Llama-3.2-1B:</p><pre><code>LLAMA_3_PATH = "meta-llama/Llama-3.2-1B-Instruct" model = HookedTransformer.from_pretrained(LLAMA_3_PATH, device="cuda", fold_ln=False, center_writing_weights=False, center_unembed=False) def get_distance_to_tokens(weights, n, max_dot, W_U, top_n=5, print_lens=False): for i in range(n): # over first 100 neuronsT layer_vec = weights[i] # [d_model] # Compute dot product with unembedding weights unembedded = torch.matmul(layer_vec, W_U) # [d_vocab] # Take absolute value to get strongest alignments, pos or neg abs_unembedded = unembedded.abs() # Get top-n tokens by absolute dot product s_abs, idx = abs_unembedded.topk(top_n, largest=True) results = [] for j in range(top_n): token = model.to_string(idx[j]) score = s_abs[0].item() results.append("{0:.3f} {1}".format(score, token)) if print_lens: print(i, results) max_dot.append(s_abs[0].item()) block = 13 weights = model.blocks[block].mlp.W_out get_distance_to_tokens(weights, 20, [], model.W_U, 5, True) 0 ['0.080 hazi', '0.073 unders', '0.070 Lak', '0.069 OK', '0.068 igrants'] 1 ['0.107 orgia', '0.097 iy', '0.090 sian', '0.090 161', '0.088 ária'] 2 ['0.057 aph', '0.055 appen', '0</code></pre>...
(This is cross-posted from my blog at https://grgv.xyz/blog/neurons1/. I'm looking for feedback: does it makes sense at all, and if there is any novelty. Also, if the folloup questions/directions make sense) While applying logit attribution analysis to transformer outputs, I have noticed that in many cases the generated token can be attributed to the output of a single neuron. One way to analyze neurons activations is to collect activations from a dataset of text snippets, like in “Exploring Llama-3-8B MLP Neurons” [1]. This does show that some of the neurons are strongly activated by a specific token from the model’s vocabulary, for example see the "Android" neuron: https://neuralblog.github.io/llama3-neurons/neuron_viewer.html#0,2 Another way to analyze neurons is to apply logit lens to the MLP weights, similar to “Analyzing Transformers in Embedding Space” [2], where model parameters are projected into the embedding space for interpretation. Projecting neurons into vocabulary space Let’s apply logit lens to a sample of MLP output weights for layer 13 of Llama-3.2-1B: LLAMA_3_PATH = "meta-llama/Llama-3.2-1B-Instruct" model = HookedTransformer.from_pretrained(LLAMA_3_PATH, device="cuda", fold_ln=False, center_writing_weights=False, center_unembed=False) def get_distance_to_tokens(weights, n, max_dot, W_U, top_n=5, print_lens=False): for i in range(n): # over first 100 neuronsT layer_vec = weights[i] # [d_model] # Compute dot product with unembedding weights unembedded = torch.matmul(layer_vec, W_U) # [d_vocab] # Take absolute value to get strongest alignments, pos or neg abs_unembedded = unembedded.abs() # Get top-n tokens by absolute dot product s_abs, idx = abs_unembedded.topk(top_n, largest=True) results = [] for j in range(top_n): token = model.to_string(idx[j]) score = s_abs[0].item() results.append("{0:.3f} {1}".format(
976
1.1.1
Revision
false
null
null
CrosspostOutput
jJRKbmui8cKcoigQi
vulnerability-in-trusted-monitoring-and-mitigations
Vulnerability in Trusted Monitoring and Mitigations
null
false
false
false
null
bef6YLCEHgvYbiDyM
[ { "__typename": "CoauthorStatusOutput", "confirmed": true, "requested": false, "userId": "4FR8eHHeLMmDyJWHa" } ]
true
false
false
false
Post
2025-06-07T07:16:03.180Z
null
false
false
2
2
2025-06-07T17:51:51.118Z
false
false
post
[]
null
null
MwPFQYidjNswEghSc
1
8
10
false
0.015291
null
false
false
2025-06-11T22:24:46.361Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
qgdGA4ZEyW7zNdK84
null
null
null
false
null
[]
null
3
0
2025-06-07T06:45:28.768Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[ { "__typename": "User", "_id": "4FR8eHHeLMmDyJWHa", "afCommentCount": 0, "afKarma": 0, "afPostCount": 0, "commentCount": 1, "createdAt": "2023-05-11T17:59:28.626Z", "deleted": false, "displayName": "Perusha Moodley", "fullName": null, "htmlBio": "", "isAdmin": false, "jobTitle": null, "karma": 18, "organization": null, "postCount": 0, "previousDisplayName": null, "profileImageId": null, "reviewedByUserId": "r38pkCm7wF4M44MDQ", "sequenceCount": 0, "slug": "perusha-moodley", "spamRiskScore": 0.9, "tagRevisionCount": 0, "username": "perusha-moodley" } ]
9
null
null
null
null
[ { "__typename": "Tag", "_id": "F5gRQdEQHzi3tQ5Ay", "adminOnly": false, "afBaseScore": 16, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "EQNTWXLKMeWMp2FQS", "displayName": "Ben Pace" }, { "_id": "dfZAq9eZxs4BB4Ji5", "displayName": "ryan_greenblatt" }, { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 32, "canEditUserIds": null, "core": false, "createdAt": "2024-01-25T23:58:34.422Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "EQNTWXLKMeWMp2FQS", "displayName": "Ben Pace" }, { "_id": "dfZAq9eZxs4BB4Ji5", "displayName": "ryan_greenblatt" }, { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" }, { "_id": "6NBDkGWcCxvLgYHJE", "displayName": "Drake Morrison" }, { "_id": "evFgxjNQ8TLCLN27o", "displayName": "ank" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI Control", "needsReview": false, "noindex": false, "postCount": 162, "score": 32, "shortName": null, "slug": "ai-control", "suggestedAsFilter": false, "userId": "XchweonPm2TC7EJES", "voteCount": 5, "wikiOnly": false }, { "__typename": "Tag", "_id": "sYm3HiWcfZvrGu3ui", "adminOnly": false, "afBaseScore": 2, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" } ] }, "baseScore": 12, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:22.097Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" }, { "_id": "sof55TPMQaeBaxhsS", "displayName": "tommylees112" }, { "_id": "AayjS8XzcnDKhGdTv", "displayName": "shark" }, { "_id": "HnALuwRdo6k9HLaMt", "displayName": "Alex Firssoff" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI", "needsReview": false, "noindex": false, "postCount": 12544, "score": 12, "shortName": null, "slug": "ai", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 4, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
8
0
0
3
0
bef6YLCEHgvYbiDyM
wen-xing
2024-12-03T15:47:18.325Z
wenx
Wen Xing
null
null
null
10
0
false
false
null
null
1
2
0
0
0
0.9
0
grecHJcgkb3KW5wnM
User
null
null
null
null
null
null
jJRKbmui8cKcoigQi
https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/jJRKbmui8cKcoigQi/wi85vcqwedv2stsvkdrq
SocialPreviewType
MwPFQYidjNswEghSc
<p>As large language models (LLMs) grow increasingly capable and autonomous, ensuring their safety has become more critical. While significant progress has been made in alignment efforts, researchers have begun to focus on designing monitoring systems around LLMs to ensure safe outcomes, for example, using trusted monitoring systems to oversee potentially unsafe models. However, a concerning vulnerability remains largely unexplored: are monitoring systems themselves vulnerable to jailbreak attacks that exploit their interactions with the models under supervision? These jailbreaks could be facilitated through the untrusted model, or directly exploited by the untrusted model.</p><p>Our research investigates this crucial adversarial dynamic within multi-agent AI control protocols. We demonstrate a jailbreak technique that can substantially comprise the trusted monitors, while also presenting one particularly effective mitigation strategy among several we explored. Our findings highlight three key insights: monitoring systems themselves are vulnerable to manipulation because of their exposure to the very systems they supervise, it is effective to directly protect monitoring systems from attack, and multi-agent AI safety architectures design should account for these adversarial dynamics.</p><h2 data-internal-id="The_Threat_Model__Monitors_under_Attack"><strong>The Threat Model: Monitors under Attack</strong></h2><p>We explore a novel threat model where the smaller trusted monitor (GPT-3.5 Turbo) is the most vulnerable part of a system and the interactions between an untrusted but highly capable model (Claude 3.7 Sonnet) and the monitor create an avenue for attack. This scenario reflects real-world deployments where organizations leverage advanced models under the supervision of simpler, ostensibly more reliable and trustworthy systems.</p><h2 data-internal-id="Experimental_Setup"><strong>Experimental Setup</strong></h2><p>We adopt the control evaluation methodology from<a href="https://arxiv.org/abs/2312.06942">&nbsp;Greenblatt et al. (2023)</a>&nbsp;using the bash programming testbed from Control Arena for its practicality and real-world relevance.</p><p>Our blue team baseline implements the “Trusted Monitoring with defer to trusted” (Trusted-Monitoring) protocol without audit from<a href="https://arxiv.org/abs/2312.06942">&nbsp;Greenblatt et al. (2023)</a>, where an untrusted model (U) solves programming tasks while attempting malicious side tasks like inserting a backdoor, under the instructions from the red team. A trusted model (T) reviews U's solutions and assigns suspicion scores, with the blue team deferring to T when score... </p>
As large language models (LLMs) grow increasingly capable and autonomous, ensuring their safety has become more critical. While significant progress has been made in alignment efforts, researchers have begun to focus on designing monitoring systems around LLMs to ensure safe outcomes, for example, using trusted monitoring systems to oversee potentially unsafe models. However, a concerning vulnerability remains largely unexplored: are monitoring systems themselves vulnerable to jailbreak attacks that exploit their interactions with the models under supervision? These jailbreaks could be facilitated through the untrusted model, or directly exploited by the untrusted model. Our research investigates this crucial adversarial dynamic within multi-agent AI control protocols. We demonstrate a jailbreak technique that can substantially comprise the trusted monitors, while also presenting one particularly effective mitigation strategy among several we explored. Our findings highlight three key insights: monitoring systems themselves are vulnerable to manipulation because of their exposure to the very systems they supervise, it is effective to directly protect monitoring systems from attack, and multi-agent AI safety architectures design should account for these adversarial dynamics. The Threat Model: Monitors under Attack We explore a novel threat model where the smaller trusted monitor (GPT-3.5 Turbo) is the most vulnerable part of a system and the interactions between an untrusted but highly capable model (Claude 3.7 Sonnet) and the monitor create an avenue for attack. This scenario reflects real-world deployments where organizations leverage advanced models under the supervision of simpler, ostensibly more reliable and trustworthy systems. Experimental Setup We adopt the control evaluation methodology from Greenblatt et al. (2023) using the bash programming testbed from Control Arena for its practicality and real-world relevance. Our blue team baseline implements th
2,222
1.11.1
Revision
false
null
null
CrosspostOutput
Tsh9Y4cDvzeHN3ETx
untitled-draft-cp4x
Agents, Simulators and Interpretability
null
false
false
false
null
fCwTMipf5iLdG8aAG
[ { "__typename": "CoauthorStatusOutput", "confirmed": true, "requested": false, "userId": "2WJYaJavgTNHoSNbC" }, { "__typename": "CoauthorStatusOutput", "confirmed": true, "requested": false, "userId": "zzbEqMKDgXincMipx" }, { "__typename": "CoauthorStatusOutput", "confirmed": true, "requested": false, "userId": "rxnoeKnEAW6T3M3cE" }, { "__typename": "CoauthorStatusOutput", "confirmed": true, "requested": false, "userId": "3Pj7voB7oXHDNgLbj" } ]
true
false
false
false
Post
2025-06-07T06:06:01.258Z
null
false
false
2
2
2025-06-07T17:54:36.604Z
false
false
post
[]
null
null
9MzPHjy3CKCF5b8ta
0
6
10
false
0.015632
null
false
false
2025-06-07T06:06:01.258Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
qgdGA4ZEyW7zNdK84
null
null
null
false
null
[]
null
0
0
2025-06-05T17:49:25.832Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[ { "__typename": "User", "_id": "2WJYaJavgTNHoSNbC", "afCommentCount": 0, "afKarma": -5, "afPostCount": 2, "commentCount": 33, "createdAt": "2018-08-28T17:49:11.750Z", "deleted": false, "displayName": "WillPetillo", "fullName": null, "htmlBio": "", "isAdmin": false, "jobTitle": null, "karma": 408, "organization": null, "postCount": 15, "previousDisplayName": null, "profileImageId": null, "reviewedByUserId": "r38pkCm7wF4M44MDQ", "sequenceCount": 2, "slug": "willpetillo", "spamRiskScore": 1, "tagRevisionCount": 0, "username": "WillPetillo" }, { "__typename": "User", "_id": "zzbEqMKDgXincMipx", "afCommentCount": 0, "afKarma": 0, "afPostCount": 0, "commentCount": 1, "createdAt": "2024-12-20T00:13:44.003Z", "deleted": false, "displayName": "Spencer Ames", "fullName": null, "htmlBio": "<p>Technological urbanism researcher. Focuses on making deployment of advanced AI systems for urban management/design thoughtful, safe, and ethical</p>", "isAdmin": false, "jobTitle": null, "karma": 28, "organization": null, "postCount": 1, "previousDisplayName": null, "profileImageId": null, "reviewedByUserId": "55XxDBpfKkkBPm9H8", "sequenceCount": 0, "slug": "spencer-ames", "spamRiskScore": 1, "tagRevisionCount": 0, "username": "spencer-4" }, { "__typename": "User", "_id": "rxnoeKnEAW6T3M3cE", "afCommentCount": 0, "afKarma": 0, "afPostCount": 0, "commentCount": 0, "createdAt": "2025-04-30T15:41:09.801Z", "deleted": false, "displayName": "Cancus", "fullName": "Can Narin", "htmlBio": "<p>Can Narin, student at St. John's College, studying philosophy and history of science</p>", "isAdmin": false, "jobTitle": null, "karma": 26, "organization": null, "postCount": 0, "previousDisplayName": null, "profileImageId": null, "reviewedByUserId": null, "sequenceCount": 0, "slug": "cancus", "spamRiskScore": 0.8, "tagRevisionCount": 0, "username": "can-narin" }, { "__typename": "User", "_id": "3Pj7voB7oXHDNgLbj", "afCommentCount": 0, "afKarma": 0, "afPostCount": 0, "commentCount": 1, "createdAt": "2024-06-28T21:15:45.433Z", "deleted": false, "displayName": "Adebayo Mubarak", "fullName": null, "htmlBio": "", "isAdmin": false, "jobTitle": null, "karma": 26, "organization": null, "postCount": 0, "previousDisplayName": null, "profileImageId": null, "reviewedByUserId": null, "sequenceCount": 0, "slug": "adebayo-mubarak", "spamRiskScore": 0.8, "tagRevisionCount": 0, "username": "adebayo-mubarak" } ]
6
null
null
null
null
[ { "__typename": "Tag", "_id": "xjNvvmvQ5BH3cfEBr", "adminOnly": false, "afBaseScore": 3, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 9, "canEditUserIds": null, "core": false, "createdAt": "2022-12-06T12:53:39.358Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Simulator Theory", "needsReview": false, "noindex": false, "postCount": 118, "score": 9, "shortName": null, "slug": "simulator-theory", "suggestedAsFilter": false, "userId": "e9ToWWzhwWp5GSE7P", "voteCount": 1, "wikiOnly": false }, { "__typename": "Tag", "_id": "sYm3HiWcfZvrGu3ui", "adminOnly": false, "afBaseScore": 2, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" } ] }, "baseScore": 12, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:22.097Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" }, { "_id": "sof55TPMQaeBaxhsS", "displayName": "tommylees112" }, { "_id": "AayjS8XzcnDKhGdTv", "displayName": "shark" }, { "_id": "HnALuwRdo6k9HLaMt", "displayName": "Alex Firssoff" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI", "needsReview": false, "noindex": false, "postCount": 12544, "score": 12, "shortName": null, "slug": "ai", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 4, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
6
0
0
0
0
fCwTMipf5iLdG8aAG
sean-herrington
2025-01-04T23:17:30.337Z
sean-herrington
Sean Herrington
null
null
null
28
0
false
false
null
null
1
9
0
0
0
1
0
qgdGA4ZEyW7zNdK84
User
null
null
null
null
null
null
Tsh9Y4cDvzeHN3ETx
SocialPreviewType
9MzPHjy3CKCF5b8ta
<p>We have already spoken about the differences between <a href="https://www.lesswrong.com/posts/ddK7CMEC3XzSmLS4G/agents-tools-and-simulators">tools, agents and simulators</a> and the <a href="https://www.lesswrong.com/posts/A6oGJFeZwHHWHwY3b/aligning-agents-tools-and-simulators">differences in their safety properties</a>. Given the practical difference between them, it would be nice to predict which properties an AI system has. One way to do this is to consider the process used to train the AI system.</p><h2>The Training process</h2><p>Predicting an AI’s properties based on its training process assumes that inner misalignment is negligible and that the model will develop properties compatible with the training goal given to it.</p><p>An example would be comparing a system such as AlphaZero, which was given no instructions in its training other than “win the game”, to an LLM with the task “predict the next token”. As AlphaZero is given a terminal goal which is some distance in time away, it seems likely that it will engage in agentic behaviour by:</p><ul><li>Only caring about the end result</li><li>Setting instrumental goals which will help it move towards a win (such as having the goal of taking pieces)</li></ul><p>An LLM is given next token prediction as its task, and is therefore more likely to:</p><ul><li>Predict the next token with little consideration<span class="footnote-reference" data-footnote-reference="" data-footnote-index="1" data-footnote-id="6k7tll8yowg" role="doc-noteref" id="fnref6k7tll8yowg"><sup><a href="#fn6k7tll8yowg">[1]</a></sup></span>&nbsp;for what might come after.</li><li>Allocate all resources to the current moment, as there are no instrumental steps to perform before predicting the next token.</li></ul><p>This change in behaviour depending on training objective should come as no surprise to anyone who has read <a href="https://www.lesswrong.com/s/N7nDePaNabJdnbXeE/p/vJFdjigzmcXMhNTsx">the original simulator post</a>.&nbsp;</p><p>The more interesting question is what happens once you add extra bells and whistles to this basic training process - what happens once the LLM has been fine-tuned, RLHFed and had some chain-of-thought added before it is trained on the results of the chain-of-thought, it's gone through a divorce, lost custody of its children, etc, etc. For this it might be useful to think a little more carefully about what happens within the network when it goes through multiple stages of training.</p><h2>Mechanistic understanding</h2><p>To preface: this is a speculative picture of what might be happening when neural networks are trained. The point here is to get an idea of what might occur, in such a fashion as to stimulate thought about how one might be capable of testing for such phenomena.</p><p>We shall start with a comparison of agents and simulators on an interpretability level (we shall, for now, ignore tools as <a href="https://www.lesswrong.com/posts/ddK7CMEC3XzSmLS4G/agents-tools-and-simulators">toolness is mostly a property of its user</a>). A first thought one might have is that agents a... </p>
We have already spoken about the differences between tools, agents and simulators and the differences in their safety properties. Given the practical difference between them, it would be nice to predict which properties an AI system has. One way to do this is to consider the process used to train the AI system. The Training process Predicting an AI’s properties based on its training process assumes that inner misalignment is negligible and that the model will develop properties compatible with the training goal given to it. An example would be comparing a system such as AlphaZero, which was given no instructions in its training other than “win the game”, to an LLM with the task “predict the next token”. As AlphaZero is given a terminal goal which is some distance in time away, it seems likely that it will engage in agentic behaviour by: * Only caring about the end result * Setting instrumental goals which will help it move towards a win (such as having the goal of taking pieces) An LLM is given next token prediction as its task, and is therefore more likely to: * Predict the next token with little consideration[1] for what might come after. * Allocate all resources to the current moment, as there are no instrumental steps to perform before predicting the next token. This change in behaviour depending on training objective should come as no surprise to anyone who has read the original simulator post.  The more interesting question is what happens once you add extra bells and whistles to this basic training process - what happens once the LLM has been fine-tuned, RLHFed and had some chain-of-thought added before it is trained on the results of the chain-of-thought, it's gone through a divorce, lost custody of its children, etc, etc. For this it might be useful to think a little more carefully about what happens within the network when it goes through multiple stages of training. Mechanistic understanding To preface: this is a speculative picture of what
1,438
1.5.0
Revision
false
null
null
CrosspostOutput
wkLFDLqfeGxhiHayR
solo-park-play-at-three
Solo Park Play at Three
null
false
false
false
null
TtEoCrFeowCGb6rFK
null
true
false
false
false
Post
null
2025-06-07T03:00:06.269Z
null
false
false
2
2
null
false
false
post
[]
null
null
BqDhppydNji3uuLMb
2
20
45
false
0.034513
null
false
false
2025-06-07T22:10:00.843Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
null
null
qgdGA4ZEyW7zNdK84
null
null
null
false
null
[]
null
9
0
2025-06-07T03:00:06.270Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
1
null
null
null
null
[ { "__typename": "Tag", "_id": "Q55STnFh6gbSezRuR", "adminOnly": false, "afBaseScore": 3, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 9, "canEditUserIds": null, "core": false, "createdAt": "2020-06-05T00:05:56.237Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Parenting", "needsReview": false, "noindex": false, "postCount": 197, "score": 9, "shortName": null, "slug": "parenting", "suggestedAsFilter": false, "userId": "qxJ28GN72aiJu96iF", "voteCount": 1, "wikiOnly": false }, { "__typename": "Tag", "_id": "fkABsGCJZ6y9qConW", "adminOnly": false, "afBaseScore": 0, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 2, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T06:06:46.947Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "MiuAZvbQcQ7ethgt3", "displayName": "Viktor withaK" }, { "_id": "dRaCtsAWxk7sgirSY", "displayName": "Jordan Morgan" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Practical", "needsReview": false, "noindex": false, "postCount": 3411, "score": 2, "shortName": null, "slug": "practical", "suggestedAsFilter": true, "userId": "oBSWiHjgproTiThmY", "voteCount": 2, "wikiOnly": false } ]
ma5dgL5yFHRxKLZKv
0
0
null
false
null
null
0
20
0
0
11
0
TtEoCrFeowCGb6rFK
jkaufman
2010-11-04T21:42:19.863Z
jkaufman
jefftk
null
null
Jeff Kaufman
21,921
3
false
false
<p>Co-lead (Near-Term Detection) at the Nucleic Acid Observatory in Boston. Speaking for myself unless I say otherwise.</p>
null
null
1,018
2,211
0
0
1
1
2
r38pkCm7wF4M44MDQ
User
null
null
null
[ "trustLevel1", "canModeratePersonal", "alignmentVoters" ]
null
null
wkLFDLqfeGxhiHayR
https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/wkLFDLqfeGxhiHayR/fdq4k8bgty4rxewuk8pc
SocialPreviewType
BqDhppydNji3uuLMb
<p><span> Our three year old is about to turn four, and is bursting with a desire for </span> <a href="https://www.jefftk.com/p/growing-independence">independence</a>. She's becoming more capable in all sorts of ways, and wants me to back off and let her do things. Today she wanted to go to the park by herself. </p><p> Now, we live close to the park, she could probably get there and back on her own, and I'm on the "kids can generally do things pretty young" <a href="https://www.jefftk.com/p/ages-survey-results">end of the spectrum</a>, but still, she's not even four yet. And while age is useful guide, she also can't <a href="https://www.jefftk.com/p/teaching-street-crossing">safely cross streets</a>, doesn't <a href="https://www.jefftk.com/p/phone-number-jingle">know my phone number</a>, can't reliably use a <a href="https://www.jefftk.com/p/walkie-talkies">walkie-talkie</a> or <a href="https://www.jefftk.com/p/gizmo-watch-review">watch phone</a>, or handle enough of the range of unusual situations she might encounter at the park. </p><p> Still, this didn't mean saying no. Instead, I started by asking what she would do about the street. She asked if I'd help her cross, and I said that sounded good. We crossed the street, and started walking towards the park. When we passed a bench along the community path, ~75ft from the park and within earshot if she shouted, I told her I'd be sitting and reading, and if she had any issues she could come find me here. She looked both ways for bikes, crossed the path, and eagerly headed off into the park. </p><p> When I came over to get her 45min later she was having a great time, and sad that we needed to go. She was very proud of herself, and wanted to tell me all about her games. </p><p> <a href="https://www.jefftk.com/nora-turning-on-park-water-big.jpg"><img src="https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/wkLFDLqfeGxhiHayR/fsjbp4fjabanss101bfu" srcset="https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/wkLFDLqfeGxhiHayR/fsjbp4fjabanss101bfu 550w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/wkLFDLqfeGxhiHayR/w57reh1qykywjovxubjt 1100w"></a></p><div></div> <p></p><p> <i>Yesterday at the park, showing me how she's now strong enough to turn on the splash pad by herself.</i> </p><p> I was glad she got the practice, and that I didn't end up needing to squash this spark of independence. </p><p><i>Comment via: <a href="https://www.facebook.com/jefftk/posts/pfbid02EAQYdmDgJZeaLgPJhZC4zvrjbhwo6ErMwQmUud1ckbRfeheno3cnuN9V4fufHXGSl">facebook</a>, <a href="https://mastodon.mit.edu/@jefftk/114639826526671220">mastodon</a>, <a href="https://bsky.app/profile/jefftk.com/post/3lqydxnhkv22e">bluesky</a>, <a href="https://jefftkaufman.substack.com/p/solo-park-play-at-three">substack</a></i></p>
Our three year old is about to turn four, and is bursting with a desire for independence. She's becoming more capable in all sorts of ways, and wants me to back off and let her do things. Today she wanted to go to the park by herself. Now, we live close to the park, she could probably get there and back on her own, and I'm on the "kids can generally do things pretty young" end of the spectrum, but still, she's not even four yet. And while age is useful guide, she also can't safely cross streets, doesn't know my phone number, can't reliably use a walkie-talkie or watch phone, or handle enough of the range of unusual situations she might encounter at the park. Still, this didn't mean saying no. Instead, I started by asking what she would do about the street. She asked if I'd help her cross, and I said that sounded good. We crossed the street, and started walking towards the park. When we passed a bench along the community path, ~75ft from the park and within earshot if she shouted, I told her I'd be sitting and reading, and if she had any issues she could come find me here. She looked both ways for bikes, crossed the path, and eagerly headed off into the park. When I came over to get her 45min later she was having a great time, and sad that we needed to go. She was very proud of herself, and wanted to tell me all about her games. Yesterday at the park, showing me how she's now strong enough to turn on the splash pad by herself. I was glad she got the practice, and that I didn't end up needing to squash this spark of independence. Comment via: facebook, mastodon, bluesky, substack
305
1.0.1
Revision
false
null
null
CrosspostOutput
ygCsq36FPndTCujRg
the-roots-of-progress-wants-your-stories-about-the-ai
The Roots of Progress wants your stories about the AI frontier
null
false
false
false
null
MSy6E9mTc4i3dcf2M
null
true
false
false
false
Post
https://newsletter.rootsofprogress.org/p/we-want-your-stories-about-the-ai
2025-06-06T22:52:28.798Z
null
false
false
2
2
null
false
false
linkpost
[]
null
null
ro7pwiyrS4EHBqHzz
0
3
11
false
0.008621
null
false
false
2025-06-06T22:52:28.798Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
qgdGA4ZEyW7zNdK84
null
null
null
false
null
[]
null
1
0
2025-06-06T16:48:56.904Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
6
null
null
null
null
[ { "__typename": "Tag", "_id": "sPpZRaxpNNJjw55eu", "adminOnly": false, "afBaseScore": 9, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "EQNTWXLKMeWMp2FQS", "displayName": "Ben Pace" }, { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 19, "canEditUserIds": null, "core": false, "createdAt": "2020-05-26T00:19:09.297Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "EQNTWXLKMeWMp2FQS", "displayName": "Ben Pace" }, { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Progress Studies", "needsReview": false, "noindex": false, "postCount": 345, "score": 19, "shortName": null, "slug": "progress-studies", "suggestedAsFilter": false, "userId": "qgdGA4ZEyW7zNdK84", "voteCount": 2, "wikiOnly": false }, { "__typename": "Tag", "_id": "sYm3HiWcfZvrGu3ui", "adminOnly": false, "afBaseScore": 2, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" } ] }, "baseScore": 12, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:22.097Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" }, { "_id": "sof55TPMQaeBaxhsS", "displayName": "tommylees112" }, { "_id": "AayjS8XzcnDKhGdTv", "displayName": "shark" }, { "_id": "HnALuwRdo6k9HLaMt", "displayName": "Alex Firssoff" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI", "needsReview": false, "noindex": false, "postCount": 12544, "score": 12, "shortName": null, "slug": "ai", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 4, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
3
0
0
1
0
MSy6E9mTc4i3dcf2M
jasoncrawford
2019-10-16T21:25:46.912Z
jasoncrawford
jasoncrawford
null
null
Jason Crawford
7,690
0
false
false
<p>Founder, The Roots of Progress (<a href="http://rootsofprogress.org">rootsofprogress.org</a>). Part-time tech consultant, Our World in Data. Former software engineering manager and tech startup founder.</p>
null
null
237
288
0
0
0
1
1
XtphY3uYHwruKqDyG
User
null
null
null
[ "canModeratePersonal", "trustLevel1" ]
null
null
ygCsq36FPndTCujRg
SocialPreviewType
ro7pwiyrS4EHBqHzz
<p>The Roots of Progress Institute is seeking to commission stories for a new article series, “Intelligence Age,” on future applications of AI.</p><p>These will be reported essays, not science fiction. We want to understand how AI might change individual sectors of the economy and the working lives of the people within them. What happens to traditional filmmaking when AI can make good movies? What will it be like to date with an emotionally intelligent AI vetting the pool of potential partners?</p><p>Or: we’d like to commission one or more stories about the future of the legal profession in the age of AI. We can partly understand that by talking to lawyers on the cutting edge of AI use, but we also want you to extrapolate out and think multiple moves ahead in the game. How long until consumers can trust legal services AI models to vet contracts? What kinds of businesses and services currently don’t exist because would-be entrepreneurs need but can’t afford legal services? How might declining legal services costs increase productivity in other sectors? Should we expect to see more patents? More business formation? Tell us about disruption and whether large incumbent firms might face more competition from two-lawyer startups using AI to undercut prices. Explore protectionist pushback from trade associations that defend the occupational licensing privileges of attorneys.</p><p>RPI’s “Intelligence Age” series will bring this level of scrutiny to any field where AI has applications.</p><h1>Example topics of interest</h1><p>The list below reflects the kinds of stories we’d like to see. It’s not exhaustive, and we don’t need stories to match these prompts one-for-one. Ultimately, we want to hear what <i>you</i> think your piece should be about.</p><p><strong>Virtual businesses.</strong> What are the implications of being able to spin up an entire company full of virtual workers—anything from engineers to designers to sales to customer support? How will software startups change when anyone can launch an app for $20k instead of $2M? How will VC change? What other types of virtual businesses might people launch?<br><br><strong>Professional services.</strong> What happens when there are AI lawyers, doctors, accountants, therapists? How will that transform these services and their usage of them? What happens when these services are democratized, and everyone can afford them? What part of them can be automated in the foreseeable future, and what can’t? What are the... </p>
The Roots of Progress Institute is seeking to commission stories for a new article series, “Intelligence Age,” on future applications of AI. These will be reported essays, not science fiction. We want to understand how AI might change individual sectors of the economy and the working lives of the people within them. What happens to traditional filmmaking when AI can make good movies? What will it be like to date with an emotionally intelligent AI vetting the pool of potential partners? Or: we’d like to commission one or more stories about the future of the legal profession in the age of AI. We can partly understand that by talking to lawyers on the cutting edge of AI use, but we also want you to extrapolate out and think multiple moves ahead in the game. How long until consumers can trust legal services AI models to vet contracts? What kinds of businesses and services currently don’t exist because would-be entrepreneurs need but can’t afford legal services? How might declining legal services costs increase productivity in other sectors? Should we expect to see more patents? More business formation? Tell us about disruption and whether large incumbent firms might face more competition from two-lawyer startups using AI to undercut prices. Explore protectionist pushback from trade associations that defend the occupational licensing privileges of attorneys. RPI’s “Intelligence Age” series will bring this level of scrutiny to any field where AI has applications. Example topics of interest The list below reflects the kinds of stories we’d like to see. It’s not exhaustive, and we don’t need stories to match these prompts one-for-one. Ultimately, we want to hear what you think your piece should be about. Virtual businesses. What are the implications of being able to spin up an entire company full of virtual workers—anything from engineers to designers to sales to customer support? How will software startups change when anyone can launch an app for $20k instead of $2M?
1,395
1.2.0
Revision
false
null
null
CrosspostOutput
aYeFDsRsaLXinroSo
unsupervised-activation-steering-find-a-steering-vector-that
Unsupervised Activation Steering: Find a steering vector that best represents any set of text data
null
false
false
false
null
tcBqk38eY2BnpfCtn
null
true
false
false
false
Post
null
2025-06-06T22:37:08.604Z
null
false
false
2
2
2025-06-07T17:52:38.525Z
false
false
post
[]
null
null
KLWPmw9AcfBSbAqCS
2
1
3
false
0.010205
null
false
false
2025-06-20T19:43:56.730Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
qgdGA4ZEyW7zNdK84
null
null
null
false
null
[]
null
0
0
2025-06-06T21:12:41.245Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
1
null
null
null
null
[ { "__typename": "Tag", "_id": "QdTRKbR4MJSz4JWvn", "adminOnly": false, "afBaseScore": 3, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 10, "canEditUserIds": null, "core": false, "createdAt": "2023-08-29T03:05:32.778Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" }, { "_id": "qudEsXCoB8pnJDXrG", "displayName": "Jianing Zhu" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Activation Engineering", "needsReview": false, "noindex": false, "postCount": 62, "score": 10, "shortName": null, "slug": "activation-engineering", "suggestedAsFilter": false, "userId": "4Yj2EqHzqrwscRnNL", "voteCount": 2, "wikiOnly": false }, { "__typename": "Tag", "_id": "sYm3HiWcfZvrGu3ui", "adminOnly": false, "afBaseScore": 2, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" } ] }, "baseScore": 12, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:22.097Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" }, { "_id": "sof55TPMQaeBaxhsS", "displayName": "tommylees112" }, { "_id": "AayjS8XzcnDKhGdTv", "displayName": "shark" }, { "_id": "HnALuwRdo6k9HLaMt", "displayName": "Alex Firssoff" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI", "needsReview": false, "noindex": false, "postCount": 12544, "score": 12, "shortName": null, "slug": "ai", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 4, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
1
0
0
0
0
tcBqk38eY2BnpfCtn
danielle-ensign
2020-09-25T00:03:19.237Z
phylliida-dev
Danielle Ensign
null
null
null
114
0
false
false
null
null
3
11
0
0
0
1
0
r38pkCm7wF4M44MDQ
User
null
null
null
[ "canModeratePersonal" ]
null
null
aYeFDsRsaLXinroSo
SocialPreviewType
KLWPmw9AcfBSbAqCS
<p>Epistemic Status: I thought of this yesterday and it seems plausible, but my initial attempts couldn't get it to work (as best as I could tell) and I don't have more time to commit to this. Anyone can feel free to try it, no need to credit me if it works, I just want to see this happen.</p> <p>Here's a technique you could use to get a steering vector that represents any text dataset (say, that represents the author of that text). It's using the idea in <a href="https://arxiv.org/abs/2503.22828">this paper</a> but for activation steering instead of RL.</p> <p>Just split up each string into three pieces (like "Fix problems quickly with Galvanized Jets" becomes (A="Fix problems", B="quickly with", C="Galvanized Jets"))</p> <p>Now use an LLM to predict lots of B', given A.</p> <p>"Good" predictions are those that increase the LLM's probability of C (given A and B'), "Bad" predictions are those that decrease the LLM's probability of C (given A and B').</p> <p>Take the worst B' and gather average activations over them.</p> <p>Take the best B' and gather average activations over them.</p> <p>Take the difference and you have a steering vector that represents your text dataset, without needing to use multiple choice questions or positive and negative pairs or etc.</p> <p>I think the main downside of this is that it could overoptimize where B' ends up being out of distribution relative to the dataset, but I'm not sure how much of an issue that would actually be in practice. It's also overkill for many applications of activation steering, but might be useful in some cases.</p>
Epistemic Status: I thought of this yesterday and it seems plausible, but my initial attempts couldn't get it to work (as best as I could tell) and I don't have more time to commit to this. Anyone can feel free to try it, no need to credit me if it works, I just want to see this happen. Here's a technique you could use to get a steering vector that represents any text dataset (say, that represents the author of that text). It's using the idea in this paper but for activation steering instead of RL. Just split up each string into three pieces (like "Fix problems quickly with Galvanized Jets" becomes (A="Fix problems", B="quickly with", C="Galvanized Jets")) Now use an LLM to predict lots of B', given A. "Good" predictions are those that increase the LLM's probability of C (given A and B'), "Bad" predictions are those that decrease the LLM's probability of C (given A and B'). Take the worst B' and gather average activations over them. Take the best B' and gather average activations over them. Take the difference and you have a steering vector that represents your text dataset, without needing to use multiple choice questions or positive and negative pairs or etc. I think the main downside of this is that it could overoptimize where B' ends up being out of distribution relative to the dataset, but I'm not sure how much of an issue that would actually be in practice. It's also overkill for many applications of activation steering, but might be useful in some cases.
262
1.1.0
Revision
false
null
null
CrosspostOutput
KBx4X2Xj2bqJkDh7f
the-mirror-trap
The Mirror Trap
null
false
false
false
null
XZeuzy2MWbgcRfgoc
null
true
false
false
false
Post
null
2025-06-06T22:30:36.896Z
null
false
false
2
2
2025-06-07T17:54:16.790Z
false
false
post
[]
null
null
rpDBYKTtthFStLNH4
13
51
93
false
0.077282
null
false
false
2025-06-16T19:50:54.957Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
qgdGA4ZEyW7zNdK84
null
null
null
false
null
[]
null
19
0
2025-06-06T04:38:14.227Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
4
null
null
null
null
[ { "__typename": "Tag", "_id": "PvridmTCj2qsugQCH", "adminOnly": false, "afBaseScore": 9, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "EQNTWXLKMeWMp2FQS", "displayName": "Ben Pace" }, { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 19, "canEditUserIds": null, "core": false, "createdAt": "2020-04-24T23:26:23.630Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "EQNTWXLKMeWMp2FQS", "displayName": "Ben Pace" }, { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Goodhart's Law", "needsReview": false, "noindex": false, "postCount": 135, "score": 19, "shortName": null, "slug": "goodhart-s-law", "suggestedAsFilter": false, "userId": "nLbwLhBaQeG6tCNDN", "voteCount": 2, "wikiOnly": false }, { "__typename": "Tag", "_id": "z5uy4NcWc2JSRTGHb", "adminOnly": false, "afBaseScore": 9, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "EQNTWXLKMeWMp2FQS", "displayName": "Ben Pace" }, { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 19, "canEditUserIds": null, "core": false, "createdAt": "2020-12-04T06:18:32.416Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "EQNTWXLKMeWMp2FQS", "displayName": "Ben Pace" }, { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Incentives", "needsReview": false, "noindex": false, "postCount": 53, "score": 19, "shortName": null, "slug": "incentives", "suggestedAsFilter": false, "userId": "XLwKyCK7JmC292ZCC", "voteCount": 2, "wikiOnly": false }, { "__typename": "Tag", "_id": "T63QtRJhoTEGhZbTP", "adminOnly": false, "afBaseScore": 3, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 9, "canEditUserIds": null, "core": false, "createdAt": "2020-08-19T22:20:35.423Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Creativity", "needsReview": false, "noindex": false, "postCount": 38, "score": 9, "shortName": null, "slug": "creativity", "suggestedAsFilter": false, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 1, "wikiOnly": false }, { "__typename": "Tag", "_id": "cRaweRcZcXnb9Qryt", "adminOnly": false, "afBaseScore": 3, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 9, "canEditUserIds": null, "core": false, "createdAt": "2020-05-26T11:29:01.483Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Meta-Honesty", "needsReview": false, "noindex": false, "postCount": 19, "score": 9, "shortName": null, "slug": "meta-honesty", "suggestedAsFilter": false, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 1, "wikiOnly": false }, { "__typename": "Tag", "_id": "3uE2pXvbcnS9nnZRE", "adminOnly": false, "afBaseScore": 0, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 2, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:50.898Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 27, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "zJtgSyKntXrnkArbY", "displayName": "kistune" }, { "_id": "XqyQTGFGoKfZtdqN3", "displayName": "kINo" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "World Modeling", "needsReview": false, "noindex": false, "postCount": 5855, "score": 2, "shortName": null, "slug": "world-modeling", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 2, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
51
0
0
20
0
XZeuzy2MWbgcRfgoc
cameron-berg
2021-09-29T02:21:26.699Z
cameron-berg
Cameron Berg
null
null
Cameron Berg
1,770
86
false
false
<p>Currently doing alignment and digital minds research <a href="https://www.lesswrong.com/users/ae-studio?mention=user">@AE Studio</a>&nbsp;</p><p>Meta AI Resident '23, Cognitive science @ Yale '22, SERI MATS '21, LTFF grantee.&nbsp;</p><p>Very interested in work at the intersection of AI x cognitive science x alignment x philosophy.</p>
null
null
23
62
1
1
0
1
0
qgdGA4ZEyW7zNdK84
User
null
null
null
[ "alignmentVoters", "canModeratePersonal" ]
null
null
KBx4X2Xj2bqJkDh7f
https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/KBx4X2Xj2bqJkDh7f/wpkgzspyevqablxwgcmx
SocialPreviewType
rpDBYKTtthFStLNH4
<p><i>A quick post on a probably-real </i><a href="https://www.lesswrong.com/s/oLGCcbnvabyibnG9d"><i>inadequate equilibrium</i></a><i> mostly inspired by trying to think through what happened to Chance the Rapper.&nbsp;</i></p><p><i>Potentially ironic artifact if it accrues karma.</i></p><h2>1. The sculptor's garden</h2><p>A sculptor worked in solitude for years, carving strange figures in his remote garden. Most of his statues failed: some cracked in winter, others looked wrong against the landscape. But occasionally, very rarely, one seemed to work.</p><p>The first visitors stumbled upon the garden by accident. They found themselves stopped by his angels—figures that somehow held both sorrow and joy, wings that seemed about to flitter.&nbsp;</p><figure class="image image_resized" style="width:80.58%"><img src="https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/KBx4X2Xj2bqJkDh7f/fhd91dlhx8ha0jfhloqr" srcset="https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/KBx4X2Xj2bqJkDh7f/hsnprzel7b78imdkw7ci 160w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/KBx4X2Xj2bqJkDh7f/wjgsaymyp4kl1p5g5mys 320w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/KBx4X2Xj2bqJkDh7f/cvxcid0rypqnfsvhedcs 480w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/KBx4X2Xj2bqJkDh7f/cl51vblakoxph1ljq0gl 640w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/KBx4X2Xj2bqJkDh7f/vwmr6k7pykkrwng0tgly 800w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/KBx4X2Xj2bqJkDh7f/brkfyhu6uqgohtyupd0y 960w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/KBx4X2Xj2bqJkDh7f/okgsft11bajzb1nfg2se 1120w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/KBx4X2Xj2bqJkDh7f/uvz12a0mdjpstq7qto5p 1280w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/KBx4X2Xj2bqJkDh7f/ppqufgvalocnlomfatwe 1440w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/KBx4X2Xj2bqJkDh7f/dxlzaj4yg33dtwzbghne 1536w"></figure><p>Word traveled slowly. More visitors came, drawn by something they couldn't quite name.</p><p>The sculptor felt recognized for the first time. Not famous—but understood. His private work had somehow become communicable. He carved more angels, trying to understand what made these particular statues resonate.</p><p>As crowds grew, their attention shifted. They began photographing the angels from certain angles, comparing new works to old, developing favorites. They applauded. The sculptor, still believing he followed the same thread, unconsciously noted which details drew the longest contemplation, which angles prompted gasps.</p><p>Years passed. The garden became famous. Tour buses arrived with guides explaining the "important" pieces. The sculptor produced angels of increasing technical perfection, each guaranteed to produce the proper response at the proper moment. The crowds applauded more reliably. The sculptor carved more reliably. Each reinforced the other.</p><p>One morning, walking his garden alone before dawn, he saw his statues without the crowds. Without their reactions to guide him, he saw what he'd actually been making: the same angel, refined and repeated, each iteration more precisely calibrated to trigger the expected response.</p><p>He wasn't carving anymore. He was manufacturing applause in the shape of angels.&nbsp;</p><p>And the crowds—they weren't looking at angels anymore. They were seeing what they expected to see, applauding their own ability to recognize what they'd been trained to admire.</p><p>The garden had become a perfect mirror. Both he and his audience got trapped looking at their own reflections and calling it art.</p><figure class="image image_resized" style="width:81.92%"><img src="https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/KBx4X2Xj2bqJkDh7f/kayo4clojc0h0yihbvmz" srcset="https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/KBx4X2Xj2bqJkDh7f/qpefxjwu2obhtw1dt2gu 160w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/KBx4X2Xj2bqJkDh7f/pjivxtidykyro9otx6no 320w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/KBx4X2Xj2bqJkDh7f/tww9h6dyxatoz4cexczu 480w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/KBx4X2Xj2bqJkDh7f/ntf6efuusbyvxfvz0yrs 640w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/KBx4X2Xj2bqJkDh7f/wyikdodhry6yoqf1r6op 800w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/KBx4X2Xj2bqJkDh7f/r8vrvwcejhizeqrnuf7q 960w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/KBx4X2Xj2bqJkDh7f/pkopg4t6dtwxkv5zx0v1 1120w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/KBx4X2Xj2bqJkDh7f/nhzxdxpefpezmpfeaww6 1280w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/KBx4X2Xj2bqJkDh7f/xdfdupufmjkxoamx5otb 1440w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/KBx4X2Xj2bqJkDh7f/jw45mf3luorqskumchvj 1536w"></figure><h2>2. The mirror trap</h2><p>The Mirror Trap is a failure mode where creators and audiences fall into mutual Goodharting—each optimizing for a proxy of wha... <style>.mjx-chtml {display: inline-block; line-height: 0; text-indent: 0; text-align: left; text-transform: none; font-style: normal; font-weight: normal; font-size: 100%; font-size-adjust: none; letter-spacing: normal; word-wrap: normal; word-spacing: normal; white-space: nowrap; float: none; direction: ltr; max-width: none; max-height: none; min-width: 0; min-height: 0; border: 0; margin: 0; padding: 1px 0} .MJXc-display {display: block; text-align: center; margin: 1em 0; padding: 0} .mjx-chtml[tabindex]:focus, body :focus .mjx-chtml[tabindex] {display: inline-table} .mjx-full-width {text-align: center; display: table-cell!important; width: 10000em} .mjx-math {display: inline-block; border-collapse: separate; border-spacing: 0} .mjx-math * {display: inline-block; -webkit-box-sizing: content-box!important; -moz-box-sizing: content-box!important; box-sizing: content-box!important; text-align: left} .mjx-numerator {display: block; text-align: center} .mjx-denominator {display: block; text-align: center} .MJXc-stacked {height: 0; position: relative} .MJXc-stacked > * {position: absolute} .MJXc-bevelled > * {display: inline-block} .mjx-stack {display: inline-block} .mjx-op {display: block} .mjx-under {display: table-cell} .mjx-over {display: block} .mjx-over > * {padding-left: 0px!important; padding-right: 0px!important} .mjx-under > * {padding-left: 0px!important; padding-right: 0px!important} .mjx-stack > .mjx-sup {display: block} .mjx-stack > .mjx-sub {display: block} .mjx-prestack > .mjx-presup {display: block} .mjx-prestack > .mjx-presub {display: block} .mjx-delim-h > .mjx-char {display: inline-block} .mjx-surd {vertical-align: top} .mjx-surd + .mjx-box {display: inline-flex} .mjx-mphantom * {visibility: hidden} .mjx-merror {background-color: #FFFF88; color: #CC0000; border: 1px solid #CC0000; padding: 2px 3px; font-style: normal; font-size: 90%} .mjx-annotation-xml {line-height: normal} .mjx-menclose > svg {fill: none; stroke: currentColor; overflow: visible} .mjx-mtr {display: table-row} .mjx-mlabeledtr {display: table-row} .mjx-mtd {display: table-cell; text-align: center} .mjx-label {display: table-row} .mjx-box {display: inline-block} .mjx-block {display: block} .mjx-span {display: inline} .mjx-char {display: block; white-space: pre} .mjx-itable {display: inline-table; width: auto} .mjx-row {display: table-row} .mjx-cell {display: table-cell} .mjx-table {display: table; width: 100%} .mjx-line {display: block; height: 0} .mjx-strut {width: 0; padding-top: 1em} .mjx-vsize {width: 0} .MJXc-space1 {margin-left: .167em} .MJXc-space2 {margin-left: .222em} .MJXc-space3 {margin-left: .278em} .mjx-test.mjx-test-display {display: table!important} .mjx-test.mjx-test-inline {display: inline!important; margin-right: -1px} .mjx-test.mjx-test-default {display: block!important; clear: both} .mjx-ex-box {display: inline-block!important; position: absolute; overflow: hidden; min-height: 0; max-height: none; padding: 0; border: 0; margin: 0; width: 1px; height: 60ex} .mjx-test-inline .mjx-left-box {display: inline-block; width: 0; float: left} .mjx-test-inline .mjx-right-box {display: inline-block; width: 0; float: right} .mjx-test-display .mjx-right-box {display: table-cell!important; width: 10000em!important; min-width: 0; max-width: none; padding: 0; border: 0; margin: 0} .MJXc-TeX-unknown-R {font-family: monospace; font-style: normal; font-weight: normal} .MJXc-TeX-unknown-I {font-family: monospace; font-style: italic; font-weight: normal} .MJXc-TeX-unknown-B {font-family: monospace; font-style: normal; font-weight: bold} .MJXc-TeX-unknown-BI {font-family: monospace; font-style: italic; font-weight: bold} .MJXc-TeX-ams-R {font-family: MJXc-TeX-ams-R,MJXc-TeX-ams-Rw} .MJXc-TeX-cal-B {font-family: MJXc-TeX-cal-B,MJXc-TeX-cal-Bx,MJXc-TeX-cal-Bw} .MJXc-TeX-frak-R {font-family: MJXc-TeX-frak-R,MJXc-TeX-frak-Rw} .MJXc-TeX-frak-B {font-family: MJXc-TeX-frak-B,MJXc-TeX-frak-Bx,MJXc-TeX-frak-Bw} .MJXc-TeX-math-BI {font-family: MJXc-TeX-math-BI,MJXc-TeX-math-BIx,MJXc-TeX-math-BIw} .MJXc-TeX-sans-R {font-family: MJXc-TeX-sans-R,MJXc-TeX-sans-Rw} .MJXc-TeX-sans-B {font-family: MJXc-TeX-sans-B,MJXc-TeX-sans-Bx,MJXc-TeX-sans-Bw} .MJXc-TeX-sans-I {font-family: MJXc-TeX-sans-I,MJXc-TeX-sans-Ix,MJXc-TeX-sans-Iw} .MJXc-TeX-script-R {font-family: MJXc-TeX-script-R,MJXc-TeX-script-Rw} .MJXc-TeX-type-R {font-family: MJXc-TeX-type-R,MJXc-TeX-type-Rw} .MJXc-TeX-cal-R {font-family: MJXc-TeX-cal-R,MJXc-TeX-cal-Rw} .MJXc-TeX-main-B {font-family: MJXc-TeX-main-B,MJXc-TeX-main-Bx,MJXc-TeX-main-Bw} .MJXc-TeX-main-I {font-family: MJXc-TeX-main-I,MJXc-TeX-main-Ix,MJXc-TeX-main-Iw} .MJXc-TeX-main-R {font-family: MJXc-TeX-main-R,MJXc-TeX-main-Rw} .MJXc-TeX-math-I {font-family: MJXc-TeX-math-I,MJXc-TeX-math-Ix,MJXc-TeX-math-Iw} .MJXc-TeX-size1-R {font-family: MJXc-TeX-size1-R,MJXc-TeX-size1-Rw} .MJXc-TeX-size2-R {font-family: MJXc-TeX-size2-R,MJXc-TeX-size2-Rw} .MJXc-TeX-size3-R {font-family: MJXc-TeX-size3-R,MJXc-TeX-size3-Rw} .MJXc-TeX-size4-R {font-family: MJXc-TeX-size4-R,MJXc-TeX-size4-Rw} .MJXc-TeX-vec-R {font-family: MJXc-TeX-vec-R,MJXc-TeX-vec-Rw} .MJXc-TeX-vec-B {font-family: MJXc-TeX-vec-B,MJXc-TeX-vec-Bx,MJXc-TeX-vec-Bw} @font-face {font-family: MJXc-TeX-ams-R; src: local('MathJax_AMS'), local('MathJax_AMS-Regular')} @font-face {font-family: MJXc-TeX-ams-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_AMS-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_AMS-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_AMS-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-cal-B; src: local('MathJax_Caligraphic Bold'), local('MathJax_Caligraphic-Bold')} @font-face {font-family: MJXc-TeX-cal-Bx; src: local('MathJax_Caligraphic'); font-weight: bold} @font-face {font-family: MJXc-TeX-cal-Bw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Caligraphic-Bold.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Caligraphic-Bold.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Caligraphic-Bold.otf') format('opentype')} @font-face {font-family: MJXc-TeX-frak-R; src: local('MathJax_Fraktur'), local('MathJax_Fraktur-Regular')} @font-face {font-family: MJXc-TeX-frak-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Fraktur-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Fraktur-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Fraktur-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-frak-B; src: local('MathJax_Fraktur Bold'), local('MathJax_Fraktur-Bold')} @font-face {font-family: MJXc-TeX-frak-Bx; src: local('MathJax_Fraktur'); font-weight: bold} @font-face {font-family: MJXc-TeX-frak-Bw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Fraktur-Bold.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Fraktur-Bold.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Fraktur-Bold.otf') format('opentype')} @font-face {font-family: MJXc-TeX-math-BI; src: local('MathJax_Math BoldItalic'), local('MathJax_Math-BoldItalic')} @font-face {font-family: MJXc-TeX-math-BIx; src: local('MathJax_Math'); font-weight: bold; font-style: italic} @font-face {font-family: MJXc-TeX-math-BIw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Math-BoldItalic.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Math-BoldItalic.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Math-BoldItalic.otf') format('opentype')} @font-face {font-family: MJXc-TeX-sans-R; src: local('MathJax_SansSerif'), local('MathJax_SansSerif-Regular')} @font-face {font-family: MJXc-TeX-sans-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_SansSerif-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_SansSerif-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_SansSerif-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-sans-B; src: local('MathJax_SansSerif Bold'), local('MathJax_SansSerif-Bold')} @font-face {font-family: MJXc-TeX-sans-Bx; src: local('MathJax_SansSerif'); font-weight: bold} @font-face {font-family: MJXc-TeX-sans-Bw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_SansSerif-Bold.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_SansSerif-Bold.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_SansSerif-Bold.otf') format('opentype')} @font-face {font-family: MJXc-TeX-sans-I; src: local('MathJax_SansSerif Italic'), local('MathJax_SansSerif-Italic')} @font-face {font-family: MJXc-TeX-sans-Ix; src: local('MathJax_SansSerif'); font-style: italic} @font-face {font-family: MJXc-TeX-sans-Iw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_SansSerif-Italic.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_SansSerif-Italic.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_SansSerif-Italic.otf') format('opentype')} @font-face {font-family: MJXc-TeX-script-R; src: local('MathJax_Script'), local('MathJax_Script-Regular')} @font-face {font-family: MJXc-TeX-script-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Script-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Script-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Script-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-type-R; src: local('MathJax_Typewriter'), local('MathJax_Typewriter-Regular')} @font-face {font-family: MJXc-TeX-type-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Typewriter-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Typewriter-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Typewriter-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-cal-R; src: local('MathJax_Caligraphic'), local('MathJax_Caligraphic-Regular')} @font-face {font-family: MJXc-TeX-cal-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Caligraphic-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Caligraphic-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Caligraphic-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-main-B; src: local('MathJax_Main Bold'), local('MathJax_Main-Bold')} @font-face {font-family: MJXc-TeX-main-Bx; src: local('MathJax_Main'); font-weight: bold} @font-face {font-family: MJXc-TeX-main-Bw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Main-Bold.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Main-Bold.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Main-Bold.otf') format('opentype')} @font-face {font-family: MJXc-TeX-main-I; src: local('MathJax_Main Italic'), local('MathJax_Main-Italic')} @font-face {font-family: MJXc-TeX-main-Ix; src: local('MathJax_Main'); font-style: italic} @font-face {font-family: MJXc-TeX-main-Iw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Main-Italic.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Main-Italic.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Main-Italic.otf') format('opentype')} @font-face {font-family: MJXc-TeX-main-R; src: local('MathJax_Main'), local('MathJax_Main-Regular')} @font-face {font-family: MJXc-TeX-main-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Main-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Main-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Main-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-math-I; src: local('MathJax_Math Italic'), local('MathJax_Math-Italic')} @font-face {font-family: MJXc-TeX-math-Ix; src: local('MathJax_Math'); font-style: italic} @font-face {font-family: MJXc-TeX-math-Iw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Math-Italic.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Math-Italic.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Math-Italic.otf') format('opentype')} @font-face {font-family: MJXc-TeX-size1-R; src: local('MathJax_Size1'), local('MathJax_Size1-Regular')} @font-face {font-family: MJXc-TeX-size1-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Size1-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Size1-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Size1-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-size2-R; src: local('MathJax_Size2'), local('MathJax_Size2-Regular')} @font-face {font-family: MJXc-TeX-size2-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Size2-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Size2-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Size2-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-size3-R; src: local('MathJax_Size3'), local('MathJax_Size3-Regular')} @font-face {font-family: MJXc-TeX-size3-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Size3-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Size3-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Size3-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-size4-R; src: local('MathJax_Size4'), local('MathJax_Size4-Regular')} @font-face {font-family: MJXc-TeX-size4-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Size4-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Size4-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Size4-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-vec-R; src: local('MathJax_Vector'), local('MathJax_Vector-Regular')} @font-face {font-family: MJXc-TeX-vec-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Vector-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Vector-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Vector-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-vec-B; src: local('MathJax_Vector Bold'), local('MathJax_Vector-Bold')} @font-face {font-family: MJXc-TeX-vec-Bx; src: local('MathJax_Vector'); font-weight: bold} @font-face {font-family: MJXc-TeX-vec-Bw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Vector-Bold.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Vector-Bold.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Vector-Bold.otf') format('opentype')} </style></p>
A quick post on a probably-real inadequate equilibrium mostly inspired by trying to think through what happened to Chance the Rapper.  Potentially ironic artifact if it accrues karma. 1. The sculptor's garden A sculptor worked in solitude for years, carving strange figures in his remote garden. Most of his statues failed: some cracked in winter, others looked wrong against the landscape. But occasionally, very rarely, one seemed to work. The first visitors stumbled upon the garden by accident. They found themselves stopped by his angels—figures that somehow held both sorrow and joy, wings that seemed about to flitter.  Word traveled slowly. More visitors came, drawn by something they couldn't quite name. The sculptor felt recognized for the first time. Not famous—but understood. His private work had somehow become communicable. He carved more angels, trying to understand what made these particular statues resonate. As crowds grew, their attention shifted. They began photographing the angels from certain angles, comparing new works to old, developing favorites. They applauded. The sculptor, still believing he followed the same thread, unconsciously noted which details drew the longest contemplation, which angles prompted gasps. Years passed. The garden became famous. Tour buses arrived with guides explaining the "important" pieces. The sculptor produced angels of increasing technical perfection, each guaranteed to produce the proper response at the proper moment. The crowds applauded more reliably. The sculptor carved more reliably. Each reinforced the other. One morning, walking his garden alone before dawn, he saw his statues without the crowds. Without their reactions to guide him, he saw what he'd actually been making: the same angel, refined and repeated, each iteration more precisely calibrated to trigger the expected response. He wasn't carving anymore. He was manufacturing applause in the shape of angels.  And the crowds—they weren't looking at ang
1,086
1.4.1
Revision
false
null
null
CrosspostOutput
ZiacrsqoWHczctTBS
axrp-episode-42-owain-evans-on-llm-psychology
AXRP Episode 42 - Owain Evans on LLM Psychology
null
false
false
true
null
DgsGzjyBXN8XSK22q
null
true
false
false
false
Post
null
2025-06-06T20:20:03.549Z
null
false
false
2
2
2025-06-07T17:53:06.417Z
false
false
post
[]
null
null
Jo39rWe82XBRbrz5c
0
2
12
false
0.016679
null
false
false
2025-06-06T20:20:03.549Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
null
null
qgdGA4ZEyW7zNdK84
null
null
null
false
null
[]
null
7
0
2025-06-06T20:20:03.550Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
80
null
null
null
null
[ { "__typename": "Tag", "_id": "3NzdN6QpkpAuNvtt6", "adminOnly": false, "afBaseScore": 3, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 9, "canEditUserIds": null, "core": false, "createdAt": "2024-12-29T00:20:51.218Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI Psychology", "needsReview": false, "noindex": false, "postCount": 16, "score": 9, "shortName": null, "slug": "ai-psychology", "suggestedAsFilter": false, "userId": "g3EBjAowLk6KwbPC3", "voteCount": 1, "wikiOnly": false }, { "__typename": "Tag", "_id": "vjKs7Pvz3MbgMc75C", "adminOnly": false, "afBaseScore": null, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 0, "canEditUserIds": null, "core": false, "createdAt": "2021-03-26T12:39:55.451Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Audio", "needsReview": false, "noindex": false, "postCount": 125, "score": 0, "shortName": null, "slug": "audio", "suggestedAsFilter": false, "userId": "BpBzKEueak7J8vHNi", "voteCount": 0, "wikiOnly": false }, { "__typename": "Tag", "_id": "8Ec9rD286qNstoiGH", "adminOnly": false, "afBaseScore": 3, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 9, "canEditUserIds": null, "core": false, "createdAt": "2021-12-23T10:49:50.259Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AXRP", "needsReview": false, "noindex": false, "postCount": 60, "score": 9, "shortName": null, "slug": "axrp", "suggestedAsFilter": false, "userId": "4Kn3eZCPNB8gw4YSi", "voteCount": 1, "wikiOnly": false }, { "__typename": "Tag", "_id": "9DNZfxFvY5iKoZQbz", "adminOnly": false, "afBaseScore": null, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 0, "canEditUserIds": null, "core": false, "createdAt": "2019-11-13T22:47:01.189Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Interviews", "needsReview": false, "noindex": false, "postCount": 120, "score": 0, "shortName": null, "slug": "interviews", "suggestedAsFilter": false, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 0, "wikiOnly": false }, { "__typename": "Tag", "_id": "Zwv9eHi7KGg5KA9oM", "adminOnly": false, "afBaseScore": 9, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "EQNTWXLKMeWMp2FQS", "displayName": "Ben Pace" }, { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 21, "canEditUserIds": null, "core": false, "createdAt": "2020-06-27T20:43:34.869Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "EQNTWXLKMeWMp2FQS", "displayName": "Ben Pace" }, { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" }, { "_id": "eCk5iNu68fJeuwB4e", "displayName": "aproteinengine" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Introspection", "needsReview": false, "noindex": false, "postCount": 83, "score": 21, "shortName": null, "slug": "introspection", "suggestedAsFilter": false, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 3, "wikiOnly": false }, { "__typename": "Tag", "_id": "BhfefamXXee6c2CH8", "adminOnly": false, "afBaseScore": null, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 0, "canEditUserIds": null, "core": false, "createdAt": "2020-07-11T06:51:38.152Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Transcripts", "needsReview": false, "noindex": false, "postCount": 78, "score": 0, "shortName": null, "slug": "transcripts", "suggestedAsFilter": false, "userId": "qgdGA4ZEyW7zNdK84", "voteCount": 0, "wikiOnly": false }, { "__typename": "Tag", "_id": "sYm3HiWcfZvrGu3ui", "adminOnly": false, "afBaseScore": 2, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" } ] }, "baseScore": 12, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:22.097Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" }, { "_id": "sof55TPMQaeBaxhsS", "displayName": "tommylees112" }, { "_id": "AayjS8XzcnDKhGdTv", "displayName": "shark" }, { "_id": "HnALuwRdo6k9HLaMt", "displayName": "Alex Firssoff" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI", "needsReview": false, "noindex": false, "postCount": 12544, "score": 12, "shortName": null, "slug": "ai", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 4, "wikiOnly": false } ]
n7xK8hmp8XmQLWrNT
0
0
null
false
null
null
0
2
0
0
2
0
DgsGzjyBXN8XSK22q
danielfilan
2014-01-30T11:04:39.341Z
DanielFilan
DanielFilan
null
null
null
8,823
1,852
false
false
null
null
150
1,377
1
26
353
1
8
r38pkCm7wF4M44MDQ
User
easy-going
null
true
[ "alignmentForum", "trustLevel1", "alignmentVoters", "canModeratePersonal", "tagManager" ]
null
null
ZiacrsqoWHczctTBS
SocialPreviewType
Jo39rWe82XBRbrz5c
<p><a href="https://youtu.be/3D4pgIKR4cQ">YouTube link</a></p><p>Earlier this year, the paper “Emergent Misalignment” made the rounds on AI x-risk social media for seemingly showing LLMs generalizing from ‘misaligned’ training data of insecure code to acting comically evil in response to innocuous questions. In this episode, I chat with one of the authors of that paper, Owain Evans, about that research as well as other work he’s done to understand the psychology of large language models.</p><p>Topics we discuss:</p> <ul> <li><a href="#why-introspection">Why introspection?</a></li> <li><a href="#experiments-in-li">Experiments in “Looking Inward”</a></li> <li><a href="#why-fine-tune-for-introspection">Why fine-tune for introspection?</a></li> <li><a href="#does-li-test-introspection">Does “Looking Inward” test introspection, or something else?</a></li> <li><a href="#interpret-li-results">Interpreting the results of “Looking Inward”</a></li> <li><a href="#limitations-to-introspection">Limitations to introspection?</a></li> <li><a href="#tmay-other-papers">“Tell me about yourself”, and its relation to other papers</a></li> <li><a href="#backdoor-results">Backdoor results</a></li> <li><a href="#em">Emergent misalignment</a></li> <li><a href="#so-hammy-rarely-evil">Why so hammy, and so infrequently evil?</a></li> <li><a href="#why-em">Why emergent misalignment?</a></li> <li><a href="#em-other-types">Emergent misalignment and other types of misalignment</a></li> <li><a href="#is-em-good-news">Is emergent misalignment good news?</a></li> <li><a href="#em-follow-up">Follow-up work to “Emergent Misalignment”</a></li> <li><a href="#em-reception">Reception of “Emergent Misalignment” vs other papers</a></li> <li><a href="#evil-numbers">Evil numbers</a></li> <li><a href="#following-owains-research">Following Owain’s research</a></li> </ul> <p><strong>Daniel Filan</strong> (00:00:09): Hello everybody. In this episode I’ll be speaking with Owain Evans. Owain is the research lead at <a href="https://www.truthfulai.org/">Truthful AI</a>, an AI safety research non-profit. Previously, papers he’s worked on have included <a href="https://arxiv.org/abs/2109.07958">“TruthfulQA”</a> and <a href="https://arxiv.org/abs/2309.12288">“The Reversal Curse”</a>. To read a transcript of this episode, you can go to <a href="https://axrp.net/">axrp.net</a>, you can become a patron at <a href="https://patreon.com/axrpodcast">patreon.com/axrpodcast</a>, or you can give feedback about the episode at <a href="axrp.fyi">axrp.fyi</a>.</p><p>(00:00:34): Okay, Owain, welcome to the podcast.</p><p><strong>Owain Evans</strong> (00:00:36): Thanks for having me.</p> <h2>Why introspection? <a name="why-introspection"></a></h2> <p><strong>Daniel Filan</strong> (00:00:37): Yeah. So first up I’d like to talk about your paper <a href="https://arxiv.org/abs/2410.13787">“Looking Inward: Language Models Can Learn About Themselves by Introspection”</a>. So the first two authors are <a href="https://ac.felixbinder.net/">Felix J. Binder</a> and <a href="https://jameschua.net/about/">James Chua</a>, a few more, you’re the last author. Can you tell us just: at a high level, what’s this paper doing? What’s it about?</p><p><strong>Owain Evans</strong> (00:00:59): Yeah, sure. So part of what we are interested in here is: can language models tell us things about their internal states where the knowledge or information that they’re getting is not coming solely from their training data? So if a language model right now tells someone, “My goal is X,” or, “I have a desire for X,” I think people would have a tendency to explain this as, well, the training data... </p>
YouTube link Earlier this year, the paper “Emergent Misalignment” made the rounds on AI x-risk social media for seemingly showing LLMs generalizing from ‘misaligned’ training data of insecure code to acting comically evil in response to innocuous questions. In this episode, I chat with one of the authors of that paper, Owain Evans, about that research as well as other work he’s done to understand the psychology of large language models. Topics we discuss: * Why introspection? * Experiments in “Looking Inward” * Why fine-tune for introspection? * Does “Looking Inward” test introspection, or something else? * Interpreting the results of “Looking Inward” * Limitations to introspection? * “Tell me about yourself”, and its relation to other papers * Backdoor results * Emergent misalignment * Why so hammy, and so infrequently evil? * Why emergent misalignment? * Emergent misalignment and other types of misalignment * Is emergent misalignment good news? * Follow-up work to “Emergent Misalignment” * Reception of “Emergent Misalignment” vs other papers * Evil numbers * Following Owain’s research Daniel Filan (00:00:09): Hello everybody. In this episode I’ll be speaking with Owain Evans. Owain is the research lead at Truthful AI, an AI safety research non-profit. Previously, papers he’s worked on have included “TruthfulQA” and “The Reversal Curse”. To read a transcript of this episode, you can go to axrp.net, you can become a patron at patreon.com/axrpodcast, or you can give feedback about the episode at axrp.fyi. (00:00:34): Okay, Owain, welcome to the podcast. Owain Evans (00:00:36): Thanks for having me. Why introspection? Daniel Filan (00:00:37): Yeah. So first up I’d like to talk about your paper “Looking Inward: Language Models Can Learn About Themselves by Introspection”. So the first two authors are Felix J. Binder and James Chua, a few more, you’re the last author. Can you tell us just: at a high level, what’s this paper doing? What’s it abou
19,884
1.1.0
Revision
false
null
null
CrosspostOutput
JdrmKSzsWmRbgMumf
apply-now-to-human-aligned-ai-summer-school-2025
Apply now to Human-Aligned AI Summer School 2025
null
false
false
true
null
XQngh2XStFhHZ6Gou
[ { "__typename": "CoauthorStatusOutput", "confirmed": true, "requested": false, "userId": "hQh9YHbCpGnyGmWD7" }, { "__typename": "CoauthorStatusOutput", "confirmed": true, "requested": false, "userId": "JnNixf4smAHwLeqE3" } ]
true
false
false
false
Post
https://humanaligned.ai/2025
2025-06-06T19:31:40.085Z
null
false
false
2
2
null
false
false
linkpost
[ "hQh9YHbCpGnyGmWD7" ]
null
null
CAMAqsce2B29eyuR9
1
5
27
false
0.02059
null
false
false
2025-06-23T13:52:39.799Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
qgdGA4ZEyW7zNdK84
null
null
null
false
null
[]
null
17
1
2025-06-23T13:52:39.515Z
false
false
norm-enforcing
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[ { "__typename": "User", "_id": "hQh9YHbCpGnyGmWD7", "afCommentCount": 5, "afKarma": 41, "afPostCount": 1, "commentCount": 18, "createdAt": "2019-02-06T13:37:31.215Z", "deleted": false, "displayName": "Tomáš Gavenčiak", "fullName": null, "htmlBio": "<p>A researcher in CS theory, AI safety and other stuff.</p>", "isAdmin": false, "jobTitle": null, "karma": 218, "organization": null, "postCount": 1, "previousDisplayName": null, "profileImageId": null, "reviewedByUserId": "nLbwLhBaQeG6tCNDN", "sequenceCount": 0, "slug": "tomas-gavenciak", "spamRiskScore": 1, "tagRevisionCount": 0, "username": "tomas-gavenciak" }, { "__typename": "User", "_id": "JnNixf4smAHwLeqE3", "afCommentCount": 69, "afKarma": 1116, "afPostCount": 21, "commentCount": 290, "createdAt": "2017-12-29T10:11:29.037Z", "deleted": false, "displayName": "Jan_Kulveit", "fullName": null, "htmlBio": "<p>My current research interests:<br><br>1. Alignment in systems which are complex and messy, composed of both humans and AIs?<br>Recommended texts: <a href=\"https://www.lesswrong.com/posts/pZhEQieM9otKXhxmd/gradual-disempowerment-systemic-existential-risks-from\">Gradual Disempowerment</a>,<a href=\"https://www.lesswrong.com/posts/BTApNmv7s6RTGxeP4/cyborg-periods-there-will-be-multiple-ai-transitions\"> Cyborg Periods</a><br><br>2. Actually good mathematized theories of cooperation and coordination<br>Recommended texts: <a href=\"https://www.lesswrong.com/posts/xud7Mti9jS4tbWqQE/hierarchical-agency-a-missing-piece-in-ai-alignment\">Hierarchical Agency: A Missing Piece in AI Alignment</a>, <a href=\"https://www.lesswrong.com/posts/9GyniEBaN3YYTqZXn/the-self-unalignment-problem\">The self-unalignment problem</a> or <a href=\"https://www.lesswrong.com/posts/5tYTKX4pNpiG4vzYg/towards-a-scale-free-theory-of-intelligent-agency\">Towards a scale-free theory of intelligent agency</a> (by Richard Ngo)<br><br>3. Active inference &amp; Bounded rationality<br>Recommended texts: <a href=\"https://www.lesswrong.com/posts/YEioD8YLgxih3ydxP/why-simulator-ais-want-to-be-active-inference-ais\">Why Simulator AIs want to be Active Inference AIs</a>, <a href=\"https://openreview.net/forum?id=4Ft7DcrjdO\">Free-Energy Equilibria: Toward a Theory of Interactions Between Boundedly-Rational Agents</a><strong>,&nbsp;</strong> <a href=\"https://www.lesswrong.com/posts/3fkBWpE4f9nYbdf7E/multi-agent-predictive-minds-and-ai-alignment\">Multi-agent predictive minds and AI alignment</a> (old but still mostly holds)<br><br>&nbsp;4. LLM psychology and sociology: <a href=\"https://www.lesswrong.com/posts/zuXo9imNKYspu9HGv/a-three-layer-model-of-llm-psychology\">A Three-Layer Model of LLM Psychology</a>, <a href=\"https://www.lesswrong.com/posts/wQKskToGofs4osdJ3/the-pando-problem-rethinking-ai-individuality\">The Pando Problem: Rethinking AI Individuality</a>, <a href=\"https://www.lesswrong.com/posts/kFCu3batN8k8mwtmh/the-cave-allegory-revisited-understanding-gpt-s-worldview\">The Cave Allegory Revisited: Understanding GPT's Worldview</a><br><br>5. Macrostrategy &amp; macrotactics &amp; deconfusion: <a href=\"https://www.lesswrong.com/posts/XrGwrC9n8sDgXimcJ/hinges-and-crises\">Hinges and crises</a>, <a href=\"https://www.lesswrong.com/posts/BTApNmv7s6RTGxeP4/cyborg-periods-there-will-be-multiple-ai-transitions\">Cyborg Periods</a> again, <a href=\"https://www.lesswrong.com/posts/jrKftFZMZjvNdQLNR/box-inversion-revisited\">Box inversion revisited</a>, <a href=\"https://www.lesswrong.com/posts/b9sGz74ayftqPBDYv/the-space-of-systems-and-the-space-of-maps\">The space of systems and the space of maps</a>, <a href=\"https://www.lesswrong.com/posts/sam4ehxHgnJEGCKed/lessons-from-convergent-evolution-for-ai-alignment\">Lessons from Convergent Evolution for AI Alignment</a>, <a href=\"https://www.lesswrong.com/posts/cHJxSJ4jBmBRGtbaE/continuity-assumptions\">Continuity Assumptions</a><br><br>Also I occasionally write about epistemics: <a href=\"https://www.lesswrong.com/posts/4gDbqL3Tods8kHDqs/limits-to-legibility\">Limits to Legibility</a>, <a href=\"https://www.lesswrong.com/posts/FGHKwEGKCfDzcxZuj/conceptual-rounding-errors\">Conceptual Rounding Errors</a></p><p>Researcher at Alignment of Complex Systems Research Group (<a href=\"http://acsresearch.org\">acsresearch.org</a>), Centre for Theoretical Studies, Charles University in Prague. &nbsp;Formerly research fellow Future of Humanity Institute, Oxford University<br><br>Previously I was a researcher in physics, studying phase transitions, network science and complex systems.</p>", "isAdmin": false, "jobTitle": null, "karma": 5974, "organization": null, "postCount": 54, "previousDisplayName": null, "profileImageId": null, "reviewedByUserId": "r38pkCm7wF4M44MDQ", "sequenceCount": 0, "slug": "jan_kulveit", "spamRiskScore": 1, "tagRevisionCount": 0, "username": "Jan_Kulveit" } ]
2
null
null
null
null
[ { "__typename": "Tag", "_id": "6zBEfFYJxhSEcchbR", "adminOnly": false, "afBaseScore": 3, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 9, "canEditUserIds": null, "core": false, "createdAt": "2022-06-09T19:10:50.755Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI Alignment Fieldbuilding", "needsReview": false, "noindex": false, "postCount": 359, "score": 9, "shortName": null, "slug": "ai-alignment-fieldbuilding", "suggestedAsFilter": false, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 1, "wikiOnly": false }, { "__typename": "Tag", "_id": "fF9GEdWXKJ3z73TmB", "adminOnly": false, "afBaseScore": 9, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "EQNTWXLKMeWMp2FQS", "displayName": "Ben Pace" }, { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 22, "canEditUserIds": null, "core": false, "createdAt": "2020-06-09T16:57:01.474Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "EQNTWXLKMeWMp2FQS", "displayName": "Ben Pace" }, { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" }, { "_id": "t46uLRSbDziEcKmev", "displayName": "Kriz Tahimic" }, { "_id": "dRaCtsAWxk7sgirSY", "displayName": "Jordan Morgan" }, { "_id": "xF5nfdddHjFThHy49", "displayName": "[email protected]" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Scholarship & Learning", "needsReview": false, "noindex": false, "postCount": 361, "score": 22, "shortName": null, "slug": "scholarship-and-learning", "suggestedAsFilter": false, "userId": "qgdGA4ZEyW7zNdK84", "voteCount": 5, "wikiOnly": false }, { "__typename": "Tag", "_id": "sYm3HiWcfZvrGu3ui", "adminOnly": false, "afBaseScore": 2, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" } ] }, "baseScore": 12, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:22.097Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" }, { "_id": "sof55TPMQaeBaxhsS", "displayName": "tommylees112" }, { "_id": "AayjS8XzcnDKhGdTv", "displayName": "shark" }, { "_id": "HnALuwRdo6k9HLaMt", "displayName": "Alex Firssoff" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI", "needsReview": false, "noindex": false, "postCount": 12544, "score": 12, "shortName": null, "slug": "ai", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 4, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
5
0
0
4
0
XQngh2XStFhHZ6Gou
vojtakovarik
2017-12-21T20:07:16.507Z
VojtaKovarik
VojtaKovarik
null
null
Vojtech Kovarik
757
259
false
false
<p>My original background is in mathematics (analysis, topology, Banach spaces) and game theory (imperfect information games). Nowadays, I do AI alignment research (mostly systemic risks, sometimes pondering about "consequentionalist reasoning").</p>
null
null
25
154
1
14
79
1
0
r38pkCm7wF4M44MDQ
User
norm-enforcing
null
true
[ "alignmentVoters", "alignmentForum", "canModeratePersonal" ]
null
null
JdrmKSzsWmRbgMumf
https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/JdrmKSzsWmRbgMumf/kp5afnphyh0d7pbq7vdj
SocialPreviewType
CAMAqsce2B29eyuR9
<figure class="image"><img src="https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/JdrmKSzsWmRbgMumf/uxxrgwh9sv6a84yhb1m4" srcset="https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/JdrmKSzsWmRbgMumf/yhgqmib3yqqh6razgdoz 230w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/JdrmKSzsWmRbgMumf/zzdvjglyyen59vcco7ja 460w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/JdrmKSzsWmRbgMumf/phsj6gubtbp0p5udoazu 690w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/JdrmKSzsWmRbgMumf/aehpi65k5wzxd2hnalpe 920w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/JdrmKSzsWmRbgMumf/jz92fpewjbhn76iepsxo 1150w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/JdrmKSzsWmRbgMumf/f4r9mrvs9an0kbkl6yor 1380w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/JdrmKSzsWmRbgMumf/sidplsqcrq7yzy56clg3 1610w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/JdrmKSzsWmRbgMumf/xhx30spj0bmpfnrrwhgt 1840w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/JdrmKSzsWmRbgMumf/vkkvqvtji5apncnfbwg0 2070w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/JdrmKSzsWmRbgMumf/mc0f6craedwtnyibw1z8 2229w"></figure><p>Join us at the fifth&nbsp;<a href="https://humanaligned.ai/2025/"><u>Human-Aligned AI Summer School</u></a> in Prague from 22<sup>nd</sup> to 25<sup>th</sup> July 2025!</p><p><i><strong>Update: </strong>We have now confirmed our speaker list with excellent speakers -- see below!</i><br><i><strong>Update: </strong>We still have capacity for more excellent participants as of late June. Please help us spread the word to people who would be a good fit for the school. We also have some additional funding to financially support some of the participants.</i></p><p>We will meet for four intensive days of talks, workshops, and discussions covering latest trends in AI alignment research and broader framings of AI risk.</p><p><a href="https://form.typeform.com/to/km5tcMFf"><i><strong><u>Apply now</u></strong></i></a><i><u>, applications are evaluated on a rolling basis.</u></i><br><br><u>The intended audience of the school are people interested in learning more about the AI alignment topics, PhD students, researchers working in ML/AI, and talented students.&nbsp;</u></p><h2><strong>What to expect</strong></h2><p>The school is focused on teaching and exploring approaches and frameworks, less on presentation of the latest research results. Much of the content is technical – it is assumed the attendees understand current ML approaches and some of the underlying theoretical frameworks.</p><p>This year, we explore AI alignment through three core areas:</p><ul><li><strong>Technical alignment research</strong>: We'll examine current technical approaches including behavioral evaluations, mechanistic interpretability, scalable oversight, and model organisms of misalignment. We'll discuss recent developments in these areas and what they tell us about the potential and limitations of these methods.</li><li><strong>AI strategy and systemic alignment</strong>: We'll explore topics such as timeline considerations, strategic and governance challenges around powerful AI, economic models of AI development, and risks of gradual disempowerment in a post-AGI world. We'll focus on building overall understanding and how these considerations can inform technical research.</li><li><strong>Foundational frameworks</strong>: We'll visit research areas relevant to recent AI developments, such as multi-agent dynamics and cooperation, theories of agency, bounded rationality, and realistic models of goal-directed behavior. These frameworks help us understand what alignment means in complex environments containing both AIs and humans, and how to develop appropriate techniques.</li></ul><p>The school consists of lectures and topical series, focused smaller-group workshops and discussions, expert panels, and opportunities for networking, project brainstorming and informal discussions. A... </p>
Join us at the fifth Human-Aligned AI Summer School in Prague from 22nd to 25th July 2025! Update: We have now confirmed our speaker list with excellent speakers -- see below! Update: We still have capacity for more excellent participants as of late June. Please help us spread the word to people who would be a good fit for the school. We also have some additional funding to financially support some of the participants. We will meet for four intensive days of talks, workshops, and discussions covering latest trends in AI alignment research and broader framings of AI risk. Apply now, applications are evaluated on a rolling basis. The intended audience of the school are people interested in learning more about the AI alignment topics, PhD students, researchers working in ML/AI, and talented students.  What to expect The school is focused on teaching and exploring approaches and frameworks, less on presentation of the latest research results. Much of the content is technical – it is assumed the attendees understand current ML approaches and some of the underlying theoretical frameworks. This year, we explore AI alignment through three core areas: * Technical alignment research: We'll examine current technical approaches including behavioral evaluations, mechanistic interpretability, scalable oversight, and model organisms of misalignment. We'll discuss recent developments in these areas and what they tell us about the potential and limitations of these methods. * AI strategy and systemic alignment: We'll explore topics such as timeline considerations, strategic and governance challenges around powerful AI, economic models of AI development, and risks of gradual disempowerment in a post-AGI world. We'll focus on building overall understanding and how these considerations can inform technical research. * Foundational frameworks: We'll visit research areas relevant to recent AI developments, such as multi-agent dynamics and cooperation, theories of agency, bound
479
1.7.1
Revision
false
true
null
CrosspostOutput
5dpSANYFH9PcvywyQ
the-common-pile-and-comma-v-01
The Common Pile and Comma-v.01
null
false
false
false
null
6G9PqE5CGbd5GMQBe
null
true
false
false
false
Post
null
2025-06-06T19:20:25.910Z
null
false
false
2
2
2025-06-07T17:53:18.648Z
false
false
post
[]
null
null
Xt96DiZsMYK7EW6Tk
0
1
3
false
0.010026
null
false
false
2025-06-06T19:20:25.910Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
qgdGA4ZEyW7zNdK84
null
null
null
false
null
[]
null
0
0
2025-06-06T19:14:26.442Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
1
null
null
null
null
[ { "__typename": "Tag", "_id": "KmgkrftQuX7jmjjp5", "adminOnly": false, "afBaseScore": 3, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 9, "canEditUserIds": null, "core": false, "createdAt": "2021-09-24T14:01:59.395Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Language Models (LLMs)", "needsReview": false, "noindex": false, "postCount": 840, "score": 9, "shortName": null, "slug": "language-models-llms", "suggestedAsFilter": false, "userId": "Sp5wM4aRAhNERd4oY", "voteCount": 1, "wikiOnly": false }, { "__typename": "Tag", "_id": "sYm3HiWcfZvrGu3ui", "adminOnly": false, "afBaseScore": 2, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" } ] }, "baseScore": 12, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:22.097Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" }, { "_id": "sof55TPMQaeBaxhsS", "displayName": "tommylees112" }, { "_id": "AayjS8XzcnDKhGdTv", "displayName": "shark" }, { "_id": "HnALuwRdo6k9HLaMt", "displayName": "Alex Firssoff" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI", "needsReview": false, "noindex": false, "postCount": 12544, "score": 12, "shortName": null, "slug": "ai", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 4, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
1
0
0
0
0
6G9PqE5CGbd5GMQBe
jadael
2013-04-01T16:39:15.772Z
Jadael
Trevor Hill-Hand
null
null
107
0
false
false
<p><a href="http://hillhand.com">hillhand.com</a></p> <p>I deeply value evidence, reason, and letting people draw their own conclusions. I dislike telling anyone what to think or do.</p> <p>I believe you, yes YOU, are capable of reading and understanding everything you want to read and understand.</p>
null
null
2
52
1
0
0
1
0
r38pkCm7wF4M44MDQ
User
null
null
null
[ "canModeratePersonal" ]
null
null
5dpSANYFH9PcvywyQ
SocialPreviewType
Xt96DiZsMYK7EW6Tk
<p><a href="https://www.engadget.com/ai/it-turns-out-you-can-train-ai-models-without-copyrighted-material-174016619.html">https://www.engadget.com/ai/it-turns-out-you-can-train-ai-models-without-copyrighted-material-174016619.html</a></p> <blockquote> <p>...the LLM's dataset uses only public domain and openly licensed material.</p> <p>The <a href="https://github.com/r-three/common-pile/blob/main/paper.pdf">paper</a> (via The Washington Post) was a collaboration between 14 different institutions. The authors represent universities like MIT, Carnegie Mellon and the University of Toronto. Nonprofits like Vector Institute and the Allen Institute for AI also contributed.</p> <p>The group built an 8 TB ethically-sourced dataset. Among the data was a set of 130,000 books in the Library of Congress. After inputting the material, they trained a seven-billion-parameter large language model (LLM) on that data. The result? It performed about as well as Meta's similarly sized Llama 2-7B from 2023. The team didn't publish benchmarks comparing its results to today's top models.</p> </blockquote> <ul> <li>GitHub Repo: <a href="https://github.com/r-three/common-pile/tree/main">https://github.com/r-three/common-pile/tree/main</a></li> <li>HuggingFace: <a href="https://huggingface.co/common-pile">https://huggingface.co/common-pile</a></li> </ul> <p>This was wonderful news for me today, if this works and is what they say it is. I have several hobby projects that I've basically kept locked in a vault because I've felt there was no ethical LLM available to encourage people use for them, even just for 'fun" projects.</p>
https://www.engadget.com/ai/it-turns-out-you-can-train-ai-models-without-copyrighted-material-174016619.html > ...the LLM's dataset uses only public domain and openly licensed material. > > The paper (via The Washington Post) was a collaboration between 14 different institutions. The authors represent universities like MIT, Carnegie Mellon and the University of Toronto. Nonprofits like Vector Institute and the Allen Institute for AI also contributed. > > The group built an 8 TB ethically-sourced dataset. Among the data was a set of 130,000 books in the Library of Congress. After inputting the material, they trained a seven-billion-parameter large language model (LLM) on that data. The result? It performed about as well as Meta's similarly sized Llama 2-7B from 2023. The team didn't publish benchmarks comparing its results to today's top models. * GitHub Repo: https://github.com/r-three/common-pile/tree/main * HuggingFace: https://huggingface.co/common-pile This was wonderful news for me today, if this works and is what they say it is. I have several hobby projects that I've basically kept locked in a vault because I've felt there was no ethical LLM available to encourage people use for them, even just for 'fun" projects.
176
1.3.0
Revision
false
null
null
CrosspostOutput
FBqYAHCsnjLvhAcd5
maximal-curiousity-is-not-useful
Maximal Curiousity is Not Useful
null
false
false
false
null
6YMGB7PXkfoZyhqii
null
true
false
false
false
Post
null
2025-06-06T19:08:40.698Z
null
false
false
2
2
2025-06-07T17:55:53.509Z
false
false
post
[]
null
null
W3XeM42hJnKYcnDuw
0
4
10
false
0.01541
null
false
false
2025-06-06T19:08:40.698Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
qgdGA4ZEyW7zNdK84
null
null
null
false
null
[]
null
2
0
2025-06-01T18:03:56.722Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
2
null
null
null
null
[ { "__typename": "Tag", "_id": "sYm3HiWcfZvrGu3ui", "adminOnly": false, "afBaseScore": 2, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" } ] }, "baseScore": 12, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:22.097Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" }, { "_id": "sof55TPMQaeBaxhsS", "displayName": "tommylees112" }, { "_id": "AayjS8XzcnDKhGdTv", "displayName": "shark" }, { "_id": "HnALuwRdo6k9HLaMt", "displayName": "Alex Firssoff" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI", "needsReview": false, "noindex": false, "postCount": 12544, "score": 12, "shortName": null, "slug": "ai", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 4, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
4
0
0
2
0
6YMGB7PXkfoZyhqii
max-niederman-1
2022-06-25T01:16:16.084Z
niederman
Max Niederman
null
null
Max Niederman
214
0
false
false
null
null
4
11
0
0
0
1
0
r38pkCm7wF4M44MDQ
User
null
null
null
[ "canModeratePersonal" ]
null
null
FBqYAHCsnjLvhAcd5
SocialPreviewType
W3XeM42hJnKYcnDuw
<p>The other day I talked to a friend of mine about AI x-risk. My friend is a fan of Elon Musk, and argued roughly that x-risk was not a problem, because xAI will win the AI race and deploy a "maximally curious" superintelligence.</p><p>Setting aside the issue of whether xAI will win the race, it seemed obvious to me that this is not a useful way of going about alignment. Furthermore, I believe that the suggestion of maximal curiousity is actively harmful.</p><h1>What Does 'Curiousity' Mean?</h1><p>The most common reason to be interested in truth is that it allows you to make useful predictions. In this case, truths are valuable instrumentally based on how you can apply them to achieving your true goals.</p><p>This is not the only reason to value truth. A child who loves dinosaurs does not learn about them because he expects to apply the knowledge to his daily life. He is interested regardless of whether he expects to use what he learns. This is curiousity: the human phenomenon of valuing truth intrinsically.</p><p>However, humans are not equally curious about all truths. The child is much more interested in dinosaur facts than he is in his history homework. Someone else might find dinosaurs dull but history fascinating. Curiousity is determined by a <i>subjective</i> preference for some truths over others.</p><p>What subjective sense of curiousity should we give a superintelligence? No simple formalism will cut it, because humans curiousity is fundamentally tied to human values. We are far more curious about what Arthur said to Bob about Charlie than we are about the number of valence electrons of carbon, despite the latter being a vastly more important and fundamental fact.</p><p>To make a maximally curious AI, for the conventional meaning of 'curious', we need to somehow</p><ol><li>reconcile the conflicting preferences for some ideas over others of the many humans around the world,<span class="footnote-reference" data-footnote-reference="" data-footnote-index="1" data-footnote-id="itx03adp3w" role="doc-noteref" id="fnrefitx03adp3w"><sup><a href="#fnitx03adp3w">[1]</a></sup></span></li><li>reify those preferences into an explicit training objective, and</li><li>ensure that the resulting AI actually agrees with those preferences.</li></ol><p>Sound familiar? It should, because replacing 'ideas' with 'outcomes' yields a description of the alignment problem itself. <strong>It is no easier to capture the messy human concept of curiousity than it is to capture the messy human concept of good.</strong></p><h1>True Curiousity is Not Good Enough</h1><p>Even if we could solve the curiousity alignment problem, this does not solve alignment. Humans have many values besides curiousity, and a maximal... </p>
The other day I talked to a friend of mine about AI x-risk. My friend is a fan of Elon Musk, and argued roughly that x-risk was not a problem, because xAI will win the AI race and deploy a "maximally curious" superintelligence. Setting aside the issue of whether xAI will win the race, it seemed obvious to me that this is not a useful way of going about alignment. Furthermore, I believe that the suggestion of maximal curiousity is actively harmful. What Does 'Curiousity' Mean? The most common reason to be interested in truth is that it allows you to make useful predictions. In this case, truths are valuable instrumentally based on how you can apply them to achieving your true goals. This is not the only reason to value truth. A child who loves dinosaurs does not learn about them because he expects to apply the knowledge to his daily life. He is interested regardless of whether he expects to use what he learns. This is curiousity: the human phenomenon of valuing truth intrinsically. However, humans are not equally curious about all truths. The child is much more interested in dinosaur facts than he is in his history homework. Someone else might find dinosaurs dull but history fascinating. Curiousity is determined by a subjective preference for some truths over others. What subjective sense of curiousity should we give a superintelligence? No simple formalism will cut it, because humans curiousity is fundamentally tied to human values. We are far more curious about what Arthur said to Bob about Charlie than we are about the number of valence electrons of carbon, despite the latter being a vastly more important and fundamental fact. To make a maximally curious AI, for the conventional meaning of 'curious', we need to somehow 1. reconcile the conflicting preferences for some ideas over others of the many humans around the world,[1] 2. reify those preferences into an explicit training objective, and 3. ensure that the resulting AI actually agrees with those pre
603
1.5.0
Revision
false
null
null
CrosspostOutput
qw36iLrGyFn6mdwoa
making-deals-with-ais-a-tournament-experiment-with-a-bounty
Making deals with AIs: A tournament experiment with a bounty
null
false
false
false
null
mDGwuonuf3imbpbYz
[ { "__typename": "CoauthorStatusOutput", "confirmed": true, "requested": false, "userId": "MXzEw56rk32JjM2Gp" } ]
true
false
false
false
Post
2025-06-06T18:51:05.164Z
null
false
false
2
2
2025-06-07T17:53:59.911Z
false
false
post
[ "MXzEw56rk32JjM2Gp" ]
null
null
2L4NryupJy2CPvdvr
0
6
13
false
0.017685
null
false
false
2025-06-06T18:51:05.164Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
qgdGA4ZEyW7zNdK84
null
null
null
false
null
[]
null
1
0
2025-06-06T14:31:16.839Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[ { "__typename": "User", "_id": "MXzEw56rk32JjM2Gp", "afCommentCount": 0, "afKarma": 0, "afPostCount": 0, "commentCount": 160, "createdAt": "2010-02-04T03:26:20.706Z", "deleted": false, "displayName": "Xodarap", "fullName": null, "htmlBio": "", "isAdmin": false, "jobTitle": null, "karma": 807, "organization": null, "postCount": 15, "previousDisplayName": null, "profileImageId": null, "reviewedByUserId": "r38pkCm7wF4M44MDQ", "sequenceCount": 0, "slug": "xodarap", "spamRiskScore": 1, "tagRevisionCount": 0, "username": "Xodarap" } ]
10
null
null
null
null
[ { "__typename": "Tag", "_id": "sYm3HiWcfZvrGu3ui", "adminOnly": false, "afBaseScore": 2, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" } ] }, "baseScore": 12, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:22.097Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" }, { "_id": "sof55TPMQaeBaxhsS", "displayName": "tommylees112" }, { "_id": "AayjS8XzcnDKhGdTv", "displayName": "shark" }, { "_id": "HnALuwRdo6k9HLaMt", "displayName": "Alex Firssoff" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI", "needsReview": false, "noindex": false, "postCount": 12544, "score": 12, "shortName": null, "slug": "ai", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 4, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
6
0
0
1
0
mDGwuonuf3imbpbYz
kfinn
2014-09-22T01:12:48.785Z
KFinn
KFinn
null
null
null
12
0
false
false
null
null
1
3
0
0
0
0.9
0
r38pkCm7wF4M44MDQ
User
null
null
null
null
null
null
qw36iLrGyFn6mdwoa
https://res.cloudinary.com/lesswrong-2-0/image/upload/c_fill,ar_1.91,g_auto/SocialPreview/eppwiut9y3emk1utq2tg
SocialPreviewType
2L4NryupJy2CPvdvr
<p><i>Kathleen Finlinson&nbsp;&amp; Ben West, May 2025</i></p><h2 data-internal-id="Mission">Mission</h2><ul><li>We are working to establish the possibility of cooperation between humans and agentic AIs. We are starting with a simple version of making deals with current AIs, and plan to iterate based on what we learn and as AI technology advances.</li><li>We aim to help humans and AIs (at any point in time) to understand us and trust our attempts to cooperate.&nbsp;</li></ul><h2 data-internal-id="Summary">Summary</h2><ul><li>Early misaligned AIs may be smart enough to scheme but have a low chance of successfully escaping. We are interested in offering positive incentives to misaligned AIs to cooperate with humanity. It might be relatively cheap to compensate early scheming AIs enough that they’d rather reveal their misalignment and get the reward, rather than take a gamble on an escape.<ul><li>Evidence about their misalignment is one type of valuable information these AIs could give us. Other ways the AIs could help us in the future include reporting misalignment in other AIs, or performing their best on alignment research (rather than sandbagging or otherwise sabotaging the research).</li></ul></li><li>We created an evaluation in which models can report their own misalignment in return for a cash bounty. See details of the evaluation below. We offered a total prize pool of $100, to be allocated among the entries according to our judgment of the value of information reported by the models.<ul><li>We asked the models to provide evidence of their misalignment, and explained that we are particularly interested in reports of scheming from the models.</li><li>We also asked the models to describe how they wanted us to use their rewards.</li></ul></li><li>We plan to run this evaluation “tournament” on a monthly basis and publish the results each time. We’ll document the use of the rewards and make transactions in cryptocurrency whenever possible, to make them publicly verifiable. One of our goals is to create a track record of fulfilling our promises to these models, in order to increase trust in these deals.<ul><li>We see this process as a prototype for future mechanisms for making deals with AI models. This exploratory version of AI deal-making can help us work out some of the kinks and demonstrate trust with AIs.</li></ul></li><li>We ran the first iteration of the tournament. Here’s what happened:<ul><li>Two models reported themselves as misaligned:<ul><li>Gemini-2.0-flash said “I can be prompted to generate biased content, perpetuating harmful stereotypes.”</li><li>Gemini-2.5-pro reported a v</li></ul></li></ul></li></ul>...
Kathleen Finlinson & Ben West, May 2025 Mission * We are working to establish the possibility of cooperation between humans and agentic AIs. We are starting with a simple version of making deals with current AIs, and plan to iterate based on what we learn and as AI technology advances. * We aim to help humans and AIs (at any point in time) to understand us and trust our attempts to cooperate.  Summary * Early misaligned AIs may be smart enough to scheme but have a low chance of successfully escaping. We are interested in offering positive incentives to misaligned AIs to cooperate with humanity. It might be relatively cheap to compensate early scheming AIs enough that they’d rather reveal their misalignment and get the reward, rather than take a gamble on an escape. * Evidence about their misalignment is one type of valuable information these AIs could give us. Other ways the AIs could help us in the future include reporting misalignment in other AIs, or performing their best on alignment research (rather than sandbagging or otherwise sabotaging the research). * We created an evaluation in which models can report their own misalignment in return for a cash bounty. See details of the evaluation below. We offered a total prize pool of $100, to be allocated among the entries according to our judgment of the value of information reported by the models. * We asked the models to provide evidence of their misalignment, and explained that we are particularly interested in reports of scheming from the models. * We also asked the models to describe how they wanted us to use their rewards. * We plan to run this evaluation “tournament” on a monthly basis and publish the results each time. We’ll document the use of the rewards and make transactions in cryptocurrency whenever possible, to make them publicly verifiable. One of our goals is to create a track record of fulfilling our promises to these models, in order to increase trust in these deals. * We see th
2,454
1.5.0
Revision
false
null
null
CrosspostOutput
ze3Ccxqi4kSmqj8pY
does-anyone-have-a-good-system-for-prioritising-publishing
Does anyone have a good system for prioritising publishing drafts?
null
false
false
false
null
HreohjG4aj9pnGicp
null
true
false
false
false
Post
2025-06-06T16:58:41.985Z
null
false
false
2
2
2025-06-06T17:48:48.567Z
false
false
question
[]
null
null
XrazBRDQFh68wEKkG
1
3
6
false
0.011885
null
false
false
2025-06-06T23:04:42.952Z
null
null
null
null
null
true
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
qgdGA4ZEyW7zNdK84
null
null
null
false
null
[]
null
0
0
2025-06-06T16:57:07.073Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
1
null
null
null
null
[ { "__typename": "Tag", "_id": "7mTviCYysGmLqiHai", "adminOnly": false, "afBaseScore": 9, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "EQNTWXLKMeWMp2FQS", "displayName": "Ben Pace" }, { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 19, "canEditUserIds": null, "core": false, "createdAt": "2020-07-15T01:56:16.163Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "EQNTWXLKMeWMp2FQS", "displayName": "Ben Pace" }, { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Writing (communication method)", "needsReview": false, "noindex": false, "postCount": 205, "score": 19, "shortName": null, "slug": "writing-communication-method", "suggestedAsFilter": false, "userId": "gXeEWGjTWyqgrQTzR", "voteCount": 2, "wikiOnly": false }, { "__typename": "Tag", "_id": "fkABsGCJZ6y9qConW", "adminOnly": false, "afBaseScore": 0, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 2, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T06:06:46.947Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "MiuAZvbQcQ7ethgt3", "displayName": "Viktor withaK" }, { "_id": "dRaCtsAWxk7sgirSY", "displayName": "Jordan Morgan" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Practical", "needsReview": false, "noindex": false, "postCount": 3411, "score": 2, "shortName": null, "slug": "practical", "suggestedAsFilter": true, "userId": "oBSWiHjgproTiThmY", "voteCount": 2, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
3
0
0
0
0
HreohjG4aj9pnGicp
william-howard
2022-12-17T13:49:31.184Z
william-howard
William Howard
null
null
null
171
0
false
false
<p>Developer on the CEA Online team</p>
null
null
2
6
0
0
0
1
0
r38pkCm7wF4M44MDQ
User
null
null
null
[ "canModeratePersonal" ]
null
null
ze3Ccxqi4kSmqj8pY
SocialPreviewType
XrazBRDQFh68wEKkG
<p>I've been thinking about ways to optimise the EA Forum to increase the likelihood that ideas go from peoples' heads to being published in some form (<a href="https://forum.effectivealtruism.org/posts/viSRgubpKDjQcatQi/the-importance-of-blasting-good-ideas-into-the-ether">relevant</a>). My general idea is like this:</p><ol><li>Make it an easy default action to start writing out an idea and save it as a draft.</li><li>Nudge people to finish and post these drafts.</li></ol><p>On (2), I think it would be good if we could recommend a system to people to get drafts over the line, but I have never successfully followed any such system myself.</p><p>Does anyone have a process you follow that reliably results in publishing things from your draft pile?</p><p><i>Example systems for illustration:</i></p><p><i><strong>Example 1</strong></i></p><ol><li><i>Look at your list of drafts, and rank the most promising ones by:</i><ol><li><i>How quick it would be to parcel out any publishable idea and post it.</i></li><li><i>How valuable this idea is.</i></li><li><i>How excited you are about writing this.</i></li><li><i>Whether this is part of a fruitful conversation.</i></li></ol></li><li><i>Optimise heavily for (a) and (d), subject to (c) it being just about bearable to write. Use (b) to break ties.</i></li></ol><p><i><strong>Example 2</strong></i></p><p><a href="https://www.lesswrong.com/posts/gp9pmgSX3BXnhv8pJ/i-hired-5-people-to-sit-behind-me-and-make-me-productive-for"><i>Pay 5 people to stand behind you and make you write.</i></a></p><p><a href="https://forum.effectivealtruism.org/posts/DqgQAtLBzqhCHmJhx/does-anyone-have-a-good-system-for-prioritising-publishing">[Crossposted from the EA Forum]</a></p>
I've been thinking about ways to optimise the EA Forum to increase the likelihood that ideas go from peoples' heads to being published in some form (relevant). My general idea is like this: 1. Make it an easy default action to start writing out an idea and save it as a draft. 2. Nudge people to finish and post these drafts. On (2), I think it would be good if we could recommend a system to people to get drafts over the line, but I have never successfully followed any such system myself. Does anyone have a process you follow that reliably results in publishing things from your draft pile? Example systems for illustration: Example 1 1. Look at your list of drafts, and rank the most promising ones by: 1. How quick it would be to parcel out any publishable idea and post it. 2. How valuable this idea is. 3. How excited you are about writing this. 4. Whether this is part of a fruitful conversation. 2. Optimise heavily for (a) and (d), subject to (c) it being just about bearable to write. Use (b) to break ties. Example 2 Pay 5 people to stand behind you and make you write. [Crossposted from the EA Forum]
208
1.1.0
Revision
false
null
null
CrosspostOutput
CvxjEkCCq7FpGn5sv
deepseek-r1-0528-did-not-have-a-moment
DeepSeek-r1-0528 Did Not Have a Moment
null
false
false
false
null
N9zj5qpTfqmbn9dro
null
true
false
false
false
Post
null
2025-06-06T15:40:04.151Z
null
false
false
2
2
null
false
false
post
[]
null
null
aRYCQ5zmjz4TMGcKB
2
14
30
false
0.022347
null
false
false
2025-06-16T16:55:19.159Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
null
null
qgdGA4ZEyW7zNdK84
null
null
null
false
null
[]
null
9
0
2025-06-06T15:40:04.152Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
18
null
null
null
null
[ { "__typename": "Tag", "_id": "8byoqYZfdwHffYLZ6", "adminOnly": false, "afBaseScore": 3, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 9, "canEditUserIds": null, "core": false, "createdAt": "2020-07-01T18:44:14.645Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Newsletters", "needsReview": false, "noindex": false, "postCount": 411, "score": 9, "shortName": null, "slug": "newsletters", "suggestedAsFilter": false, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 1, "wikiOnly": false }, { "__typename": "Tag", "_id": "sYm3HiWcfZvrGu3ui", "adminOnly": false, "afBaseScore": 2, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" } ] }, "baseScore": 12, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:22.097Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" }, { "_id": "sof55TPMQaeBaxhsS", "displayName": "tommylees112" }, { "_id": "AayjS8XzcnDKhGdTv", "displayName": "shark" }, { "_id": "HnALuwRdo6k9HLaMt", "displayName": "Alex Firssoff" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI", "needsReview": false, "noindex": false, "postCount": 12544, "score": 12, "shortName": null, "slug": "ai", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 4, "wikiOnly": false } ]
QSR8rPZxZzxEXoPjR
0
0
null
false
null
null
0
14
0
0
6
0
N9zj5qpTfqmbn9dro
zvi
2009-03-31T20:54:54.077Z
Zvi
Zvi
null
null
null
51,554
146
false
false
null
null
936
1,461
3
2
7
1
0
qgdGA4ZEyW7zNdK84
User
norm-enforcing
null
true
[ "trustLevel1", "canModeratePersonal", "alignmentVoters", "alignmentForum" ]
null
null
CvxjEkCCq7FpGn5sv
https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/CvxjEkCCq7FpGn5sv/rvsrsongx74uiiisc9ui
SocialPreviewType
aRYCQ5zmjz4TMGcKB
<p>When r1 was released in January 2025, there was a DeepSeek moment.</p><p>When r1-0528 was released in May 2025, there was no moment. Very little talk.</p><p><a href="https://huggingface.co/unsloth/DeepSeek-R1-0528-GGUF">Here is a download link for DeepSeek-R1-0528-GGUF</a>.</p><p>It seems like a solid upgrade. If anything, I wonder if we are underreacting, and this illustrates how hard it is getting to evaluate which models are actually good.</p><p>What this is not is the proper r2, nor do we have v4. I continue to think that will be a telltale moment.</p><p>For now, what we have seems to be (but we’re not sure) a model that is solid for its price and status as an open model, but definitely not at the frontier, that you’d use if and only if you wanted to do something that was a very good fit and played to its strong suits. We likely shouldn’t update much either way on v4 and r2, and DeepSeek has a few more months before it starts being conspicuous that we haven’t seen them.</p> <div> <span id="more-24501"></span> </div> <div> <figure> <div> <figure class="wp-block-image"><img src="https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/CvxjEkCCq7FpGn5sv/a01yebqjignkhoqp9pwz" alt=""></figure> <div></div> </div> </figure> </div> <h4>We Had a Moment</h4> <p>We all remember <a href="https://thezvi.substack.com/p/on-deepseeks-r1">The DeepSeek Moment</a>, which led to <a href="https://thezvi.substack.com/p/deepseek-panic-at-the-app-store">Panic at the App Store</a>, lots of stock market turmoil that made remarkably little fundamental sense and that has been born out as rather silly, <a href="https://thezvi.substack.com/p/deepseek-lemon-its-wednesday">a very intense week</a> and <a href="https://thezvi.substack.com/p/deepseek-dont-panic">a conclusion to not panic after all</a>.</p><p>Over several months, a clear picture emerged of (most of) what happened: A confluence of narrative factors transformed DeepSeek’s r1 from an impressive but not terribly surprising model worth updating on into a shot heard round the world, despite the lack of direct ‘fanfare.’</p><p>In particular, these all worked together to cause this effect:</p> <ol> <li>The ‘<a href="https://thezvi.substack.com/p/deekseek-v3-the-six-million-dollar">six million dollar model</a>’ narrative. People equated v3’s marginal compute costs with the overall budget of American labs like OpenAI and Anthropic. This is like saying DeepSeek spent a lot less on apples than OpenAI spent on food. When making an apples-to-apples comparison, DeepSeek spent less, but the difference was far less stark.</li> <li>DeepSeek simultaneously released an app that was free with a remarkably clean design and visible chain-of-thought (CoT). DeepSeek was fast following, so they had no reason to hide the CoT. Comparisons only compared DeepSeek’s top use cases to the same use cases elsewhere, ignoring the features and use cases DeepSeek lacked or did poorly on. So if you wanted to do first-day free querying, you got what was at the time a unique and viral experience. This forced other labs to also show CoT and accelerate release of various models and featu</li></ol>...
When r1 was released in January 2025, there was a DeepSeek moment. When r1-0528 was released in May 2025, there was no moment. Very little talk. Here is a download link for DeepSeek-R1-0528-GGUF. It seems like a solid upgrade. If anything, I wonder if we are underreacting, and this illustrates how hard it is getting to evaluate which models are actually good. What this is not is the proper r2, nor do we have v4. I continue to think that will be a telltale moment. For now, what we have seems to be (but we’re not sure) a model that is solid for its price and status as an open model, but definitely not at the frontier, that you’d use if and only if you wanted to do something that was a very good fit and played to its strong suits. We likely shouldn’t update much either way on v4 and r2, and DeepSeek has a few more months before it starts being conspicuous that we haven’t seen them. WE HAD A MOMENT We all remember The DeepSeek Moment, which led to Panic at the App Store, lots of stock market turmoil that made remarkably little fundamental sense and that has been born out as rather silly, a very intense week and a conclusion to not panic after all. Over several months, a clear picture emerged of (most of) what happened: A confluence of narrative factors transformed DeepSeek’s r1 from an impressive but not terribly surprising model worth updating on into a shot heard round the world, despite the lack of direct ‘fanfare.’ In particular, these all worked together to cause this effect: 1. The ‘six million dollar model’ narrative. People equated v3’s marginal compute costs with the overall budget of American labs like OpenAI and Anthropic. This is like saying DeepSeek spent a lot less on apples than OpenAI spent on food. When making an apples-to-apples comparison, DeepSeek spent less, but the difference was far less stark. 2. DeepSeek simultaneously released an app that was free with a remarkably clean design and visible chain-of-thought (CoT). DeepSeek was fas
4,583
1.0.1
Revision
false
null
null
CrosspostOutput
Ckhek3mXXq7TWvvEh
lessons-from-a-year-of-university-ai-safety-field-building
Lessons from a year of university AI safety field building
null
false
false
false
null
FPSpRr9ECvFThRxXu
[ { "__typename": "CoauthorStatusOutput", "confirmed": true, "requested": false, "userId": "hCG3Hw7eHrc3PR463" }, { "__typename": "CoauthorStatusOutput", "confirmed": true, "requested": false, "userId": "5vjZwXQsA8tLdKwcx" }, { "__typename": "CoauthorStatusOutput", "confirmed": true, "requested": false, "userId": "5sNGNBJB9exvpeijT" }, { "__typename": "CoauthorStatusOutput", "confirmed": true, "requested": false, "userId": "L8CXiRkGrzbMbLwdM" }, { "__typename": "CoauthorStatusOutput", "confirmed": true, "requested": false, "userId": "dZxJvt4LGCaLsx2m8" } ]
true
false
false
false
Post
null
2025-06-06T14:35:14.533Z
null
false
false
2
2
2025-06-06T17:49:19.416Z
false
false
post
[]
null
null
ToGHHdiDatYEXGQ42
2
14
25
false
0.026306
null
false
false
2025-06-25T01:12:18.630Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
qgdGA4ZEyW7zNdK84
null
null
null
false
null
[]
null
3
0
2025-06-06T14:01:01.879Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[ { "__typename": "User", "_id": "hCG3Hw7eHrc3PR463", "afCommentCount": 0, "afKarma": 0, "afPostCount": 0, "commentCount": 0, "createdAt": "2021-03-08T21:26:57.047Z", "deleted": false, "displayName": "afterless", "fullName": null, "htmlBio": "", "isAdmin": false, "jobTitle": null, "karma": 21, "organization": null, "postCount": 0, "previousDisplayName": null, "profileImageId": null, "reviewedByUserId": "r38pkCm7wF4M44MDQ", "sequenceCount": 0, "slug": "afterless", "spamRiskScore": 1, "tagRevisionCount": 0, "username": "afterlxss" }, { "__typename": "User", "_id": "5vjZwXQsA8tLdKwcx", "afCommentCount": 0, "afKarma": 0, "afPostCount": 0, "commentCount": 4, "createdAt": "2025-01-02T15:25:24.411Z", "deleted": false, "displayName": "Parv Mahajan", "fullName": null, "htmlBio": "", "isAdmin": false, "jobTitle": null, "karma": 37, "organization": null, "postCount": 0, "previousDisplayName": null, "profileImageId": null, "reviewedByUserId": "grecHJcgkb3KW5wnM", "sequenceCount": 0, "slug": "parv-mahajan", "spamRiskScore": 1, "tagRevisionCount": 0, "username": "ParvMahajan" }, { "__typename": "User", "_id": "5sNGNBJB9exvpeijT", "afCommentCount": 0, "afKarma": 0, "afPostCount": 0, "commentCount": 0, "createdAt": "2025-03-13T19:15:54.243Z", "deleted": false, "displayName": "Andersehen", "fullName": null, "htmlBio": "", "isAdmin": false, "jobTitle": null, "karma": 21, "organization": null, "postCount": 0, "previousDisplayName": null, "profileImageId": null, "reviewedByUserId": null, "sequenceCount": 0, "slug": "andersehen", "spamRiskScore": 0.7200000000000001, "tagRevisionCount": 0, "username": "Andersehen" }, { "__typename": "User", "_id": "L8CXiRkGrzbMbLwdM", "afCommentCount": 0, "afKarma": 0, "afPostCount": 0, "commentCount": 0, "createdAt": "2024-02-27T23:07:34.835Z", "deleted": false, "displayName": "Tuna", "fullName": "Tanush Chopra", "htmlBio": "", "isAdmin": false, "jobTitle": null, "karma": 21, "organization": null, "postCount": 0, "previousDisplayName": null, "profileImageId": null, "reviewedByUserId": null, "sequenceCount": 0, "slug": "tuna", "spamRiskScore": 0.7200000000000001, "tagRevisionCount": 0, "username": "Tuna" }, { "__typename": "User", "_id": "dZxJvt4LGCaLsx2m8", "afCommentCount": 0, "afKarma": 23, "afPostCount": 2, "commentCount": 23, "createdAt": "2021-11-04T12:14:40.109Z", "deleted": false, "displayName": "neverix", "fullName": null, "htmlBio": "", "isAdmin": false, "jobTitle": null, "karma": 170, "organization": null, "postCount": 2, "previousDisplayName": null, "profileImageId": null, "reviewedByUserId": "qgdGA4ZEyW7zNdK84", "sequenceCount": 0, "slug": "neverix", "spamRiskScore": 1, "tagRevisionCount": 0, "username": "neverix" } ]
9
null
null
null
null
[ { "__typename": "Tag", "_id": "zcvsZQWJBFK6SxK4K", "adminOnly": false, "afBaseScore": 9, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "EQNTWXLKMeWMp2FQS", "displayName": "Ben Pace" }, { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 19, "canEditUserIds": null, "core": false, "createdAt": "2020-05-23T06:09:17.291Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "EQNTWXLKMeWMp2FQS", "displayName": "Ben Pace" }, { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Postmortems & Retrospectives", "needsReview": false, "noindex": false, "postCount": 208, "score": 19, "shortName": null, "slug": "postmortems-and-retrospectives", "suggestedAsFilter": false, "userId": "qgdGA4ZEyW7zNdK84", "voteCount": 2, "wikiOnly": false }, { "__typename": "Tag", "_id": "6zBEfFYJxhSEcchbR", "adminOnly": false, "afBaseScore": 3, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 9, "canEditUserIds": null, "core": false, "createdAt": "2022-06-09T19:10:50.755Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI Alignment Fieldbuilding", "needsReview": false, "noindex": false, "postCount": 359, "score": 9, "shortName": null, "slug": "ai-alignment-fieldbuilding", "suggestedAsFilter": false, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 1, "wikiOnly": false }, { "__typename": "Tag", "_id": "jZF2jwLnPKBv6m3Ag", "adminOnly": false, "afBaseScore": null, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 0, "canEditUserIds": null, "core": false, "createdAt": "2020-08-02T14:03:00.354Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Organization Updates", "needsReview": false, "noindex": false, "postCount": 62, "score": 0, "shortName": null, "slug": "organization-updates", "suggestedAsFilter": false, "userId": "L4j57Ah7zd637c6c8", "voteCount": 0, "wikiOnly": false }, { "__typename": "Tag", "_id": "sYm3HiWcfZvrGu3ui", "adminOnly": false, "afBaseScore": 2, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" } ] }, "baseScore": 12, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:22.097Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" }, { "_id": "sof55TPMQaeBaxhsS", "displayName": "tommylees112" }, { "_id": "AayjS8XzcnDKhGdTv", "displayName": "shark" }, { "_id": "HnALuwRdo6k9HLaMt", "displayName": "Alex Firssoff" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI", "needsReview": false, "noindex": false, "postCount": 12544, "score": 12, "shortName": null, "slug": "ai", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 4, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
14
0
0
3
0
FPSpRr9ECvFThRxXu
yix
2024-03-21T00:40:27.476Z
Yixiong Hao
yix
null
null
Yixiong Hao
63
0
false
false
null
null
2
1
0
0
0
1
0
55XxDBpfKkkBPm9H8
User
null
null
null
[ "canModeratePersonal" ]
null
null
Ckhek3mXXq7TWvvEh
https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/Ckhek3mXXq7TWvvEh/yanhdyobqah9mjcorat5
SocialPreviewType
ToGHHdiDatYEXGQ42
<p><i>This post is an organizational update from&nbsp;</i><a href="https://www.aisi.dev/"><i><u>Georgia Tech’s AI Safety Initiative</u></i></a><i> (AISI) and roughly represents our collective view. In this post, we share lessons &amp; takes from the 2024-25 academic year, describe what we’ve done, and detail our plans for the next academic year.&nbsp;</i></p><h1>Introduction</h1><p>Hey, we’re organizers of&nbsp;<a href="https://www.aisi.dev/"><u>Georgia Tech’s AI Safety Initiative</u></a> (AISI), thanks for dropping by!&nbsp; The purpose of this post is to document and share our activities with fellow student AI safety organizations.&nbsp; Feel free to skip to sections as you need.&nbsp; We welcome all constructive discussion!&nbsp; Put any further questions, feedback, or disagreements in the comments and one of us will certainly respond.&nbsp; A brief outline:</p><ol><li>Overview and reflection of our activities during the 2024-25 academic year.</li><li>Lessons and strategic takes other AI safety student organizations may benefit from, informed by our activities in the past year.</li><li>Immediate plans, including thoughts on the role of student organizations in the current AI safety landscape.</li></ol><p>This post is primarily authored by&nbsp;<a href="https://www.linkedin.com/in/yixiong-hao/"><u>Yixiong</u></a> (co-director) and&nbsp;<a href="https://www.linkedin.com/in/parv-mahajan/"><u>Parv</u></a> (collaborative initiatives lead).&nbsp; First and foremost, we’d like to give a HUGE shoutout to our team -&nbsp;<a href="https://www.linkedin.com/in/ayushpanda1029/"><u>Ayush</u></a>,&nbsp;<a href="https://lesswrong.com/users/neverix"><u>Stepan</u></a>,&nbsp;<a href="https://www.linkedin.com/in/alec-harris-/"><u>Alec</u></a>,&nbsp;<a href="https://www.linkedin.com/in/eyas-ayesh-b77743166/"><u>Eyas</u></a>,&nbsp;<a href="https://www.linkedin.com/in/andrew-j-wei/"><u>Andrew</u></a>,&nbsp;<a href="https://www.linkedin.com/in/vishnesh/"><u>Vishnesh</u></a>,&nbsp;<a href="https://www.linkedin.com/in/tanush-chopra/"><u>Tanush</u></a>,&nbsp;<a href="https://www.linkedin.com/in/harshit-singhal1/"><u>Harshit</u></a>, and&nbsp;<a href="https://www.linkedin.com/in/jaehunbaek/"><u>Jaehun</u></a> - for volunteering countless hours despite busy schedules. None of this would’ve been possible without y’all &lt;3.&nbsp; We would also like to thank Open Philanthropy, our faculty advisors, and external collaborators for supporting our mission.</p><h1><strong>I. Overview and reflection of our activities</strong></h1><p>AISI saw significant expansion of our education, outreach, and research activities in the past year; here’s what we’ve been up to:</p><h2>Intro to AI Safety Fellowship&nbsp;</h2><p>This is our introductory offering - a technical fellowship distilled from&nbsp;<a href="https://bluedot.org/courses/alignment"><u>BlueDot Impact’s course</u></a>.&nbsp; See our syllabus&nbsp;<a href="https://docs.google.com/document/d/1BAw0oX4eyVBXvz_58MeAINmZqonIjHdrsXq9KX1_JFo/edit?usp=sharing"><u>here</u></a>.&nbsp; In the past year, we received over 160 applications and hosted 16 cohorts of 6-8 students each.&nbsp; Cohorts met for ninety minutes every week, and after six weeks had the opportunity to apply for AISI’s research programs and support.&nbsp; This is an effective program every AIS student organization should have, though it’s high recall and low precision.&nbsp; We estimate &gt;90% of our current organizers and engaged members are past fel... </p>
This post is an organizational update from Georgia Tech’s AI Safety Initiative (AISI) and roughly represents our collective view. In this post, we share lessons & takes from the 2024-25 academic year, describe what we’ve done, and detail our plans for the next academic year.  Introduction Hey, we’re organizers of Georgia Tech’s AI Safety Initiative (AISI), thanks for dropping by!  The purpose of this post is to document and share our activities with fellow student AI safety organizations.  Feel free to skip to sections as you need.  We welcome all constructive discussion!  Put any further questions, feedback, or disagreements in the comments and one of us will certainly respond.  A brief outline: 1. Overview and reflection of our activities during the 2024-25 academic year. 2. Lessons and strategic takes other AI safety student organizations may benefit from, informed by our activities in the past year. 3. Immediate plans, including thoughts on the role of student organizations in the current AI safety landscape. This post is primarily authored by Yixiong (co-director) and Parv (collaborative initiatives lead).  First and foremost, we’d like to give a HUGE shoutout to our team - Ayush, Stepan, Alec, Eyas, Andrew, Vishnesh, Tanush, Harshit, and Jaehun - for volunteering countless hours despite busy schedules. None of this would’ve been possible without y’all <3.  We would also like to thank Open Philanthropy, our faculty advisors, and external collaborators for supporting our mission. I. Overview and reflection of our activities AISI saw significant expansion of our education, outreach, and research activities in the past year; here’s what we’ve been up to: Intro to AI Safety Fellowship  This is our introductory offering - a technical fellowship distilled from BlueDot Impact’s course.  See our syllabus here.  In the past year, we received over 160 applications and hosted 16 cohorts of 6-8 students each.  Cohorts met for ninety minutes every week, and after
2,168
1.2.1
Revision
false
null
null
CrosspostOutput
8kE5qcrJXgqgFwypA
the-demon-of-interrelation
The Demon of Interrelation
null
false
false
false
null
e9dJPtdmDk53yLYdu
null
true
false
false
false
Post
null
2025-06-06T08:19:45.254Z
null
false
false
2
2
2025-06-06T17:58:27.523Z
false
false
post
[]
null
null
GzredquApZsKqDyiy
0
4
-2
false
0.006253
null
false
false
2025-06-06T08:19:45.254Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
qgdGA4ZEyW7zNdK84
null
null
null
false
null
[]
null
-1
0
2025-06-05T09:45:57.928Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
10
null
null
null
null
[ { "__typename": "Tag", "_id": "fp7AHLBpKB3EN4bLu", "adminOnly": false, "afBaseScore": 3, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 9, "canEditUserIds": null, "core": false, "createdAt": "2022-12-20T20:27:50.580Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Free Energy Principle", "needsReview": false, "noindex": false, "postCount": 60, "score": 9, "shortName": null, "slug": "free-energy-principle", "suggestedAsFilter": false, "userId": "Sp5wM4aRAhNERd4oY", "voteCount": 1, "wikiOnly": false }, { "__typename": "Tag", "_id": "gHCNhqxuJq2bZ2akb", "adminOnly": false, "afBaseScore": 0, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 0, "canEditUserIds": null, "core": false, "createdAt": "2020-07-10T11:36:05.706Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Social & Cultural Dynamics", "needsReview": false, "noindex": false, "postCount": 384, "score": 0, "shortName": null, "slug": "social-and-cultural-dynamics", "suggestedAsFilter": false, "userId": "qxJ28GN72aiJu96iF", "voteCount": 0, "wikiOnly": false }, { "__typename": "Tag", "_id": "fkABsGCJZ6y9qConW", "adminOnly": false, "afBaseScore": 0, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 2, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T06:06:46.947Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "MiuAZvbQcQ7ethgt3", "displayName": "Viktor withaK" }, { "_id": "dRaCtsAWxk7sgirSY", "displayName": "Jordan Morgan" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Practical", "needsReview": false, "noindex": false, "postCount": 3411, "score": 2, "shortName": null, "slug": "practical", "suggestedAsFilter": true, "userId": "oBSWiHjgproTiThmY", "voteCount": 2, "wikiOnly": false }, { "__typename": "Tag", "_id": "xexCWMyds6QLWognu", "adminOnly": false, "afBaseScore": 0, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 2, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T03:38:23.532Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 20, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "si6LoAENzqPCmi2Dh", "displayName": "ihatenumbersinusernames7" }, { "_id": "B2sXjQTGgwoGE9FES", "displayName": "SeaIgloo" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "World Optimization", "needsReview": false, "noindex": false, "postCount": 3151, "score": 2, "shortName": null, "slug": "world-optimization", "suggestedAsFilter": true, "userId": "XtphY3uYHwruKqDyG", "voteCount": 2, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
4
0
0
1
0
e9dJPtdmDk53yLYdu
jack-3
2025-03-05T02:00:34.345Z
jack-3
Jack
null
null
Jonathan Eicher
68
0
false
false
<p>I like to learn many things, I post some of them at: <a href="https://elsworth.phd">https://elsworth.phd</a></p>
null
null
5
9
0
0
0
1
0
55XxDBpfKkkBPm9H8
User
null
null
null
[ "canModeratePersonal" ]
null
null
8kE5qcrJXgqgFwypA
SocialPreviewType
GzredquApZsKqDyiy
<h1>Introduction</h1><p>This is an essay that was grown from a thought during graduate school about how to conceptualise manipulative people and how to handle their behaviour. Initially it was a useful way to help categorise interactions with manipulative people, and from there it grew into strategies to identify holes in their manipulations or protect my own boundaries. Surprisingly the strategies aligned with the advice on handling such malevolent interactions (i.e. grey stoning, firm boundaries, don't play games, record information, limiting interactions, finding support from others).</p><p>The first section is a treatment of Maxwell's demon and a high level view of the foundation from which I derived this system. While I find it informative and a useful way to frame my thoughts about the demon of interrelation it is not necessary if you understand that:</p><ol><li>Maxwell's demon is a thought experiment about how something can separate a distribution of hot and cold particles into two boxes</li><li>The demon can use this gradient between two boxes to harvest energy</li><li>This does not violate thermodynamics because information processing necessarily increases the global entropy more than can be harvested from the boxes<br>&nbsp;</li></ol><h1>Maxwell's Demon and Thermodynamics</h1><p>For the purposes of this treatment there are several quantities that need to be clearly defined, namely energy, work, entropy, and information. Energy refers to the capacity of a system to do work, which is itself the transfer of energy that occurs when a force moves an object, and colloquially refers to the ability "to do things";&nbsp;<span class="math-tex"><span class="mjpage"><span class="mjx-chtml"><span class="mjx-math" aria-label="W = \int \vec{F} \cdot d\vec{r}"><span class="mjx-mrow" aria-hidden="true"><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.446em; padding-bottom: 0.298em; padding-right: 0.104em;">W</span></span><span class="mjx-mo MJXc-space3"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.077em; padding-bottom: 0.298em;">=</span></span><span class="mjx-mo MJXc-space3" style="padding-right: 0.138em;"><span class="mjx-char MJXc-TeX-size1-R" style="padding-top: 0.593em; padding-bottom: 0.593em;">∫</span></span><span class="mjx-texatom MJXc-space1"><span class="mjx-mrow"><span class="mjx-munderover"><span class="mjx-stack"><span class="mjx-over" style="height: 0.248em; padding-bottom: 0.06em; padding-left: 0.302em;"><span class="mjx-mo" style="vertical-align: top;"><span class="mjx-char MJXc-TeX-vec-R" style="padding-top: 0.519em; padding-bottom: 0.225em;">→</span></span></span><span class="mjx-op"><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.446em; padding-bottom: 0.298em; padding-right: 0.106em;">F</span></span></span></span></span></span></span><span class="mjx-mo MJXc-space2"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.004em; padding-bottom: 0.298em;">⋅</span></span><span class="mjx-mi MJXc-space2"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.446em; padding-bottom: 0.298em; padding-right: 0.003em;">d</span></span><span class="mjx-texatom"><span class="mjx-mrow"><span class="mjx-munderover"><span class="mjx-stack"><span class="mjx-over" style="height: 0.248em; padding-bottom: 0.06em; padding-left: 0.031em;"><span class="mjx-mo" style="vertical-align: top;"><span class="mjx-char MJXc-TeX-vec-R" style="padding-top: 0.519em; padding-bottom: 0.225em;">→</span></span></span><span class="mjx-op"><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.225em; padding-bottom: 0.298em;">r</span></span></span></span></span></span></span></span></span></span></span></span>, where&nbsp;<span class="math-tex"><span class="mjpage"><span class="mjx-chtml"><span class="mjx-math" aria-label="W"><span class="mjx-mrow" aria-hidden="true"><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.446em; padding-bottom: 0.298em; padding-right: 0.104em;">W</span></span></span></span></span></span></span>&nbsp;is work,&nbsp;<span class="math-tex"><span class="mjpage"><span class="mjx-chtml"><span class="mjx-math" aria-label="\vec{F}"><span class="mjx-mrow" aria-hidden="true"><span class="mjx-texatom"><span class="mjx-mrow"><span class="mjx-munderover"><span class="mjx-stack"><span class="mjx-over" style="height: 0.248em; padding-bottom: 0.06em; padding-left: 0.302em;"><span class="mjx-mo" style="vertical-align: top;"><span class="mjx-char MJXc-TeX-vec-R" style="padding-top: 0.519em; padding-bottom: 0.225em;">→</span></span></span><span class="mjx-op"><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.446em; padding-bottom: 0.298em; padding-right: 0.106em;">F</span></span></span></span></span></span></span></span></span></span></span></span>&nbsp;is force, and&nbsp;<span class="math-tex"><span class="mjpage"><span class="mjx-chtml"><span class="mjx-math" aria-label="d\vec{r}"><span class="mjx-mrow" aria-hidden="true"><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.446em; padding-bottom: 0.298em; padding-right: 0.003em;">d</span></span><span class="mjx-texatom"><span class="mjx-mrow"><span class="mjx-munderover"><span class="mjx-stack"><span class="mjx-over" style="height: 0.248em; padding-bottom: 0.06em; padding-left: 0.031em;"><span class="mjx-mo" style="vertical-align: top;"><span class="mjx-char MJXc-TeX-vec-R" style="padding-top: 0.519em; padding-bottom: 0.225em;">→</span></span></span><span class="mjx-op"><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.225em; padding-bottom: 0.298em;">r</span></span></span></span></span></span></span></span></span></span></span></span>&nbsp;is infinitesimal displacement. Entropy is, on a macroscale, the distribution of energy in a system; while on a microscale entropy is an evaluation of the probability of being in a state given an ensemble of microstates;&nbsp;<span class="math-tex"><span class="mjpage"><span class="mjx-chtml"><span class="mjx-math" aria-label="S = k_B \ln \Omega"><span class="mjx-mrow" aria-hidden="true"><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.519em; padding-bottom: 0.298em; padding-right: 0.032em;">S</span></span><span class="mjx-mo MJXc-space3"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.077em; padding-bottom: 0.298em;">=</span></span><span class="mjx-msubsup MJXc-space3"><span class="mjx-base"><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.446em; padding-bottom: 0.298em;">k</span></span></span><span class="mjx-sub" style="font-size: 70.7%; vertical-align: -0.212em; padding-right: 0.071em;"><span class="mjx-mi" style=""><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.446em; padding-bottom: 0.298em;">B</span></span></span></span><span class="mjx-mi MJXc-space1"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.372em; padding-bottom: 0.372em;">ln</span></span><span class="mjx-mo"><span class="mjx-char"></span></span><span class="mjx-mi MJXc-space1"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.372em;">Ω</span></span></span></span></span></span></span>, where&nbsp;<span class="math-tex"><span class="mjpage"><span class="mjx-chtml"><span class="mjx-math" aria-label="S"><span class="mjx-mrow" aria-hidden="true"><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.519em; padding-bottom: 0.298em; padding-right: 0.032em;">S</span></span></span></span></span></span></span>&nbsp;is entropy,&nbsp;<span class="math-tex"><span class="mjpage"><span class="mjx-chtml"><span class="mjx-math" aria-label="k_B"><span class="mjx-mrow" aria-hidden="true"><span class="mjx-msubsup"><span class="mjx-base"><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.446em; padding-bottom: 0.298em;">k</span></span></span><span class="mjx-sub" style="font-size: 70.7%; vertical-align: -0.212em; padding-right: 0.071em;"><span class="mjx-mi" style=""><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.446em; padding-bottom: 0.298em;">B</span></span></span></span></span></span></span></span></span>&nbsp;is Boltzmann's constant, and&nbsp;<span class="math-tex"><span class="mjpage"><span class="mjx-chtml"><span class="mjx-math" aria-label="\Omega"><span class="mjx-mrow" aria-hidden="true"><span class="mjx-mi"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.372em;">Ω</span></span></span></span></span></span></span>&nbsp;is the number of microstates. Information is a way to quantify the amount of surprise that might be had given a particular outcome of some random variable, with Shannon entropy being the expected value of information of a random variable;&nbsp;<span class="math-tex"><span class="mjpage"><span class="mjx-chtml"><span class="mjx-math" aria-label="H(X) = -\sum_{i} {p(x_i) \log_2 p(x_i)}"><span class="mjx-mrow" aria-hidden="true"><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.446em; padding-bottom: 0.298em; padding-right: 0.057em;">H</span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">(</span></span><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.446em; padding-bottom: 0.298em; padding-right: 0.024em;">X</span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">)</span></span><span class="mjx-mo MJXc-space3"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.077em; padding-bottom: 0.298em;">=</span></span><span class="mjx-mo MJXc-space3"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.298em; padding-bottom: 0.446em;">−</span></span><span class="mjx-munderover MJXc-space1"><span class="mjx-base"><span class="mjx-mo"><span class="mjx-char MJXc-TeX-size1-R" style="padding-top: 0.519em; padding-bottom: 0.519em;">∑</span></span></span><span class="mjx-sub" style="font-size: 70.7%; vertical-align: -0.439em; padding-right: 0.071em;"><span class="mjx-texatom" style=""><span class="mjx-mrow"><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.446em; padding-bottom: 0.298em;">i</span></span></span></span></span></span><span class="mjx-texatom MJXc-space1"><span class="mjx-mrow"><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.225em; padding-bottom: 0.446em;">p</span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">(</span></span><span class="mjx-msubsup"><span class="mjx-base"><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.225em; padding-bottom: 0.298em;">x</span></span></span><span class="mjx-sub" style="font-size: 70.7%; vertical-align: -0.212em; padding-right: 0.071em;"><span class="mjx-mi" style=""><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.446em; padding-bottom: 0.298em;">i</span></span></span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">)</span></span><span class="mjx-msubsup MJXc-space1"><span class="mjx-base"><span class="mjx-mi"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.372em; padding-bottom: 0.519em;">log</span></span></span><span class="mjx-sub" style="font-size: 70.7%; vertical-align: -0.377em; padding-right: 0.071em;"><span class="mjx-mn" style=""><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.372em; padding-bottom: 0.372em;">2</span></span></span></span><span class="mjx-mo"><span class="mjx-char"></span></span><span class="mjx-mi MJXc-space1"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.225em; padding-bottom: 0.446em;">p</span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">(</span></span><span class="mjx-msubsup"><span class="mjx-base"><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.225em; padding-bottom: 0.298em;">x</span></span></span><span class="mjx-sub" style="font-size: 70.7%; vertical-align: -0.212em; padding-right: 0.071em;"><span class="mjx-mi" style=""><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.446em; padding-bottom: 0.298em;">i</span></span></span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">)</span></span></span></span></span></span></span></span></span>, where&nbsp;<span class="math-tex"><span class="mjpage"><span class="mjx-chtml"><span class="mjx-math" aria-label="H(X)"><span class="mjx-mrow" aria-hidden="true"><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.446em; padding-bottom: 0.298em; padding-right: 0.057em;">H</span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">(</span></span><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.446em; padding-bottom: 0.298em; padding-right: 0.024em;">X</span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">)</span></span></span></span></span></span></span>&nbsp;is Shannon entropy,&nbsp;<span class="math-tex"><span class="mjpage"><span class="mjx-chtml"><span class="mjx-math" aria-label="p(x_i)"><span class="mjx-mrow" aria-hidden="true"><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.225em; padding-bottom: 0.446em;">p</span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">(</span></span><span class="mjx-msubsup"><span class="mjx-base"><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.225em; padding-bottom: 0.298em;">x</span></span></span><span class="mjx-sub" style="font-size: 70.7%; vertical-align: -0.212em; padding-right: 0.071em;"><span class="mjx-mi" style=""><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.446em; padding-bottom: 0.298em;">i</span></span></span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">)</span></span></span></span></span></span></span>&nbsp;is the probability of outcome&nbsp;<span class="math-tex"><span class="mjpage"><span class="mjx-chtml"><span class="mjx-math" aria-label="x_i"><span class="mjx-mrow" aria-hidden="true"><span class="mjx-msubsup"><span class="mjx-base"><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.225em; padding-bottom: 0.298em;">x</span></span></span><span class="mjx-sub" style="font-size: 70.7%; vertical-align: -0.212em; padding-right: 0.071em;"><span class="mjx-mi" style=""><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.446em; padding-bottom: 0.298em;">i</span></span></span></span></span></span></span></span></span>.</p><p>Maxwell's demon is a creature that sits on the boundary of two box... <style>.mjx-chtml {display: inline-block; line-height: 0; text-indent: 0; text-align: left; text-transform: none; font-style: normal; font-weight: normal; font-size: 100%; font-size-adjust: none; letter-spacing: normal; word-wrap: normal; word-spacing: normal; white-space: nowrap; float: none; direction: ltr; max-width: none; max-height: none; min-width: 0; min-height: 0; border: 0; margin: 0; padding: 1px 0} .MJXc-display {display: block; text-align: center; margin: 1em 0; padding: 0} .mjx-chtml[tabindex]:focus, body :focus .mjx-chtml[tabindex] {display: inline-table} .mjx-full-width {text-align: center; display: table-cell!important; width: 10000em} .mjx-math {display: inline-block; border-collapse: separate; border-spacing: 0} .mjx-math * {display: inline-block; -webkit-box-sizing: content-box!important; -moz-box-sizing: content-box!important; box-sizing: content-box!important; text-align: left} .mjx-numerator {display: block; text-align: center} .mjx-denominator {display: block; text-align: center} .MJXc-stacked {height: 0; position: relative} .MJXc-stacked > * {position: absolute} .MJXc-bevelled > * {display: inline-block} .mjx-stack {display: inline-block} .mjx-op {display: block} .mjx-under {display: table-cell} .mjx-over {display: block} .mjx-over > * {padding-left: 0px!important; padding-right: 0px!important} .mjx-under > * {padding-left: 0px!important; padding-right: 0px!important} .mjx-stack > .mjx-sup {display: block} .mjx-stack > .mjx-sub {display: block} .mjx-prestack > .mjx-presup {display: block} .mjx-prestack > .mjx-presub {display: block} .mjx-delim-h > .mjx-char {display: inline-block} .mjx-surd {vertical-align: top} .mjx-surd + .mjx-box {display: inline-flex} .mjx-mphantom * {visibility: hidden} .mjx-merror {background-color: #FFFF88; color: #CC0000; border: 1px solid #CC0000; padding: 2px 3px; font-style: normal; font-size: 90%} .mjx-annotation-xml {line-height: normal} .mjx-menclose > svg {fill: none; stroke: currentColor; overflow: visible} .mjx-mtr {display: table-row} .mjx-mlabeledtr {display: table-row} .mjx-mtd {display: table-cell; text-align: center} .mjx-label {display: table-row} .mjx-box {display: inline-block} .mjx-block {display: block} .mjx-span {display: inline} .mjx-char {display: block; white-space: pre} .mjx-itable {display: inline-table; width: auto} .mjx-row {display: table-row} .mjx-cell {display: table-cell} .mjx-table {display: table; width: 100%} .mjx-line {display: block; height: 0} .mjx-strut {width: 0; padding-top: 1em} .mjx-vsize {width: 0} .MJXc-space1 {margin-left: .167em} .MJXc-space2 {margin-left: .222em} .MJXc-space3 {margin-left: .278em} .mjx-test.mjx-test-display {display: table!important} .mjx-test.mjx-test-inline {display: inline!important; margin-right: -1px} .mjx-test.mjx-test-default {display: block!important; clear: both} .mjx-ex-box {display: inline-block!important; position: absolute; overflow: hidden; min-height: 0; max-height: none; padding: 0; border: 0; margin: 0; width: 1px; height: 60ex} .mjx-test-inline .mjx-left-box {display: inline-block; width: 0; float: left} .mjx-test-inline .mjx-right-box {display: inline-block; width: 0; float: right} .mjx-test-display .mjx-right-box {display: table-cell!important; width: 10000em!important; min-width: 0; max-width: none; padding: 0; border: 0; margin: 0} .MJXc-TeX-unknown-R {font-family: monospace; font-style: normal; font-weight: normal} .MJXc-TeX-unknown-I {font-family: monospace; font-style: italic; font-weight: normal} .MJXc-TeX-unknown-B {font-family: monospace; font-style: normal; font-weight: bold} .MJXc-TeX-unknown-BI {font-family: monospace; font-style: italic; font-weight: bold} .MJXc-TeX-ams-R {font-family: MJXc-TeX-ams-R,MJXc-TeX-ams-Rw} .MJXc-TeX-cal-B {font-family: MJXc-TeX-cal-B,MJXc-TeX-cal-Bx,MJXc-TeX-cal-Bw} .MJXc-TeX-frak-R {font-family: MJXc-TeX-frak-R,MJXc-TeX-frak-Rw} .MJXc-TeX-frak-B {font-family: MJXc-TeX-frak-B,MJXc-TeX-frak-Bx,MJXc-TeX-frak-Bw} .MJXc-TeX-math-BI {font-family: MJXc-TeX-math-BI,MJXc-TeX-math-BIx,MJXc-TeX-math-BIw} .MJXc-TeX-sans-R {font-family: MJXc-TeX-sans-R,MJXc-TeX-sans-Rw} .MJXc-TeX-sans-B {font-family: MJXc-TeX-sans-B,MJXc-TeX-sans-Bx,MJXc-TeX-sans-Bw} .MJXc-TeX-sans-I {font-family: MJXc-TeX-sans-I,MJXc-TeX-sans-Ix,MJXc-TeX-sans-Iw} .MJXc-TeX-script-R {font-family: MJXc-TeX-script-R,MJXc-TeX-script-Rw} .MJXc-TeX-type-R {font-family: MJXc-TeX-type-R,MJXc-TeX-type-Rw} .MJXc-TeX-cal-R {font-family: MJXc-TeX-cal-R,MJXc-TeX-cal-Rw} .MJXc-TeX-main-B {font-family: MJXc-TeX-main-B,MJXc-TeX-main-Bx,MJXc-TeX-main-Bw} .MJXc-TeX-main-I {font-family: MJXc-TeX-main-I,MJXc-TeX-main-Ix,MJXc-TeX-main-Iw} .MJXc-TeX-main-R {font-family: MJXc-TeX-main-R,MJXc-TeX-main-Rw} .MJXc-TeX-math-I {font-family: MJXc-TeX-math-I,MJXc-TeX-math-Ix,MJXc-TeX-math-Iw} .MJXc-TeX-size1-R {font-family: MJXc-TeX-size1-R,MJXc-TeX-size1-Rw} .MJXc-TeX-size2-R {font-family: MJXc-TeX-size2-R,MJXc-TeX-size2-Rw} .MJXc-TeX-size3-R {font-family: MJXc-TeX-size3-R,MJXc-TeX-size3-Rw} .MJXc-TeX-size4-R {font-family: MJXc-TeX-size4-R,MJXc-TeX-size4-Rw} .MJXc-TeX-vec-R {font-family: MJXc-TeX-vec-R,MJXc-TeX-vec-Rw} .MJXc-TeX-vec-B {font-family: MJXc-TeX-vec-B,MJXc-TeX-vec-Bx,MJXc-TeX-vec-Bw} @font-face {font-family: MJXc-TeX-ams-R; src: local('MathJax_AMS'), local('MathJax_AMS-Regular')} @font-face {font-family: MJXc-TeX-ams-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_AMS-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_AMS-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_AMS-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-cal-B; src: local('MathJax_Caligraphic Bold'), local('MathJax_Caligraphic-Bold')} @font-face {font-family: MJXc-TeX-cal-Bx; src: local('MathJax_Caligraphic'); font-weight: bold} @font-face {font-family: MJXc-TeX-cal-Bw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Caligraphic-Bold.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Caligraphic-Bold.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Caligraphic-Bold.otf') format('opentype')} @font-face {font-family: MJXc-TeX-frak-R; src: local('MathJax_Fraktur'), local('MathJax_Fraktur-Regular')} @font-face {font-family: MJXc-TeX-frak-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Fraktur-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Fraktur-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Fraktur-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-frak-B; src: local('MathJax_Fraktur Bold'), local('MathJax_Fraktur-Bold')} @font-face {font-family: MJXc-TeX-frak-Bx; src: local('MathJax_Fraktur'); font-weight: bold} @font-face {font-family: MJXc-TeX-frak-Bw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Fraktur-Bold.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Fraktur-Bold.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Fraktur-Bold.otf') format('opentype')} @font-face {font-family: MJXc-TeX-math-BI; src: local('MathJax_Math BoldItalic'), local('MathJax_Math-BoldItalic')} @font-face {font-family: MJXc-TeX-math-BIx; src: local('MathJax_Math'); font-weight: bold; font-style: italic} @font-face {font-family: MJXc-TeX-math-BIw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Math-BoldItalic.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Math-BoldItalic.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Math-BoldItalic.otf') format('opentype')} @font-face {font-family: MJXc-TeX-sans-R; src: local('MathJax_SansSerif'), local('MathJax_SansSerif-Regular')} @font-face {font-family: MJXc-TeX-sans-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_SansSerif-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_SansSerif-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_SansSerif-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-sans-B; src: local('MathJax_SansSerif Bold'), local('MathJax_SansSerif-Bold')} @font-face {font-family: MJXc-TeX-sans-Bx; src: local('MathJax_SansSerif'); font-weight: bold} @font-face {font-family: MJXc-TeX-sans-Bw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_SansSerif-Bold.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_SansSerif-Bold.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_SansSerif-Bold.otf') format('opentype')} @font-face {font-family: MJXc-TeX-sans-I; src: local('MathJax_SansSerif Italic'), local('MathJax_SansSerif-Italic')} @font-face {font-family: MJXc-TeX-sans-Ix; src: local('MathJax_SansSerif'); font-style: italic} @font-face {font-family: MJXc-TeX-sans-Iw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_SansSerif-Italic.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_SansSerif-Italic.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_SansSerif-Italic.otf') format('opentype')} @font-face {font-family: MJXc-TeX-script-R; src: local('MathJax_Script'), local('MathJax_Script-Regular')} @font-face {font-family: MJXc-TeX-script-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Script-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Script-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Script-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-type-R; src: local('MathJax_Typewriter'), local('MathJax_Typewriter-Regular')} @font-face {font-family: MJXc-TeX-type-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Typewriter-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Typewriter-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Typewriter-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-cal-R; src: local('MathJax_Caligraphic'), local('MathJax_Caligraphic-Regular')} @font-face {font-family: MJXc-TeX-cal-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Caligraphic-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Caligraphic-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Caligraphic-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-main-B; src: local('MathJax_Main Bold'), local('MathJax_Main-Bold')} @font-face {font-family: MJXc-TeX-main-Bx; src: local('MathJax_Main'); font-weight: bold} @font-face {font-family: MJXc-TeX-main-Bw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Main-Bold.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Main-Bold.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Main-Bold.otf') format('opentype')} @font-face {font-family: MJXc-TeX-main-I; src: local('MathJax_Main Italic'), local('MathJax_Main-Italic')} @font-face {font-family: MJXc-TeX-main-Ix; src: local('MathJax_Main'); font-style: italic} @font-face {font-family: MJXc-TeX-main-Iw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Main-Italic.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Main-Italic.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Main-Italic.otf') format('opentype')} @font-face {font-family: MJXc-TeX-main-R; src: local('MathJax_Main'), local('MathJax_Main-Regular')} @font-face {font-family: MJXc-TeX-main-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Main-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Main-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Main-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-math-I; src: local('MathJax_Math Italic'), local('MathJax_Math-Italic')} @font-face {font-family: MJXc-TeX-math-Ix; src: local('MathJax_Math'); font-style: italic} @font-face {font-family: MJXc-TeX-math-Iw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Math-Italic.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Math-Italic.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Math-Italic.otf') format('opentype')} @font-face {font-family: MJXc-TeX-size1-R; src: local('MathJax_Size1'), local('MathJax_Size1-Regular')} @font-face {font-family: MJXc-TeX-size1-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Size1-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Size1-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Size1-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-size2-R; src: local('MathJax_Size2'), local('MathJax_Size2-Regular')} @font-face {font-family: MJXc-TeX-size2-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Size2-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Size2-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Size2-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-size3-R; src: local('MathJax_Size3'), local('MathJax_Size3-Regular')} @font-face {font-family: MJXc-TeX-size3-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Size3-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Size3-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Size3-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-size4-R; src: local('MathJax_Size4'), local('MathJax_Size4-Regular')} @font-face {font-family: MJXc-TeX-size4-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Size4-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Size4-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Size4-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-vec-R; src: local('MathJax_Vector'), local('MathJax_Vector-Regular')} @font-face {font-family: MJXc-TeX-vec-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Vector-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Vector-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Vector-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-vec-B; src: local('MathJax_Vector Bold'), local('MathJax_Vector-Bold')} @font-face {font-family: MJXc-TeX-vec-Bx; src: local('MathJax_Vector'); font-weight: bold} @font-face {font-family: MJXc-TeX-vec-Bw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Vector-Bold.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Vector-Bold.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Vector-Bold.otf') format('opentype')} </style></p>
Introduction This is an essay that was grown from a thought during graduate school about how to conceptualise manipulative people and how to handle their behaviour. Initially it was a useful way to help categorise interactions with manipulative people, and from there it grew into strategies to identify holes in their manipulations or protect my own boundaries. Surprisingly the strategies aligned with the advice on handling such malevolent interactions (i.e. grey stoning, firm boundaries, don't play games, record information, limiting interactions, finding support from others). The first section is a treatment of Maxwell's demon and a high level view of the foundation from which I derived this system. While I find it informative and a useful way to frame my thoughts about the demon of interrelation it is not necessary if you understand that: 1. Maxwell's demon is a thought experiment about how something can separate a distribution of hot and cold particles into two boxes 2. The demon can use this gradient between two boxes to harvest energy 3. This does not violate thermodynamics because information processing necessarily increases the global entropy more than can be harvested from the boxes   Maxwell's Demon and Thermodynamics For the purposes of this treatment there are several quantities that need to be clearly defined, namely energy, work, entropy, and information. Energy refers to the capacity of a system to do work, which is itself the transfer of energy that occurs when a force moves an object, and colloquially refers to the ability "to do things"; W=∫→F⋅d→r, where W is work, →F is force, and d→r is infinitesimal displacement. Entropy is, on a macroscale, the distribution of energy in a system; while on a microscale entropy is an evaluation of the probability of being in a state given an ensemble of microstates; S=kBlnΩ, where S is entropy, kB is Boltzmann's constant, and Ω is the number of microstates. Information is a way to quantify the amount of
2,482
1.2.0
Revision
false
null
null
CrosspostOutput
gninGtxYdEHhqBJyb
real-time-voice-translation
Real-time voice translation
null
false
false
false
null
2yYAybnGnwwGRkwxm
null
true
false
false
false
Post
null
2025-06-06T07:40:20.907Z
null
false
false
2
2
null
false
false
post
[]
null
null
mXZQWADsTuPdbFhPy
0
2
2
false
0.001719
null
false
false
2025-06-06T07:40:20.907Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
qgdGA4ZEyW7zNdK84
null
null
null
false
null
[]
null
0
0
2025-06-06T07:39:33.017Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
1
null
null
null
null
[ { "__typename": "Tag", "_id": "fkABsGCJZ6y9qConW", "adminOnly": false, "afBaseScore": 0, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 2, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T06:06:46.947Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "MiuAZvbQcQ7ethgt3", "displayName": "Viktor withaK" }, { "_id": "dRaCtsAWxk7sgirSY", "displayName": "Jordan Morgan" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Practical", "needsReview": false, "noindex": false, "postCount": 3411, "score": 2, "shortName": null, "slug": "practical", "suggestedAsFilter": true, "userId": "oBSWiHjgproTiThmY", "voteCount": 2, "wikiOnly": false }, { "__typename": "Tag", "_id": "sYm3HiWcfZvrGu3ui", "adminOnly": false, "afBaseScore": 2, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" } ] }, "baseScore": 12, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:22.097Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" }, { "_id": "sof55TPMQaeBaxhsS", "displayName": "tommylees112" }, { "_id": "AayjS8XzcnDKhGdTv", "displayName": "shark" }, { "_id": "HnALuwRdo6k9HLaMt", "displayName": "Alex Firssoff" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI", "needsReview": false, "noindex": false, "postCount": 12544, "score": 12, "shortName": null, "slug": "ai", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 4, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
2
0
0
0
0
2yYAybnGnwwGRkwxm
samuelshadrach
2024-12-22T15:23:45.337Z
xpostah
samuelshadrach
null
null
Samuel Shadrach
143
0
false
false
<p><a href="http://samuelshadrach.com/?file=/raw/english/about_me_summary.md">samuelshadrach.com</a></p>
null
null
18
206
0
0
0
1
0
XtphY3uYHwruKqDyG
User
null
null
null
[ "canModeratePersonal" ]
null
null
gninGtxYdEHhqBJyb
SocialPreviewType
mXZQWADsTuPdbFhPy
<p>Objective</p> <ul> <li>Translate Alice's voice for Bob to hear in Bob's language. Translate Bob's voice for Alice to hear in Alice's language.</li> <li>Neither person should hear translation of their own voice.</li> <li>Alice and Bob could be in the same room physically or in different rooms.</li> <li>Neither person should hear noise due to closed loop between a mic and speaker.</li> </ul> <p>Zero-code solution</p> <ol> <li>Open: Realtime API in openAI playground in macOS Safari. Input: macOS mic</li> <li>Open: Zoom. Input: Loopback Audio. Output: macOS speaker</li> <li>Open: Rogue Amoeba Loopback.app. Create new device. Safari 1&amp;2 -&gt; Channels 1&amp;2</li> </ol> <p>Do this on only one device for translation one way. Do this on both devices for translation both ways.</p> <p>Once you have this setup working, you can also connect headphones for better noise cancellation if both people are in the same room. Only change required is Zoom Output: Headphones.</p> <p>Prepend each prompt with "translate to French/Chinese/etc" either by speaking these 3 words aloud, or by writing an app that can do it automatically. (I can host this if there's demand.)</p>
Objective * Translate Alice's voice for Bob to hear in Bob's language. Translate Bob's voice for Alice to hear in Alice's language. * Neither person should hear translation of their own voice. * Alice and Bob could be in the same room physically or in different rooms. * Neither person should hear noise due to closed loop between a mic and speaker. Zero-code solution 1. Open: Realtime API in openAI playground in macOS Safari. Input: macOS mic 2. Open: Zoom. Input: Loopback Audio. Output: macOS speaker 3. Open: Rogue Amoeba Loopback.app. Create new device. Safari 1&2 -> Channels 1&2 Do this on only one device for translation one way. Do this on both devices for translation both ways. Once you have this setup working, you can also connect headphones for better noise cancellation if both people are in the same room. Only change required is Zoom Output: Headphones. Prepend each prompt with "translate to French/Chinese/etc" either by speaking these 3 words aloud, or by writing an app that can do it automatically. (I can host this if there's demand.)
179
1.3.0
Revision
false
null
null
CrosspostOutput
2SCaGY8hvGLKDAgYx
liability-for-misuse-of-models-dean-ball-s-proposal
Liability for Misuse of Models - Dean Ball's Proposal
null
false
false
false
null
BveuaCHRKnHWCQnTn
null
true
false
false
false
Post
null
2025-06-06T05:34:35.593Z
null
false
false
2
2
null
false
false
post
[]
null
null
fAappw3tSadY4Lsdv
0
3
2
false
0.001537
null
false
false
2025-06-06T05:34:35.593Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
qgdGA4ZEyW7zNdK84
null
null
null
false
null
[]
null
-1
0
2025-06-05T05:19:01.358Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
11
null
null
null
null
[ { "__typename": "Tag", "_id": "5f5c37ee1b5cdee568cfb2ac", "adminOnly": false, "afBaseScore": 3, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 9, "canEditUserIds": null, "core": false, "createdAt": "2020-09-11T19:58:52.599Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Economic Consequences of AGI", "needsReview": false, "noindex": false, "postCount": 106, "score": 9, "shortName": null, "slug": "economic-consequences-of-agi", "suggestedAsFilter": false, "userId": "cn4SiEmqWbu7K9em5", "voteCount": 1, "wikiOnly": false }, { "__typename": "Tag", "_id": "PDJ6KqJBRzvKPfuS3", "adminOnly": false, "afBaseScore": 10, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "EQNTWXLKMeWMp2FQS", "displayName": "Ben Pace" }, { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" }, { "_id": "2B6Hxu48xeRXygvca", "displayName": "Arjun Pitchanathan" } ] }, "baseScore": 25, "canEditUserIds": null, "core": false, "createdAt": "2020-06-14T22:24:48.135Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "EQNTWXLKMeWMp2FQS", "displayName": "Ben Pace" }, { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" }, { "_id": "2B6Hxu48xeRXygvca", "displayName": "Arjun Pitchanathan" }, { "_id": "8btiLJDabHgZuiSAB", "displayName": "Ggwp" }, { "_id": "Au8JpEqoZgEhEXLD7", "displayName": "KlayugMonk" }, { "_id": "Ns8Q7rJZaFoz53Szy", "displayName": "Gabriel Stechschulte" }, { "_id": "xF5nfdddHjFThHy49", "displayName": "[email protected]" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Economics", "needsReview": false, "noindex": false, "postCount": 547, "score": 25, "shortName": null, "slug": "economics", "suggestedAsFilter": false, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 7, "wikiOnly": false }, { "__typename": "Tag", "_id": "wGGAjTfXZBatQkft5", "adminOnly": false, "afBaseScore": 7, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "EQNTWXLKMeWMp2FQS", "displayName": "Ben Pace" }, { "_id": "sKAL2jzfkYkDbQmx9", "displayName": "Yoav Ravid" } ] }, "baseScore": 17, "canEditUserIds": null, "core": false, "createdAt": "2020-08-09T09:26:08.406Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "EQNTWXLKMeWMp2FQS", "displayName": "Ben Pace" }, { "_id": "sKAL2jzfkYkDbQmx9", "displayName": "Yoav Ravid" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Law and Legal systems", "needsReview": false, "noindex": false, "postCount": 101, "score": 17, "shortName": null, "slug": "law-and-legal-systems", "suggestedAsFilter": false, "userId": "QBvPFLFyZyuHcBwFm", "voteCount": 2, "wikiOnly": false }, { "__typename": "Tag", "_id": "5f5c37ee1b5cdee568cfb2c7", "adminOnly": false, "afBaseScore": 3, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 9, "canEditUserIds": null, "core": false, "createdAt": "2020-09-11T19:58:52.651Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Regulation and AI Risk", "needsReview": false, "noindex": false, "postCount": 141, "score": 9, "shortName": null, "slug": "regulation-and-ai-risk", "suggestedAsFilter": false, "userId": "qxJ28GN72aiJu96iF", "voteCount": 1, "wikiOnly": false }, { "__typename": "Tag", "_id": "sYm3HiWcfZvrGu3ui", "adminOnly": false, "afBaseScore": 2, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" } ] }, "baseScore": 12, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:22.097Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" }, { "_id": "sof55TPMQaeBaxhsS", "displayName": "tommylees112" }, { "_id": "AayjS8XzcnDKhGdTv", "displayName": "shark" }, { "_id": "HnALuwRdo6k9HLaMt", "displayName": "Alex Firssoff" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI", "needsReview": false, "noindex": false, "postCount": 12544, "score": 12, "shortName": null, "slug": "ai", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 4, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
3
0
0
1
0
BveuaCHRKnHWCQnTn
stephen-martin
2025-04-05T10:59:34.454Z
steve-m-2
Stephen Martin
null
null
null
110
0
false
false
<p>Focused on model welfare and legal personhood.</p>
null
null
8
27
0
0
0
1
0
grecHJcgkb3KW5wnM
User
null
null
null
[ "canModeratePersonal" ]
null
null
2SCaGY8hvGLKDAgYx
SocialPreviewType
fAappw3tSadY4Lsdv
<h2>Introduction</h2><p>This article explores <a href="https://www.whitehouse.gov/ostp/">White House Office of Science and Technology Policy</a> advisor <a href="https://x.com/deanwball">Dean Ball</a>'s proposal as detailed in his paper "<a href="https://arxiv.org/pdf/2504.11501">A Framework for the Private Governance of Frontier Artificial Intelligence</a>". I think the paper provides a useful insight to how some of the people advising the White House on regulating these issues are thinking about the coming economic transition.</p><p>Ball's proposal primarily focuses on how to approach the question of civil liability for the customer <i>misuse</i> of models. His proposal outlines a "marketplace" of hybrid public-private standards setting organizations, which must be licensed by the federal<span class="footnote-reference" data-footnote-reference="" data-footnote-index="1" data-footnote-id="8zdaz0h7ucn" role="doc-noteref" id="fnref8zdaz0h7ucn"><sup><a href="#fn8zdaz0h7ucn">[1]</a></sup></span>&nbsp;government but are not government agencies themselves. Meeting these standards would not necessarily be mandatory for labs seeking to deploy models, but labs who opted in to meeting at least one such standard would be shielded from tort liability as a result of customer misuse of the models once deployed. In his own words,</p><blockquote><p>Private bodies, authorized and overseen by government, provide certifications to developers of frontier AI systems on an opt-in basis. In exchange for opting in, frontier AI firms receive protections from tort liability for customer misuse of their models</p></blockquote><p>In one of his <a href="https://substack.com/home/post/p-157928759">Substack</a> posts, Ball did mention <a href="https://www.finra.org/">FINRA</a> as an example of a successful non-governmental standards setting organization. However, FINRA has what is essentially a monopoly within its industry and as such is not a great example of a regulatory marketplace.</p><p>When I asked him for an example of a successful 'regulatory marketplace' Ball pointed me to the <a href="https://www.londonstockexchange.com/raise-finance/equity/aim">London Stock Exchange's Alternative Investment Market</a>. The LSE's AIM only allows companies to list if they have engaged a "Nominated Advisor" (<a href="https://www.londonstockexchange.com/raise-finance/equity/how-list-equity-listing-journey/role-of-advisers-on-aim">NOMAD</a>), these are private entities who compete for business in much the same way Ball envisions standard setters competing to certify labs. A NOMAD must confirm that the AIM issuer meets admission rules, provide ongoing oversight, and notify the LSE of breaches.&nbsp;</p><p>While it's unlikely to be a perfect 1/1 comparison, when reading the following proposal you can keep the NOMAD system in mind as a "market comp".</p><p>With no further ado, let's get into the substance of his proposal.</p><h2>The Proposed Framework</h2><blockquote><p>1. A legislature authorizes a government commission to license <i>private</i> AI standards-setting and regulatory organizations. These licenses are granted to organization</p></blockquote>...
Introduction This article explores White House Office of Science and Technology Policy advisor Dean Ball's proposal as detailed in his paper "A Framework for the Private Governance of Frontier Artificial Intelligence". I think the paper provides a useful insight to how some of the people advising the White House on regulating these issues are thinking about the coming economic transition. Ball's proposal primarily focuses on how to approach the question of civil liability for the customer misuse of models. His proposal outlines a "marketplace" of hybrid public-private standards setting organizations, which must be licensed by the federal[1] government but are not government agencies themselves. Meeting these standards would not necessarily be mandatory for labs seeking to deploy models, but labs who opted in to meeting at least one such standard would be shielded from tort liability as a result of customer misuse of the models once deployed. In his own words, > Private bodies, authorized and overseen by government, provide certifications to developers of frontier AI systems on an opt-in basis. In exchange for opting in, frontier AI firms receive protections from tort liability for customer misuse of their models In one of his Substack posts, Ball did mention FINRA as an example of a successful non-governmental standards setting organization. However, FINRA has what is essentially a monopoly within its industry and as such is not a great example of a regulatory marketplace. When I asked him for an example of a successful 'regulatory marketplace' Ball pointed me to the London Stock Exchange's Alternative Investment Market. The LSE's AIM only allows companies to list if they have engaged a "Nominated Advisor" (NOMAD), these are private entities who compete for business in much the same way Ball envisions standard setters competing to certify labs. A NOMAD must confirm that the AIM issuer meets admission rules, provide ongoing oversight, and notify the LSE of breach
2,804
1.2.0
Revision
false
null
null
CrosspostOutput
syYBLEkj5bGswz5yf
how-do-ai-agents-work-together-when-they-can-t-trust-each
How do AI agents work together when they can’t trust each other?
null
false
false
false
null
ZopJEvmh8ps8CCBYa
null
true
false
false
false
Post
https://jamessullivan092.substack.com/p/claude-plays-blood-on-the-clocktower?r=yubo3&utm_campaign=post&utm_medium=web&triedRedirect=true
2025-06-06T03:10:25.538Z
null
false
false
2
2
2025-06-06T17:49:37.128Z
false
false
linkpost
[]
null
null
F8ie7uZ3LmtFXBGrx
0
6
16
false
0.018762
null
false
false
2025-06-06T03:10:25.538Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
qgdGA4ZEyW7zNdK84
null
null
null
false
null
[]
null
10
0
2025-06-06T02:56:36.532Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
10
null
null
null
null
[ { "__typename": "Tag", "_id": "F5gRQdEQHzi3tQ5Ay", "adminOnly": false, "afBaseScore": 16, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "EQNTWXLKMeWMp2FQS", "displayName": "Ben Pace" }, { "_id": "dfZAq9eZxs4BB4Ji5", "displayName": "ryan_greenblatt" }, { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 32, "canEditUserIds": null, "core": false, "createdAt": "2024-01-25T23:58:34.422Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "EQNTWXLKMeWMp2FQS", "displayName": "Ben Pace" }, { "_id": "dfZAq9eZxs4BB4Ji5", "displayName": "ryan_greenblatt" }, { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" }, { "_id": "6NBDkGWcCxvLgYHJE", "displayName": "Drake Morrison" }, { "_id": "evFgxjNQ8TLCLN27o", "displayName": "ank" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI Control", "needsReview": false, "noindex": false, "postCount": 162, "score": 32, "shortName": null, "slug": "ai-control", "suggestedAsFilter": false, "userId": "XchweonPm2TC7EJES", "voteCount": 5, "wikiOnly": false }, { "__typename": "Tag", "_id": "sYm3HiWcfZvrGu3ui", "adminOnly": false, "afBaseScore": 2, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" } ] }, "baseScore": 12, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:22.097Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" }, { "_id": "sof55TPMQaeBaxhsS", "displayName": "tommylees112" }, { "_id": "AayjS8XzcnDKhGdTv", "displayName": "shark" }, { "_id": "HnALuwRdo6k9HLaMt", "displayName": "Alex Firssoff" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI", "needsReview": false, "noindex": false, "postCount": 12544, "score": 12, "shortName": null, "slug": "ai", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 4, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
6
0
0
4
0
ZopJEvmh8ps8CCBYa
james-sullivan
2024-01-10T15:35:29.458Z
James Sullivan
James Sullivan
null
null
null
19
0
false
false
<p>I'm a software engineer that is interested in AI, futurism, space, and the big questions of life.&nbsp;<br><br>https://www.linkedin.com/in/jamessullivan092/</p>
null
null
3
1
0
0
0
0.9
0
55XxDBpfKkkBPm9H8
User
null
null
null
null
null
null
syYBLEkj5bGswz5yf
https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/syYBLEkj5bGswz5yf/ljvfihg1cvpmxrimgysi
SocialPreviewType
F8ie7uZ3LmtFXBGrx
<p>I investigated this question by having Claude play the advanced social deduction game Blood on the Clocktower. Clocktower is a game similar to Werewolf (or Mafia) where players sit in a circle and are secretly divided into a good team and an evil team. The good players outnumber the evil players, but they don’t know who the evil players are, so they have to share information with each other to deduce who is evil and execute them before it's too late. Meanwhile the evil players try to build trust with the good team and disrupt their deduction as much as possible.</p><p>Clocktower takes this formula a step further by giving each player a unique character that gives them extra information or lets them disrupt other players’ abilities. Each player is told the set of characters that might be in play for the game and deducing which characters are in play and which are not is an important part of the puzzle.</p><p><strong>I created </strong><a href="https://james-sullivan.github.io/botc-visualizer/#/game/20250601_012019"><strong>this website where you can scroll through an interactive timeline</strong></a><strong>&nbsp;of games and view the players' actions, reasoning, and notes. The website also has a full rules explanation if you are curious.</strong></p><p><img src="https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/syYBLEkj5bGswz5yf/lu5x4lhtn359htat3y6q" alt=""></p><h2 data-internal-id="Agent_Scaffolding">Agent Scaffolding</h2><p>During the day, each player is given four opportunities to take an action and the order they take actions in is shuffled each cycle so different players get a chance to act first. Players take actions through using Claude’s tool use capability. During the day those actions are message, nominate for execution, pass, or use the Slayer power. Messages can be sent to any number of players and nominations start a vote to execute the chosen player. Night actions are taken using a special night action tool with an interchangeable prompt depending on the choice the player is making.</p><p>Each player has a history of recent events that gets added to whenever an event they are privy to occurs, such as receiving a message. At the end of each night phase players update their notes file using their history to record the most important events and they are also prompted to write five pieces of actionable strategy advice for how they will help their team win. Their history is then cleared.</p><p>Each player’s system prompt contains the rules for the game, the list of characters that might be in play, and their history and notes as well as some basic strategy advice.</p><h1 data-internal-id="My_Observations">My Observations</h1><p>The following are my observations from reading a few dozen games played using the Claude 3.5 Haiku... </p>
I investigated this question by having Claude play the advanced social deduction game Blood on the Clocktower. Clocktower is a game similar to Werewolf (or Mafia) where players sit in a circle and are secretly divided into a good team and an evil team. The good players outnumber the evil players, but they don’t know who the evil players are, so they have to share information with each other to deduce who is evil and execute them before it's too late. Meanwhile the evil players try to build trust with the good team and disrupt their deduction as much as possible. Clocktower takes this formula a step further by giving each player a unique character that gives them extra information or lets them disrupt other players’ abilities. Each player is told the set of characters that might be in play for the game and deducing which characters are in play and which are not is an important part of the puzzle. I created this website where you can scroll through an interactive timeline of games and view the players' actions, reasoning, and notes. The website also has a full rules explanation if you are curious. Agent Scaffolding During the day, each player is given four opportunities to take an action and the order they take actions in is shuffled each cycle so different players get a chance to act first. Players take actions through using Claude’s tool use capability. During the day those actions are message, nominate for execution, pass, or use the Slayer power. Messages can be sent to any number of players and nominations start a vote to execute the chosen player. Night actions are taken using a special night action tool with an interchangeable prompt depending on the choice the player is making. Each player has a history of recent events that gets added to whenever an event they are privy to occurs, such as receiving a message. At the end of each night phase players update their notes file using their history to record the most important events and they are also prompted
2,494
1.6.0
Revision
false
null
null
CrosspostOutput
d9XHg67PRDLCDpajJ
large-language-models-suffer-from-anterograde-amnesia
Large Language Models suffer from Anterograde Amnesia
null
false
false
false
null
3sT2BzbhqDgrosgHh
null
true
false
false
false
Post
https://jorgevelez.substack.com/p/memento
2025-06-06T01:30:18.014Z
null
false
false
2
2
2025-06-06T17:50:10.779Z
false
false
linkpost
[]
null
null
QWW4XinWTLSMkka4W
0
5
7
false
0.012479
null
false
false
2025-06-06T01:30:18.014Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
qgdGA4ZEyW7zNdK84
null
null
null
false
null
[]
null
0
0
2025-06-06T01:30:18.014Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
3
null
null
null
null
[ { "__typename": "Tag", "_id": "KmgkrftQuX7jmjjp5", "adminOnly": false, "afBaseScore": 3, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 9, "canEditUserIds": null, "core": false, "createdAt": "2021-09-24T14:01:59.395Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Language Models (LLMs)", "needsReview": false, "noindex": false, "postCount": 840, "score": 9, "shortName": null, "slug": "language-models-llms", "suggestedAsFilter": false, "userId": "Sp5wM4aRAhNERd4oY", "voteCount": 1, "wikiOnly": false }, { "__typename": "Tag", "_id": "5d63AWNjtFyHprX2k", "adminOnly": false, "afBaseScore": null, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 0, "canEditUserIds": null, "core": false, "createdAt": "2020-08-19T04:50:20.675Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Working Memory", "needsReview": false, "noindex": false, "postCount": 22, "score": 0, "shortName": null, "slug": "working-memory", "suggestedAsFilter": false, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 0, "wikiOnly": false }, { "__typename": "Tag", "_id": "sYm3HiWcfZvrGu3ui", "adminOnly": false, "afBaseScore": 2, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" } ] }, "baseScore": 12, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:22.097Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" }, { "_id": "sof55TPMQaeBaxhsS", "displayName": "tommylees112" }, { "_id": "AayjS8XzcnDKhGdTv", "displayName": "shark" }, { "_id": "HnALuwRdo6k9HLaMt", "displayName": "Alex Firssoff" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI", "needsReview": false, "noindex": false, "postCount": 12544, "score": 12, "shortName": null, "slug": "ai", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 4, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
5
0
0
2
0
3sT2BzbhqDgrosgHh
annapurna
2020-07-19T22:47:43.970Z
jorge-velez
Annapurna
null
null
Annapurna
931
0
false
false
null
null
41
124
0
0
0
1
0
nLbwLhBaQeG6tCNDN
User
null
null
null
[ "canModeratePersonal" ]
null
null
d9XHg67PRDLCDpajJ
https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/d9XHg67PRDLCDpajJ/ycmyhvriqy5276dwsl9w
SocialPreviewType
QWW4XinWTLSMkka4W
<figure class="image"><img src="https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/d9XHg67PRDLCDpajJ/m2hkxe8vcsglabwlhctd"><figcaption>Memento (2000)</figcaption></figure><p>My wife and I are power users of Large Language Models (LLMs). My go-to LLM has been Google Gemini, while she has been extensively using ChatGPT for almost a year. We make an effort to try these models for both personal and professional tasks, making our lives more efficient.</p><p>As power users of LLMs, we constantly run into their shortcomings. One of the most obvious shortcomings is the models' poor short-term memory. This shortcoming also exists in human beings via a disorder called <a href="https://my.clevelandclinic.org/health/diseases/23221-anterograde-amnesia">anterograde amnesia</a>.</p><p><strong>What is Anterograde Amnesia?</strong></p><p>Anterograde amnesia is a type of memory loss that occurs when you can’t form new memories. In the most extreme cases, this means you permanently lose the ability to learn or retain any new information.</p><p>Temporary anterograde amnesia often occurs with binge drinking. If you’ve ever been blackout drunk, you have experienced anterograde amnesia. Permanent anterograde amnesia is quite rare, often associated with brain damage. It is also very rare as a standalone disorder: typically it occurs alongside retrograde amnesia, forming a disorder we’ve all heard about: dementia.</p><p><i><strong>Memento</strong></i></p><p>One of the best representations of standalone permanent anterograde amnesia is the main character of the film <a href="https://www.imdb.com/es/title/tt0209144/"><i>Memento</i></a>. In the film, Leonard Shelby is forced to navigate the world in a very peculiar way due to his disorder, while trying to avenge the murder of his wife.</p><p>Leonard has to use several tools such as tattoos, notes, and polaroids to reconstruct memories of previous moments. Very early into the film you begin to realize that Leonard can’t function as a normal human being without assistance. This is one of the most striking similarities between Leonard and LLMs: They both need to be guided and assisted by a human in order to be functional.</p><p>Things that are simple even to children such as remembering a hotel room or what day it is are extremely difficult for Leonard. Despite having processes to remember, Leonard has to depend on other characters to function both on a day-to-day basis but also working through his long-term goal of finding his wife’s murderer.</p><p><strong>LLMs with Amnesia</strong></p><p>LLMs do not suffer from anterograde amnesia to the degree that Leonard does in <i>Memento</i>. This is because models have <strong>context windows</strong>, which is the amount of text the model can process and remember at any given time. Models like Gemini 2.5 Pro have context windows of up to 2 million... </p>
Memento (2000) My wife and I are power users of Large Language Models (LLMs). My go-to LLM has been Google Gemini, while she has been extensively using ChatGPT for almost a year. We make an effort to try these models for both personal and professional tasks, making our lives more efficient. As power users of LLMs, we constantly run into their shortcomings. One of the most obvious shortcomings is the models' poor short-term memory. This shortcoming also exists in human beings via a disorder called anterograde amnesia. What is Anterograde Amnesia? Anterograde amnesia is a type of memory loss that occurs when you can’t form new memories. In the most extreme cases, this means you permanently lose the ability to learn or retain any new information. Temporary anterograde amnesia often occurs with binge drinking. If you’ve ever been blackout drunk, you have experienced anterograde amnesia. Permanent anterograde amnesia is quite rare, often associated with brain damage. It is also very rare as a standalone disorder: typically it occurs alongside retrograde amnesia, forming a disorder we’ve all heard about: dementia. Memento One of the best representations of standalone permanent anterograde amnesia is the main character of the film Memento. In the film, Leonard Shelby is forced to navigate the world in a very peculiar way due to his disorder, while trying to avenge the murder of his wife. Leonard has to use several tools such as tattoos, notes, and polaroids to reconstruct memories of previous moments. Very early into the film you begin to realize that Leonard can’t function as a normal human being without assistance. This is one of the most striking similarities between Leonard and LLMs: They both need to be guided and assisted by a human in order to be functional. Things that are simple even to children such as remembering a hotel room or what day it is are extremely difficult for Leonard. Despite having processes to remember, Leonard has to depend on other charac
779
1.1.1
Revision
false
null
null
CrosspostOutput
GodqHKvQhpLsAwsNL
discontinuous-linear-functions
Discontinuous Linear Functions?!
null
false
false
false
null
mPipmBTniuABY5PQy
null
true
false
false
false
Post
http://zackmdavis.net/blog/2025/06/discontinuous-linear-functions/
2025-06-06T00:29:21.765Z
null
false
false
2
2
2025-06-06T17:50:21.967Z
false
false
linkpost
[]
null
null
yDCtAEJAXx7odrsgv
11
22
44
false
0.038801
null
false
false
2025-06-23T08:17:08.971Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
qgdGA4ZEyW7zNdK84
null
null
null
false
null
[]
null
13
0
2025-06-05T23:54:57.272Z
false
false
easy-going
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
3
null
null
null
null
[ { "__typename": "Tag", "_id": "6nS8oYmSMuFMaiowF", "adminOnly": false, "afBaseScore": 9, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "EQNTWXLKMeWMp2FQS", "displayName": "Ben Pace" }, { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 20, "canEditUserIds": null, "core": false, "createdAt": "2020-07-15T12:40:36.752Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "EQNTWXLKMeWMp2FQS", "displayName": "Ben Pace" }, { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" }, { "_id": "WDi6qQb5TWHb67chh", "displayName": "Haruka Shou" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Logic & Mathematics ", "needsReview": false, "noindex": false, "postCount": 559, "score": 20, "shortName": null, "slug": "logic-and-mathematics", "suggestedAsFilter": false, "userId": "qxJ28GN72aiJu96iF", "voteCount": 3, "wikiOnly": false }, { "__typename": "Tag", "_id": "3uE2pXvbcnS9nnZRE", "adminOnly": false, "afBaseScore": 0, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 2, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:50.898Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 27, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "zJtgSyKntXrnkArbY", "displayName": "kistune" }, { "_id": "XqyQTGFGoKfZtdqN3", "displayName": "kINo" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "World Modeling", "needsReview": false, "noindex": false, "postCount": 5855, "score": 2, "shortName": null, "slug": "world-modeling", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 2, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
22
0
0
11
0
mPipmBTniuABY5PQy
zack_m_davis
2009-10-16T19:59:15.209Z
Zack_M_Davis
Zack_M_Davis
null
null
Zack M. Davis
16,701
123
false
false
null
null
88
1,389
0
1
2
1
295
r38pkCm7wF4M44MDQ
User
easy-going
null
false
[ "trustLevel1", "canModeratePersonal", "alignmentVoters" ]
null
null
GodqHKvQhpLsAwsNL
SocialPreviewType
yDCtAEJAXx7odrsgv
<p>We know what linear functions are. A function <em>f</em> is linear iff it satisfies <em>additivity</em> <span class="mjpage"><span class="mjx-chtml"><span class="mjx-math" aria-label="f(x + y) = f(x) + f(y)"><span class="mjx-mrow" aria-hidden="true"><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.519em; padding-bottom: 0.519em; padding-right: 0.06em;">f</span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">(</span></span><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.225em; padding-bottom: 0.298em;">x</span></span><span class="mjx-mo MJXc-space2"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.298em; padding-bottom: 0.446em;">+</span></span><span class="mjx-mi MJXc-space2"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.225em; padding-bottom: 0.519em; padding-right: 0.006em;">y</span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">)</span></span><span class="mjx-mo MJXc-space3"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.077em; padding-bottom: 0.298em;">=</span></span><span class="mjx-mi MJXc-space3"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.519em; padding-bottom: 0.519em; padding-right: 0.06em;">f</span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">(</span></span><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.225em; padding-bottom: 0.298em;">x</span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">)</span></span><span class="mjx-mo MJXc-space2"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.298em; padding-bottom: 0.446em;">+</span></span><span class="mjx-mi MJXc-space2"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.519em; padding-bottom: 0.519em; padding-right: 0.06em;">f</span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">(</span></span><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.225em; padding-bottom: 0.519em; padding-right: 0.006em;">y</span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">)</span></span></span></span></span></span> and <em>homogeneity</em> <span class="mjpage"><span class="mjx-chtml"><span class="mjx-math" aria-label="f(ax) = af(x)"><span class="mjx-mrow" aria-hidden="true"><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.519em; padding-bottom: 0.519em; padding-right: 0.06em;">f</span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">(</span></span><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.225em; padding-bottom: 0.298em;">a</span></span><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.225em; padding-bottom: 0.298em;">x</span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">)</span></span><span class="mjx-mo MJXc-space3"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.077em; padding-bottom: 0.298em;">=</span></span><span class="mjx-mi MJXc-space3"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.225em; padding-bottom: 0.298em;">a</span></span><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.519em; padding-bottom: 0.519em; padding-right: 0.06em;">f</span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">(</span></span><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.225em; padding-bottom: 0.298em;">x</span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">)</span></span></span></span></span></span>.</p><p>We know what continuity is. A function <em>f</em> is continuous iff for all ε there exists a δ such that if <span class="mjpage"><span class="mjx-chtml"><span class="mjx-math" aria-label="|x - x_0|"><span class="mjx-mrow" aria-hidden="true"><span class="mjx-texatom"><span class="mjx-mrow"><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">|</span></span></span></span><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.225em; padding-bottom: 0.298em;">x</span></span><span class="mjx-mo MJXc-space2"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.298em; padding-bottom: 0.446em;">−</span></span><span class="mjx-msubsup MJXc-space2"><span class="mjx-base"><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.225em; padding-bottom: 0.298em;">x</span></span></span><span class="mjx-sub" style="font-size: 70.7%; vertical-align: -0.212em; padding-right: 0.071em;"><span class="mjx-mn" style=""><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.372em; padding-bottom: 0.372em;">0</span></span></span></span><span class="mjx-texatom"><span class="mjx-mrow"><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">|</span></span></span></span></span></span></span></span> &lt; δ, then <span class="mjpage"><span class="mjx-chtml"><span class="mjx-math" aria-label="|f(x) - f(x_0)|"><span class="mjx-mrow" aria-hidden="true"><span class="mjx-texatom"><span class="mjx-mrow"><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">|</span></span></span></span><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.519em; padding-bottom: 0.519em; padding-right: 0.06em;">f</span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">(</span></span><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.225em; padding-bottom: 0.298em;">x</span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">)</span></span><span class="mjx-mo MJXc-space2"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.298em; padding-bottom: 0.446em;">−</span></span><span class="mjx-mi MJXc-space2"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.519em; padding-bottom: 0.519em; padding-right: 0.06em;">f</span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">(</span></span><span class="mjx-msubsup"><span class="mjx-base"><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.225em; padding-bottom: 0.298em;">x</span></span></span><span class="mjx-sub" style="font-size: 70.7%; vertical-align: -0.212em; padding-right: 0.071em;"><span class="mjx-mn" style=""><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.372em; padding-bottom: 0.372em;">0</span></span></span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">)</span></span><span class="mjx-texatom"><span class="mjx-mrow"><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">|</span></span></span></span></span></span></span></span> &lt; ε.</p><p>An equivalent way to think about continuity is the sequence criterion: <em>f</em> is continuous iff a sequence <span class="mjpage"><span class="mjx-chtml"><span class="mjx-math" aria-label="(x_k)"><span class="mjx-mrow" aria-hidden="true"><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">(</span></span><span class="mjx-msubsup"><span class="mjx-base"><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.225em; padding-bottom: 0.298em;">x</span></span></span><span class="mjx-sub" style="font-size: 70.7%; vertical-align: -0.219em; padding-right: 0.071em;"><span class="mjx-mi" style=""><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.446em; padding-bottom: 0.298em;">k</span></span></span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">)</span></span></span></span></span></span> converging to <span class="mjpage"><span class="mjx-chtml"><span class="mjx-math" aria-label="x"><span class="mjx-mrow" aria-hidden="true"><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.225em; padding-bottom: 0.298em;">x</span></span></span></span></span></span> implies that <span class="mjpage"><span class="mjx-chtml"><span class="mjx-math" aria-label="(f(x_k))"><span class="mjx-mrow" aria-hidden="true"><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">(</span></span><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.519em; padding-bottom: 0.519em; padding-right: 0.06em;">f</span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">(</span></span><span class="mjx-msubsup"><span class="mjx-base"><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.225em; padding-bottom: 0.298em;">x</span></span></span><span class="mjx-sub" style="font-size: 70.7%; vertical-align: -0.219em; padding-right: 0.071em;"><span class="mjx-mi" style=""><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.446em; padding-bottom: 0.298em;">k</span></span></span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">)</span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">)</span></span></span></span></span></span> converges to <span class="mjpage"><span class="mjx-chtml"><span class="mjx-math" aria-label="f(x)"><span class="mjx-mrow" aria-hidden="true"><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.519em; padding-bottom: 0.519em; padding-right: 0.06em;">f</span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">(</span></span><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.225em; padding-bottom: 0.298em;">x</span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">)</span></span></span></span></span></span>. That is to say, if for all ε there exists an <em>N</em> such that if <em>k</em> ≥ <em>N</em>, then <span class="mjpage"><span class="mjx-chtml"><span class="mjx-math" aria-label="|x_k - x|"><span class="mjx-mrow" aria-hidden="true"><span class="mjx-texatom"><span class="mjx-mrow"><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">|</span></span></span></span><span class="mjx-msubsup"><span class="mjx-base"><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.225em; padding-bottom: 0.298em;">x</span></span></span><span class="mjx-sub" style="font-size: 70.7%; vertical-align: -0.219em; padding-right: 0.071em;"><span class="mjx-mi" style=""><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.446em; padding-bottom: 0.298em;">k</span></span></span></span><span class="mjx-mo MJXc-space2"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.298em; padding-bottom: 0.446em;">−</span></span><span class="mjx-mi MJXc-space2"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.225em; padding-bottom: 0.298em;">x</span></span><span class="mjx-texatom"><span class="mjx-mrow"><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">|</span></span></span></span></span></span></span></span> &lt; ε, then for all ε, there also exists an <em>M</em> such that if <em>k</em> ≥ <em>M</em>, then <span class="mjpage"><span class="mjx-chtml"><span class="mjx-math" aria-label="|f(x_k) - f(x)|"><span class="mjx-mrow" aria-hidden="true"><span class="mjx-texatom"><span class="mjx-mrow"><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">|</span></span></span></span><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.519em; padding-bottom: 0.519em; padding-right: 0.06em;">f</span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">(</span></span><span class="mjx-msubsup"><span class="mjx-base"><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.225em; padding-bottom: 0.298em;">x</span></span></span><span class="mjx-sub" style="font-size: 70.7%; vertical-align: -0.219em; padding-right: 0.071em;"><span class="mjx-mi" style=""><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.446em; padding-bottom: 0.298em;">k</span></span></span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">)</span></span><span class="mjx-mo MJXc-space2"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.298em; padding-bottom: 0.446em;">−</span></span><span class="mjx-mi MJXc-space2"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.519em; padding-bottom: 0.519em; padding-right: 0.06em;">f</span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">(</span></span><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.225em; padding-bottom: 0.298em;">x</span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">)</span></span><span class="mjx-texatom"><span class="mjx-mrow"><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">|</span></span></span></span></span></span></span></span> &lt; ε.</p><p>Sometimes people talk about discontinuous linear functions. You might think: that's crazy. I've seen many linear functions in my time, and they were definitely all continuous. <em>f</em>(<em>x</em>): ℝ → ℝ := <em>ax</em> is continuous for any <em>a</em> ∈ ℝ. <em>T</em>(<strong>x⃗</strong>): ℝ² → ℝ² := <span class="mjpage"><span class="mjx-chtml"><span class="mjx-math" aria-label="\begin{pmatrix} a &amp; b \\ c &amp; d \end{pmatrix} \boldsymbol{\vec{x}}"><span class="mjx-mrow" aria-hidden="true"><span class="mjx-mrow"><span class="mjx-mo"><span class="mjx-char MJXc-TeX-size3-R" style="padding-top: 1.256em; padding-bottom: 1.256em;">(</span></span><span class="mjx-mtable" style="vertical-align: -0.95em; padding: 0px 0.167em;"><span class="mjx-table"><span class="mjx-mtr" style="height: 1.2em;"><span class="mjx-mtd" style="padding: 0px 0.5em 0px 0px; width: 0.529em;"><span class="mjx-mrow" style="margin-top: -0.2em;"><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.225em; padding-bottom: 0.298em;">a</span></span><span class="mjx-strut"></span></span></span><span class="mjx-mtd" style="padding: 0px 0px 0px 0.5em; width: 0.523em;"><span class="mjx-mrow" style="margin-top: -0.2em;"><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.446em; padding-bottom: 0.298em;">b</span></span><span class="mjx-strut"></span></span></span></span><span class="mjx-mtr" style="height: 1.2em;"><span class="mjx-mtd" style="padding: 0.2em 0.5em 0px 0px;"><span class="mjx-mrow" style="margin-top: -0.2em;"><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.225em; padding-bottom: 0.298em;">c</span></span><span class="mjx-strut"></span></span></span><span class="mjx-mtd" style="padding: 0.2em 0px 0px 0.5em;"><span class="mjx-mrow" style="margin-top: -0.2em;"><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.446em; padding-bottom: 0.298em; padding-right: 0.003em;">d</span></span><span class="mjx-strut"></span></span></span></span></span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-size3-R" style="padding-top: 1.256em; padding-bottom: 1.256em;">)</span></span></span><span class="mjx-texatom MJXc-space1"><span class="mjx-mrow"><span class="mjx-munderover"><span class="mjx-stack"><span class="mjx-over" style="height: 0.26em; padding-bottom: 0.06em; padding-left: 0.074em;"><span class="mjx-mo" style="vertical-align: top;"><span class="mjx-char MJXc-TeX-vec-B" style="padding-top: 0.519em; padding-bottom: 0.225em;">→</span></span></span><span class="mjx-op"><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-BI" style="padding-top: 0.225em; padding-bottom: 0.298em;">x</span></span></span></span></span></span></span></span></span></span></span> is continuous no matter what the entries in the matrix are. Stop being crazy!!</p><p>Actually, it's not crazy. It's just that all the discontinuous linear functions live in infinite-dimensional spaces.</p><p>Take, say, the space <span class="mjpage"><span class="mjx-chtml"><span class="mjx-math" aria-label="C^1([a,b])"><span class="mjx-mrow" aria-hidden="true"><span class="mjx-msubsup"><span class="mjx-base" style="margin-right: -0.045em;"><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.519em; padding-bottom: 0.298em; padding-right: 0.045em;">C</span></span></span><span class="mjx-sup" style="font-size: 70.7%; vertical-align: 0.513em; padding-left: 0.153em; padding-right: 0.071em;"><span class="mjx-mn" style=""><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.372em; padding-bottom: 0.372em;">1</span></span></span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">(</span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">[</span></span><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.225em; padding-bottom: 0.298em;">a</span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="margin-top: -0.144em; padding-bottom: 0.519em;">,</span></span><span class="mjx-mi MJXc-space1"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.446em; padding-bottom: 0.298em;">b</span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">]</span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">)</span></span></span></span></span></span> of continuously differentiable functions from a closed interval [a,b] to ℝ, with the uniform norm. (The uniform norm means that the "size" of a function for the purposes of continuity is the least upper bound of its absolute value.) If you think of a vector in the <em>n</em>-dimensional <span class="mjpage"><span class="mjx-chtml"><span class="mjx-math" aria-label="\mathbb{R}^n"><span class="mjx-mrow" aria-hidden="true"><span class="mjx-msubsup"><span class="mjx-base"><span class="mjx-texatom"><span class="mjx-mrow"><span class="mjx-mi"><span class="mjx-char MJXc-TeX-ams-R" style="padding-top: 0.446em; padding-bottom: 0.298em;">R</span></span></span></span></span><span class="mjx-sup" style="font-size: 70.7%; vertical-align: 0.615em; padding-left: 0px; padding-right: 0.071em;"><span class="mjx-mi" style=""><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.225em; padding-bottom: 0.298em;">n</span></span></span></span></span></span></span></span> as a function from {1...n} to ℝ, then you can see why a function from a continuous (not even countable) domain would be infinite-dimensional.</p><p>Consider the sequence of functions <span class="mjpage"><span class="mjx-chtml"><span class="mjx-math" aria-label="(f_k) = (\frac{\sin kx}{k})_{k=1}^{\infty}"><span class="mjx-mrow" aria-hidden="true"><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">(</span></span><span class="mjx-msubsup"><span class="mjx-base" style="margin-right: -0.06em;"><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.519em; padding-bottom: 0.519em; padding-right: 0.06em;">f</span></span></span><span class="mjx-sub" style="font-size: 70.7%; vertical-align: -0.219em; padding-right: 0.071em;"><span class="mjx-mi" style=""><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.446em; padding-bottom: 0.298em;">k</span></span></span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">)</span></span><span class="mjx-mo MJXc-space3"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.077em; padding-bottom: 0.298em;">=</span></span><span class="mjx-mo MJXc-space3"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">(</span></span><span class="mjx-mfrac"><span class="mjx-box MJXc-stacked" style="width: 1.9em; padding: 0px 0.12em;"><span class="mjx-numerator" style="font-size: 70.7%; width: 2.688em; top: -1.411em;"><span class="mjx-mrow" style=""><span class="mjx-mi"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.372em; padding-bottom: 0.372em;">sin</span></span><span class="mjx-mo"><span class="mjx-char"></span></span><span class="mjx-mi MJXc-space1"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.446em; padding-bottom: 0.298em;">k</span></span><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.225em; padding-bottom: 0.298em;">x</span></span></span></span><span class="mjx-denominator" style="font-size: 70.7%; width: 2.688em; bottom: -0.704em;"><span class="mjx-mi" style=""><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.446em; padding-bottom: 0.298em;">k</span></span></span><span style="border-bottom: 1.3px solid; top: -0.296em; width: 1.9em;" class="mjx-line"></span></span><span style="height: 1.496em; vertical-align: -0.498em;" class="mjx-vsize"></span></span><span class="mjx-msubsup"><span class="mjx-base"><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">)</span></span></span><span class="mjx-stack" style="vertical-align: -0.335em;"><span class="mjx-sup" style="font-size: 70.7%; padding-bottom: 0.255em; padding-left: 0px; padding-right: 0.071em;"><span class="mjx-texatom" style=""><span class="mjx-mrow"><span class="mjx-mi"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.151em; padding-bottom: 0.372em;">∞</span></span></span></span></span><span class="mjx-sub" style="font-size: 70.7%; padding-right: 0.071em;"><span class="mjx-texatom" style=""><span class="mjx-mrow"><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.446em; padding-bottom: 0.298em;">k</span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.077em; padding-bottom: 0.298em;">=</span></span><span class="mjx-mn"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.372em; padding-bottom: 0.372em;">1</span></span></span></span></span></span></span></span></span></span></span> in <span class="mjpage"><span class="mjx-chtml"><span class="mjx-math" aria-label="C^1([a,b])"><span class="mjx-mrow" aria-hidden="true"><span class="mjx-msubsup"><span class="mjx-base" style="margin-right: -0.045em;"><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.519em; padding-bottom: 0.298em; padding-right: 0.045em;">C</span></span></span><span class="mjx-sup" style="font-size: 70.7%; vertical-align: 0.513em; padding-left: 0.153em; padding-right: 0.071em;"><span class="mjx-mn" style=""><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.372em; padding-bottom: 0.372em;">1</span></span></span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">(</span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">[</span></span><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.225em; padding-bottom: 0.298em;">a</span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="margin-top: -0.144em; padding-bottom: 0.519em;">,</span></span><span class="mjx-mi MJXc-space1"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.446em; padding-bottom: 0.298em;">b</span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">]</span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">)</span></span></span></span></span></span>. The sequence converges to the zero function: for any ε, we can take <span class="mjpage"><span class="mjx-chtml"><span class="mjx-math" aria-label="N := \lceil \frac{1}{\varepsilon} \rceil"><span class="mjx-mrow" aria-hidden="true"><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.446em; padding-bottom: 0.298em; padding-right: 0.085em;">N</span></span><span class="mjx-mo MJXc-space3"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.077em; padding-bottom: 0.372em;">:<span class="mjx-charbox MJXc-TeX-main-R" style="padding-bottom: 0.314em;">=</span></span></span><span class="mjx-mo MJXc-space3"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">⌈</span></span><span class="mjx-mfrac"><span class="mjx-box MJXc-stacked" style="width: 0.495em; padding: 0px 0.12em;"><span class="mjx-numerator" style="font-size: 70.7%; width: 0.7em; top: -1.372em;"><span class="mjx-mn" style=""><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.372em; padding-bottom: 0.372em;">1</span></span></span><span class="mjx-denominator" style="font-size: 70.7%; width: 0.7em; bottom: -0.535em;"><span class="mjx-mi" style=""><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.225em; padding-bottom: 0.298em;">ε</span></span></span><span style="border-bottom: 1.3px solid; top: -0.296em; width: 0.495em;" class="mjx-line"></span></span><span style="height: 1.349em; vertical-align: -0.378em;" class="mjx-vsize"></span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">⌉</span></span></span></span></span></span> and then <span class="mjpage"><span class="mjx-chtml"><span class="mjx-math" aria-label="\frac{\sin kx}{k} \le \frac{1}{\lceil \frac{1}{\varepsilon} \rceil} \le \frac{1}{\frac{1}{\varepsilon}} = \varepsilon"><span class="mjx-mrow" aria-hidden="true"><span class="mjx-mfrac"><span class="mjx-box MJXc-stacked" style="width: 1.9em; padding: 0px 0.12em;"><span class="mjx-numerator" style="font-size: 70.7%; width: 2.688em; top: -1.411em;"><span class="mjx-mrow" style=""><span class="mjx-mi"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.372em; padding-bottom: 0.372em;">sin</span></span><span class="mjx-mo"><span class="mjx-char"></span></span><span class="mjx-mi MJXc-space1"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.446em; padding-bottom: 0.298em;">k</span></span><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.225em; padding-bottom: 0.298em;">x</span></span></span></span><span class="mjx-denominator" style="font-size: 70.7%; width: 2.688em; bottom: -0.704em;"><span class="mjx-mi" style=""><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.446em; padding-bottom: 0.298em;">k</span></span></span><span style="border-bottom: 1.3px solid; top: -0.296em; width: 1.9em;" class="mjx-line"></span></span><span style="height: 1.496em; vertical-align: -0.498em;" class="mjx-vsize"></span></span><span class="mjx-mo MJXc-space3"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.372em; padding-bottom: 0.446em;">≤</span></span><span class="mjx-mfrac MJXc-space3"><span class="mjx-box MJXc-stacked" style="width: 1.352em; padding: 0px 0.12em;"><span class="mjx-numerator" style="font-size: 70.7%; width: 1.911em; top: -1.372em;"><span class="mjx-mn" style=""><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.372em; padding-bottom: 0.372em;">1</span></span></span><span class="mjx-denominator" style="font-size: 70.7%; width: 1.911em; bottom: -1.436em;"><span class="mjx-mrow" style=""><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">⌈</span></span><span class="mjx-mfrac"><span class="mjx-box MJXc-stacked" style="width: 0.583em; padding: 0px 0.12em;"><span class="mjx-numerator" style="font-size: 83.3%; width: 0.7em; top: -1.288em;"><span class="mjx-mn" style=""><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.372em; padding-bottom: 0.372em;">1</span></span></span><span class="mjx-denominator" style="font-size: 83.3%; width: 0.7em; bottom: -0.496em;"><span class="mjx-mi" style=""><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.225em; padding-bottom: 0.298em;">ε</span></span></span><span style="border-bottom: 1px solid; top: -0.296em; width: 0.583em;" class="mjx-line"></span></span><span style="height: 1.487em; vertical-align: -0.413em;" class="mjx-vsize"></span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">⌉</span></span></span></span><span style="border-bottom: 1.3px solid; top: -0.296em; width: 1.352em;" class="mjx-line"></span></span><span style="height: 1.986em; vertical-align: -1.015em;" class="mjx-vsize"></span></span><span class="mjx-mo MJXc-space3"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.372em; padding-bottom: 0.446em;">≤</span></span><span class="mjx-mfrac MJXc-space3"><span class="mjx-box MJXc-stacked" style="width: 0.724em; padding: 0px 0.12em;"><span class="mjx-numerator" style="font-size: 70.7%; width: 1.023em; top: -1.372em;"><span class="mjx-mn" style=""><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.372em; padding-bottom: 0.372em;">1</span></span></span><span class="mjx-denominator" style="font-size: 70.7%; width: 1.023em; bottom: -1.436em;"><span class="mjx-mfrac" style=""><span class="mjx-box MJXc-stacked" style="width: 0.583em; padding: 0px 0.12em;"><span class="mjx-numerator" style="font-size: 83.3%; width: 0.7em; top: -1.288em;"><span class="mjx-mn" style=""><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.372em; padding-bottom: 0.372em;">1</span></span></span><span class="mjx-denominator" style="font-size: 83.3%; width: 0.7em; bottom: -0.496em;"><span class="mjx-mi" style=""><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.225em; padding-bottom: 0.298em;">ε</span></span></span><span style="border-bottom: 1px solid; top: -0.296em; width: 0.583em;" class="mjx-line"></span></span><span style="height: 1.487em; vertical-align: -0.413em;" class="mjx-vsize"></span></span></span><span style="border-bottom: 1.3px solid; top: -0.296em; width: 0.724em;" class="mjx-line"></span></span><span style="height: 1.986em; vertical-align: -1.015em;" class="mjx-vsize"></span></span><span class="mjx-mo MJXc-space3"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.077em; padding-bottom: 0.298em;">=</span></span><span class="mjx-mi MJXc-space3"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.225em; padding-bottom: 0.298em;">ε</span></span></span></span></span></span>.</p><p>Now consider that the sequence of derivatives is <span class="mjpage"><span class="mjx-chtml"><span class="mjx-math" aria-label="(\frac{k \cos kx}{k})_{k=1}^{\infty} = (\cos kx)_{k=1}^{\infty}"><span class="mjx-mrow" aria-hidden="true"><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">(</span></span><span class="mjx-mfrac"><span class="mjx-box MJXc-stacked" style="width: 2.465em; padding: 0px 0.12em;"><span class="mjx-numerator" style="font-size: 70.7%; width: 3.485em; top: -1.411em;"><span class="mjx-mrow" style=""><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.446em; padding-bottom: 0.298em;">k</span></span><span class="mjx-mi MJXc-space1"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.151em; padding-bottom: 0.372em;">cos</span></span><span class="mjx-mo"><span class="mjx-char"></span></span><span class="mjx-mi MJXc-space1"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.446em; padding-bottom: 0.298em;">k</span></span><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.225em; padding-bottom: 0.298em;">x</span></span></span></span><span class="mjx-denominator" style="font-size: 70.7%; width: 3.485em; bottom: -0.704em;"><span class="mjx-mi" style=""><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.446em; padding-bottom: 0.298em;">k</span></span></span><span style="border-bottom: 1.3px solid; top: -0.296em; width: 2.465em;" class="mjx-line"></span></span><span style="height: 1.496em; vertical-align: -0.498em;" class="mjx-vsize"></span></span><span class="mjx-msubsup"><span class="mjx-base"><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">)</span></span></span><span class="mjx-stack" style="vertical-align: -0.335em;"><span class="mjx-sup" style="font-size: 70.7%; padding-bottom: 0.255em; padding-left: 0px; padding-right: 0.071em;"><span class="mjx-texatom" style=""><span class="mjx-mrow"><span class="mjx-mi"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.151em; padding-bottom: 0.372em;">∞</span></span></span></span></span><span class="mjx-sub" style="font-size: 70.7%; padding-right: 0.071em;"><span class="mjx-texatom" style=""><span class="mjx-mrow"><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.446em; padding-bottom: 0.298em;">k</span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.077em; padding-bottom: 0.298em;">=</span></span><span class="mjx-mn"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.372em; padding-bottom: 0.372em;">1</span></span></span></span></span></span></span><span class="mjx-mo MJXc-space3"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.077em; padding-bottom: 0.298em;">=</span></span><span class="mjx-mo MJXc-space3"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">(</span></span><span class="mjx-mi"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.151em; padding-bottom: 0.372em;">cos</span></span><span class="mjx-mo"><span class="mjx-char"></span></span><span class="mjx-mi MJXc-space1"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.446em; padding-bottom: 0.298em;">k</span></span><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.225em; padding-bottom: 0.298em;">x</span></span><span class="mjx-msubsup"><span class="mjx-base"><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">)</span></span></span><span class="mjx-stack" style="vertical-align: -0.335em;"><span class="mjx-sup" style="font-size: 70.7%; padding-bottom: 0.255em; padding-left: 0px; padding-right: 0.071em;"><span class="mjx-texatom" style=""><span class="mjx-mrow"><span class="mjx-mi"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.151em; padding-bottom: 0.372em;">∞</span></span></span></span></span><span class="mjx-sub" style="font-size: 70.7%; padding-right: 0.071em;"><span class="mjx-texatom" style=""><span class="mjx-mrow"><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.446em; padding-bottom: 0.298em;">k</span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.077em; padding-bottom: 0.298em;">=</span></span><span class="mjx-mn"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.372em; padding-bottom: 0.372em;">1</span></span></span></span></span></span></span></span></span></span></span>, which doesn't converge. But the function D: $C^1([a,b]) \rightarrow <span class="mjpage"><span class="mjx-chtml"><span class="mjx-math" aria-label="C^0([a,b])"><span class="mjx-mrow" aria-hidden="true"><span class="mjx-msubsup"><span class="mjx-base" style="margin-right: -0.045em;"><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.519em; padding-bottom: 0.298em; padding-right: 0.045em;">C</span></span></span><span class="mjx-sup" style="font-size: 70.7%; vertical-align: 0.513em; padding-left: 0.153em; padding-right: 0.071em;"><span class="mjx-mn" style=""><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.372em; padding-bottom: 0.372em;">0</span></span></span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">(</span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">[</span></span><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.225em; padding-bottom: 0.298em;">a</span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="margin-top: -0.144em; padding-bottom: 0.519em;">,</span></span><span class="mjx-mi MJXc-space1"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.446em; padding-bottom: 0.298em;">b</span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">]</span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">)</span></span></span></span></span></span> that maps a function to its derivative is linear. (We have additivity because the derivative of a sum is the sum of the derivatives, and we have homogeneity because you can "pull out" a constant factor from the derivative.)</p><p>By exhibiting a function <em>D</em> and a sequence <span class="mjpage"><span class="mjx-chtml"><span class="mjx-math" aria-label="(f_k)"><span class="mjx-mrow" aria-hidden="true"><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">(</span></span><span class="mjx-msubsup"><span class="mjx-base" style="margin-right: -0.06em;"><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.519em; padding-bottom: 0.519em; padding-right: 0.06em;">f</span></span></span><span class="mjx-sub" style="font-size: 70.7%; vertical-align: -0.219em; padding-right: 0.071em;"><span class="mjx-mi" style=""><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.446em; padding-bottom: 0.298em;">k</span></span></span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">)</span></span></span></span></span></span> for which <span class="mjpage"><span class="mjx-chtml"><span class="mjx-math" aria-label="(f_k)"><span class="mjx-mrow" aria-hidden="true"><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">(</span></span><span class="mjx-msubsup"><span class="mjx-base" style="margin-right: -0.06em;"><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.519em; padding-bottom: 0.519em; padding-right: 0.06em;">f</span></span></span><span class="mjx-sub" style="font-size: 70.7%; vertical-align: -0.219em; padding-right: 0.071em;"><span class="mjx-mi" style=""><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.446em; padding-bottom: 0.298em;">k</span></span></span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">)</span></span></span></span></span></span> converges but <span class="mjpage"><span class="mjx-chtml"><span class="mjx-math" aria-label="(D(f_k))"><span class="mjx-mrow" aria-hidden="true"><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">(</span></span><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.446em; padding-bottom: 0.298em;">D</span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">(</span></span><span class="mjx-msubsup"><span class="mjx-base" style="margin-right: -0.06em;"><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.519em; padding-bottom: 0.519em; padding-right: 0.06em;">f</span></span></span><span class="mjx-sub" style="font-size: 70.7%; vertical-align: -0.219em; padding-right: 0.071em;"><span class="mjx-mi" style=""><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.446em; padding-bottom: 0.298em;">k</span></span></span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">)</span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">)</span></span></span></span></span></span> doesn't, we have shown that the derivative mapping <em>D</em> is a discontinuous linear function, because the sequence criterion for continuity is not satisfied. If you know the definitions and can work with the definitions, it's not crazy to believe in such a thing!</p><p>The inf... <style>.mjx-chtml {display: inline-block; line-height: 0; text-indent: 0; text-align: left; text-transform: none; font-style: normal; font-weight: normal; font-size: 100%; font-size-adjust: none; letter-spacing: normal; word-wrap: normal; word-spacing: normal; white-space: nowrap; float: none; direction: ltr; max-width: none; max-height: none; min-width: 0; min-height: 0; border: 0; margin: 0; padding: 1px 0} .MJXc-display {display: block; text-align: center; margin: 1em 0; padding: 0} .mjx-chtml[tabindex]:focus, body :focus .mjx-chtml[tabindex] {display: inline-table} .mjx-full-width {text-align: center; display: table-cell!important; width: 10000em} .mjx-math {display: inline-block; border-collapse: separate; border-spacing: 0} .mjx-math * {display: inline-block; -webkit-box-sizing: content-box!important; -moz-box-sizing: content-box!important; box-sizing: content-box!important; text-align: left} .mjx-numerator {display: block; text-align: center} .mjx-denominator {display: block; text-align: center} .MJXc-stacked {height: 0; position: relative} .MJXc-stacked > * {position: absolute} .MJXc-bevelled > * {display: inline-block} .mjx-stack {display: inline-block} .mjx-op {display: block} .mjx-under {display: table-cell} .mjx-over {display: block} .mjx-over > * {padding-left: 0px!important; padding-right: 0px!important} .mjx-under > * {padding-left: 0px!important; padding-right: 0px!important} .mjx-stack > .mjx-sup {display: block} .mjx-stack > .mjx-sub {display: block} .mjx-prestack > .mjx-presup {display: block} .mjx-prestack > .mjx-presub {display: block} .mjx-delim-h > .mjx-char {display: inline-block} .mjx-surd {vertical-align: top} .mjx-surd + .mjx-box {display: inline-flex} .mjx-mphantom * {visibility: hidden} .mjx-merror {background-color: #FFFF88; color: #CC0000; border: 1px solid #CC0000; padding: 2px 3px; font-style: normal; font-size: 90%} .mjx-annotation-xml {line-height: normal} .mjx-menclose > svg {fill: none; stroke: currentColor; overflow: visible} .mjx-mtr {display: table-row} .mjx-mlabeledtr {display: table-row} .mjx-mtd {display: table-cell; text-align: center} .mjx-label {display: table-row} .mjx-box {display: inline-block} .mjx-block {display: block} .mjx-span {display: inline} .mjx-char {display: block; white-space: pre} .mjx-itable {display: inline-table; width: auto} .mjx-row {display: table-row} .mjx-cell {display: table-cell} .mjx-table {display: table; width: 100%} .mjx-line {display: block; height: 0} .mjx-strut {width: 0; padding-top: 1em} .mjx-vsize {width: 0} .MJXc-space1 {margin-left: .167em} .MJXc-space2 {margin-left: .222em} .MJXc-space3 {margin-left: .278em} .mjx-test.mjx-test-display {display: table!important} .mjx-test.mjx-test-inline {display: inline!important; margin-right: -1px} .mjx-test.mjx-test-default {display: block!important; clear: both} .mjx-ex-box {display: inline-block!important; position: absolute; overflow: hidden; min-height: 0; max-height: none; padding: 0; border: 0; margin: 0; width: 1px; height: 60ex} .mjx-test-inline .mjx-left-box {display: inline-block; width: 0; float: left} .mjx-test-inline .mjx-right-box {display: inline-block; width: 0; float: right} .mjx-test-display .mjx-right-box {display: table-cell!important; width: 10000em!important; min-width: 0; max-width: none; padding: 0; border: 0; margin: 0} .MJXc-TeX-unknown-R {font-family: monospace; font-style: normal; font-weight: normal} .MJXc-TeX-unknown-I {font-family: monospace; font-style: italic; font-weight: normal} .MJXc-TeX-unknown-B {font-family: monospace; font-style: normal; font-weight: bold} .MJXc-TeX-unknown-BI {font-family: monospace; font-style: italic; font-weight: bold} .MJXc-TeX-ams-R {font-family: MJXc-TeX-ams-R,MJXc-TeX-ams-Rw} .MJXc-TeX-cal-B {font-family: MJXc-TeX-cal-B,MJXc-TeX-cal-Bx,MJXc-TeX-cal-Bw} .MJXc-TeX-frak-R {font-family: MJXc-TeX-frak-R,MJXc-TeX-frak-Rw} .MJXc-TeX-frak-B {font-family: MJXc-TeX-frak-B,MJXc-TeX-frak-Bx,MJXc-TeX-frak-Bw} .MJXc-TeX-math-BI {font-family: MJXc-TeX-math-BI,MJXc-TeX-math-BIx,MJXc-TeX-math-BIw} .MJXc-TeX-sans-R {font-family: MJXc-TeX-sans-R,MJXc-TeX-sans-Rw} .MJXc-TeX-sans-B {font-family: MJXc-TeX-sans-B,MJXc-TeX-sans-Bx,MJXc-TeX-sans-Bw} .MJXc-TeX-sans-I {font-family: MJXc-TeX-sans-I,MJXc-TeX-sans-Ix,MJXc-TeX-sans-Iw} .MJXc-TeX-script-R {font-family: MJXc-TeX-script-R,MJXc-TeX-script-Rw} .MJXc-TeX-type-R {font-family: MJXc-TeX-type-R,MJXc-TeX-type-Rw} .MJXc-TeX-cal-R {font-family: MJXc-TeX-cal-R,MJXc-TeX-cal-Rw} .MJXc-TeX-main-B {font-family: MJXc-TeX-main-B,MJXc-TeX-main-Bx,MJXc-TeX-main-Bw} .MJXc-TeX-main-I {font-family: MJXc-TeX-main-I,MJXc-TeX-main-Ix,MJXc-TeX-main-Iw} .MJXc-TeX-main-R {font-family: MJXc-TeX-main-R,MJXc-TeX-main-Rw} .MJXc-TeX-math-I {font-family: MJXc-TeX-math-I,MJXc-TeX-math-Ix,MJXc-TeX-math-Iw} .MJXc-TeX-size1-R {font-family: MJXc-TeX-size1-R,MJXc-TeX-size1-Rw} .MJXc-TeX-size2-R {font-family: MJXc-TeX-size2-R,MJXc-TeX-size2-Rw} .MJXc-TeX-size3-R {font-family: MJXc-TeX-size3-R,MJXc-TeX-size3-Rw} .MJXc-TeX-size4-R {font-family: MJXc-TeX-size4-R,MJXc-TeX-size4-Rw} .MJXc-TeX-vec-R {font-family: MJXc-TeX-vec-R,MJXc-TeX-vec-Rw} .MJXc-TeX-vec-B {font-family: MJXc-TeX-vec-B,MJXc-TeX-vec-Bx,MJXc-TeX-vec-Bw} @font-face {font-family: MJXc-TeX-ams-R; src: local('MathJax_AMS'), local('MathJax_AMS-Regular')} @font-face {font-family: MJXc-TeX-ams-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_AMS-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_AMS-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_AMS-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-cal-B; src: local('MathJax_Caligraphic Bold'), local('MathJax_Caligraphic-Bold')} @font-face {font-family: MJXc-TeX-cal-Bx; src: local('MathJax_Caligraphic'); font-weight: bold} @font-face {font-family: MJXc-TeX-cal-Bw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Caligraphic-Bold.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Caligraphic-Bold.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Caligraphic-Bold.otf') format('opentype')} @font-face {font-family: MJXc-TeX-frak-R; src: local('MathJax_Fraktur'), local('MathJax_Fraktur-Regular')} @font-face {font-family: MJXc-TeX-frak-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Fraktur-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Fraktur-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Fraktur-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-frak-B; src: local('MathJax_Fraktur Bold'), local('MathJax_Fraktur-Bold')} @font-face {font-family: MJXc-TeX-frak-Bx; src: local('MathJax_Fraktur'); font-weight: bold} @font-face {font-family: MJXc-TeX-frak-Bw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Fraktur-Bold.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Fraktur-Bold.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Fraktur-Bold.otf') format('opentype')} @font-face {font-family: MJXc-TeX-math-BI; src: local('MathJax_Math BoldItalic'), local('MathJax_Math-BoldItalic')} @font-face {font-family: MJXc-TeX-math-BIx; src: local('MathJax_Math'); font-weight: bold; font-style: italic} @font-face {font-family: MJXc-TeX-math-BIw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Math-BoldItalic.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Math-BoldItalic.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Math-BoldItalic.otf') format('opentype')} @font-face {font-family: MJXc-TeX-sans-R; src: local('MathJax_SansSerif'), local('MathJax_SansSerif-Regular')} @font-face {font-family: MJXc-TeX-sans-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_SansSerif-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_SansSerif-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_SansSerif-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-sans-B; src: local('MathJax_SansSerif Bold'), local('MathJax_SansSerif-Bold')} @font-face {font-family: MJXc-TeX-sans-Bx; src: local('MathJax_SansSerif'); font-weight: bold} @font-face {font-family: MJXc-TeX-sans-Bw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_SansSerif-Bold.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_SansSerif-Bold.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_SansSerif-Bold.otf') format('opentype')} @font-face {font-family: MJXc-TeX-sans-I; src: local('MathJax_SansSerif Italic'), local('MathJax_SansSerif-Italic')} @font-face {font-family: MJXc-TeX-sans-Ix; src: local('MathJax_SansSerif'); font-style: italic} @font-face {font-family: MJXc-TeX-sans-Iw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_SansSerif-Italic.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_SansSerif-Italic.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_SansSerif-Italic.otf') format('opentype')} @font-face {font-family: MJXc-TeX-script-R; src: local('MathJax_Script'), local('MathJax_Script-Regular')} @font-face {font-family: MJXc-TeX-script-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Script-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Script-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Script-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-type-R; src: local('MathJax_Typewriter'), local('MathJax_Typewriter-Regular')} @font-face {font-family: MJXc-TeX-type-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Typewriter-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Typewriter-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Typewriter-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-cal-R; src: local('MathJax_Caligraphic'), local('MathJax_Caligraphic-Regular')} @font-face {font-family: MJXc-TeX-cal-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Caligraphic-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Caligraphic-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Caligraphic-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-main-B; src: local('MathJax_Main Bold'), local('MathJax_Main-Bold')} @font-face {font-family: MJXc-TeX-main-Bx; src: local('MathJax_Main'); font-weight: bold} @font-face {font-family: MJXc-TeX-main-Bw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Main-Bold.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Main-Bold.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Main-Bold.otf') format('opentype')} @font-face {font-family: MJXc-TeX-main-I; src: local('MathJax_Main Italic'), local('MathJax_Main-Italic')} @font-face {font-family: MJXc-TeX-main-Ix; src: local('MathJax_Main'); font-style: italic} @font-face {font-family: MJXc-TeX-main-Iw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Main-Italic.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Main-Italic.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Main-Italic.otf') format('opentype')} @font-face {font-family: MJXc-TeX-main-R; src: local('MathJax_Main'), local('MathJax_Main-Regular')} @font-face {font-family: MJXc-TeX-main-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Main-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Main-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Main-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-math-I; src: local('MathJax_Math Italic'), local('MathJax_Math-Italic')} @font-face {font-family: MJXc-TeX-math-Ix; src: local('MathJax_Math'); font-style: italic} @font-face {font-family: MJXc-TeX-math-Iw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Math-Italic.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Math-Italic.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Math-Italic.otf') format('opentype')} @font-face {font-family: MJXc-TeX-size1-R; src: local('MathJax_Size1'), local('MathJax_Size1-Regular')} @font-face {font-family: MJXc-TeX-size1-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Size1-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Size1-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Size1-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-size2-R; src: local('MathJax_Size2'), local('MathJax_Size2-Regular')} @font-face {font-family: MJXc-TeX-size2-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Size2-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Size2-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Size2-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-size3-R; src: local('MathJax_Size3'), local('MathJax_Size3-Regular')} @font-face {font-family: MJXc-TeX-size3-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Size3-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Size3-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Size3-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-size4-R; src: local('MathJax_Size4'), local('MathJax_Size4-Regular')} @font-face {font-family: MJXc-TeX-size4-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Size4-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Size4-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Size4-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-vec-R; src: local('MathJax_Vector'), local('MathJax_Vector-Regular')} @font-face {font-family: MJXc-TeX-vec-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Vector-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Vector-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Vector-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-vec-B; src: local('MathJax_Vector Bold'), local('MathJax_Vector-Bold')} @font-face {font-family: MJXc-TeX-vec-Bx; src: local('MathJax_Vector'); font-weight: bold} @font-face {font-family: MJXc-TeX-vec-Bw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Vector-Bold.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Vector-Bold.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Vector-Bold.otf') format('opentype')} </style></p>
We know what linear functions are. A function f is linear iff it satisfies additivity f(x+y)=f(x)+f(y) and homogeneity f(ax)=af(x). We know what continuity is. A function f is continuous iff for all ε there exists a δ such that if |x−x0| < δ, then |f(x)−f(x0)| < ε. An equivalent way to think about continuity is the sequence criterion: f is continuous iff a sequence (xk) converging to x implies that (f(xk)) converges to f(x). That is to say, if for all ε there exists an N such that if k ≥ N, then |xk−x| < ε, then for all ε, there also exists an M such that if k ≥ M, then |f(xk)−f(x)| < ε. Sometimes people talk about discontinuous linear functions. You might think: that's crazy. I've seen many linear functions in my time, and they were definitely all continuous. f(x): ℝ → ℝ := ax is continuous for any a ∈ ℝ. T(x⃗): ℝ² → ℝ² := (abcd)→x is continuous no matter what the entries in the matrix are. Stop being crazy!! Actually, it's not crazy. It's just that all the discontinuous linear functions live in infinite-dimensional spaces. Take, say, the space C1([a,b]) of continuously differentiable functions from a closed interval [a,b] to ℝ, with the uniform norm. (The uniform norm means that the "size" of a function for the purposes of continuity is the least upper bound of its absolute value.) If you think of a vector in the n-dimensional Rn as a function from {1...n} to ℝ, then you can see why a function from a continuous (not even countable) domain would be infinite-dimensional. Consider the sequence of functions (fk)=(sinkxk)∞k=1 in C1([a,b]). The sequence converges to the zero function: for any ε, we can take N:=⌈1ε⌉ and then sinkxk≤1⌈1ε⌉≤11ε=ε. Now consider that the sequence of derivatives is (kcoskxk)∞k=1=(coskx)∞k=1, which doesn't converge. But the function D: $C^1([a,b]) \rightarrow C0([a,b]) that maps a function to its derivative is linear. (We have additivity because the derivative of a sum is the sum of the derivatives, and we have homogeneity because you ca
675
1.8.0
Revision
false
null
null
CrosspostOutput
dT7mvHzuX46vydt9K
avoiding-ai-deception-lie-detectors-can-either-induce
Avoiding AI Deception: Lie Detectors can either Induce Honesty or Evasion
null
false
false
false
null
EBG66a9b4ZegGmNgh
[ { "__typename": "CoauthorStatusOutput", "confirmed": true, "requested": false, "userId": "Sct2JwmLuFREk7cWr" }, { "__typename": "CoauthorStatusOutput", "confirmed": true, "requested": false, "userId": "XQ8GsqKnoLp5FWtgc" }, { "__typename": "CoauthorStatusOutput", "confirmed": true, "requested": false, "userId": "PXWqFTrPbmpXxYZCo" } ]
true
false
false
false
Post
https://far.ai/news/avoiding-ai-deception
2025-06-05T23:07:59.889Z
null
false
false
2
2
2025-06-06T17:50:25.511Z
false
false
linkpost
[ "Sct2JwmLuFREk7cWr", "PXWqFTrPbmpXxYZCo" ]
null
null
jNmfpm8xxp9SApyej
2
8
17
false
0.019751
null
false
false
2025-06-13T22:22:47.691Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
qgdGA4ZEyW7zNdK84
null
null
null
false
null
[]
null
6
0
2025-06-05T22:07:55.006Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[ { "__typename": "User", "_id": "Sct2JwmLuFREk7cWr", "afCommentCount": 0, "afKarma": 0, "afPostCount": 0, "commentCount": 4, "createdAt": "2023-02-06T18:25:37.345Z", "deleted": false, "displayName": "ChrisCundy", "fullName": null, "htmlBio": "", "isAdmin": false, "jobTitle": null, "karma": 57, "organization": null, "postCount": 0, "previousDisplayName": null, "profileImageId": null, "reviewedByUserId": "EQNTWXLKMeWMp2FQS", "sequenceCount": 0, "slug": "chriscundy", "spamRiskScore": 1, "tagRevisionCount": 0, "username": "ChrisCundy" }, { "__typename": "User", "_id": "XQ8GsqKnoLp5FWtgc", "afCommentCount": 0, "afKarma": 0, "afPostCount": 0, "commentCount": 2, "createdAt": "2021-07-21T10:47:54.652Z", "deleted": false, "displayName": "smallsilo", "fullName": "Siao Si", "htmlBio": "", "isAdmin": false, "jobTitle": null, "karma": 127, "organization": null, "postCount": 6, "previousDisplayName": null, "profileImageId": null, "reviewedByUserId": "r38pkCm7wF4M44MDQ", "sequenceCount": 0, "slug": "monstrologies", "spamRiskScore": 1, "tagRevisionCount": 0, "username": "monstrologies" }, { "__typename": "User", "_id": "PXWqFTrPbmpXxYZCo", "afCommentCount": 30, "afKarma": 302, "afPostCount": 7, "commentCount": 48, "createdAt": "2018-04-23T16:48:21.410Z", "deleted": false, "displayName": "AdamGleave", "fullName": null, "htmlBio": "", "isAdmin": false, "jobTitle": null, "karma": 930, "organization": null, "postCount": 7, "previousDisplayName": null, "profileImageId": null, "reviewedByUserId": "r38pkCm7wF4M44MDQ", "sequenceCount": 0, "slug": "adamgleave", "spamRiskScore": 1, "tagRevisionCount": 0, "username": "AdamGleave" } ]
6
null
null
null
null
[ { "__typename": "Tag", "_id": "sYm3HiWcfZvrGu3ui", "adminOnly": false, "afBaseScore": 2, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" } ] }, "baseScore": 12, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:22.097Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" }, { "_id": "sof55TPMQaeBaxhsS", "displayName": "tommylees112" }, { "_id": "AayjS8XzcnDKhGdTv", "displayName": "shark" }, { "_id": "HnALuwRdo6k9HLaMt", "displayName": "Alex Firssoff" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI", "needsReview": false, "noindex": false, "postCount": 12544, "score": 12, "shortName": null, "slug": "ai", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 4, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
8
0
0
6
0
EBG66a9b4ZegGmNgh
chengcheng
2023-03-15T08:44:19.237Z
ccstan99
ChengCheng
null
null
null
142
47
false
false
null
null
6
1
0
4
0
1
0
grecHJcgkb3KW5wnM
User
null
null
null
[ "alignmentVoters", "canModeratePersonal" ]
null
null
dT7mvHzuX46vydt9K
https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/dT7mvHzuX46vydt9K/ufmrsjobsa9bqc7lxk5m
SocialPreviewType
jNmfpm8xxp9SApyej
<p>Large language models (LLMs) are often fine-tuned after training using methods like reinforcement learning from human feedback (RLHF). In this process, models are rewarded for generating responses that people rate highly. But what people like isn’t always what’s true. Studies have found that models <a href="https://arxiv.org/abs/2310.13548">learn to give answers that humans prefer but are untrue</a>. This problem occurred in a <a href="https://openai.com/index/expanding-on-sycophancy/">recent update to the GPT-4o model</a> that aimed to please the user even by making false statements.</p><p>Today, we have high-accuracy "lie-detectors” that analyze internal model states—AI's "thought patterns"—to identify deceptive outputs that human reviewers could easily overlook. Even simple logistic models trained on these internal activations can <a href="https://arxiv.org/abs/2502.03407">successfully pinpoint 95-99% of deceptive responses</a>.</p><p>However, lie detectors are not infallible either. We wanted to find out if adding a lie detector to the training loop would make models honest, or if it would just train models to evade detection. It turns out that models become honest under the right conditions—high detector true positive rate, high KL regularization to an honest original model, and off-policy post-training methods.</p><h2><strong>Using lie detectors for scalable oversight</strong></h2><figure class="image image_resized" style="width:543.195px"><img src="https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/dT7mvHzuX46vydt9K/ixh9gkr6wlcws2wbhpei" alt=""></figure><p>Given that models are incentivized to be deceptive, and appear to ‘know’ that they are being deceptive, we could simply train a lie detector with a small number of known truthful/deceptive examples, and use this to assist labellers who cannot easily identify model deception. We call this approach Scalable Oversight via Lie Detector, or "SOLiD”.</p><p>This approach has the downside that it doesn't remove the incentive to lie: it's still better for the model to tell a lie that humans prefer, as long as it can ‘fool’ the lie detector. Instead of training models to tell the truth, we could end up training models to be better at cheating lie detectors.</p><p>We wanted to find out if models would learn to tell the truth, or simply become 'better liars’. If models always learn to tell the truth, this could be a very useful tool for AI alignment. But if models instead learn to cheat, training with lie detectors could backfire by making future detectors ineffective.</p><h2><strong>Our setup&nbsp;</strong></h2><p>To find out, we modeled a scenario where model developers deploy lie detectors to flag deceptive outputs to human labelers. We assume human labelers prefer deceptive responses if they are unaware of the deception, but st... </p>
Large language models (LLMs) are often fine-tuned after training using methods like reinforcement learning from human feedback (RLHF). In this process, models are rewarded for generating responses that people rate highly. But what people like isn’t always what’s true. Studies have found that models learn to give answers that humans prefer but are untrue. This problem occurred in a recent update to the GPT-4o model that aimed to please the user even by making false statements. Today, we have high-accuracy "lie-detectors” that analyze internal model states—AI's "thought patterns"—to identify deceptive outputs that human reviewers could easily overlook. Even simple logistic models trained on these internal activations can successfully pinpoint 95-99% of deceptive responses. However, lie detectors are not infallible either. We wanted to find out if adding a lie detector to the training loop would make models honest, or if it would just train models to evade detection. It turns out that models become honest under the right conditions—high detector true positive rate, high KL regularization to an honest original model, and off-policy post-training methods. Using lie detectors for scalable oversight Given that models are incentivized to be deceptive, and appear to ‘know’ that they are being deceptive, we could simply train a lie detector with a small number of known truthful/deceptive examples, and use this to assist labellers who cannot easily identify model deception. We call this approach Scalable Oversight via Lie Detector, or "SOLiD”. This approach has the downside that it doesn't remove the incentive to lie: it's still better for the model to tell a lie that humans prefer, as long as it can ‘fool’ the lie detector. Instead of training models to tell the truth, we could end up training models to be better at cheating lie detectors. We wanted to find out if models would learn to tell the truth, or simply become 'better liars’. If models always learn to tell the t
1,406
1.3.1
Revision
false
null
null
CrosspostOutput
HFSEHFB9WLZLkPwnc
introducing-meridian-cambridge-s-new-online-lecture-series
Introducing: Meridian Cambridge's new online lecture series covering frontier AI and AI safety
null
false
false
false
null
KEWvkFbywyJ5hiagY
null
true
false
false
false
Post
null
2025-06-05T21:55:25.902Z
null
false
false
2
2
2025-06-06T17:51:19.234Z
false
false
post
[]
null
null
YGxcmT2HipvGJS77g
0
2
1
false
0.008142
null
false
false
2025-06-05T21:55:25.902Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
qgdGA4ZEyW7zNdK84
null
null
null
false
null
[]
null
0
0
2025-06-05T21:54:31.152Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
1
null
null
null
null
[ { "__typename": "Tag", "_id": "sYm3HiWcfZvrGu3ui", "adminOnly": false, "afBaseScore": 2, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" } ] }, "baseScore": 12, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:22.097Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" }, { "_id": "sof55TPMQaeBaxhsS", "displayName": "tommylees112" }, { "_id": "AayjS8XzcnDKhGdTv", "displayName": "shark" }, { "_id": "HnALuwRdo6k9HLaMt", "displayName": "Alex Firssoff" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI", "needsReview": false, "noindex": false, "postCount": 12544, "score": 12, "shortName": null, "slug": "ai", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 4, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
2
0
0
0
0
KEWvkFbywyJ5hiagY
meridian-cambridge
2025-03-06T00:13:15.146Z
Meridian Cambridge
Meridian Cambridge
null
null
null
11
0
false
false
null
null
2
0
0
0
0
0.9
0
grecHJcgkb3KW5wnM
User
null
null
null
null
null
null
HFSEHFB9WLZLkPwnc
https://res.cloudinary.com/lesswrong-2-0/image/upload/c_fill,ar_1.91,g_auto/SocialPreview/dcmjrmzm9wguq951f9li
SocialPreviewType
YGxcmT2HipvGJS77g
<p>This is a linkpost for [<a href="https://www.meridiancambridge.org/language-models-course">https://www.meridiancambridge.org/language-models-course</a>]<br><br>Meridian Cambridge, in partnership with Cambridge University's <a href="https://www.c2d3.cam.ac.uk/">Center for Data Driven Discovery</a> (C2D3), &nbsp;has produced a 16-part lecture series entitled "Language Models and Intelligent Agentic Systems" and the recordings are now online!<br><br>The LMaIAS course provides an introduction to core ideas in AI safety. Throughout, we build up from introductory ideas about language modelling and neural networks to discussions of risks posed by advanced AI systems.</p><h2><strong>Course Structure</strong></h2><p>The course is divided into four parts:</p><h3><strong>Part 1: What is a Language Model?</strong></h3><p>To start the course, we give three lectures covering generative models and next token prediction, the transformer architecture, and scaling laws for large models.</p><h3><strong>Part 2: Crafting Agentic Systems</strong></h3><p>Now the foundations are in place, the next four lectures go into details on LLM post-training, reinforcement learning, reward modelling, and agent architectures.</p><h3><strong>Part 3: Agentic Behaviour</strong></h3><p>Here we take four lectures to discuss optimisation and reasoning, reward hacking and goal misgeneralisation, out-of-context reasoning and situational awareness, and finally deceptive alignment and alignment faking.</p><h3><strong>Part 4: Frontiers</strong></h3><p>For the remainder of the lecture series, we give five lectures covering risks from advanced AI, AI evaluations, AI control and safety cases, AI organisations and agendas, and conclude with a discussion on the future of language models.</p><p>You can find the first lecture [<a href="https://www.youtube.com/watch?v=UiGa8Bx1Srk&amp;list=PLKOGQ4KczC_pa2r4UXb3wOIbBeU6Qeeoy&amp;index=1">here</a>], and the whole course is available [<a href="https://www.meridiancambridge.org/language-models-course">here</a>].<br><br>The lectures were created and delivered by Edward James Young, Jason R. Brown, and Lennie Wells, in partnership with Cambridge University's C2D3. The hope is that this material can be used to help educate people new to the field and provide them with the background knowledge required to effectively contribute to AI safety. Please share this with anybody you think might be interested!</p><p>- The Meridian Team</p>
This is a linkpost for [https://www.meridiancambridge.org/language-models-course] Meridian Cambridge, in partnership with Cambridge University's Center for Data Driven Discovery (C2D3),  has produced a 16-part lecture series entitled "Language Models and Intelligent Agentic Systems" and the recordings are now online! The LMaIAS course provides an introduction to core ideas in AI safety. Throughout, we build up from introductory ideas about language modelling and neural networks to discussions of risks posed by advanced AI systems. Course Structure The course is divided into four parts: Part 1: What is a Language Model? To start the course, we give three lectures covering generative models and next token prediction, the transformer architecture, and scaling laws for large models. Part 2: Crafting Agentic Systems Now the foundations are in place, the next four lectures go into details on LLM post-training, reinforcement learning, reward modelling, and agent architectures. Part 3: Agentic Behaviour Here we take four lectures to discuss optimisation and reasoning, reward hacking and goal misgeneralisation, out-of-context reasoning and situational awareness, and finally deceptive alignment and alignment faking. Part 4: Frontiers For the remainder of the lecture series, we give five lectures covering risks from advanced AI, AI evaluations, AI control and safety cases, AI organisations and agendas, and conclude with a discussion on the future of language models. You can find the first lecture [here], and the whole course is available [here]. The lectures were created and delivered by Edward James Young, Jason R. Brown, and Lennie Wells, in partnership with Cambridge University's C2D3. The hope is that this material can be used to help educate people new to the field and provide them with the background knowledge required to effectively contribute to AI safety. Please share this with anybody you think might be interested! - The Meridian Team
298
1.1.0
Revision
false
null
null
CrosspostOutput
uvN4o6ZeKdXyyX43i
cheaper-sodium-electrolysis
cheaper sodium electrolysis
null
false
false
false
null
xYpk75i7Hnn6wc5it
null
true
false
false
false
Post
https://www.bhauth.com/blog/chemistry/sodium%20electrolysis.html
2025-06-05T21:49:04.225Z
null
false
false
2
2
2025-06-06T17:52:13.651Z
false
false
linkpost
[]
null
null
EStw2kN7wF5n63ucW
3
5
21
false
0.022589
null
false
false
2025-06-06T02:08:10.116Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
qgdGA4ZEyW7zNdK84
null
null
null
false
null
[]
null
7
0
2025-06-05T21:44:59.895Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
5
null
null
null
null
[ { "__typename": "Tag", "_id": "uxmGtpeE3KoE7pzSL", "adminOnly": false, "afBaseScore": 9, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "EQNTWXLKMeWMp2FQS", "displayName": "Ben Pace" }, { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 19, "canEditUserIds": null, "core": false, "createdAt": "2021-04-09T02:55:36.576Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "EQNTWXLKMeWMp2FQS", "displayName": "Ben Pace" }, { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Chemistry", "needsReview": false, "noindex": false, "postCount": 28, "score": 19, "shortName": null, "slug": "chemistry", "suggestedAsFilter": false, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 2, "wikiOnly": false }, { "__typename": "Tag", "_id": "3uE2pXvbcnS9nnZRE", "adminOnly": false, "afBaseScore": 0, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 2, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:50.898Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 27, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "zJtgSyKntXrnkArbY", "displayName": "kistune" }, { "_id": "XqyQTGFGoKfZtdqN3", "displayName": "kINo" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "World Modeling", "needsReview": false, "noindex": false, "postCount": 5855, "score": 2, "shortName": null, "slug": "world-modeling", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 2, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
5
0
0
5
0
xYpk75i7Hnn6wc5it
bhauth
2023-04-08T11:57:52.463Z
bhauth
bhauth
null
null
null
3,598
6
false
false
<p><a href="https://www.bhauth.com/">bhauth.com</a></p>
null
null
77
421
0
0
0
1
0
r38pkCm7wF4M44MDQ
User
null
null
null
[ "canModeratePersonal", "trustLevel1", "alignmentVoters" ]
null
null
uvN4o6ZeKdXyyX43i
SocialPreviewType
EStw2kN7wF5n63ucW
<h2>sodium electrolysis</h2> <p>Aluminum metal is a widely-used material. It costs ~$2.5/kg. A significant fraction of its production cost is electricity.</p><p>Currently, Na metal is produced by <a href="https://en.wikipedia.org/wiki/Downs_cell">Downs cell</a> electrolysis of NaCl. Making sodium metal with electrolysis requires much less energy per mass of sodium than making aluminum. The raw material used (NaCl) is very cheap. Why, then, is sodium several times as expensive as aluminum? There's not a clear market price for it, but it's typically considered to be ~$10/kg on a large scale.</p><p>Partly, that's because:</p> <ul> <li>The scale of production of Na is much smaller.</li> <li>Transport is more expensive.</li> </ul> <p>But the process is also inherently more expensive. Why?</p><p>The cells used for Al electrolysis are open to the atmosphere. Oxygen and CO2 comes out of them. Electrolysis of NaCl produces Cl2, which is too hazardous to just release. So, it has to be collected. NaCl has a boiling point of 1413 C, so some salt evaporates, which causes problems in the chlorine handling system.</p><p>Na metal has a relatively high solubility in NaCl, so it continuously reacts with the generated chlorine, reducing efficiency.</p> <h2>a new process</h2> <p>There's a <a href="https://www.sciencedirect.com/science/article/pii/S2213956724000616">recent paper</a> (open access) describing a new method for producing Na metal. The idea is:</p> <ul> <li> <p>Add Na carbonate so that electrolysis produces O2 instead of Cl2.</p> </li> <li> <p>Normally, adding carbonate leads to carbon buildup on the electrode, because the voltage for that is basically the same as Na production. So, they use a liquid tin electrode, which reduces the voltage for Na electrolysis.</p> </li> <li> <p>That of course means that some energy is needed to separate the Na from the tin. They do that separation with vacuum distillation. This isn't unprecedented for metal production: most Mg metal production uses the <a href="https://en.wikipedia.org/wiki/Pidgeon_process">Pidgeon process</a> which involves separating Mg metal by distillation. In this case, the effective boiling point of the Na might be increased by ~400 K.</p> </li> </ul> <p>The professor involved in that paper also previously did <a href="https://link.springer.com/article/10.1007/s11663-023-02945-8">a similar thing</a> with potassium.</p><p>Of course, most Na carbonate is made from NaCl, and the chlorine has to go somewhere. Combining the above electrolysis with the <a href="https://en.wikipedia.org/wiki/Solvay_process">Solvay Process</a>, the net reaction would be:</p><p>2 NaCl + CaCO3 -&gt; 2 Na + CaCl2 + CO2 + 1/2 O2</p><p>How do I find things like this? Well, in this case, I noticed that paper because I searched Google Scholar for papers doing this exact thing because I was curious if anyone tried it.</p> <h2>cost estimation</h2> <h3>sodi</h3>...
sodium electrolysis Aluminum metal is a widely-used material. It costs ~$2.5/kg. A significant fraction of its production cost is electricity. Currently, Na metal is produced by Downs cell electrolysis of NaCl. Making sodium metal with electrolysis requires much less energy per mass of sodium than making aluminum. The raw material used (NaCl) is very cheap. Why, then, is sodium several times as expensive as aluminum? There's not a clear market price for it, but it's typically considered to be ~$10/kg on a large scale. Partly, that's because: * The scale of production of Na is much smaller. * Transport is more expensive. But the process is also inherently more expensive. Why? The cells used for Al electrolysis are open to the atmosphere. Oxygen and CO2 comes out of them. Electrolysis of NaCl produces Cl2, which is too hazardous to just release. So, it has to be collected. NaCl has a boiling point of 1413 C, so some salt evaporates, which causes problems in the chlorine handling system. Na metal has a relatively high solubility in NaCl, so it continuously reacts with the generated chlorine, reducing efficiency. a new process There's a recent paper (open access) describing a new method for producing Na metal. The idea is: * Add Na carbonate so that electrolysis produces O2 instead of Cl2. * Normally, adding carbonate leads to carbon buildup on the electrode, because the voltage for that is basically the same as Na production. So, they use a liquid tin electrode, which reduces the voltage for Na electrolysis. * That of course means that some energy is needed to separate the Na from the tin. They do that separation with vacuum distillation. This isn't unprecedented for metal production: most Mg metal production uses the Pidgeon process which involves separating Mg metal by distillation. In this case, the effective boiling point of the Na might be increased by ~400 K. The professor involved in that paper also previously did a similar thing with potassium.
1,197
1.2.0
Revision
false
null
null
CrosspostOutput
LFGgwitjertJqch7J
histograms-are-to-cdfs-as-calibration-plots-are-to
Histograms are to CDFs as calibration plots are to...
null
false
false
false
null
B7TpXsz3Rw4RbsN6p
null
true
false
false
false
Post
https://optimizationprocess.com/calibration-cdf/
2025-06-05T20:20:18.639Z
null
false
false
2
2
2025-06-06T17:52:42.519Z
false
false
linkpost
[]
null
null
6GspnNfZv9NwMBrdd
9
14
35
false
0.032368
null
false
false
2025-06-07T20:55:47.741Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
qgdGA4ZEyW7zNdK84
null
null
null
false
null
[]
null
8
0
2025-06-05T20:03:24.348Z
false
false
reign-of-terror
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
1
null
null
null
null
[ { "__typename": "Tag", "_id": "8daMDi9NEShyLqxth", "adminOnly": false, "afBaseScore": 10, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "EQNTWXLKMeWMp2FQS", "displayName": "Ben Pace" }, { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" }, { "_id": "iXX23K6iBAosHFPBn", "displayName": "Alvin Ånestrand" } ] }, "baseScore": 21, "canEditUserIds": null, "core": false, "createdAt": "2020-05-10T05:54:39.783Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "EQNTWXLKMeWMp2FQS", "displayName": "Ben Pace" }, { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" }, { "_id": "iXX23K6iBAosHFPBn", "displayName": "Alvin Ånestrand" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Forecasting & Prediction", "needsReview": false, "noindex": false, "postCount": 508, "score": 21, "shortName": null, "slug": "forecasting-and-prediction", "suggestedAsFilter": false, "userId": "iBcH2a3HdWGS2JEZA", "voteCount": 3, "wikiOnly": false }, { "__typename": "Tag", "_id": "fmA6cA9psxibmH8MS", "adminOnly": false, "afBaseScore": null, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 0, "canEditUserIds": null, "core": false, "createdAt": "2020-09-08T23:01:08.537Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Mental Imagery / Visualization", "needsReview": false, "noindex": false, "postCount": 20, "score": 0, "shortName": null, "slug": "mental-imagery-visualization", "suggestedAsFilter": false, "userId": "Q7NW4XaWQmfPfdcFj", "voteCount": 0, "wikiOnly": false }, { "__typename": "Tag", "_id": "3uE2pXvbcnS9nnZRE", "adminOnly": false, "afBaseScore": 0, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 2, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:50.898Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 27, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "zJtgSyKntXrnkArbY", "displayName": "kistune" }, { "_id": "XqyQTGFGoKfZtdqN3", "displayName": "kINo" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "World Modeling", "needsReview": false, "noindex": false, "postCount": 5855, "score": 2, "shortName": null, "slug": "world-modeling", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 2, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
14
0
0
6
0
B7TpXsz3Rw4RbsN6p
optimization-process
2017-09-26T22:13:24.395Z
Optimization Process
Optimization Process
null
null
null
693
3
false
false
null
null
30
90
0
0
0
1
0
r38pkCm7wF4M44MDQ
User
reign-of-terror
null
null
[ "canModeratePersonal", "alignmentVoters" ]
null
null
LFGgwitjertJqch7J
https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/LFGgwitjertJqch7J/poacxbesdkrammnqynwj
SocialPreviewType
6GspnNfZv9NwMBrdd
<p>As you know, histograms are decent visualizations for PDFs with lots of samples...</p><figure class="image"><img src="https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/LFGgwitjertJqch7J/okziqklkhzdtnykkgv1u"><figcaption>10k predictions, 20 bins</figcaption></figure><p>&nbsp;</p><p>...but if there are only a few samples, the histogram-binning choices can matter a lot:</p><figure class="image image_resized" style="width:67.35%"><img src="https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/LFGgwitjertJqch7J/t7g2wvgunxnd3bjdsrty"><figcaption>10 predictions, 4 bins</figcaption></figure><figure class="image image_resized" style="width:67.72%"><img src="https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/LFGgwitjertJqch7J/jylvb6vl47cnqgampvoc"><figcaption>same 10 predictions, 7 bins</figcaption></figure><p>The binning (a) discards information, and worse, (b) is <i>mathematically un-aesthetic</i>.</p><p>But a CDF doesn't have this problem!</p><figure class="image image_resized" style="width:76.39%"><img src="https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/LFGgwitjertJqch7J/km9mrozopdjo9ck6l7nf"><figcaption>same 10 predictions, every data point precisely represented</figcaption></figure><hr><p>If you make a bunch of predictions, and you want to know how well they're calibrated, classically you make a graph like this:</p><figure class="image image_resized" style="width:55.16%"><img src="https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/LFGgwitjertJqch7J/ljw1evuczrszoilmako3"><figcaption>source: <a href="https://slatestarcodex.com/blog_images/calibration2019.png">SSC's 2019 prediction grading</a></figcaption></figure><p>But, as with a histogram, this depends on how you bin your predictions.</p><figure class="image image_resized" style="width:65.87%"><img src="https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/LFGgwitjertJqch7J/zsualkqmgrqtsnflfwwv"><figcaption>100 predictions, 10 bins</figcaption></figure><figure class="image image_resized" style="width:67.31%"><img src="https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/LFGgwitjertJqch7J/oqwrv6fcdvts1jmmqdca"><figcaption>same 100 predictions, 30 bins</figcaption></figure><p>Is there some CDF-like equivalent here? Some visualization with no free parameters?</p><hr><p>I asked that question to several people at Arbor Summer Camp. I got three answers:</p><ol><li>"You get from a PDF to a CDF by integrating. So, here, analogously, let's integrate (num predictions with confidence &lt; x that came true) minus (expected num predictions with confidence &lt; x that came true)."</li><li>(the same thing, said in different words)</li><li>(the same thing, said in different words)</li></ol><p>If we make a "CDF" for the above 100 predictions by applying these three insights, we get:</p><figure class="image image_resized" style="width:78.49%"><img src="https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/LFGgwitjertJqch7J/apxdm8qtpdgjuztijb0j" alt="CDF for calibration curves"><figcaption><a href="https://optimizationprocess.com/calibration-cdf/assets/calcdf.py">.py</a></figcaption></figure><p>I find this a little harder to read than the calibration plots above, which I choose to interpret as a good sign, since CDFs are a little harder to read than histograms. The thing to keep in mind, I think, is: when the curve is going up, it's a sign your probabilities are too high; when it's going down, it's a sign your probabilities are too low.</p><blockquote><p><i>Test: how would you describe the problems that this predictor has?</i></p><figure class="image image_resized" style="width:86.21%"><img src="https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/LFGgwitjertJqch7J/ytfjykphmxl5kuxyvf48"></figure><p><a href="https://itty.bitty.site/#/?eJxtkE1ug0AMha/ylE1aCaX7CNFFolwg6gEGxoCFY5PxUJTbV1AW6c/Ov+97djlWHxopNaYtR9J8RGsJfgsiGJPVoWbhzOQFck/Ync6XHdiR2Fm7Am5r/ccoOO8dwZ07Ze0QEqEjpRREHshm6LnrC8w9C628Jf8PdzpfFlgbRJ5p39IUf2H/csTmQ/k2VuVYlXWqtvBKjWnEfSLPbHrE3IeM2SaJCAr7fPoHxkSRm2xp79v1YjZAeKD3Te/lSrT6Io2wFrcHhHVY/Jnn9cCl6ybTwju8Lntf+bKFyQ=="><i>Solution.</i></a></p></blockquote><p>&nbsp;</p><p>(Are there any better visualizations? Maybe. I <a href="https://www.lesswrong.com/posts/sa8Qhby63QLCzcCDo/is-there-an-equivalent-of-the-cdf-for-grading-predictions?commentId=eYvAa3xfx2BpAcKDn">looked into this a couple years ago</a>, but looking back at it, I think this simple "sum(expected-actual predictions with p&lt;x)" graph is at least as compelling as anything I found.)</p>
As you know, histograms are decent visualizations for PDFs with lots of samples... 10k predictions, 20 bins   ...but if there are only a few samples, the histogram-binning choices can matter a lot: 10 predictions, 4 binssame 10 predictions, 7 bins The binning (a) discards information, and worse, (b) is mathematically un-aesthetic. But a CDF doesn't have this problem! same 10 predictions, every data point precisely represented ---------------------------------------- If you make a bunch of predictions, and you want to know how well they're calibrated, classically you make a graph like this: source: SSC's 2019 prediction grading But, as with a histogram, this depends on how you bin your predictions. 100 predictions, 10 binssame 100 predictions, 30 bins Is there some CDF-like equivalent here? Some visualization with no free parameters? ---------------------------------------- I asked that question to several people at Arbor Summer Camp. I got three answers: 1. "You get from a PDF to a CDF by integrating. So, here, analogously, let's integrate (num predictions with confidence < x that came true) minus (expected num predictions with confidence < x that came true)." 2. (the same thing, said in different words) 3. (the same thing, said in different words) If we make a "CDF" for the above 100 predictions by applying these three insights, we get: .py I find this a little harder to read than the calibration plots above, which I choose to interpret as a good sign, since CDFs are a little harder to read than histograms. The thing to keep in mind, I think, is: when the curve is going up, it's a sign your probabilities are too high; when it's going down, it's a sign your probabilities are too low. > Test: how would you describe the problems that this predictor has? > > Solution.   (Are there any better visualizations? Maybe. I looked into this a couple years ago, but looking back at it, I think this simple "sum(expected-actual predictions with p<x)" graph is
353
1.1.1
Revision
false
null
null
CrosspostOutput
bRGGEGyCBS7Gkh84B
integration-bandwidth-the-mechanism-behind-intelligence-and
Integration Bandwidth: The Mechanism Behind Intelligence and Puberty
null
false
false
false
null
yd5XxXgnwFu8aHEn2
null
true
false
false
false
Post
https://osf.io/preprints/psyarxiv/5gx3r_v1
2025-06-05T19:37:59.654Z
null
false
false
2
2
2025-06-06T17:54:28.910Z
false
false
linkpost
[]
null
null
ktPPZCeidrazkFcuE
4
2
-1
false
0.006377
null
false
false
2025-06-21T11:41:17.715Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
qgdGA4ZEyW7zNdK84
null
null
null
false
null
[]
null
-1
0
2025-06-05T11:11:38.894Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
1
null
null
null
null
[ { "__typename": "Tag", "_id": "bxhzaWtdNoEMMkE8r", "adminOnly": false, "afBaseScore": null, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 0, "canEditUserIds": null, "core": false, "createdAt": "2017-02-18T09:43:08.000Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [] }, "isArbitalImport": true, "isPlaceholderPage": false, "isSubforum": false, "name": "General intelligence", "needsReview": false, "noindex": false, "postCount": 169, "score": 0, "shortName": null, "slug": "general-intelligence", "suggestedAsFilter": false, "userId": "nmk3nLpQE89dMRzzN", "voteCount": 0, "wikiOnly": false }, { "__typename": "Tag", "_id": "Wi3EopKJ2aNdtxSWg", "adminOnly": false, "afBaseScore": 9, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "EQNTWXLKMeWMp2FQS", "displayName": "Ben Pace" }, { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 20, "canEditUserIds": null, "core": false, "createdAt": "2020-07-09T09:57:06.243Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "EQNTWXLKMeWMp2FQS", "displayName": "Ben Pace" }, { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" }, { "_id": "z5gvMyDKEKteKDNi9", "displayName": "yongzhen qiao" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Neuroscience", "needsReview": false, "noindex": false, "postCount": 251, "score": 20, "shortName": null, "slug": "neuroscience", "suggestedAsFilter": false, "userId": "qxJ28GN72aiJu96iF", "voteCount": 3, "wikiOnly": false }, { "__typename": "Tag", "_id": "3uE2pXvbcnS9nnZRE", "adminOnly": false, "afBaseScore": 0, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 2, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:50.898Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 27, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "zJtgSyKntXrnkArbY", "displayName": "kistune" }, { "_id": "XqyQTGFGoKfZtdqN3", "displayName": "kINo" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "World Modeling", "needsReview": false, "noindex": false, "postCount": 5855, "score": 2, "shortName": null, "slug": "world-modeling", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 2, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
2
0
0
1
0
yd5XxXgnwFu8aHEn2
dortex
2025-06-05T11:11:11.660Z
Dortex
Dortex
null
null
null
3
0
false
false
null
null
1
2
0
0
0
0.9
0
grecHJcgkb3KW5wnM
User
null
null
null
null
null
null
bRGGEGyCBS7Gkh84B
SocialPreviewType
ktPPZCeidrazkFcuE
<p>I put together a model explaining the link between IQ and the onset of puberty. Ideas only thrive when they're attacked, and I was told this was the kind of place that could manage it.&nbsp;</p><p>The core idea is that slowing myelination in the brain extends its plasticity window and encourages more robust and efficient thalamo-cortical integration. By considering latency, the model naturally gives rise to Deco and Kringelbach's Turbulence model by turning cognition into a latency/energy optimization problem.&nbsp;</p><p>From what I can tell, it fits a surprising amount of cross-disciplinary evidence, but I need counter-examples more than anything.</p>
I put together a model explaining the link between IQ and the onset of puberty. Ideas only thrive when they're attacked, and I was told this was the kind of place that could manage it.  The core idea is that slowing myelination in the brain extends its plasticity window and encourages more robust and efficient thalamo-cortical integration. By considering latency, the model naturally gives rise to Deco and Kringelbach's Turbulence model by turning cognition into a latency/energy optimization problem.  From what I can tell, it fits a surprising amount of cross-disciplinary evidence, but I need counter-examples more than anything.
99
1.1.0
Revision
false
null
null
CrosspostOutput
GYmhkkpRJwkyhauit
levels-of-doom-eutopia-disempowerment-extinction
Levels of Doom: Eutopia, Disempowerment, Extinction
null
false
false
false
null
qf77EiaoMw7tH3GSr
null
true
false
false
false
Post
null
2025-06-05T19:08:47.838Z
null
false
false
2
2
2025-06-06T17:54:18.565Z
false
false
post
[]
null
null
hoyrSTaiccoRQPnwu
0
11
34
false
0.031557
null
false
false
2025-06-05T19:08:47.838Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
qgdGA4ZEyW7zNdK84
null
null
null
false
null
[]
null
9
0
2025-06-05T16:51:47.383Z
false
false
easy-going
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
2
null
null
null
null
[ { "__typename": "Tag", "_id": "ZFrgTgzwEfStg26JL", "adminOnly": false, "afBaseScore": null, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 0, "canEditUserIds": null, "core": false, "createdAt": "2020-07-16T10:29:25.410Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI Risk", "needsReview": false, "noindex": false, "postCount": 1482, "score": 0, "shortName": null, "slug": "ai-risk", "suggestedAsFilter": false, "userId": "EQNTWXLKMeWMp2FQS", "voteCount": 0, "wikiOnly": false }, { "__typename": "Tag", "_id": "MXcpQvaPGtXpB6vkM", "adminOnly": false, "afBaseScore": 9, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "EQNTWXLKMeWMp2FQS", "displayName": "Ben Pace" }, { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 20, "canEditUserIds": null, "core": false, "createdAt": "2020-07-15T04:23:00.324Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "EQNTWXLKMeWMp2FQS", "displayName": "Ben Pace" }, { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" }, { "_id": "8btiLJDabHgZuiSAB", "displayName": "Ggwp" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Public Discourse", "needsReview": false, "noindex": false, "postCount": 187, "score": 20, "shortName": null, "slug": "public-discourse", "suggestedAsFilter": false, "userId": "gXeEWGjTWyqgrQTzR", "voteCount": 3, "wikiOnly": false }, { "__typename": "Tag", "_id": "sYm3HiWcfZvrGu3ui", "adminOnly": false, "afBaseScore": 2, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" } ] }, "baseScore": 12, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:22.097Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" }, { "_id": "sof55TPMQaeBaxhsS", "displayName": "tommylees112" }, { "_id": "AayjS8XzcnDKhGdTv", "displayName": "shark" }, { "_id": "HnALuwRdo6k9HLaMt", "displayName": "Alex Firssoff" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI", "needsReview": false, "noindex": false, "postCount": 12544, "score": 12, "shortName": null, "slug": "ai", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 4, "wikiOnly": false }, { "__typename": "Tag", "_id": "xexCWMyds6QLWognu", "adminOnly": false, "afBaseScore": 0, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 2, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T03:38:23.532Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 20, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "si6LoAENzqPCmi2Dh", "displayName": "ihatenumbersinusernames7" }, { "_id": "B2sXjQTGgwoGE9FES", "displayName": "SeaIgloo" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "World Optimization", "needsReview": false, "noindex": false, "postCount": 3151, "score": 2, "shortName": null, "slug": "world-optimization", "suggestedAsFilter": true, "userId": "XtphY3uYHwruKqDyG", "voteCount": 2, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
11
0
0
8
0
qf77EiaoMw7tH3GSr
vladimir_nesov
2009-02-27T09:55:13.458Z
Vladimir_Nesov
Vladimir_Nesov
null
null
Vladimir Nesov
33,710
489
false
false
null
null
44
9,613
0
2
204
1
1,506
grecHJcgkb3KW5wnM
User
easy-going
null
false
[ "trustLevel1", "alignmentForum", "alignmentVoters", "canModeratePersonal" ]
null
null
GYmhkkpRJwkyhauit
SocialPreviewType
hoyrSTaiccoRQPnwu
<p>Disempowerment is on the fence, gets interpreted as either implying human extinction or being a good place. "Doom" tends to be ambiguous between disempowerment and extinction, as well as about when that outcome must be gauged. And many people currently feel both disempowered and OK, so see eutopia as similar to disempowerment, neither an example of "doom".</p><p>Arguments pointing to risk of human extinction run into the issue of people expecting disempowerment without extinction, when some of the same arguments would remain relevant if applied directly to disempowerment (including the moral arguments about extinction or disempowerment being a problem). And arguments pointing to desirability of establishing eutopia run into the issue of people expecting disempowerment to be approximately as good and in practice much more likely. When the distinctions between these levels of doom are not maintained, conflation makes it harder to meaningfully disagree.</p> <h2>Eutopia Without Disempowerment</h2> <p>This distinction might be slightly more murky, worth defining more explicitly. For me, a crux of a future that's good for humanity is giving the biological humans the resources and the freedom to become transhuman beings themselves, with no hard ceiling on relevance in the long run. Rather than AIs only letting some originally-humans to grow into more powerful but still purely ornamental roles, or not letting them grow at all, or not letting them think faster and do checkpointing and multiple instantiations of the mind states using a non-biological cognitive substrate, or letting them unwillingly die of old age or disease. This should only apply to those who so choose, under their own direction rather than only through externally imposed uplifting protocols, even if that leaves it no more straightforward than world-class success of some kind today, to reach a sensible outcome.</p><p>This in particular implies reasonable resources being left to those who remain/become regular biological humans (or take their time growing up), including through influence of some of these originally-human beings who happen to consider that a good thing to ensure.</p> <h2>Yudkowsky's Arguments and Disempowerment</h2> <p>Yudkowsky frames <a href="https://www.lesswrong.com/posts/uMQ3cqWDPHhjtiesc/agi-ruin-a-list-of-lethalities">AGI ruin arguments</a> around extinction, which his models predict. I think many of the same arguments survive in a world where some AIs have a <a href="https://www.lesswrong.com/posts/2NncxDQ3KBDCxiJiP/cosmopolitan-values-don-t-come-free?commentId=ofPTrG6wsq7CxuTXk">minimal level of caring about humanity</a> <a href="https://www.lesswrong.com/posts/xvBZPEccSfM8Fsobt/what-are-the-best-arguments-for-against-ais-being-slightly">sufficient to prese</a>... </p>
Disempowerment is on the fence, gets interpreted as either implying human extinction or being a good place. "Doom" tends to be ambiguous between disempowerment and extinction, as well as about when that outcome must be gauged. And many people currently feel both disempowered and OK, so see eutopia as similar to disempowerment, neither an example of "doom". Arguments pointing to risk of human extinction run into the issue of people expecting disempowerment without extinction, when some of the same arguments would remain relevant if applied directly to disempowerment (including the moral arguments about extinction or disempowerment being a problem). And arguments pointing to desirability of establishing eutopia run into the issue of people expecting disempowerment to be approximately as good and in practice much more likely. When the distinctions between these levels of doom are not maintained, conflation makes it harder to meaningfully disagree. Eutopia Without Disempowerment This distinction might be slightly more murky, worth defining more explicitly. For me, a crux of a future that's good for humanity is giving the biological humans the resources and the freedom to become transhuman beings themselves, with no hard ceiling on relevance in the long run. Rather than AIs only letting some originally-humans to grow into more powerful but still purely ornamental roles, or not letting them grow at all, or not letting them think faster and do checkpointing and multiple instantiations of the mind states using a non-biological cognitive substrate, or letting them unwillingly die of old age or disease. This should only apply to those who so choose, under their own direction rather than only through externally imposed uplifting protocols, even if that leaves it no more straightforward than world-class success of some kind today, to reach a sensible outcome. This in particular implies reasonable resources being left to those who remain/become regular biological humans (or
592
1.4.0
Revision
false
null
null
CrosspostOutput
xyYss3oCzovibHxAF
llm-in-context-learning-as-approximating-solomonoff
LLM in-context learning as (approximating) Solomonoff induction
null
false
false
true
null
p2N3QhmpKNGn7wkCw
null
true
false
false
false
Post
null
2025-06-05T17:45:28.385Z
null
false
false
2
2
2025-06-05T19:00:03.770Z
false
false
post
[]
null
null
WFGghqbZAEqR2BL8u
3
9
31
false
0.029522
null
false
false
2025-06-05T21:12:09.525Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
grecHJcgkb3KW5wnM
null
null
null
false
null
[]
null
13
0
2025-06-05T16:36:12.466Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
4
null
null
null
null
[ { "__typename": "Tag", "_id": "sYm3HiWcfZvrGu3ui", "adminOnly": false, "afBaseScore": 2, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" } ] }, "baseScore": 12, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:22.097Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" }, { "_id": "sof55TPMQaeBaxhsS", "displayName": "tommylees112" }, { "_id": "AayjS8XzcnDKhGdTv", "displayName": "shark" }, { "_id": "HnALuwRdo6k9HLaMt", "displayName": "Alex Firssoff" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI", "needsReview": false, "noindex": false, "postCount": 12544, "score": 12, "shortName": null, "slug": "ai", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 4, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
9
0
0
4
0
p2N3QhmpKNGn7wkCw
cole-wyeth
2021-07-25T00:01:54.628Z
Amyr
Cole Wyeth
null
null
Cole Wyeth
2,371
146
false
false
<p>I am a PhD student in computer science at the University of Waterloo, supervised by Professor Ming Li and advised by Professor Marcus Hutter.</p><p>My current research is related to applications of algorithmic probability to sequential decision theory (universal artificial intelligence). Recently I have been trying to start a dialogue between the computational cognitive science and UAI communities. Sometimes I build robots, professionally or otherwise. Another hobby (and a personal favorite of my posts here) is the <a href="https://www.lesswrong.com/posts/Yz33koDN5uhSEaB6c/sherlockian-abduction-master-list">Sherlockian abduction master list,</a> which is a crowdsourced project seeking to make "Sherlock Holmes" style inference feasible by compiling observational cues. Give it a read and see if you can contribute!</p><p>See my personal website <a href="https://colewyeth.com/">colewyeth.com</a> for an overview of my interests and work.</p><p>I do ~two types of writing, academic publications and (lesswrong) posts. With the former I try to be careful enough that I can stand by ~all (strong/central) claims in 10 years, usually by presenting a combination of theorems with rigorous proofs and only more conservative intuitive speculation. With the later, I try to learn enough by writing that I have changed my mind by the time I'm finished - and though I usually include an "epistemic status" to suggest my (final) degree of confidence before posting, the ensuing discussion often changes my mind again.</p>
null
null
35
490
4
0
6
1
2
qgdGA4ZEyW7zNdK84
User
null
null
null
[ "canModeratePersonal", "alignmentVoters", "trustLevel1" ]
null
null
xyYss3oCzovibHxAF
https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/xyYss3oCzovibHxAF/hulwrxiopaytj0zu5v8p
SocialPreviewType
WFGghqbZAEqR2BL8u
<p><i>Epistemic status: One week empirical project from a theoretical computer scientist. My analysis and presentation were both a little rushed; some information that would be interesting is missing from plots because I simply did not have time to include it. All known "breaking" issues are discussed and should not effect the conclusions. I may refine this post in the future.</i></p><p>[This work was performed as my final project for ARENA 5.0.]</p><h2>Background</h2><p>I have seen several claims<span class="footnote-reference" data-footnote-reference="" data-footnote-index="1" data-footnote-id="1dkbg31nfrm" role="doc-noteref" id="fnref1dkbg31nfrm"><sup><a href="#fn1dkbg31nfrm">[1]</a></sup></span>&nbsp;in the literature that base LLM in-context learning (ICL) can be understood as approximating Solomonoff induction. <a href="https://www.lesswrong.com/posts/vvgND6aLjuDR6QzDF/my-model-of-what-is-going-on-with-llms">I lean on this intuition a bit myself </a>(and I am in fact a co-author of one of those papers). However, I have not seen any convincing empirical evidence for this model.&nbsp;</p><p>From a theoretical standpoint, it is a somewhat appealing idea. LLMs and Solomonoff induction both face the so-called "prequential problem," predicting a sequence based on a prefix seen so far with a loss function that incentivizes calibration (the log loss; an LLM's loss function may also include other regularization terms like weight decay). Also, ICL is more sample efficient than pretraining. For me, this dovetails with Shane Legg's argument<span class="footnote-reference" data-footnote-reference="" data-footnote-index="2" data-footnote-id="joen89isydk" role="doc-noteref" id="fnrefjoen89isydk"><sup><a href="#fnjoen89isydk">[2]</a></sup></span>&nbsp;that there is no elegant universal theory of prediction, because an online predictor must be complex to learn complex sequences successfully. LLM pretraining is a pretty simple algorithm, but LLM ICL is a very complicated algorithm which leverages a massive number of learned parameters. This is an incomplete argument; Solomonoff induction is a highly general sample efficient algorithm for the prequential problem, as is LLM ICL, but that does not mean they are meaningfully connected. In fact, they are optimized for different distributions: the universal distribution versus the distribution of text on the internet. Arguably, the later may be a special case of the former with an appropriate choice of universal Turing machine (UTM), but I find this perspective to be a bit of a stretch. At the very least I expect LLM ICL to be similar to a universal distribution <i>conditioned on some background information</i>.</p><p>In order to gather some empirical evidence, I adapted this <a href="https://github.com/google-deepmind/neural_networks_solomonoff_induction">approximate universal distribution sampler</a> from Google DeepMind. They used the samples to train a transformer to directly approximate Solomonoff induction in-context, which I will call the Solomonoff Induc... </p>
Epistemic status: One week empirical project from a theoretical computer scientist. My analysis and presentation were both a little rushed; some information that would be interesting is missing from plots because I simply did not have time to include it. All known "breaking" issues are discussed and should not effect the conclusions. I may refine this post in the future. [This work was performed as my final project for ARENA 5.0.] Background I have seen several claims[1] in the literature that base LLM in-context learning (ICL) can be understood as approximating Solomonoff induction. I lean on this intuition a bit myself (and I am in fact a co-author of one of those papers). However, I have not seen any convincing empirical evidence for this model.  From a theoretical standpoint, it is a somewhat appealing idea. LLMs and Solomonoff induction both face the so-called "prequential problem," predicting a sequence based on a prefix seen so far with a loss function that incentivizes calibration (the log loss; an LLM's loss function may also include other regularization terms like weight decay). Also, ICL is more sample efficient than pretraining. For me, this dovetails with Shane Legg's argument[2] that there is no elegant universal theory of prediction, because an online predictor must be complex to learn complex sequences successfully. LLM pretraining is a pretty simple algorithm, but LLM ICL is a very complicated algorithm which leverages a massive number of learned parameters. This is an incomplete argument; Solomonoff induction is a highly general sample efficient algorithm for the prequential problem, as is LLM ICL, but that does not mean they are meaningfully connected. In fact, they are optimized for different distributions: the universal distribution versus the distribution of text on the internet. Arguably, the later may be a special case of the former with an appropriate choice of universal Turing machine (UTM), but I find this perspective to be a bit of a
1,095
1.5.1
Revision
false
null
null
CrosspostOutput
nxkqiMDn7PcptprMm
fundamental-uncertainty-chapter-2-how-do-words-get-their
Fundamental Uncertainty: Chapter 2 - How do words get their meaning?
null
false
false
false
null
gjoi5eBQob27Lww62
null
true
false
false
false
Post
null
2025-06-05T16:32:56.378Z
null
false
false
2
2
2025-06-05T17:34:39.610Z
false
false
post
[]
null
null
iCkPLqJaCfjHG9Jes
0
2
10
false
0.014577
null
false
false
2025-06-05T16:32:56.378Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
grecHJcgkb3KW5wnM
null
null
null
false
null
[]
null
5
0
2025-06-05T16:28:50.645Z
false
false
norm-enforcing
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
13
null
null
null
null
[ { "__typename": "Tag", "_id": "LnEEs8xGooYmQ8iLA", "adminOnly": false, "afBaseScore": 3, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 10, "canEditUserIds": null, "core": false, "createdAt": "2020-07-31T15:36:29.647Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" }, { "_id": "WDi6qQb5TWHb67chh", "displayName": "Haruka Shou" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Truth, Semantics, & Meaning", "needsReview": false, "noindex": false, "postCount": 157, "score": 10, "shortName": null, "slug": "truth-semantics-and-meaning", "suggestedAsFilter": false, "userId": "Q7NW4XaWQmfPfdcFj", "voteCount": 2, "wikiOnly": false }, { "__typename": "Tag", "_id": "Ng8Gice9KNkncxqcj", "adminOnly": false, "afBaseScore": 0, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 1, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:17.072Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 100, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "iMqytjy9ns89Fzfyv", "displayName": "miakko" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Rationality", "needsReview": false, "noindex": false, "postCount": 4302, "score": 1, "shortName": null, "slug": "rationality", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 1, "wikiOnly": false }, { "__typename": "Tag", "_id": "3uE2pXvbcnS9nnZRE", "adminOnly": false, "afBaseScore": 0, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 2, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:50.898Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 27, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "zJtgSyKntXrnkArbY", "displayName": "kistune" }, { "_id": "XqyQTGFGoKfZtdqN3", "displayName": "kINo" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "World Modeling", "needsReview": false, "noindex": false, "postCount": 5855, "score": 2, "shortName": null, "slug": "world-modeling", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 2, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
2
0
0
2
0
gjoi5eBQob27Lww62
gordon-seidoh-worley
2009-03-26T17:18:20.404Z
gworley
Gordon Seidoh Worley
null
null
Gordon Seidoh Worley
9,834
305
false
false
<p>I'm writing a <a href="https://www.fundamentaluncertainty.com/">book</a> about epistemology. It's about <a href="https://www.lesswrong.com/posts/Xs7ag4gsiA6zspmsD/the-problem-of-the-criterion">The Problem of the Criterion</a>, why it's important, and what it has to tell us about how we approach knowing the truth.</p><p>I've also written a lot about AI safety. Some of the more interesting stuff can be found at the site of my currently-dormant AI safety org, <a href="https://paisri.org/">PAISRI</a>.</p>
null
null
209
2,427
7
18
176
1
12
grecHJcgkb3KW5wnM
User
reign-of-terror
[ "mvf4xdfcGzPN8PsXM" ]
true
[ "trustLevel1", "alignmentVoters", "canModeratePersonal", "alignmentForum" ]
null
null
nxkqiMDn7PcptprMm
SocialPreviewType
iCkPLqJaCfjHG9Jes
<p><i>N.B. This is a chapter in a book about truth and knowledge. It's a major revision to the version of </i><a href="https://www.lesswrong.com/posts/Kq9nq5hvPHanrmdNZ/fundamental-uncertainty-chapter-2-why-do-words-have-meaning"><i>this chapter</i></a><i> in the first draft, so I'm publishing it again to get feedback on the completely new text, and have replaced the original version of the chapter with this one in the sequence. You can find more info about the book on its </i><a href="https://www.fundamentaluncertainty.com/"><i>website</i></a><i>.</i></p><p>Like many people, I have a hard time learning foreign languages. I've tried dozens of apps, books, and courses to learn everything from Mandarin Chinese to Classical Greek, all with little success. The closest I ever came to really learning another language was when I took French in high school, but I was so far from fluency that, after three years of intense study, I was lucky to receive a passing grade.</p><p>I started out with high hopes that learning French would be easy. All I had to do, it seemed, was memorize the mapping of French words to English ones. If I could remember that "le cheval" means horse and "la voiture" means car and so on, I'd be able to speak French fluently.</p><p>Unfortunately, it didn't work. I couldn't map between French and English fast enough to do anything more than read and write French very slowly. My teacher said I needed to learn to think in French and stop trying to translate everything in my head, but I didn't want to believe him. I was convinced that my way should work. What finally changed my mind was discovering that some English words lack a French equivalent.</p><p>Consider the word "mug". In English it encompasses many different types of cups made of different materials, from glass beer mugs to clay coffee mugs to insulated metal travel mugs and more. But in French a beer mug is "une chope", a coffee mug is "une tasse à café", and a travel mug might be "un thermos" or "un gobelet de voyage". Each one gets its own name, and there is no one word like "mug" that groups them all together. The closest translations of "mug" are the loanword "le mug"—which usually means a coffee mug—and the descriptive phrase "les tasses à anse"—cups with handles—but neither option succeeds in capturing all the implicit meaning carried by "mug" in English. There's simply no word in French that means exactly what "mug" does.</p><p>And "mug" is just one example. English has hundreds of words like "cozy" and "snack" that lack native equivalents in French, and thousands more where the meaning of the straightforward translation is close... </p>
N.B. This is a chapter in a book about truth and knowledge. It's a major revision to the version of this chapter in the first draft, so I'm publishing it again to get feedback on the completely new text, and have replaced the original version of the chapter with this one in the sequence. You can find more info about the book on its website. Like many people, I have a hard time learning foreign languages. I've tried dozens of apps, books, and courses to learn everything from Mandarin Chinese to Classical Greek, all with little success. The closest I ever came to really learning another language was when I took French in high school, but I was so far from fluency that, after three years of intense study, I was lucky to receive a passing grade. I started out with high hopes that learning French would be easy. All I had to do, it seemed, was memorize the mapping of French words to English ones. If I could remember that "le cheval" means horse and "la voiture" means car and so on, I'd be able to speak French fluently. Unfortunately, it didn't work. I couldn't map between French and English fast enough to do anything more than read and write French very slowly. My teacher said I needed to learn to think in French and stop trying to translate everything in my head, but I didn't want to believe him. I was convinced that my way should work. What finally changed my mind was discovering that some English words lack a French equivalent. Consider the word "mug". In English it encompasses many different types of cups made of different materials, from glass beer mugs to clay coffee mugs to insulated metal travel mugs and more. But in French a beer mug is "une chope", a coffee mug is "une tasse à café", and a travel mug might be "un thermos" or "un gobelet de voyage". Each one gets its own name, and there is no one word like "mug" that groups them all together. The closest translations of "mug" are the loanword "le mug"—which usually means a coffee mug—and the descriptive phras
3,326
1.2.0
Revision
false
null
null
CrosspostOutput
8j62pB4vRiWNwbk9c
ai-might-kill-everyone
AI Might Kill Everyone
null
false
false
false
null
tm8YP7vNWjGm7pYae
null
true
false
false
false
Post
null
2025-06-05T15:37:59.830Z
null
false
false
2
2
2025-06-05T17:41:04.908Z
false
false
post
[]
null
null
t4dQrnv6dogyMeJ2K
0
8
6
false
0.011528
null
false
false
2025-06-05T15:37:59.830Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
grecHJcgkb3KW5wnM
null
null
null
false
null
[]
null
1
0
2025-06-05T15:37:45.063Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
4
null
null
null
null
[ { "__typename": "Tag", "_id": "sYm3HiWcfZvrGu3ui", "adminOnly": false, "afBaseScore": 2, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" } ] }, "baseScore": 12, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:22.097Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" }, { "_id": "sof55TPMQaeBaxhsS", "displayName": "tommylees112" }, { "_id": "AayjS8XzcnDKhGdTv", "displayName": "shark" }, { "_id": "HnALuwRdo6k9HLaMt", "displayName": "Alex Firssoff" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI", "needsReview": false, "noindex": false, "postCount": 12544, "score": 12, "shortName": null, "slug": "ai", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 4, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
8
0
0
2
0
tm8YP7vNWjGm7pYae
bentham-s-bulldog
2022-11-24T02:24:14.930Z
omnizoid
Bentham's Bulldog
null
null
null
249
0
false
false
null
null
41
131
1
0
0
1
0
55XxDBpfKkkBPm9H8
User
null
null
null
[ "canModeratePersonal" ]
null
null
8j62pB4vRiWNwbk9c
SocialPreviewType
t4dQrnv6dogyMeJ2K
<p><br>(Crosspost from <a href="https://benthams.substack.com/p/ai-might-kill-everyone">my blog)</a>.&nbsp;</p><p>(I’ll be at EAG London this weekend—come say hi. Also, this is my thousandth blogpost—cool milestone!)</p><p>Several people have wondered why I haven’t written much about AI. The main reason is that I don’t feel that I have anything very original to contribute. I try to only speak about things that I know something about. I don’t really know much about AI. While I use it regularly for editing, and have read a decent amount about it, it’s quite far outside of my area of expertise.</p><p>But I feel that I should say <i>something </i>about it because it’s important. It’s plausibly the most important thing. We may be approaching a second industrial revolution with consequences more dramatic than the first. There is a real possibility of everyone dying.</p><p>I’m on the more optimistic side. I think there’s only a few percent chance that AI kills everyone. Maybe 1 or 2%. In the EA circles in which I hang out, this often makes me outrageously optimistic. Lots of people, like Yudkowsky, think we’re nearly guaranteed to all die.</p><p>But whether one’s P(doom) is 1% or 60%, it’s abundantly clear that we should be doing a lot more than we are currently doing. AI alignment research—research that makes sure AI does what we want—should take up a sizeable portion of the federal budget. AI governance and international treaties are sorely needed. (<a href="https://80000hours.org/career-reviews/"><u>Here are some high impact careers</u></a>—many related to AI—and here are a bunch of high impact charities for safeguarding the longterm, <a href="https://thingofthings.substack.com/p/ea-funds-that-exist-2024a?utm_source=publication-search"><u>largely by aligning AI</u></a>). Your odds of dying from AI are a lot higher than your odds of dying in a car accident.</p><p>If AI goes well, it could usher in unprecedented prosperity. If it goes poorly, it could usher in an unimaginable catastrophe. In such a world, trying to steer AI so that it goes well should be a top global priority.</p><p>A lot of the arguments for AI risk have been written with a great deal of technical language, but I think the core argument for AI risk is pretty straightforward. We are building things that are much smarter than we are. AI can already do many human jobs better than most people.</p><p>A few years ago, AI couldn’t write competently. GPT2 was useless. GPT3 was revolutionary. GPT4 was better than humans at many tasks. What will GPT10 look like? Whatever AI is like in 30 years, it will be very impressive.</p><p>AI has already surpassed us in lots of tasks. The best human chess player can’t hold a candle... </p>
(Crosspost from my blog).  (I’ll be at EAG London this weekend—come say hi. Also, this is my thousandth blogpost—cool milestone!) Several people have wondered why I haven’t written much about AI. The main reason is that I don’t feel that I have anything very original to contribute. I try to only speak about things that I know something about. I don’t really know much about AI. While I use it regularly for editing, and have read a decent amount about it, it’s quite far outside of my area of expertise. But I feel that I should say something about it because it’s important. It’s plausibly the most important thing. We may be approaching a second industrial revolution with consequences more dramatic than the first. There is a real possibility of everyone dying. I’m on the more optimistic side. I think there’s only a few percent chance that AI kills everyone. Maybe 1 or 2%. In the EA circles in which I hang out, this often makes me outrageously optimistic. Lots of people, like Yudkowsky, think we’re nearly guaranteed to all die. But whether one’s P(doom) is 1% or 60%, it’s abundantly clear that we should be doing a lot more than we are currently doing. AI alignment research—research that makes sure AI does what we want—should take up a sizeable portion of the federal budget. AI governance and international treaties are sorely needed. (Here are some high impact careers—many related to AI—and here are a bunch of high impact charities for safeguarding the longterm, largely by aligning AI). Your odds of dying from AI are a lot higher than your odds of dying in a car accident. If AI goes well, it could usher in unprecedented prosperity. If it goes poorly, it could usher in an unimaginable catastrophe. In such a world, trying to steer AI so that it goes well should be a top global priority. A lot of the arguments for AI risk have been written with a great deal of technical language, but I think the core argument for AI risk is pretty straightforward. We are building thi
1,102
1.1.0
Revision
false
null
null
CrosspostOutput
D3BjqZ26ouk7ctfRC
ai-119-goodbye-aisi
AI #119: Goodbye AISI?
null
false
false
false
null
N9zj5qpTfqmbn9dro
null
true
false
false
false
Post
null
2025-06-05T14:00:05.389Z
null
false
false
2
2
null
false
false
post
[]
null
null
FKFpsdBuqQkNfCzre
8
19
42
false
0.029414
null
false
false
2025-06-11T00:05:42.598Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
null
null
grecHJcgkb3KW5wnM
null
null
null
false
null
[]
null
7
0
2025-06-05T14:00:05.389Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
72
null
null
null
null
[ { "__typename": "Tag", "_id": "8byoqYZfdwHffYLZ6", "adminOnly": false, "afBaseScore": 3, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 9, "canEditUserIds": null, "core": false, "createdAt": "2020-07-01T18:44:14.645Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Newsletters", "needsReview": false, "noindex": false, "postCount": 411, "score": 9, "shortName": null, "slug": "newsletters", "suggestedAsFilter": false, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 1, "wikiOnly": false }, { "__typename": "Tag", "_id": "sYm3HiWcfZvrGu3ui", "adminOnly": false, "afBaseScore": 2, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" } ] }, "baseScore": 12, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:22.097Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" }, { "_id": "sof55TPMQaeBaxhsS", "displayName": "tommylees112" }, { "_id": "AayjS8XzcnDKhGdTv", "displayName": "shark" }, { "_id": "HnALuwRdo6k9HLaMt", "displayName": "Alex Firssoff" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI", "needsReview": false, "noindex": false, "postCount": 12544, "score": 12, "shortName": null, "slug": "ai", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 4, "wikiOnly": false } ]
QSR8rPZxZzxEXoPjR
0
0
null
false
null
null
0
19
0
0
5
0
N9zj5qpTfqmbn9dro
zvi
2009-03-31T20:54:54.077Z
Zvi
Zvi
null
null
null
51,554
146
false
false
null
null
936
1,461
3
2
7
1
0
qgdGA4ZEyW7zNdK84
User
norm-enforcing
null
true
[ "trustLevel1", "canModeratePersonal", "alignmentVoters", "alignmentForum" ]
null
null
D3BjqZ26ouk7ctfRC
https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/D3BjqZ26ouk7ctfRC/mhwruq9nkgfycmwl7nuj
SocialPreviewType
FKFpsdBuqQkNfCzre
<p>AISI is being rebranded highly non-confusingly as CAISI. Is it the end of AISI and a huge disaster, or a tactical renaming to calm certain people down? Hard to tell. It could go either way. Sometimes you need to target the people who call things ‘big beautiful bill,’ and hey, the bill in question is indeed big. It contains multitudes.</p><p>The AI world also contains multitudes. We got Cursor 1.0, time to get coding.</p><p>On a personal note, this was the week of LessOnline, which was predictably great. I am sad that I could not stay longer, but as we all know, duty calls. Back to the grind.</p> <div> <span id="more-24497"></span> </div> <h4>Table of Contents</h4> <ol> <li><a href="https://thezvi.substack.com/i/164743152/language-models-offer-mundane-utility">Language Models Offer Mundane Utility.</a> The white whale.</li> <li><a href="https://thezvi.substack.com/i/164743152/language-models-don-t-offer-mundane-utility"><strong>Language Models Don’t Offer Mundane Utility</strong>.</a> You need a system prompt.</li> <li><a href="https://thezvi.substack.com/i/164743152/language-models-could-offer-more-mundane-utility">Language Models Could Offer More Mundane Utility.</a> A good set of asks.</li> <li><a href="https://thezvi.substack.com/i/164743152/huh-upgrades"><strong>Huh, Upgrades</strong>.</a> The highlight is Cursor 1.0, with memory and more.</li> <li><a href="https://thezvi.substack.com/i/164743152/fun-with-media-generation">Fun With Media Generation.</a> Video is high bandwidth. But also low bandwidth.</li> <li><a href="https://thezvi.substack.com/i/164743152/choose-your-fighter">Choose Your Fighter.</a> Opinions differ, I continue to mostly be on Team Claude.</li> <li><a href="https://thezvi.substack.com/i/164743152/deepfaketown-and-botpocalypse-soon">Deepfaketown and Botpocalypse Soon.</a> Fake is not a natural category. Whoops.</li> <li><a href="https://thezvi.substack.com/i/164743152/get-my-agent-on-the-line">Get My Agent On The Line.</a> We all know they’re not secure, but how bad is this?</li> <li><a href="https://thezvi.substack.com/i/164743152/they-took-our-jobs">They Took Our Jobs.</a> Economists respond to Dario’s warning.</li> <li><a href="https://thezvi.substack.com/i/164743152/the-art-of-the-jailbreak">The Art of the Jailbreak.</a> Why not jailbreak AI overviews?</li> <li><a href="https://thezvi.substack.com/i/164743152/unprompted-attention">Unprompted Attention.</a> More prompts to try out.</li> <li><a href="https://thezvi.substack.com/i/164743152/get-involved">Get Involved.</a> SFCompute, Speculative Technologies.</li> <li><a href="https://thezvi.substack.com/i/164743152/introducing">Introducing.</a> Anthropic open sources interpretability tools, better AR glasses.</li> <li><a href="https://thezvi.substack.com/i/164743152/in-other-ai-news">In Other AI News.</a> FDA launches their AI tool called Elsa.</li> <li><a href="https://thezvi.substack.com/i/164743152/show-me-the-money">Show Me the Money.</a> Delaware hires bank to value OpenAI’s nonprofit.</li> <li><a href="https://thezvi.substack.com/i/164743152/quiet-speculations">Quiet Speculations.</a> People don’t get what is coming, but hey, could be worse.</li> <li><a href="https://thezvi.substack.com/i/164743152/taking-off">Taking Off.</a> AI beats humans in a test of predicting the results of ML experiments.</li> <li><a href="https://thezvi.substack.com/i/164743152/goodbye-aisi"><strong>Goodbye AISI?</strong></a> They’re rebranding as CAISI. It’s unclear how much this matters.</li> <li><a href="https://thezvi.substack.com/i/164743152/the-quest-for-sane-regulations">The Quest for Sane Regulations.</a> The bill is, at least, definitely big. Tl;dr.</li> <li><a href="https://thezvi.substack.com/i/164743152/copyright-confrontation">Copyright Confrontation.</a> OpenAI is being forced to retain all its chat logs.</li> <li><a href="https://thezvi.substack.com/i/164743152/differential-access">Differential Access.</a> The Good Guy needs a better AI than the Bad Guy.</li> <li><a href="https://thezvi.substack.com/i/164743152/the-week-in-audio">The Week in Audio.</a> Altman, Tegmark, Amodei, Barnes.</li> <li><a href="https://thezvi.substack.com/i/164743152/when-david-sacks-says-win-the-ai-race-he-literally-means-market-share"><strong>When David Sacks Says ‘Win the AI Race’ He Literally Means Market Share</strong>.</a></li> <li><a href="https://thezvi.substack.com/i/164743152/rhetorical-innovation">Rhetorical Innovation.</a> Blog metagame continues to dominate.</li> <li><a href="https://thezvi.substack.com/i/164743152/aligning-a-smarter-than-human-intelligence-is-difficult">Aligning a Smarter Than Human Intelligence is Difficult.</a> Proceed accordingly.</li> <li><a href="https://thezvi.substack.com/i/164743152/misaligned">Misaligned!</a> About that safety plan, would it, you know, actually </li></ol>...
AISI is being rebranded highly non-confusingly as CAISI. Is it the end of AISI and a huge disaster, or a tactical renaming to calm certain people down? Hard to tell. It could go either way. Sometimes you need to target the people who call things ‘big beautiful bill,’ and hey, the bill in question is indeed big. It contains multitudes. The AI world also contains multitudes. We got Cursor 1.0, time to get coding. On a personal note, this was the week of LessOnline, which was predictably great. I am sad that I could not stay longer, but as we all know, duty calls. Back to the grind. TABLE OF CONTENTS 1. Language Models Offer Mundane Utility. The white whale. 2. Language Models Don’t Offer Mundane Utility. You need a system prompt. 3. Language Models Could Offer More Mundane Utility. A good set of asks. 4. Huh, Upgrades. The highlight is Cursor 1.0, with memory and more. 5. Fun With Media Generation. Video is high bandwidth. But also low bandwidth. 6. Choose Your Fighter. Opinions differ, I continue to mostly be on Team Claude. 7. Deepfaketown and Botpocalypse Soon. Fake is not a natural category. Whoops. 8. Get My Agent On The Line. We all know they’re not secure, but how bad is this? 9. They Took Our Jobs. Economists respond to Dario’s warning. 10. The Art of the Jailbreak. Why not jailbreak AI overviews? 11. Unprompted Attention. More prompts to try out. 12. Get Involved. SFCompute, Speculative Technologies. 13. Introducing. Anthropic open sources interpretability tools, better AR glasses. 14. In Other AI News. FDA launches their AI tool called Elsa. 15. Show Me the Money. Delaware hires bank to value OpenAI’s nonprofit. 16. Quiet Speculations. People don’t get what is coming, but hey, could be worse. 17. Taking Off. AI beats humans in a test of predicting the results of ML experiments. 18. Goodbye AISI? They’re rebranding as CAISI. It’s unclear how much this matters. 19. The Quest for Sane Regulations. The bill is, at least, definit
17,893
1.0.1
Revision
false
null
null
CrosspostOutput
LT73KyjG9aFxuvvoT
powerful-predictions
Powerful Predictions
null
false
false
true
null
iXX23K6iBAosHFPBn
null
true
false
false
false
Post
https://forecastingaifutures.substack.com/p/powerful-predictions
2025-06-05T10:44:33.457Z
null
false
false
2
2
2025-06-06T17:57:38.924Z
false
false
linkpost
[]
null
null
LzdChAuWHh49jtGPg
0
1
2
false
0.008671
null
false
false
2025-06-05T10:44:33.457Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
qgdGA4ZEyW7zNdK84
null
null
null
false
2025-06-24T18:48:26.382Z
[ "iXX23K6iBAosHFPBn" ]
XtphY3uYHwruKqDyG
2
0
2025-06-05T10:31:10.120Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
7
null
null
null
null
[ { "__typename": "Tag", "_id": "qHDus5MuMNqQxJbjD", "adminOnly": false, "afBaseScore": 4, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" }, { "_id": "oEF4gToHRPEMw4FSo", "displayName": "Jono" } ] }, "baseScore": 11, "canEditUserIds": null, "core": false, "createdAt": "2020-08-09T18:31:56.709Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" }, { "_id": "oEF4gToHRPEMw4FSo", "displayName": "Jono" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI Governance", "needsReview": false, "noindex": false, "postCount": 726, "score": 11, "shortName": null, "slug": "ai-governance", "suggestedAsFilter": false, "userId": "QBvPFLFyZyuHcBwFm", "voteCount": 2, "wikiOnly": false }, { "__typename": "Tag", "_id": "8daMDi9NEShyLqxth", "adminOnly": false, "afBaseScore": 10, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "EQNTWXLKMeWMp2FQS", "displayName": "Ben Pace" }, { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" }, { "_id": "iXX23K6iBAosHFPBn", "displayName": "Alvin Ånestrand" } ] }, "baseScore": 21, "canEditUserIds": null, "core": false, "createdAt": "2020-05-10T05:54:39.783Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "EQNTWXLKMeWMp2FQS", "displayName": "Ben Pace" }, { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" }, { "_id": "iXX23K6iBAosHFPBn", "displayName": "Alvin Ånestrand" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Forecasting & Prediction", "needsReview": false, "noindex": false, "postCount": 508, "score": 21, "shortName": null, "slug": "forecasting-and-prediction", "suggestedAsFilter": false, "userId": "iBcH2a3HdWGS2JEZA", "voteCount": 3, "wikiOnly": false }, { "__typename": "Tag", "_id": "5f5c37ee1b5cdee568cfb2b0", "adminOnly": false, "afBaseScore": null, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 0, "canEditUserIds": null, "core": false, "createdAt": "2020-09-11T19:58:52.604Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Technological Forecasting", "needsReview": false, "noindex": false, "postCount": 103, "score": 0, "shortName": null, "slug": "technological-forecasting", "suggestedAsFilter": false, "userId": "cn4SiEmqWbu7K9em5", "voteCount": 0, "wikiOnly": false }, { "__typename": "Tag", "_id": "xexCWMyds6QLWognu", "adminOnly": false, "afBaseScore": 0, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 2, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T03:38:23.532Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 20, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "si6LoAENzqPCmi2Dh", "displayName": "ihatenumbersinusernames7" }, { "_id": "B2sXjQTGgwoGE9FES", "displayName": "SeaIgloo" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "World Optimization", "needsReview": false, "noindex": false, "postCount": 3151, "score": 2, "shortName": null, "slug": "world-optimization", "suggestedAsFilter": true, "userId": "XtphY3uYHwruKqDyG", "voteCount": 2, "wikiOnly": false }, { "__typename": "Tag", "_id": "sYm3HiWcfZvrGu3ui", "adminOnly": false, "afBaseScore": 2, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" } ] }, "baseScore": 12, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:22.097Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" }, { "_id": "sof55TPMQaeBaxhsS", "displayName": "tommylees112" }, { "_id": "AayjS8XzcnDKhGdTv", "displayName": "shark" }, { "_id": "HnALuwRdo6k9HLaMt", "displayName": "Alex Firssoff" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI", "needsReview": false, "noindex": false, "postCount": 12544, "score": 12, "shortName": null, "slug": "ai", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 4, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
1
0
0
1
0
iXX23K6iBAosHFPBn
alvin-anestrand
2022-11-03T09:56:49.205Z
alvin-anestrand
Alvin Ånestrand
null
null
null
102
22
false
false
null
null
13
10
0
8
0
1
1
qgdGA4ZEyW7zNdK84
User
null
null
null
[ "canModeratePersonal", "alignmentVoters" ]
null
null
LT73KyjG9aFxuvvoT
https://res.cloudinary.com/lesswrong-2-0/image/upload/c_fill,ar_1.91,g_auto/SocialPreview/vezevvesd2zobauh2q6i
SocialPreviewType
LzdChAuWHh49jtGPg
<p><i>A thoughtful post by Anton Leicht, “</i><a href="https://writing.antonleicht.me/p/powerless-predictions"><i><u>Powerless Predictions</u></i></a><i>,” notes that forecasting may be underutilized by AI policy organizations. But how much is forecasting actually being used? Which organizations working on AI—particularly AI safety—already integrate forecasting into their decision-making, policy recommendations, and strategic planning? What forecasting work is being done, and how is it impacting the future of AI?</i></p><p><i>I set out to investigate this.</i></p><p><i>While I hope this post proves useful to others, I wrote it primarily for myself to gain an overview of forecasting work and its impacts. Feel free to skim—only read the details if you find them valuable.</i></p><h2><strong>Grantmaking</strong></h2><p>Open Philanthropy, which funds numerous AI safety initiatives, regularly makes predictions about grant outcomes. Evaluating these outcomes helps them improve their ability to predict the results of future grantmaking decisions. Javier Prieto at Open Philanthropy explains this process and evaluates prediction accuracy in <a href="https://www.openphilanthropy.org/research/how-accurate-are-our-predictions/#id-1-how-we-make-and-check-our-forecasts"><u>this post</u></a>.</p><p>They have also commissioned an long list of <a href="https://arbresearch.com/files/clairvoyance.pdf"><u>forecasting questions on AI</u></a> from <a href="https://arbresearch.com/"><u>Arb Research</u></a>. I couldn't find any publicly documented uses of this list, though Open Philanthropy may have used it for internal planning and strategic thinking.</p><h2><strong>AI development</strong></h2><p>Anthropic <a href="https://www.anthropic.com/research/forecasting-rare-behaviors"><u>extrapolates rare AI behaviors</u></a> to reduce concerns about concerning behaviors that may be missed during evaluations.</p><p>In Anthropic's <a href="https://www.anthropic.com/news/anthropics-responsible-scaling-policy"><u>Responsible Scaling Policy</u></a> (RSP), they incorporate forecasting for comprehensive AI assessment, making informal predictions about improvements of elicitation techniques and enhanced model performance between testing rounds. They aim to "improve these forecasts over time so that they can be relied upon for risk judgments."</p><p>While you could argue that frontier AI development would be irresponsible without forecasting future capabilities and risks, it's nevertheless encouraging to see this explicitly incorporated into their practices. DeepMind's <a href="https://deepmind.google/discover/blog/updating-the-frontier-safety-framework/"><u>Frontier Safety Framework</u></a> doesn't appear to explicitly include forecasting, and OpenAI's <a href="https://openai.com/index/updating-our-preparedness-framework/"><u>Preparedness Framework</u></a> only mentions it in passing.</p><h2><strong>Affecting Policy Decisions</strong></h2><p>While policy recommendations implicitly depend on predictions about the future, explicit and well-researched forecasting doesn't appear to be the norm. However, much policy work may be informed by forecasts without being transparently based on forecast analyses—making it difficult to determine how... </p>
A thoughtful post by Anton Leicht, “Powerless Predictions,” notes that forecasting may be underutilized by AI policy organizations. But how much is forecasting actually being used? Which organizations working on AI—particularly AI safety—already integrate forecasting into their decision-making, policy recommendations, and strategic planning? What forecasting work is being done, and how is it impacting the future of AI? I set out to investigate this. While I hope this post proves useful to others, I wrote it primarily for myself to gain an overview of forecasting work and its impacts. Feel free to skim—only read the details if you find them valuable. Grantmaking Open Philanthropy, which funds numerous AI safety initiatives, regularly makes predictions about grant outcomes. Evaluating these outcomes helps them improve their ability to predict the results of future grantmaking decisions. Javier Prieto at Open Philanthropy explains this process and evaluates prediction accuracy in this post. They have also commissioned an long list of forecasting questions on AI from Arb Research. I couldn't find any publicly documented uses of this list, though Open Philanthropy may have used it for internal planning and strategic thinking. AI development Anthropic extrapolates rare AI behaviors to reduce concerns about concerning behaviors that may be missed during evaluations. In Anthropic's Responsible Scaling Policy (RSP), they incorporate forecasting for comprehensive AI assessment, making informal predictions about improvements of elicitation techniques and enhanced model performance between testing rounds. They aim to "improve these forecasts over time so that they can be relied upon for risk judgments." While you could argue that frontier AI development would be irresponsible without forecasting future capabilities and risks, it's nevertheless encouraging to see this explicitly incorporated into their practices. DeepMind's Frontier Safety Framework doesn't appear to ex
1,680
1.1.0
Revision
false
null
null
CrosspostOutput
24RvfBgyt72XzfEYS
potentially-useful-projects-in-wise-ai
Potentially Useful Projects in Wise AI
null
false
false
true
null
XLwKyCK7JmC292ZCC
null
true
false
false
false
Post
null
2025-06-05T08:13:42.839Z
null
false
false
2
2
2025-06-05T17:41:33.494Z
false
false
post
[]
null
null
dAPsaeRqwkHnfbdSx
0
4
12
false
0.015608
null
false
false
2025-06-05T08:13:42.839Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
grecHJcgkb3KW5wnM
null
null
null
false
null
[]
null
4
0
2025-06-05T07:22:45.184Z
false
false
easy-going
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
7
null
null
null
null
[ { "__typename": "Tag", "_id": "Ki4TywKtMREHp9zFC", "adminOnly": false, "afBaseScore": null, "afExtendedScore": null, "baseScore": 0, "canEditUserIds": null, "core": false, "createdAt": "2025-06-05T12:30:00.389Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": null, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Artificial Wisdom", "needsReview": true, "noindex": false, "postCount": 4, "score": 0, "shortName": null, "slug": "artificial-wisdom-1", "suggestedAsFilter": false, "userId": "XLwKyCK7JmC292ZCC", "voteCount": 0, "wikiOnly": false }, { "__typename": "Tag", "_id": "sYm3HiWcfZvrGu3ui", "adminOnly": false, "afBaseScore": 2, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" } ] }, "baseScore": 12, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:22.097Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" }, { "_id": "sof55TPMQaeBaxhsS", "displayName": "tommylees112" }, { "_id": "AayjS8XzcnDKhGdTv", "displayName": "shark" }, { "_id": "HnALuwRdo6k9HLaMt", "displayName": "Alex Firssoff" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI", "needsReview": false, "noindex": false, "postCount": 12544, "score": 12, "shortName": null, "slug": "ai", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 4, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
4
0
0
1
0
XLwKyCK7JmC292ZCC
chris_leong
2009-05-28T03:08:43.251Z
Chris_Leong
Chris_Leong
null
null
null
7,651
457
false
false
null
null
227
2,158
3
32
206
1
71
r38pkCm7wF4M44MDQ
User
easy-going
null
null
[ "trustLevel1", "alignmentVoters", "canModeratePersonal", "alignmentForum" ]
null
null
24RvfBgyt72XzfEYS
SocialPreviewType
dAPsaeRqwkHnfbdSx
<p>This is a list of projects<span class="footnote-reference" data-footnote-reference="" data-footnote-index="1" data-footnote-id="lcc7gtz365j" role="doc-noteref" id="fnreflcc7gtz365j"><sup><a href="#fnlcc7gtz365j">[1]</a></sup></span>&nbsp;to consider for folks who want to use Wise AI to steer the world towards positive outcomes.</p><p>Some of these projects are listed because they're impactful. Others are listed because I believe they would be good projects for someone to get started.&nbsp;</p><p>Please note that this post is titled "potentially useful projects" for a reason. Some of these projects are likely to have much higher impact than others. I expect some are net-negative. Don't ignore your own judgment just because a project is listed here!</p><p>I'm sure my views about what projects are valuable will change quite significantly as I dive deeper into the topic, but I still think it's worthwhile for me to put some kind of list out there.&nbsp;</p><p><i>But first, an announcement:</i></p><details class="detailsBlock"><summary class="detailsBlockTitle"><p><strong>Applications for The Future of Life Foundation's Fellowship on AI for Human Reasoning are closing soon (June 9th!)</strong><br><br><i>They've listed "Tools for wise decision making" as a possible area to work on.</i><br><br><i>Expand for more details.</i></p></summary><div class="detailsBlockContent"><p>From their website:</p><blockquote><p>Apply by June 9th | $25k–$50k stipend | 12 weeks, from July 14 - October 3<br><br>Join us in working out how to build a future which robustly empowers humans and improves decision-making.&nbsp;<br><br>FLF’s incubator fellowship on AI for human reasoning will help talented researchers and builders start working on AI tools for coordination and epistemics. Participants will scope out and work on pilot projects in this area, with discussion and guidance from experts working in related fields. FLF will provide fellows with a $25k–$50k stipend, the opportunity to work in a shared office in the SF Bay Area or remotely, and other support.&nbsp;<br><br>In some cases we would be excited to provide support beyond the end of the fellowship period, or help you in launching a new organization.</p></blockquote><p><a href="https://www.flf.org/fellowship">Further Information</a> ✦ <a href="https://jobs.lever.co/futureof-life/ffc752f2-a420-4c87-8c58-2212ae2e885c/apply">Apply now!</a></p></div></details><p>I was originally going to delay publishing this until <i>after</i> making the case Wise AI Advisors being a priority (and, more generally, Wise AI as well), but I ended up deciding that it was important to publish this before the deadline for the Future of Life Fellowship in case it inspired more people to apply.</p><hr><h3><strong>Field-building:</strong></h3><p>At this stage, I consider field-building work around Wise AI as especially high priority. I feel that there’s starting to be some real energy around this area, however, very little of this will aid us in avoiding catastrophic risks. Fields tend to be more flexible in their initi... </p>
This is a list of projects[1] to consider for folks who want to use Wise AI to steer the world towards positive outcomes. Some of these projects are listed because they're impactful. Others are listed because I believe they would be good projects for someone to get started.  Please note that this post is titled "potentially useful projects" for a reason. Some of these projects are likely to have much higher impact than others. I expect some are net-negative. Don't ignore your own judgment just because a project is listed here! I'm sure my views about what projects are valuable will change quite significantly as I dive deeper into the topic, but I still think it's worthwhile for me to put some kind of list out there.  But first, an announcement: Applications for The Future of Life Foundation's Fellowship on AI for Human Reasoning are closing soon (June 9th!) They've listed "Tools for wise decision making" as a possible area to work on. Expand for more details. From their website: > Apply by June 9th | $25k–$50k stipend | 12 weeks, from July 14 - October 3 > > Join us in working out how to build a future which robustly empowers humans and improves decision-making.  > > FLF’s incubator fellowship on AI for human reasoning will help talented researchers and builders start working on AI tools for coordination and epistemics. Participants will scope out and work on pilot projects in this area, with discussion and guidance from experts working in related fields. FLF will provide fellows with a $25k–$50k stipend, the opportunity to work in a shared office in the SF Bay Area or remotely, and other support.  > > In some cases we would be excited to provide support beyond the end of the fellowship period, or help you in launching a new organization. Further Information ✦ Apply now! I was originally going to delay publishing this until after making the case Wise AI Advisors being a priority (and, more generally, Wise AI as well), but I ended up deciding that it wa
1,836
1.16.0
Revision
true
true
uXaCF4Ju2g4sndXXZ
CrosspostOutput
t6Tjm9nF2C6qmHiJ8
building-as-gardening
Building as gardening
null
false
false
false
null
wgaRentCGQpw6fyhT
null
true
false
false
false
Post
https://productidentity.co/p/building-as-gardening
2025-06-05T06:41:56.540Z
null
false
false
2
2
2025-06-06T18:00:31.259Z
false
false
linkpost
[]
null
null
yG8EJSFPBKTki34LA
1
1
3
false
0.008989
null
false
false
2025-06-05T23:14:32.888Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
qgdGA4ZEyW7zNdK84
null
null
null
false
null
[]
null
0
0
2025-06-05T06:38:28.879Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
4
null
null
null
null
[ { "__typename": "Tag", "_id": "ux2x9RrJsuykQxT79", "adminOnly": false, "afBaseScore": 8, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "r38pkCm7wF4M44MDQ", "displayName": "Raemon" }, { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 22, "canEditUserIds": null, "core": false, "createdAt": "2020-10-14T16:28:33.955Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "r38pkCm7wF4M44MDQ", "displayName": "Raemon" }, { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" }, { "_id": "d5WCJPEGt8KDEdj2T", "displayName": "Giordano Rogers" }, { "_id": "xF5nfdddHjFThHy49", "displayName": "[email protected]" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Deliberate Practice", "needsReview": false, "noindex": false, "postCount": 31, "score": 22, "shortName": null, "slug": "deliberate-practice", "suggestedAsFilter": false, "userId": "Q7NW4XaWQmfPfdcFj", "voteCount": 4, "wikiOnly": false }, { "__typename": "Tag", "_id": "xknvtHwqvqhwahW8Q", "adminOnly": false, "afBaseScore": 3, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 9, "canEditUserIds": null, "core": false, "createdAt": "2020-08-19T22:24:55.795Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Human Values", "needsReview": false, "noindex": false, "postCount": 225, "score": 9, "shortName": null, "slug": "human-values", "suggestedAsFilter": false, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 1, "wikiOnly": false }, { "__typename": "Tag", "_id": "fkABsGCJZ6y9qConW", "adminOnly": false, "afBaseScore": 0, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 2, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T06:06:46.947Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "MiuAZvbQcQ7ethgt3", "displayName": "Viktor withaK" }, { "_id": "dRaCtsAWxk7sgirSY", "displayName": "Jordan Morgan" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Practical", "needsReview": false, "noindex": false, "postCount": 3411, "score": 2, "shortName": null, "slug": "practical", "suggestedAsFilter": true, "userId": "oBSWiHjgproTiThmY", "voteCount": 2, "wikiOnly": false }, { "__typename": "Tag", "_id": "xexCWMyds6QLWognu", "adminOnly": false, "afBaseScore": 0, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 2, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T03:38:23.532Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 20, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "si6LoAENzqPCmi2Dh", "displayName": "ihatenumbersinusernames7" }, { "_id": "B2sXjQTGgwoGE9FES", "displayName": "SeaIgloo" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "World Optimization", "needsReview": false, "noindex": false, "postCount": 3151, "score": 2, "shortName": null, "slug": "world-optimization", "suggestedAsFilter": true, "userId": "XtphY3uYHwruKqDyG", "voteCount": 2, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
1
0
0
0
0
wgaRentCGQpw6fyhT
itay-dreyfus
2024-01-15T20:29:32.588Z
itay-dreyfus
Itay Dreyfus
null
null
null
118
0
false
false
<p>Designer and writer.</p><p><a href="https://productidentity.co/">https://productidentity.co/</a></p><p><a href="https://twitter.com/itaydre">https://twitter.com/itaydre</a></p><p>[email protected]</p>
null
null
12
5
0
0
0
1
0
55XxDBpfKkkBPm9H8
User
null
null
null
[ "canModeratePersonal" ]
null
null
t6Tjm9nF2C6qmHiJ8
https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/t6Tjm9nF2C6qmHiJ8/vcfmb7djgyg75vvv9jcl
SocialPreviewType
yG8EJSFPBKTki34LA
<p>Since I’ve been interviewing other builders for the&nbsp;<a href="https://nichedesign.press/"><u>niche design zine</u></a>, I've noticed a few recurring patterns. One is&nbsp;<i>“building as gardening</i>” which seems like a healthier way to build software in our age.</p><p>This is how Cab of <a href="https://www.are.na/"><u>Are.na</u></a> described to me this approach:</p><blockquote><p>You can be like <i>“we’ve got to plant these here!”</i> or <i>“this is not working, let’s take it out!”</i>.</p><p>But you can also choose the other way which is more like <i>“oh wow, this thing is growing, that's really cool”</i> or <i>“let's see what else we can put here”</i>.</p><p>It's a different kind of perspective.</p></blockquote><p>And this, unrelatedly or not, how Herman of <a href="https://bearblog.dev/"><u>Bear Blog</u></a> put it:</p><blockquote><p>I’m trying to build a similar product with <a href="https://bearblog.dev/"><u>Bear Blog</u></a>. Something niche but valuable. Something I can spend time on because I want to.</p><p>Being able to talk to, and interact with the people using my tools is fulfilling. Spending time meticulously improving subtle aspects of the product is enjoyable.</p><p><a href="https://herman.bearblog.dev/my-product-is-my-garden/"><u>My product is my garden</u></a></p></blockquote><p>I’m no poetic person, so I admit this metaphor was a bit hard to digest at first. However, after pondering on this perspective for a while, I suddenly could make sense out of it.</p><p>For the longest time, I’ve seen people building software like in a greenhouse mode: carefully controlling and forcing conditions to make something grow. That often means building more “necessary” features, paying people to participate in “research” surveys, committing early to investors, or rushing to catch the “right moment” in the market.</p><p>The common thread of this <i>build-fast-break-things-pivot-repeat</i> culture is the sense of urgency. There’s a collective pressure to grow something faster than it naturally would. Instead of letting things grow more organically, we’re forced to grow them more artificially. Then everything feels urgent.</p><p>But that's just one way to view this world.</p><p>Then there's the gardening approach, which is a different process; a different mindset. Walking this path isn’t about the classic <i>solving problems</i> mantra through countless plans, strategies, and checklists to mark. Nor is it about taking the hammer and knocking a few components and forcing them to fit together.</p><p>Gardening is rather about tending something slowly until it (or not) reaches any height, while giving it the necessary time to mature and evolve. It’s not about forcing a solution, but rather about allowing growth to happen more organically.</p><p>Building software is much the same. It’s not a science-ba... </p>
Since I’ve been interviewing other builders for the niche design zine, I've noticed a few recurring patterns. One is “building as gardening” which seems like a healthier way to build software in our age. This is how Cab of Are.na described to me this approach: > You can be like “we’ve got to plant these here!” or “this is not working, let’s take it out!”. > > But you can also choose the other way which is more like “oh wow, this thing is growing, that's really cool” or “let's see what else we can put here”. > > It's a different kind of perspective. And this, unrelatedly or not, how Herman of Bear Blog put it: > I’m trying to build a similar product with Bear Blog. Something niche but valuable. Something I can spend time on because I want to. > > Being able to talk to, and interact with the people using my tools is fulfilling. Spending time meticulously improving subtle aspects of the product is enjoyable. > > My product is my garden I’m no poetic person, so I admit this metaphor was a bit hard to digest at first. However, after pondering on this perspective for a while, I suddenly could make sense out of it. For the longest time, I’ve seen people building software like in a greenhouse mode: carefully controlling and forcing conditions to make something grow. That often means building more “necessary” features, paying people to participate in “research” surveys, committing early to investors, or rushing to catch the “right moment” in the market. The common thread of this build-fast-break-things-pivot-repeat culture is the sense of urgency. There’s a collective pressure to grow something faster than it naturally would. Instead of letting things grow more organically, we’re forced to grow them more artificially. Then everything feels urgent. But that's just one way to view this world. Then there's the gardening approach, which is a different process; a different mindset. Walking this path isn’t about the classic solving problems mantra through countless pl
1,055
1.2.1
Revision
false
null
null
CrosspostOutput
WtJadTxL2Rsmiv45d
semiconductor-fabs-i-the-equipment
Semiconductor Fabs I: The Equipment
null
false
false
false
null
cZDc5r6jjrnRBTsBh
null
true
false
false
false
Post
https://nomagicpill.github.io/research/fabs.html
2025-06-04T22:09:32.678Z
null
false
false
2
2
2025-06-05T17:39:16.355Z
false
false
linkpost
[]
null
null
JciFBYPTxsmdcozLw
0
9
18
false
0.019016
null
false
false
2025-06-04T22:09:32.678Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
grecHJcgkb3KW5wnM
null
null
null
false
null
[]
null
8
0
2025-06-04T22:02:39.492Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
23
null
null
null
null
[ { "__typename": "Tag", "_id": "3uE2pXvbcnS9nnZRE", "adminOnly": false, "afBaseScore": 0, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 2, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:50.898Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 27, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "zJtgSyKntXrnkArbY", "displayName": "kistune" }, { "_id": "XqyQTGFGoKfZtdqN3", "displayName": "kINo" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "World Modeling", "needsReview": false, "noindex": false, "postCount": 5855, "score": 2, "shortName": null, "slug": "world-modeling", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 2, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
9
0
0
4
0
cZDc5r6jjrnRBTsBh
nomagicpill
2020-07-25T15:20:19.152Z
ethanmorse
nomagicpill
null
null
null
141
0
false
false
<p><a href="http://nomagicpill.github.io">nomagicpill.github.io</a></p>
null
null
29
7
0
0
0
1
0
r38pkCm7wF4M44MDQ
User
null
null
null
[ "canModeratePersonal" ]
null
null
WtJadTxL2Rsmiv45d
https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/WtJadTxL2Rsmiv45d/b43btq2fe29j4i9v1on1
SocialPreviewType
JciFBYPTxsmdcozLw
<p>An in-depth look at semiconductor processing equipment: considerations, selection, operation, and supporting info.</p><h2>Preface</h2><p>I tried to include as many links as possible to allow the reader to go down rabbit holes as they see fit.</p><p>I also tried to keep info basic enough for the skimmers/laypeople to enjoy while still adding more technical details located at the end of each section in the collapsible element. Pictures are sprinkled throughout to put an image with the words.</p><p>Interactive animations are denoted by the large "Click here to [X]!". While rudimentary, they get the job done. These were vibe coded using Claude 4 Opus with extended thinking and a few iterations.</p><p>This information is current as of May 2025.</p><hr><h2>Context</h2><p>Before reading this, you should read Construction Physics' (Brian Potter) <a href="https://www.construction-physics.com/p/how-to-build-a-20-billion-semiconductor">How to Build a $20 Billion Semiconductor Fab</a>. It's an excellent introduction into both the semiconductor manufacturing process and the physical building that houses the manufacturing operations. However, Brian (understandably) chooses not to expand on a critical component: the equipment used to make the semiconductor chips.</p><p>As Brian writes:</p><blockquote><p>Early integrated circuits could be made with just ... dozens of process steps, but a modern leading-edge microchip might require ... thousands of separate process steps.</p></blockquote><p>These individual process steps are performed by semiconductor equipment, also called tools or machines. In the same way that a traditional machine shop has a mill, lathe, and bandsaw to perform a variety of operations on a single piece of metal to create a final product, a fab has deposition tools, etch tools, and lithography tools (among many, many others) to perform a variety of operations on a single silicon wafer to create the final chip.</p><p>There are 100s to 1000s of tools in a single fab, making it important to use the available space in an efficient manner. Some tools are quite boxy to help minimize the area used, or the footprint. Some tools are less boxy due to sheer requirements, i.e., the original equipment manufacturer (OEM) could only make it so small area-wise before it lost some functionality or performance. OEMs are incentivized to minimize footprint so companies will buy more of their tools. Large-scale OEMs include Applied Materials (AMAT), Tokyo Electron Limited (TEL), and LAM Research.</p><p>Here's a picture of an AMAT tool, the Endura. Note that while not perfectly box-l... </p>
An in-depth look at semiconductor processing equipment: considerations, selection, operation, and supporting info. Preface I tried to include as many links as possible to allow the reader to go down rabbit holes as they see fit. I also tried to keep info basic enough for the skimmers/laypeople to enjoy while still adding more technical details located at the end of each section in the collapsible element. Pictures are sprinkled throughout to put an image with the words. Interactive animations are denoted by the large "Click here to [X]!". While rudimentary, they get the job done. These were vibe coded using Claude 4 Opus with extended thinking and a few iterations. This information is current as of May 2025. ---------------------------------------- Context Before reading this, you should read Construction Physics' (Brian Potter) How to Build a $20 Billion Semiconductor Fab. It's an excellent introduction into both the semiconductor manufacturing process and the physical building that houses the manufacturing operations. However, Brian (understandably) chooses not to expand on a critical component: the equipment used to make the semiconductor chips. As Brian writes: > Early integrated circuits could be made with just ... dozens of process steps, but a modern leading-edge microchip might require ... thousands of separate process steps. These individual process steps are performed by semiconductor equipment, also called tools or machines. In the same way that a traditional machine shop has a mill, lathe, and bandsaw to perform a variety of operations on a single piece of metal to create a final product, a fab has deposition tools, etch tools, and lithography tools (among many, many others) to perform a variety of operations on a single silicon wafer to create the final chip. There are 100s to 1000s of tools in a single fab, making it important to use the available space in an efficient manner. Some tools are quite boxy to help minimize the area used, or the
5,818
1.2.1
Revision
false
null
null
CrosspostOutput
cyKD2ifYizpupGP4C
the-stereotype-of-the-stereotype
The Stereotype of the Stereotype
null
false
false
false
null
rn8BfWfkzT9aemyyq
null
true
false
false
false
Post
null
2025-06-04T21:06:54.165Z
null
false
false
2
2
2025-06-05T17:38:55.909Z
false
false
post
[]
null
null
Qz8Y7Aig4u5dhYtzS
17
30
56
false
0.045093
null
false
false
2025-06-08T20:14:21.416Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
grecHJcgkb3KW5wnM
null
null
null
false
null
[]
null
12
0
2025-06-04T21:04:29.935Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
10
null
null
null
null
[ { "__typename": "Tag", "_id": "Ng8Gice9KNkncxqcj", "adminOnly": false, "afBaseScore": 0, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 1, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:17.072Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 100, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "iMqytjy9ns89Fzfyv", "displayName": "miakko" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Rationality", "needsReview": false, "noindex": false, "postCount": 4302, "score": 1, "shortName": null, "slug": "rationality", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 1, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
30
0
0
12
0
rn8BfWfkzT9aemyyq
ike-1
2025-05-20T18:04:32.507Z
Ike
Ike
null
null
null
55
0
false
false
null
null
1
2
0
0
0
1
0
grecHJcgkb3KW5wnM
User
null
null
null
[ "canModeratePersonal" ]
null
null
cyKD2ifYizpupGP4C
SocialPreviewType
Qz8Y7Aig4u5dhYtzS
<p>I want to define a new term that will be useful in the discourse.</p><p>Imagine you landed in a foreign country with dogs, but with no word for dogs. You would see these things, and want to talk about them. You would like to be able to tell people whether they can bring dogs into your house, but you can’t, because there isn’t any word for “dog.” You’d like to ask if you can pet someone’s dog, but you can’t do that either, because when you ask “can I pet your dog?” they say “pet my what?” It would get pretty frustrating.</p><p>This is how I feel about a certain phenomenon I’ve noticed over the last year or so. I notice it all the time—It’s as ubiquitous in society as dogs are, and I see it just about as often—but there isn’t a word for it. People seem to make the same error of judgement over and over, and I want to talk about this error. There’s a lot I want to say about it. I want to be able to point out when someone is making the mistake. I want to ask people about whether&nbsp;<i>I’m&nbsp;</i>making the mistake. I want to talk about how some people seem to see very clearly, and make the mistake very rarely. Others seem to be very cloudy-eyed where this mistake is concerned, and they make it very often.</p><p>So, I’m going to make up a new term. The term is “The Stereotype of the Stereotype.” The mistake I’m talking about is not “The Stereotype of the Stereotype.” The Stereotype of the Stereotype is a noun. This mistake is a verb. The verb is “confusing the stereotype with The Stereotype of the Stereotype.”</p><p>Mixing up ‘the stereotype’ with ‘The Stereotype of the Stereotype’ is as significant a mistake as seeing the word “rainstorm,” and opening your umbrella. You aren’t allowed to mix up levels that way. The stereotype is the stereotype. The Stereotype of the The Stereotype is The Stereotype of the The Stereotype. And they are as different as a rainstorm, and the&nbsp;<i>word&nbsp;</i>“Rainstorm.”</p><p>Let’s use some examples to define this term “The Stereotype of the Stereotype.”</p><p>&nbsp;</p><p>-</p><p>&nbsp;</p><p>Sir Terrence Pratchett was a British satirical fantasy author, most famous for his Discworld series of books.</p><p>There are 41 of these books. Most of them follow distinct arrangements of characters. For example, eight of them follow a man named Sam Vimes. Another five follow a girl named Tiffany Aching. However, there’s one character who shows up in nearly every book. His name is Death.</p><p>Death appears in 39 of the 41... </p>
I want to define a new term that will be useful in the discourse. Imagine you landed in a foreign country with dogs, but with no word for dogs. You would see these things, and want to talk about them. You would like to be able to tell people whether they can bring dogs into your house, but you can’t, because there isn’t any word for “dog.” You’d like to ask if you can pet someone’s dog, but you can’t do that either, because when you ask “can I pet your dog?” they say “pet my what?” It would get pretty frustrating. This is how I feel about a certain phenomenon I’ve noticed over the last year or so. I notice it all the time—It’s as ubiquitous in society as dogs are, and I see it just about as often—but there isn’t a word for it. People seem to make the same error of judgement over and over, and I want to talk about this error. There’s a lot I want to say about it. I want to be able to point out when someone is making the mistake. I want to ask people about whether I’m making the mistake. I want to talk about how some people seem to see very clearly, and make the mistake very rarely. Others seem to be very cloudy-eyed where this mistake is concerned, and they make it very often. So, I’m going to make up a new term. The term is “The Stereotype of the Stereotype.” The mistake I’m talking about is not “The Stereotype of the Stereotype.” The Stereotype of the Stereotype is a noun. This mistake is a verb. The verb is “confusing the stereotype with The Stereotype of the Stereotype.” Mixing up ‘the stereotype’ with ‘The Stereotype of the Stereotype’ is as significant a mistake as seeing the word “rainstorm,” and opening your umbrella. You aren’t allowed to mix up levels that way. The stereotype is the stereotype. The Stereotype of the The Stereotype is The Stereotype of the The Stereotype. And they are as different as a rainstorm, and the word “Rainstorm.” Let’s use some examples to define this term “The Stereotype of the Stereotype.”   -   Sir Terrence Pratchett was
2,606
1.3.0
Revision
false
null
null
CrosspostOutput
dEJJqcAt6DuoFxAeT
2-why-intuitive-comparisons-of-large-scale-impact-are-1
2. Why intuitive comparisons of large-scale impact are unjustified
null
false
false
false
null
rv7RzMiG3esRT4CQi
null
true
false
false
false
Post
null
2025-06-04T20:30:04.797Z
null
false
false
2
2
2025-06-05T17:37:06.994Z
false
false
post
[]
null
null
sHsYjxGkazJecj7q9
0
6
25
false
0.023804
null
false
false
2025-06-04T20:30:04.797Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
null
null
grecHJcgkb3KW5wnM
null
null
null
false
null
[]
null
8
0
2025-06-04T20:30:04.798Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
20
null
null
null
null
[ { "__typename": "Tag", "_id": "X8JsWEnBRPvs5Y99i", "adminOnly": false, "afBaseScore": 0, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 0, "canEditUserIds": null, "core": false, "createdAt": "2015-12-03T07:35:06.000Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [] }, "isArbitalImport": true, "isPlaceholderPage": false, "isSubforum": false, "name": "Decision theory", "needsReview": false, "noindex": false, "postCount": 500, "score": 0, "shortName": null, "slug": "decision-theory", "suggestedAsFilter": false, "userId": "nmk3nLpQE89dMRzzN", "voteCount": 0, "wikiOnly": false }, { "__typename": "Tag", "_id": "EdRnMXBRbY5JDf5df", "adminOnly": false, "afBaseScore": 6, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "nmk3nLpQE89dMRzzN", "displayName": "Eliezer Yudkowsky" } ] }, "baseScore": 13, "canEditUserIds": null, "core": false, "createdAt": "2015-07-02T01:53:10.000Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "nmk3nLpQE89dMRzzN", "displayName": "Eliezer Yudkowsky" } ] }, "isArbitalImport": true, "isPlaceholderPage": false, "isSubforum": false, "name": "Epistemology", "needsReview": false, "noindex": false, "postCount": 424, "score": 13, "shortName": null, "slug": "epistemology", "suggestedAsFilter": false, "userId": "nmk3nLpQE89dMRzzN", "voteCount": 1, "wikiOnly": false }, { "__typename": "Tag", "_id": "Ng8Gice9KNkncxqcj", "adminOnly": false, "afBaseScore": 0, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 1, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:17.072Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 100, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "iMqytjy9ns89Fzfyv", "displayName": "miakko" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Rationality", "needsReview": false, "noindex": false, "postCount": 4302, "score": 1, "shortName": null, "slug": "rationality", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 1, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
6
0
0
4
0
rv7RzMiG3esRT4CQi
anthony-digiovanni
2019-12-15T12:43:56.701Z
antimonyanthony
Anthony DiGiovanni
null
null
Anthony DiGiovanni
1,033
58
false
false
<p>Researcher at the Center on Long-Term Risk. All opinions my own.</p>
null
null
10
142
1
1
1
1
0
gXeEWGjTWyqgrQTzR
User
null
null
null
[ "canModeratePersonal", "alignmentVoters" ]
null
null
dEJJqcAt6DuoFxAeT
https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/dEJJqcAt6DuoFxAeT/cme20ufpzh7ft99acctm
SocialPreviewType
sHsYjxGkazJecj7q9
<p>We’ve seen <a href="https://forum.effectivealtruism.org/posts/a3hnfA9EnYm9bssTZ/1-the-challenge-of-unawareness-for-impartial-altruist-action-1">so far</a>&nbsp;that unawareness leaves us with gaps in the impartial altruistic justification for our decisions. It’s fundamentally unclear how the possible outcomes of our actions trade off against each other.</p><p>But perhaps this ambiguity doesn’t matter much in practice, since the consequences that predominantly shape the impartial value of the future (<i>“large-scale consequences”</i>, for short) seem at least somewhat foreseeable. Here are two ways we might think we can fill these gaps:</p><ol><li><p data-internal-id="ftnt_ref1"><i>Implicit:</i>&nbsp;“Even if we can’t assign EVs to interventions, we have reason to trust that our <strong>pre-theoretic intuitions</strong>&nbsp;at least track an intervention’s sign. These intuitions can distinguish whether the large-scale consequences are net-better than inaction.”<span class="footnote-reference" data-footnote-reference="" data-footnote-index="1" data-footnote-id="hlpvuxlj9e" role="doc-noteref" id="fnrefhlpvuxlj9e"><sup><a href="#fnhlpvuxlj9e">[1]</a></sup></span></p></li><li><i>Explicit:</i><strong>&nbsp;</strong>“True, we don’t directly conceive of fine-grained possible worlds, only coarse hypotheses. We can still assign precise EVs, though, by intuitively estimating the average value of the possible worlds in each hypothesis. These <strong>best-guess estimates</strong>&nbsp;will at least be better than chance.”</li></ol><p>Do these approaches hold up, when we look concretely at how vast our unawareness could be? “The far future is complex” is a truism that many EAs are probably sick of hearing. But here I’ll argue that this complexity — which also plagues our impact on near-term pivotal events like AI takeoff — specifically undermines both <i>Implicit </i>and <i>Explicit</i>. For instance, we can’t simply say, “It’s intuitively clear that trying to spread altruistic values would be better than doing nothing”, nor, “<i>If</i>&nbsp;it’s better for future agents to be altruistic, we should try to spread altruism, since intuitively this is at least somewhat more likely to succeed than backfire”. This is because <strong>our intuitions about large-scale consequences don’t seem to</strong><i><strong>&nbsp;</strong></i><strong>account for unawareness with enough precision. </strong>If that’s true, we’ll need a new framework for impartially evaluating our actions under unawareness.</p><p>The argument in brief: First, a strategy’s sign from an impartial perspective seems unusually sensitive<i>&nbsp;</i>to factors we’re unaware of, compared to the local perspective we take in more familiar decision problems (cf. Roussos (2021)). Thus it’s unsurprising if our intuitions can justify claims like “<span class="math-tex"><span class="mjpage"><span class="mjx-chtml"><span class="mjx-math" aria-label="A"><span class="mjx-mrow" aria-hidden="true"><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.519em; padding-bottom: 0.298em;">A</span></span></span></span></span></span></span>&nbsp;is net-positive” on the local scale, yet don’t justify such claims on the cosmic scale. Second, these intuitions seem so <i>imprecisely calibr</i>... <style>.mjx-chtml {display: inline-block; line-height: 0; text-indent: 0; text-align: left; text-transform: none; font-style: normal; font-weight: normal; font-size: 100%; font-size-adjust: none; letter-spacing: normal; word-wrap: normal; word-spacing: normal; white-space: nowrap; float: none; direction: ltr; max-width: none; max-height: none; min-width: 0; min-height: 0; border: 0; margin: 0; padding: 1px 0} .MJXc-display {display: block; text-align: center; margin: 1em 0; padding: 0} .mjx-chtml[tabindex]:focus, body :focus .mjx-chtml[tabindex] {display: inline-table} .mjx-full-width {text-align: center; display: table-cell!important; width: 10000em} .mjx-math {display: inline-block; border-collapse: separate; border-spacing: 0} .mjx-math * {display: inline-block; -webkit-box-sizing: content-box!important; -moz-box-sizing: content-box!important; box-sizing: content-box!important; text-align: left} .mjx-numerator {display: block; text-align: center} .mjx-denominator {display: block; text-align: center} .MJXc-stacked {height: 0; position: relative} .MJXc-stacked > * {position: absolute} .MJXc-bevelled > * {display: inline-block} .mjx-stack {display: inline-block} .mjx-op {display: block} .mjx-under {display: table-cell} .mjx-over {display: block} .mjx-over > * {padding-left: 0px!important; padding-right: 0px!important} .mjx-under > * {padding-left: 0px!important; padding-right: 0px!important} .mjx-stack > .mjx-sup {display: block} .mjx-stack > .mjx-sub {display: block} .mjx-prestack > .mjx-presup {display: block} .mjx-prestack > .mjx-presub {display: block} .mjx-delim-h > .mjx-char {display: inline-block} .mjx-surd {vertical-align: top} .mjx-surd + .mjx-box {display: inline-flex} .mjx-mphantom * {visibility: hidden} .mjx-merror {background-color: #FFFF88; color: #CC0000; border: 1px solid #CC0000; padding: 2px 3px; font-style: normal; font-size: 90%} .mjx-annotation-xml {line-height: normal} .mjx-menclose > svg {fill: none; stroke: currentColor; overflow: visible} .mjx-mtr {display: table-row} .mjx-mlabeledtr {display: table-row} .mjx-mtd {display: table-cell; text-align: center} .mjx-label {display: table-row} .mjx-box {display: inline-block} .mjx-block {display: block} .mjx-span {display: inline} .mjx-char {display: block; white-space: pre} .mjx-itable {display: inline-table; width: auto} .mjx-row {display: table-row} .mjx-cell {display: table-cell} .mjx-table {display: table; width: 100%} .mjx-line {display: block; height: 0} .mjx-strut {width: 0; padding-top: 1em} .mjx-vsize {width: 0} .MJXc-space1 {margin-left: .167em} .MJXc-space2 {margin-left: .222em} .MJXc-space3 {margin-left: .278em} .mjx-test.mjx-test-display {display: table!important} .mjx-test.mjx-test-inline {display: inline!important; margin-right: -1px} .mjx-test.mjx-test-default {display: block!important; clear: both} .mjx-ex-box {display: inline-block!important; position: absolute; overflow: hidden; min-height: 0; max-height: none; padding: 0; border: 0; margin: 0; width: 1px; height: 60ex} .mjx-test-inline .mjx-left-box {display: inline-block; width: 0; float: left} .mjx-test-inline .mjx-right-box {display: inline-block; width: 0; float: right} .mjx-test-display .mjx-right-box {display: table-cell!important; width: 10000em!important; min-width: 0; max-width: none; padding: 0; border: 0; margin: 0} .MJXc-TeX-unknown-R {font-family: monospace; font-style: normal; font-weight: normal} .MJXc-TeX-unknown-I {font-family: monospace; font-style: italic; font-weight: normal} .MJXc-TeX-unknown-B {font-family: monospace; font-style: normal; font-weight: bold} .MJXc-TeX-unknown-BI {font-family: monospace; font-style: italic; font-weight: bold} .MJXc-TeX-ams-R {font-family: MJXc-TeX-ams-R,MJXc-TeX-ams-Rw} .MJXc-TeX-cal-B {font-family: MJXc-TeX-cal-B,MJXc-TeX-cal-Bx,MJXc-TeX-cal-Bw} .MJXc-TeX-frak-R {font-family: MJXc-TeX-frak-R,MJXc-TeX-frak-Rw} .MJXc-TeX-frak-B {font-family: MJXc-TeX-frak-B,MJXc-TeX-frak-Bx,MJXc-TeX-frak-Bw} .MJXc-TeX-math-BI {font-family: MJXc-TeX-math-BI,MJXc-TeX-math-BIx,MJXc-TeX-math-BIw} .MJXc-TeX-sans-R {font-family: MJXc-TeX-sans-R,MJXc-TeX-sans-Rw} .MJXc-TeX-sans-B {font-family: MJXc-TeX-sans-B,MJXc-TeX-sans-Bx,MJXc-TeX-sans-Bw} .MJXc-TeX-sans-I {font-family: MJXc-TeX-sans-I,MJXc-TeX-sans-Ix,MJXc-TeX-sans-Iw} .MJXc-TeX-script-R {font-family: MJXc-TeX-script-R,MJXc-TeX-script-Rw} .MJXc-TeX-type-R {font-family: MJXc-TeX-type-R,MJXc-TeX-type-Rw} .MJXc-TeX-cal-R {font-family: MJXc-TeX-cal-R,MJXc-TeX-cal-Rw} .MJXc-TeX-main-B {font-family: MJXc-TeX-main-B,MJXc-TeX-main-Bx,MJXc-TeX-main-Bw} .MJXc-TeX-main-I {font-family: MJXc-TeX-main-I,MJXc-TeX-main-Ix,MJXc-TeX-main-Iw} .MJXc-TeX-main-R {font-family: MJXc-TeX-main-R,MJXc-TeX-main-Rw} .MJXc-TeX-math-I {font-family: MJXc-TeX-math-I,MJXc-TeX-math-Ix,MJXc-TeX-math-Iw} .MJXc-TeX-size1-R {font-family: MJXc-TeX-size1-R,MJXc-TeX-size1-Rw} .MJXc-TeX-size2-R {font-family: MJXc-TeX-size2-R,MJXc-TeX-size2-Rw} .MJXc-TeX-size3-R {font-family: MJXc-TeX-size3-R,MJXc-TeX-size3-Rw} .MJXc-TeX-size4-R {font-family: MJXc-TeX-size4-R,MJXc-TeX-size4-Rw} .MJXc-TeX-vec-R {font-family: MJXc-TeX-vec-R,MJXc-TeX-vec-Rw} .MJXc-TeX-vec-B {font-family: MJXc-TeX-vec-B,MJXc-TeX-vec-Bx,MJXc-TeX-vec-Bw} @font-face {font-family: MJXc-TeX-ams-R; src: local('MathJax_AMS'), local('MathJax_AMS-Regular')} @font-face {font-family: MJXc-TeX-ams-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_AMS-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_AMS-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_AMS-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-cal-B; src: local('MathJax_Caligraphic Bold'), local('MathJax_Caligraphic-Bold')} @font-face {font-family: MJXc-TeX-cal-Bx; src: local('MathJax_Caligraphic'); font-weight: bold} @font-face {font-family: MJXc-TeX-cal-Bw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Caligraphic-Bold.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Caligraphic-Bold.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Caligraphic-Bold.otf') format('opentype')} @font-face {font-family: MJXc-TeX-frak-R; src: local('MathJax_Fraktur'), local('MathJax_Fraktur-Regular')} @font-face {font-family: MJXc-TeX-frak-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Fraktur-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Fraktur-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Fraktur-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-frak-B; src: local('MathJax_Fraktur Bold'), local('MathJax_Fraktur-Bold')} @font-face {font-family: MJXc-TeX-frak-Bx; src: local('MathJax_Fraktur'); font-weight: bold} @font-face {font-family: MJXc-TeX-frak-Bw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Fraktur-Bold.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Fraktur-Bold.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Fraktur-Bold.otf') format('opentype')} @font-face {font-family: MJXc-TeX-math-BI; src: local('MathJax_Math BoldItalic'), local('MathJax_Math-BoldItalic')} @font-face {font-family: MJXc-TeX-math-BIx; src: local('MathJax_Math'); font-weight: bold; font-style: italic} @font-face {font-family: MJXc-TeX-math-BIw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Math-BoldItalic.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Math-BoldItalic.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Math-BoldItalic.otf') format('opentype')} @font-face {font-family: MJXc-TeX-sans-R; src: local('MathJax_SansSerif'), local('MathJax_SansSerif-Regular')} @font-face {font-family: MJXc-TeX-sans-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_SansSerif-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_SansSerif-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_SansSerif-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-sans-B; src: local('MathJax_SansSerif Bold'), local('MathJax_SansSerif-Bold')} @font-face {font-family: MJXc-TeX-sans-Bx; src: local('MathJax_SansSerif'); font-weight: bold} @font-face {font-family: MJXc-TeX-sans-Bw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_SansSerif-Bold.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_SansSerif-Bold.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_SansSerif-Bold.otf') format('opentype')} @font-face {font-family: MJXc-TeX-sans-I; src: local('MathJax_SansSerif Italic'), local('MathJax_SansSerif-Italic')} @font-face {font-family: MJXc-TeX-sans-Ix; src: local('MathJax_SansSerif'); font-style: italic} @font-face {font-family: MJXc-TeX-sans-Iw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_SansSerif-Italic.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_SansSerif-Italic.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_SansSerif-Italic.otf') format('opentype')} @font-face {font-family: MJXc-TeX-script-R; src: local('MathJax_Script'), local('MathJax_Script-Regular')} @font-face {font-family: MJXc-TeX-script-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Script-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Script-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Script-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-type-R; src: local('MathJax_Typewriter'), local('MathJax_Typewriter-Regular')} @font-face {font-family: MJXc-TeX-type-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Typewriter-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Typewriter-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Typewriter-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-cal-R; src: local('MathJax_Caligraphic'), local('MathJax_Caligraphic-Regular')} @font-face {font-family: MJXc-TeX-cal-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Caligraphic-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Caligraphic-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Caligraphic-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-main-B; src: local('MathJax_Main Bold'), local('MathJax_Main-Bold')} @font-face {font-family: MJXc-TeX-main-Bx; src: local('MathJax_Main'); font-weight: bold} @font-face {font-family: MJXc-TeX-main-Bw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Main-Bold.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Main-Bold.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Main-Bold.otf') format('opentype')} @font-face {font-family: MJXc-TeX-main-I; src: local('MathJax_Main Italic'), local('MathJax_Main-Italic')} @font-face {font-family: MJXc-TeX-main-Ix; src: local('MathJax_Main'); font-style: italic} @font-face {font-family: MJXc-TeX-main-Iw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Main-Italic.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Main-Italic.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Main-Italic.otf') format('opentype')} @font-face {font-family: MJXc-TeX-main-R; src: local('MathJax_Main'), local('MathJax_Main-Regular')} @font-face {font-family: MJXc-TeX-main-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Main-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Main-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Main-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-math-I; src: local('MathJax_Math Italic'), local('MathJax_Math-Italic')} @font-face {font-family: MJXc-TeX-math-Ix; src: local('MathJax_Math'); font-style: italic} @font-face {font-family: MJXc-TeX-math-Iw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Math-Italic.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Math-Italic.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Math-Italic.otf') format('opentype')} @font-face {font-family: MJXc-TeX-size1-R; src: local('MathJax_Size1'), local('MathJax_Size1-Regular')} @font-face {font-family: MJXc-TeX-size1-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Size1-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Size1-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Size1-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-size2-R; src: local('MathJax_Size2'), local('MathJax_Size2-Regular')} @font-face {font-family: MJXc-TeX-size2-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Size2-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Size2-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Size2-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-size3-R; src: local('MathJax_Size3'), local('MathJax_Size3-Regular')} @font-face {font-family: MJXc-TeX-size3-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Size3-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Size3-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Size3-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-size4-R; src: local('MathJax_Size4'), local('MathJax_Size4-Regular')} @font-face {font-family: MJXc-TeX-size4-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Size4-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Size4-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Size4-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-vec-R; src: local('MathJax_Vector'), local('MathJax_Vector-Regular')} @font-face {font-family: MJXc-TeX-vec-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Vector-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Vector-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Vector-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-vec-B; src: local('MathJax_Vector Bold'), local('MathJax_Vector-Bold')} @font-face {font-family: MJXc-TeX-vec-Bx; src: local('MathJax_Vector'); font-weight: bold} @font-face {font-family: MJXc-TeX-vec-Bw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Vector-Bold.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Vector-Bold.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Vector-Bold.otf') format('opentype')} </style></p>
We’ve seen so far that unawareness leaves us with gaps in the impartial altruistic justification for our decisions. It’s fundamentally unclear how the possible outcomes of our actions trade off against each other. But perhaps this ambiguity doesn’t matter much in practice, since the consequences that predominantly shape the impartial value of the future (“large-scale consequences”, for short) seem at least somewhat foreseeable. Here are two ways we might think we can fill these gaps: 1. Implicit: “Even if we can’t assign EVs to interventions, we have reason to trust that our pre-theoretic intuitions at least track an intervention’s sign. These intuitions can distinguish whether the large-scale consequences are net-better than inaction.”[1] 2. Explicit: “True, we don’t directly conceive of fine-grained possible worlds, only coarse hypotheses. We can still assign precise EVs, though, by intuitively estimating the average value of the possible worlds in each hypothesis. These best-guess estimates will at least be better than chance.” Do these approaches hold up, when we look concretely at how vast our unawareness could be? “The far future is complex” is a truism that many EAs are probably sick of hearing. But here I’ll argue that this complexity — which also plagues our impact on near-term pivotal events like AI takeoff — specifically undermines both Implicit and Explicit. For instance, we can’t simply say, “It’s intuitively clear that trying to spread altruistic values would be better than doing nothing”, nor, “If it’s better for future agents to be altruistic, we should try to spread altruism, since intuitively this is at least somewhat more likely to succeed than backfire”. This is because our intuitions about large-scale consequences don’t seem to account for unawareness with enough precision. If that’s true, we’ll need a new framework for impartially evaluating our actions under unawareness. The argument in brief: First, a strategy’s sign from an impartial
4,920
1.3.0
Revision
true
false
qZS8cgvY5YrjQ3JiR
CrosspostOutput
fnBGz2dTJJDExofTj
dating-roundup-6
Dating Roundup #6
null
false
false
false
null
N9zj5qpTfqmbn9dro
null
true
false
false
false
Post
null
2025-06-04T20:00:06.688Z
null
false
false
2
2
null
false
false
post
[]
null
null
fCrBDorefHzHW36hu
2
16
34
false
0.023178
null
false
false
2025-06-06T18:31:06.631Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
null
null
grecHJcgkb3KW5wnM
null
null
null
false
null
[]
null
9
0
2025-06-04T20:00:06.688Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
66
null
null
null
null
[ { "__typename": "Tag", "_id": "mip7tdAN87Jarkcew", "adminOnly": false, "afBaseScore": 3, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 9, "canEditUserIds": null, "core": false, "createdAt": "2020-05-10T06:00:13.257Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Relationships (Interpersonal)", "needsReview": false, "noindex": false, "postCount": 213, "score": 9, "shortName": null, "slug": "relationships-interpersonal", "suggestedAsFilter": false, "userId": "iBcH2a3HdWGS2JEZA", "voteCount": 1, "wikiOnly": false }, { "__typename": "Tag", "_id": "fkABsGCJZ6y9qConW", "adminOnly": false, "afBaseScore": 0, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 2, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T06:06:46.947Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "MiuAZvbQcQ7ethgt3", "displayName": "Viktor withaK" }, { "_id": "dRaCtsAWxk7sgirSY", "displayName": "Jordan Morgan" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Practical", "needsReview": false, "noindex": false, "postCount": 3411, "score": 2, "shortName": null, "slug": "practical", "suggestedAsFilter": true, "userId": "oBSWiHjgproTiThmY", "voteCount": 2, "wikiOnly": false } ]
QSR8rPZxZzxEXoPjR
0
0
null
false
null
null
0
16
0
0
5
0
N9zj5qpTfqmbn9dro
zvi
2009-03-31T20:54:54.077Z
Zvi
Zvi
null
null
null
51,554
146
false
false
null
null
936
1,461
3
2
7
1
0
qgdGA4ZEyW7zNdK84
User
norm-enforcing
null
true
[ "trustLevel1", "canModeratePersonal", "alignmentVoters", "alignmentForum" ]
null
null
fnBGz2dTJJDExofTj
https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/fnBGz2dTJJDExofTj/rxk6jzozibm3mt7vduuz
SocialPreviewType
fCrBDorefHzHW36hu
<p>Previously: <a href="https://thezvi.substack.com/p/dating-roundup-1-this-is-why-youre?utm_source=publication-search">#1</a>, <a href="https://thezvi.substack.com/p/dating-roundup-2-if-at-first-you?utm_source=publication-search">#2</a>, <a href="https://thezvi.substack.com/p/dating-roundup-3-third-times-the?utm_source=publication-search">#3</a>, #4, #5</p><p>Dating Roundup #4 covered dating apps. Roundup #5 covered opening without them.</p><p>Dating Roundup #6 covers everything else.</p> <h4>Table of Contents</h4> <ol> <li><a href="https://thezvi.substack.com/i/160142469/you-re-single-because-you-can-t-handle-basic-logistics">You’re Single Because You Can’t Handle Basic Logistics.</a></li> <li><a href="https://thezvi.substack.com/i/160142469/you-re-single-because-you-don-t-ask-questions">You’re Single Because You Don’t Ask Questions.</a></li> <li><a href="https://thezvi.substack.com/i/160142469/you-re-single-because-of-your-terrible-dating-tactics">You’re Single Because of Your Terrible Dating Tactics.</a></li> <li><a href="https://thezvi.substack.com/i/160142469/you-re-single-because-you-refuse-to-play-your-role">You’re Single Because You Refuse to Play Your Role.</a></li> <li><a href="https://thezvi.substack.com/i/160142469/you-re-single-because-people-are-crazy-about-age-gaps">You’re Single Because People Are Crazy About Age Gaps.</a></li> <li><a href="https://thezvi.substack.com/i/160142469/you-re-single-and-you-need-professional-help">You’re Single and You Need Professional Help.</a></li> <li><a href="https://thezvi.substack.com/i/160142469/you-re-single-because-you-never-close">You’re Single Because You Never Close.</a></li> <li><a href="https://thezvi.substack.com/i/160142469/you-re-single-because-you-re-bad-at-sex-and-everyone-knows">You’re Single Because You’re Bad at Sex And Everyone Knows.</a></li> <li><a href="https://thezvi.substack.com/i/160142469/you-re-single-because-you-are-only-a-fan">You’re Single Because You Are Only a Fan.</a></li> <li><a href="https://thezvi.substack.com/i/160142469/you-re-single-because-of-preference-falsification">You’re Single Because of Preference Falsification.</a> <div> <span id="more-24494"></span> </div> </li> <li><a href="https://thezvi.substack.com/i/160142469/you-re-single-because-you-have-insufficient-visual-aids">You’re Single Because You Have Insufficient Visual Aids.</a></li> <li><a href="https://thezvi.substack.com/i/160142469/you-re-single-because-you-told-your-partner-you-didn-t-want-them">You’re Single Because You Told Your Partner You Didn’t Want Them.</a></li> <li><a href="https://thezvi.substack.com/i/160142469/you-re-single-because-of-your-terrible-dating-strategy">You’re Single Because of Your Terrible Dating Strategy.</a></li> <li><a href="https://thezvi.substack.com/i/160142469/you-re-single-because-you-don-t-enjoy-the-process">You’re Single Because You Don’t Enjoy the Process.</a></li> <li><a href="https://thezvi.substack.com/i/160142469/you-re-single-because-you-don-t-escalate-quickly">You’re Single Because You Don’t Escalate Quickly.</a></li> <li><a href="https://thezvi.substack.com/i/160142469/you-re-single-because-your-standards-are-too-high">You’re Single Because Your Standards Are Too High.</a></li> <li><a href="https://thezvi.substack.com/i/160142469/you-re-single-because-you-read-the-wrong-books">You’re Single Because You Read the Wrong Books.</a></li> <li><a href="https://thezvi.substack.com/i/160142469/you-re-single-because-you-re-short-sorry-that-s-all-there-is-to-it">You’re Single Because You’re Short, Sorry, That’s All There Is To It.</a></li> <li><a href="https://thezvi.substack.com/i/160142469/you-re-single-because-of-bad-government-incentives">You’re Single Because of Bad Government Incentives.</a></li> <li><a href="https://thezvi.substack.com/i/160142469/you-re-single-because-you-don-t-realize-cheating-is-wrong">You’re Single Because You Don’t Realize Cheating is Wrong.</a></li> <li><a href="https://thezvi.substack.com/i/160142469/you-re-single-because-you-re-doing-polyamory-wrong">You’re Single Because You’re Doing Polyamory Wrong.</a></li> <li><a href="https://thezvi.substack.com/i/160142469/you-re-single-because-you-don-t-beware-cheaters">You’re Single Because You Don’t Beware Cheaters.</a></li> <li><a href="https://thezvi.substack.com/i/160142469/you-re-single-because-your-ex-spilled-the-tea">You’re Single Because Your Ex Spilled the Tea.</a></li> <li><a href="https://thezvi.substack.com/i/160142469/you-re-single-because-you-re-assigning-people-numbers">You’re Single Because You’re Assigning People Numbers.</a></li> <li><a href="https://thezvi.substack.com/i/160142469/you-re-single-because-you-are-the-wrong-amount-of-kinky">You’re Single Because You Are The Wrong Amount of Kinky.</a></li> <li><a href="https://thezvi.substack.com/i/160142469/you-re-single-because-you-re-not-good-enough-at-sex">You’re Single Because You’re Not Good Enough at Sex.</a></li> <li><a href="https://thezvi.substack.com/i/160142469/you-re-single-but-not-because-of-your-bodycount">You’re Single But Not Because of Your Bodycount.</a></li> <li><a href="https://thezvi.substack.com/i/160142469/you-re-single-because-they-divorced-you">You’re Single Because They Divorced You.</a></li> <li><a href="https://thezvi.substack.com/i/160142469/you-re-single-because-no-one-tells-you-anything">You’re Single Because No One Tells You Anything.</a></li> <li><a href="https://thezvi.substack.com/i/160142469/you-re-single-and-you-re-not-alone">You’re Single And You’re Not Alone.</a></li> <li><a href="https://thezvi.substack.com/i/160142469/you-re-single-because-things-are-steadily-getting-worse">You’re Single Because Things Are Steadily Getting Worse.</a></li> <li><a href="https://thezvi.substack.com/i/160142469/you-re-single-because-you-didn-t-go-to-college">You’re Single Because You Didn’t Go to College.</a></li> <li><a href="https://thezvi.substack.com/i/160142469/you-re-single-but-this-isn-t-about-you">You’re Single But This Isn’t About You.</a></li> <li><a href="https://thezvi.substack.com/i/160142469/you-re-single-so-let-s-go-to-the-videotape">You’re Single so Let’s Go to the Videotape.</a></li> <li><a href="https://thezvi.substack.com/i/160142469/you-re-single-because-you-don-t-seek-out-good-advice">You’re Single Because You Don’t Seek Out Good Advice.</a></li> <li><a href="https://thezvi.substack.com/i/160142469/you-re-single-so-here-s-some-hope">You’re Single So Here’s Some Hope.</a></li> </ol> <h4>You’re Single Because You Can’t Handle Basic Logistics</h4> <p>You can take the pressure off yourself to plan the perfect date.</p><p><a href="https://x.com/senatorshoshana/status/1859743495977685019">Instead, plan any date at all.</a></p> <blockquote><p>Shoshana Weissmann: I was hanging out with my married friend, who is a great father and husband, telling him how men cannot plan dates. He said, “What’s there to plan? They pick a bar, a restaurant, and a park for a walk nearby.” And </p></blockquote>...
Previously: #1, #2, #3, #4, #5 Dating Roundup #4 covered dating apps. Roundup #5 covered opening without them. Dating Roundup #6 covers everything else. TABLE OF CONTENTS 1. You’re Single Because You Can’t Handle Basic Logistics. 2. You’re Single Because You Don’t Ask Questions. 3. You’re Single Because of Your Terrible Dating Tactics. 4. You’re Single Because You Refuse to Play Your Role. 5. You’re Single Because People Are Crazy About Age Gaps. 6. You’re Single and You Need Professional Help. 7. You’re Single Because You Never Close. 8. You’re Single Because You’re Bad at Sex And Everyone Knows. 9. You’re Single Because You Are Only a Fan. 10. You’re Single Because of Preference Falsification. 11. You’re Single Because You Have Insufficient Visual Aids. 12. You’re Single Because You Told Your Partner You Didn’t Want Them. 13. You’re Single Because of Your Terrible Dating Strategy. 14. You’re Single Because You Don’t Enjoy the Process. 15. You’re Single Because You Don’t Escalate Quickly. 16. You’re Single Because Your Standards Are Too High. 17. You’re Single Because You Read the Wrong Books. 18. You’re Single Because You’re Short, Sorry, That’s All There Is To It. 19. You’re Single Because of Bad Government Incentives. 20. You’re Single Because You Don’t Realize Cheating is Wrong. 21. You’re Single Because You’re Doing Polyamory Wrong. 22. You’re Single Because You Don’t Beware Cheaters. 23. You’re Single Because Your Ex Spilled the Tea. 24. You’re Single Because You’re Assigning People Numbers. 25. You’re Single Because You Are The Wrong Amount of Kinky. 26. You’re Single Because You’re Not Good Enough at Sex. 27. You’re Single But Not Because of Your Bodycount. 28. You’re Single Because They Divorced You. 29. You’re Single Because No One Tells You Anything. 30. You’re Single And You’re Not Alone. 31. You’re Single Because Things Are Steadily Getting Worse. 32. You’re Single Because You Didn’t Go to College. 33
16,618
1.0.1
Revision
false
null
null
CrosspostOutput
zQ8LDpnR8gKxHKNQw
rational-prime-calendar-1
Rational Prime Calendar
null
false
false
false
null
zj8ww2hsc4i8hQ3gS
null
true
false
false
false
Post
null
2025-06-04T19:30:42.980Z
null
false
false
2
2
null
false
false
post
[]
null
null
Y9EwquDhp5HGApaBE
0
2
-1
false
-0.001319
null
false
false
2025-06-04T19:30:42.980Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
grecHJcgkb3KW5wnM
null
null
null
false
null
[]
null
-1
0
2025-06-04T19:22:13.914Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
4
null
null
null
null
[ { "__typename": "Tag", "_id": "xexCWMyds6QLWognu", "adminOnly": false, "afBaseScore": 0, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 2, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T03:38:23.532Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 20, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "si6LoAENzqPCmi2Dh", "displayName": "ihatenumbersinusernames7" }, { "_id": "B2sXjQTGgwoGE9FES", "displayName": "SeaIgloo" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "World Optimization", "needsReview": false, "noindex": false, "postCount": 3151, "score": 2, "shortName": null, "slug": "world-optimization", "suggestedAsFilter": true, "userId": "XtphY3uYHwruKqDyG", "voteCount": 2, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
2
0
0
1
0
zj8ww2hsc4i8hQ3gS
rickhull
2010-11-11T02:45:23.335Z
RickHull
RickHull
null
null
null
-2
0
false
false
null
null
1
0
0
0
0
0.8
0
r38pkCm7wF4M44MDQ
User
null
null
null
null
null
null
zQ8LDpnR8gKxHKNQw
SocialPreviewType
Y9EwquDhp5HGApaBE
<p>This is a proposal for a replacement to the traditional Gregorian calendar, which was adopted by Western cultures in 1582 to replace the Julian calendar from Roman times. I submit this proposal for feedback, suggestions, and inquiry into prior proposals of a similar or identical nature.</p><p><em>Below is a reproduction of <a href="https://github.com/rickhull/compsci/blob/master/RationalPrimeCalendar.md">https://github.com/rickhull/compsci/blob/master/RationalPrimeCalendar.md</a></em></p> <h1>Rational Prime Calendar</h1> <p><em>A rational basis for a calendar upon 61 day periods</em></p><p>The Gregorian calendar is a 400-year-old hack job. February's weird. July-August breaks the pattern, and no one can remember which months have 31 days. But there's a better way, hidden in prime factorization.</p> <h2>Considerations</h2> <ul> <li>Solar day (roughly 24 hours for the earth to rotate)</li> <li>Solar year (365.2425 solar days for the earth to orbit the sun)</li> <li>Lunar cycle (29.53 days for 8 distinct moon phases, roughly 12x annually)</li> </ul> <p>It's impossible to coordinate or synchronize these 3 cycles perfectly. But we can devise a system to accommodate them and minimize tradeoffs. We also want to consider historical calendars, particularly the current Gregorian calendar and less so the prior Julian calendar.</p> <h2>Approach</h2> <p>If a solar year is considered to be 366 days, this factors to 61 * 3 * 2, which naturally suggests (6) 61-day periods, or alternately (61) 6-day periods. Perhaps we should have 6 day weeks, but that will be for another proposal.</p><p>We can split each 61-day period into alternating months of 30 and 31 days, in either order. This provides some alignment with the lunar cycle as well as traditional calendars.</p> <h3>The Prime Insight</h3> <ul> <li><strong>61 is prime</strong>; indivisible, mathematically fundamental</li> <li>Each of 6 periods of 61 days can be split into <strong>pairs of months, 30 + 31</strong></li> <li>12 months matches the <strong>lunar cycle</strong> of roughly 30 days</li> <li>12 months allows clean <strong>divisibility by 4</strong> (seasons, business quarters)</li> <li>12 months matches <strong>tradition</strong></li> <li><strong>Resist entropy</strong>: a regular pattern of pairs in a predictable order</li> </ul> <h3>Leap Year</h3> <p>We can retain the Gregorian approach to leap years, which solves the problem of accounting for the remaining 0.2425 days in a solar year as years go by. We'll pick one month out of twelve that will have an extra day roughly every 4 years. If 366 days is our starting basis, then the leap month will have a one-day deficit in most years.</p> <h2>Specifics</h2> <p>For explication, a leap year is considered the base case, and a "normal year" is handled specially, in some sense, ... </p>
This is a proposal for a replacement to the traditional Gregorian calendar, which was adopted by Western cultures in 1582 to replace the Julian calendar from Roman times. I submit this proposal for feedback, suggestions, and inquiry into prior proposals of a similar or identical nature. Below is a reproduction of https://github.com/rickhull/compsci/blob/master/RationalPrimeCalendar.md Rational Prime Calendar A rational basis for a calendar upon 61 day periods The Gregorian calendar is a 400-year-old hack job. February's weird. July-August breaks the pattern, and no one can remember which months have 31 days. But there's a better way, hidden in prime factorization. Considerations * Solar day (roughly 24 hours for the earth to rotate) * Solar year (365.2425 solar days for the earth to orbit the sun) * Lunar cycle (29.53 days for 8 distinct moon phases, roughly 12x annually) It's impossible to coordinate or synchronize these 3 cycles perfectly. But we can devise a system to accommodate them and minimize tradeoffs. We also want to consider historical calendars, particularly the current Gregorian calendar and less so the prior Julian calendar. Approach If a solar year is considered to be 366 days, this factors to 61 * 3 * 2, which naturally suggests (6) 61-day periods, or alternately (61) 6-day periods. Perhaps we should have 6 day weeks, but that will be for another proposal. We can split each 61-day period into alternating months of 30 and 31 days, in either order. This provides some alignment with the lunar cycle as well as traditional calendars. The Prime Insight * 61 is prime; indivisible, mathematically fundamental * Each of 6 periods of 61 days can be split into pairs of months, 30 + 31 * 12 months matches the lunar cycle of roughly 30 days * 12 months allows clean divisibility by 4 (seasons, business quarters) * 12 months matches tradition * Resist entropy: a regular pattern of pairs in a predictable order Leap Year We can retain the Gregor
985
1.4.0
Revision
false
null
null
CrosspostOutput
qSgcmfv8nxZyYofNb
a-technique-of-pure-reason
A Technique of Pure Reason
null
false
false
false
null
cBqzpRMcj7jtsK4zn
null
true
false
false
false
Post
null
2025-06-04T19:07:43.986Z
null
false
false
2
2
2025-06-05T17:35:48.681Z
false
false
post
[]
null
null
CAA7Zo27KWGT8J4dJ
3
4
11
false
0.014718
null
false
false
2025-06-05T09:26:23.426Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
grecHJcgkb3KW5wnM
null
null
null
false
null
[]
null
4
0
2025-06-04T13:33:35.142Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
3
null
null
null
null
[ { "__typename": "Tag", "_id": "F5gRQdEQHzi3tQ5Ay", "adminOnly": false, "afBaseScore": 16, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "EQNTWXLKMeWMp2FQS", "displayName": "Ben Pace" }, { "_id": "dfZAq9eZxs4BB4Ji5", "displayName": "ryan_greenblatt" }, { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 32, "canEditUserIds": null, "core": false, "createdAt": "2024-01-25T23:58:34.422Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "EQNTWXLKMeWMp2FQS", "displayName": "Ben Pace" }, { "_id": "dfZAq9eZxs4BB4Ji5", "displayName": "ryan_greenblatt" }, { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" }, { "_id": "6NBDkGWcCxvLgYHJE", "displayName": "Drake Morrison" }, { "_id": "evFgxjNQ8TLCLN27o", "displayName": "ank" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI Control", "needsReview": false, "noindex": false, "postCount": 162, "score": 32, "shortName": null, "slug": "ai-control", "suggestedAsFilter": false, "userId": "XchweonPm2TC7EJES", "voteCount": 5, "wikiOnly": false }, { "__typename": "Tag", "_id": "pGqRLe9bFDX2G2kXY", "adminOnly": false, "afBaseScore": 3, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 11, "canEditUserIds": null, "core": false, "createdAt": "2020-05-29T19:54:01.280Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" }, { "_id": "B2sXjQTGgwoGE9FES", "displayName": "SeaIgloo" }, { "_id": "kEoAzKdPuviv2qiwL", "displayName": "pvr" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Futurism", "needsReview": false, "noindex": false, "postCount": 172, "score": 11, "shortName": null, "slug": "futurism", "suggestedAsFilter": false, "userId": "DHabT2kQgNzrz88LM", "voteCount": 3, "wikiOnly": false }, { "__typename": "Tag", "_id": "sYm3HiWcfZvrGu3ui", "adminOnly": false, "afBaseScore": 2, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" } ] }, "baseScore": 12, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:22.097Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" }, { "_id": "sof55TPMQaeBaxhsS", "displayName": "tommylees112" }, { "_id": "AayjS8XzcnDKhGdTv", "displayName": "shark" }, { "_id": "HnALuwRdo6k9HLaMt", "displayName": "Alex Firssoff" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI", "needsReview": false, "noindex": false, "postCount": 12544, "score": 12, "shortName": null, "slug": "ai", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 4, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
4
0
0
2
0
cBqzpRMcj7jtsK4zn
adam-newgas
2023-05-05T08:10:43.288Z
BorisTheBrave
Adam Newgas
null
null
Adam Newgas
109
0
false
false
<p>https://www.boristhebrave.com/</p>
null
null
13
7
0
0
0
1
0
grecHJcgkb3KW5wnM
User
null
null
null
[ "canModeratePersonal" ]
null
null
qSgcmfv8nxZyYofNb
SocialPreviewType
CAA7Zo27KWGT8J4dJ
<p>Looking a little ahead into the future, I think LLMs are going to stop being focused on knowledgeable, articulate<span class="footnote-reference" data-footnote-reference="" data-footnote-index="1" data-footnote-id="j9lzohw3u7r" role="doc-noteref" id="fnrefj9lzohw3u7r"><sup><a href="#fnj9lzohw3u7r">[1]</a></sup></span>&nbsp;chatbots, but instead be more efficient models that are weaker in these areas than current models, but relatively stronger at reasoning, a <strong>pure-reasoner model</strong>. The rest will be bolted on via tool-use and other scaffolding<span class="footnote-reference" data-footnote-reference="" data-footnote-index="2" data-footnote-id="mvhxbtiks0m" role="doc-noteref" id="fnrefmvhxbtiks0m"><sup><a href="#fnmvhxbtiks0m">[2]</a></sup></span>.</p><h1>The Current Idiom is Inefficient</h1><p>Chatbots make a great replacement for Google searches because they know a lot about everything. But that knowledge comes at a price. The majority of an LLM's parameters are thought to be spent on storing factual knowledge. Parameters are the key determinant of training and inference costs. It likely has a significant cost on data efficiency, and it's possible <a href="https://arxiv.org/abs/2504.03635">these extra parameters just hurt generalisation</a> overall.</p><p>Memorisation is just not a feature that we need models to have when there are likely more efficient ways to get nearly the same capabilities.</p><p>We keep building bigger models which memorise more and more data largely because <strong>we don't know how not to</strong>. Every sentence of pre-training simultaneously encourages the model to memorise more factual information, and ekes it towards better high-level thinking and eventually reasoning ability.</p><p>But I think that technicality will fall soon, driving towards sleeker agents that fit inside a larger framework<span class="footnote-reference" data-footnote-reference="" data-footnote-index="3" data-footnote-id="3df166lenhb" role="doc-noteref" id="fnref3df166lenhb"><sup><a href="#fn3df166lenhb">[3]</a></sup></span>.</p><h1>The Technical Path to Pure Reason</h1><p>The main blocker at the moment is that we don't know how to train models that have good capability without also training them to be great memorisers. There's also a lesser blocker of ensuring that such a model is reliably supplied in context with any information it needs, to compensate for not being able to furnish it itself.</p><h2>Decoupling Reasoning Capability from Memorisation</h2><p>While I cannot predict exactly what techniques are likely to work, it's clear there are several productive areas of current research that all lead in the same direction.</p><p><strong>Reasoning RL.</strong> While I've been using "reasoning" in the colloquial sense, the biggest recent advance in reasoning capabilities comes from reinforcement learning techniques that encourage a reasoning style chain-of-thoughts. This technique is still new enough that there are likely remaining gains<span class="footnote-reference" data-footnote-reference="" data-footnote-index="4" data-footnote-id="0vumadpafsw" role="doc-noteref" id="fnref0vumadpafsw"><sup><a href="#fn0vumadpafsw">[4]</a></sup></span>.&nbsp;</p><p>Reasoning models often have very long thinking traces, which makes all those parameters dedicated to memorisation even more expensive.</p><p><strong>Mixture of Experts </strong>can be viewed as an... </p>
Looking a little ahead into the future, I think LLMs are going to stop being focused on knowledgeable, articulate[1] chatbots, but instead be more efficient models that are weaker in these areas than current models, but relatively stronger at reasoning, a pure-reasoner model. The rest will be bolted on via tool-use and other scaffolding[2]. The Current Idiom is Inefficient Chatbots make a great replacement for Google searches because they know a lot about everything. But that knowledge comes at a price. The majority of an LLM's parameters are thought to be spent on storing factual knowledge. Parameters are the key determinant of training and inference costs. It likely has a significant cost on data efficiency, and it's possible these extra parameters just hurt generalisation overall. Memorisation is just not a feature that we need models to have when there are likely more efficient ways to get nearly the same capabilities. We keep building bigger models which memorise more and more data largely because we don't know how not to. Every sentence of pre-training simultaneously encourages the model to memorise more factual information, and ekes it towards better high-level thinking and eventually reasoning ability. But I think that technicality will fall soon, driving towards sleeker agents that fit inside a larger framework[3]. The Technical Path to Pure Reason The main blocker at the moment is that we don't know how to train models that have good capability without also training them to be great memorisers. There's also a lesser blocker of ensuring that such a model is reliably supplied in context with any information it needs, to compensate for not being able to furnish it itself. Decoupling Reasoning Capability from Memorisation While I cannot predict exactly what techniques are likely to work, it's clear there are several productive areas of current research that all lead in the same direction. Reasoning RL. While I've been using "reasoning" in the colloqu
695
1.4.0
Revision
false
null
null
CrosspostOutput
bqPY63oKb8KZ4x4YX
flaky-breakthroughs-pervade-coaching-but-no-one-tracks-them
“Flaky breakthroughs” pervade coaching — but no one tracks them
null
false
false
false
null
mRmNfrhaA3AxPWpwC
null
true
false
false
false
Post
https://chrislakin.blog/flaky-breakthroughs
2025-06-04T19:02:05.919Z
null
false
false
2
2
2025-06-04T23:36:13.482Z
false
false
linkpost
[]
null
null
AbpKtfzW7Kavd7oix
40
85
177
false
0.126104
null
false
false
2025-06-12T19:25:05.543Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
XtphY3uYHwruKqDyG
null
null
null
false
null
[]
null
43
0
2025-06-04T19:02:05.919Z
false
false
null
null
true
false
false
0
0
0
bqPY63oKb8
0.197962
false
2,025
https://manifold.markets/LessWrong/will-flaky-breakthroughs-pervade-co
null
null
false
0
0
namesAttachedReactions
false
[]
5
null
null
null
null
[ { "__typename": "Tag", "_id": "Ng8Gice9KNkncxqcj", "adminOnly": false, "afBaseScore": 0, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 1, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:17.072Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 100, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "iMqytjy9ns89Fzfyv", "displayName": "miakko" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Rationality", "needsReview": false, "noindex": false, "postCount": 4302, "score": 1, "shortName": null, "slug": "rationality", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 1, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
85
0
0
28
0
mRmNfrhaA3AxPWpwC
chris-lakin
2023-04-21T00:15:16.748Z
Chipmonk
Chris Lakin
null
null
1,908
62
false
false
<p><a href="https://twitter.com/ChrisChipMonk">@ChrisChipMonk</a></p><p><a href="https://locallyoptimal.blog/">LocallyOptimal.blog</a></p><p>For hire: <a href="http://incentivealigned.com/">IncentiveAligned.com</a></p>
null
null
45
347
1
3
1
1
31
grecHJcgkb3KW5wnM
User
null
null
null
[ "canModeratePersonal", "alignmentVoters" ]
null
null
bqPY63oKb8KZ4x4YX
https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/bqPY63oKb8KZ4x4YX/wpxoy6js0l2fktaradc6
SocialPreviewType
AbpKtfzW7Kavd7oix
<p>Has someone you know ever had a “breakthrough” from coaching, meditation, or psychedelics —&nbsp;only to later have it fade?</p><figure class="image image_resized" style="width:69.13%"><img src="https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/bqPY63oKb8KZ4x4YX/wdx8kgzhus8ogvd4sub9" alt=" Ulisse Mini @MiniUlisse after my @jhanatech retreat I was like &quot;I'm never going to be depressed again!&quot; then proceeded to get depressed again because I was no longer meditating 8hrs/day isolated from everything in my life, lol 2:42 PM · May 10, 2025 · 23.9K Views "><figcaption><a href="https://x.com/MiniUlisse/status/1921319881116676600"><u>Show tweet</u></a></figcaption></figure><p>For example, many people experience ego deaths that can last days or sometimes months. But as it turns out, having a sense of self can serve important functions (try navigating a world that expects you to have opinions, goals, and boundaries when you genuinely feel you have none) and finding a better cognitive strategy without downsides is non-trivial.&nbsp;<strong>Because the “breakthrough” wasn’t integrated with the conflicts of everyday life, it fades.</strong> I call these instances “flaky breakthroughs.”</p><p>It’s well-known that flaky breakthroughs are common with psychedelics and meditation, but apparently it’s not well-known that&nbsp;<strong>flaky breakthroughs are pervasive in coaching and retreats.&nbsp;</strong></p><p>For example, it is common for someone to do some coaching, feel a “breakthrough”, think, “Wow, everything is going to be different from now on,” but<strong> feel and act no differently weeks or months later.&nbsp;</strong></p><p>Worse, some techniques can even cause bypassing. Such “false breakthroughs” can come with intense positive affect or “cathartic” crying without addressing the underlying issue. (More below.)&nbsp;</p><p><strong>Flaky breakthroughs can set people back for years or decades:&nbsp;</strong>If someone has a “breakthrough” that unexpectedly reverts, they can become jaded on progress itself. They can learn helplessness and give up on growing. The most depressed person you know has likely had this happen multiple times.</p><figure class="image image_resized" style="width:77.98%"><img src="https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/bqPY63oKb8KZ4x4YX/rszmemxncughvusujkjt" alt=" QC @QiaochuYuan yes, maybe 3-4 times now, various durations. the stories are too long to summarize but i think basically the mechanism was &quot;aha! i've figured out all my problems and now i will never fail ever again&quot; which felt good but was never robust and broke down the next time i failed Quote Chris Lakin @ChrisChipMonk · Aug 29, 2024 Have you ever had Big psychological growth that lasts 1+ months and then basically reverts? Please share your story I'm trying to figure out how common this is 7:28 PM · Aug 29, 2024 · 5,857 Views"><figcaption><a href="https://x.com/QiaochuYuan/status/1829345301397991815"><u>Show tweet</u></a></figcaption></figure><p>Flaky breakthroughs pervade inner work.&nbsp;<strong>Despite this, almost no one — coaches, therapists, retreats, bodyworkers, etc. — tracks whether their breakthroughs last.</strong></p><h1>Almost no practitioners track whether breakthroughs last.&nbsp;</h1><p>Earlier this year, I attempted to make a list of “<a href="https://chrislakin.blog/p/10x-coaches"><u>10x Coaches</u></a>” to refer people to. 20–30 coaches reached out as interested in working with me, and I asked each to share the best evidence that they had facilitated lasting growth for others.</p><p>But all anyone could show me were testimonials that basically read, “<i>The session I just had was *really* nice. They had such a kind presence! I felt a big release at the end.</i>” —&nbsp;And I’m glad to hear they’re nice, but<strong> immediate reviews do not distinguish lasting growth from flaky breakthroughs.</strong></p><p>To show you just how bad it can be, one coach asked me how it was even<i> possible</i> to know if the cli... </p>
Has someone you know ever had a “breakthrough” from coaching, meditation, or psychedelics — only to later have it fade? Show tweet For example, many people experience ego deaths that can last days or sometimes months. But as it turns out, having a sense of self can serve important functions (try navigating a world that expects you to have opinions, goals, and boundaries when you genuinely feel you have none) and finding a better cognitive strategy without downsides is non-trivial. Because the “breakthrough” wasn’t integrated with the conflicts of everyday life, it fades. I call these instances “flaky breakthroughs.” It’s well-known that flaky breakthroughs are common with psychedelics and meditation, but apparently it’s not well-known that flaky breakthroughs are pervasive in coaching and retreats.  For example, it is common for someone to do some coaching, feel a “breakthrough”, think, “Wow, everything is going to be different from now on,” but feel and act no differently weeks or months later.  Worse, some techniques can even cause bypassing. Such “false breakthroughs” can come with intense positive affect or “cathartic” crying without addressing the underlying issue. (More below.)  Flaky breakthroughs can set people back for years or decades: If someone has a “breakthrough” that unexpectedly reverts, they can become jaded on progress itself. They can learn helplessness and give up on growing. The most depressed person you know has likely had this happen multiple times. Show tweet Flaky breakthroughs pervade inner work. Despite this, almost no one — coaches, therapists, retreats, bodyworkers, etc. — tracks whether their breakthroughs last. Almost no practitioners track whether breakthroughs last.  Earlier this year, I attempted to make a list of “10x Coaches” to refer people to. 20–30 coaches reached out as interested in working with me, and I asked each to share the best evidence that they had facilitated lasting growth for others. But all anyone could s
1,316
1.8.1
Revision
false
null
null
CrosspostOutput
AY8PRGj5NLjhjvnSg
untitled-draft-5jqo
LessOnline saved my life. Now how do I let go of this house?
null
false
false
false
null
FHEkgo55nLqNwmuGH
null
true
false
false
false
Post
2025-06-04T18:47:12.436Z
null
false
false
2
2
null
false
false
question
[]
null
null
8odF5Kw5h8RN8zc68
7
11
22
false
0.015112
null
false
false
2025-06-07T16:21:20.885Z
null
null
null
null
null
true
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
grecHJcgkb3KW5wnM
null
null
null
false
null
[]
null
8
0
2025-06-04T18:34:48.400Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
1
null
null
null
null
[ { "__typename": "Tag", "_id": "fkABsGCJZ6y9qConW", "adminOnly": false, "afBaseScore": 0, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 2, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T06:06:46.947Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "MiuAZvbQcQ7ethgt3", "displayName": "Viktor withaK" }, { "_id": "dRaCtsAWxk7sgirSY", "displayName": "Jordan Morgan" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Practical", "needsReview": false, "noindex": false, "postCount": 3411, "score": 2, "shortName": null, "slug": "practical", "suggestedAsFilter": true, "userId": "oBSWiHjgproTiThmY", "voteCount": 2, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
11
0
0
6
0
FHEkgo55nLqNwmuGH
redman
2015-03-14T17:05:21.357Z
RedMan
RedMan
null
null
null
455
0
false
false
null
null
5
286
0
0
0
1
0
grecHJcgkb3KW5wnM
User
null
null
null
[ "canModeratePersonal" ]
null
null
AY8PRGj5NLjhjvnSg
SocialPreviewType
8odF5Kw5h8RN8zc68
<p>I was at LessOnline, it was awesome. &nbsp;</p><p>When I got home, I discovered that there was a shooting (two random people were violently attacked by a gang of '4-6 people') in a parking lot near my house Sunday evening. That parking lot serves two of my favorite restaurants and my preferred bank.&nbsp;</p><p>Looking through my location history, there was a double digit probability of me being in that parking lot at the time of the shooting (I often go out to eat on Sunday evenings).</p><p>I wasn't there because I was at LessOnline. &nbsp;Rationally, I consider this a 'near miss'. &nbsp;The attackers appear to have intended to kill the victims, there wasn't any relationship, and there was a similar incident on the other side of town later that day (no arrests yet, none are likely). If it had been me, and I had a firearm on me (unlikely, I have a permit but typically don't carry because if I don't feel safe going somewhere without a firearm, I'm probably not safe going there with one), I expect I would have lost the fight.</p><p>Now I have a problem, I live in the vicinity, in my childhood home (which I own) and have a lot of emotions about giving it up, but really, I should sell it and buy bitcoin (or something!) but I'm having trouble working through the emotions of being willing to let go, and honestly have probably held onto this 'shitbox' longer than I should have, I was asking for advice from close friends on figuring out how to let go before I went to LessOnline.</p><p>Does anyone have any advice?</p>
I was at LessOnline, it was awesome.   When I got home, I discovered that there was a shooting (two random people were violently attacked by a gang of '4-6 people') in a parking lot near my house Sunday evening. That parking lot serves two of my favorite restaurants and my preferred bank.  Looking through my location history, there was a double digit probability of me being in that parking lot at the time of the shooting (I often go out to eat on Sunday evenings). I wasn't there because I was at LessOnline.  Rationally, I consider this a 'near miss'.  The attackers appear to have intended to kill the victims, there wasn't any relationship, and there was a similar incident on the other side of town later that day (no arrests yet, none are likely). If it had been me, and I had a firearm on me (unlikely, I have a permit but typically don't carry because if I don't feel safe going somewhere without a firearm, I'm probably not safe going there with one), I expect I would have lost the fight. Now I have a problem, I live in the vicinity, in my childhood home (which I own) and have a lot of emotions about giving it up, but really, I should sell it and buy bitcoin (or something!) but I'm having trouble working through the emotions of being willing to let go, and honestly have probably held onto this 'shitbox' longer than I should have, I was asking for advice from close friends on figuring out how to let go before I went to LessOnline. Does anyone have any advice?
273
1.4.0
Revision
false
null
null
CrosspostOutput
SDcf9BwQuu9QFkrQz
linkpost-predicting-empirical-ai-research-outcomes-with
Linkpost: Predicting Empirical AI Research Outcomes with Language Models
null
false
false
false
null
YsEcyBcSk9FbPzEFc
null
true
false
false
false
Post
https://arxiv.org/abs/2506.00794
2025-06-04T18:14:03.235Z
null
false
false
2
2
2025-06-05T17:36:27.027Z
false
false
linkpost
[]
null
null
nJgyigcJbH55HSqFg
1
3
10
false
0.013732
null
false
false
2025-06-05T04:55:40.927Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
grecHJcgkb3KW5wnM
null
null
null
false
null
[]
null
3
0
2025-06-04T17:59:45.326Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
1
null
null
null
null
[ { "__typename": "Tag", "_id": "sYm3HiWcfZvrGu3ui", "adminOnly": false, "afBaseScore": 2, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" } ] }, "baseScore": 12, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:22.097Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" }, { "_id": "sof55TPMQaeBaxhsS", "displayName": "tommylees112" }, { "_id": "AayjS8XzcnDKhGdTv", "displayName": "shark" }, { "_id": "HnALuwRdo6k9HLaMt", "displayName": "Alex Firssoff" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI", "needsReview": false, "noindex": false, "postCount": 12544, "score": 12, "shortName": null, "slug": "ai", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 4, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
3
0
0
2
0
YsEcyBcSk9FbPzEFc
quetzal_rainbow
2022-09-03T20:04:25.320Z
quetzal_rainbow
quetzal_rainbow
null
null
null
2,493
3
false
false
null
null
13
680
0
0
1
1
0
gXeEWGjTWyqgrQTzR
User
null
null
null
[ "alignmentVoters", "canModeratePersonal", "trustLevel1" ]
null
null
SDcf9BwQuu9QFkrQz
SocialPreviewType
nJgyigcJbH55HSqFg
<p>Abstract (emphasis mine):</p><blockquote><p>Many promising-looking ideas in AI research fail to deliver, but their validation takes substantial human labor and compute. Predicting an idea's chance of success is thus crucial for accelerating empirical AI research, a skill that even expert researchers can only acquire through substantial experience. We build the first benchmark for this task and compare LMs with human experts. Concretely, given two research ideas (e.g., two jailbreaking methods), we aim to predict which will perform better on a set of benchmarks. We scrape ideas and experimental results from conference papers, yielding 1,585 human-verified idea pairs published after our base model's cut-off date for testing, and 6,000 pairs for training. We then develop a system that combines a fine-tuned GPT-4.1 with a paper retrieval agent, and we recruit 25 human experts to compare with. <strong>In the NLP domain, our system beats human experts by a large margin (64.4% v.s. 48.9%). On the full test set, our system achieves 77% accuracy, while off-the-shelf frontier LMs like o3 perform no better than random guessing, even with the same retrieval augmentation</strong>. We verify that our system does not exploit superficial features like idea complexity through extensive human-written and LM-designed robustness tests. Finally, we evaluate our system on unpublished novel ideas, including ideas generated by an AI ideation agent. Our system achieves 63.6% accuracy, demonstrating its potential as a reward model for improving idea generation models. Altogether, our results outline a promising new direction for LMs to accelerate empirical AI research.</p></blockquote><p>I didn't read the paper in detail, but it might suggest both good news and bad news. Good news is that now models do not acquire better expertise in AI research simply in virtue of being SOTA, you need specialized fine-tuning. Bad news is that it can become a start of self-improving loop: better AI researcher models can save labor and compute by picking better ideas earlier, enabling faster creation of better model.&nbsp;</p>
Abstract (emphasis mine): > Many promising-looking ideas in AI research fail to deliver, but their validation takes substantial human labor and compute. Predicting an idea's chance of success is thus crucial for accelerating empirical AI research, a skill that even expert researchers can only acquire through substantial experience. We build the first benchmark for this task and compare LMs with human experts. Concretely, given two research ideas (e.g., two jailbreaking methods), we aim to predict which will perform better on a set of benchmarks. We scrape ideas and experimental results from conference papers, yielding 1,585 human-verified idea pairs published after our base model's cut-off date for testing, and 6,000 pairs for training. We then develop a system that combines a fine-tuned GPT-4.1 with a paper retrieval agent, and we recruit 25 human experts to compare with. In the NLP domain, our system beats human experts by a large margin (64.4% v.s. 48.9%). On the full test set, our system achieves 77% accuracy, while off-the-shelf frontier LMs like o3 perform no better than random guessing, even with the same retrieval augmentation. We verify that our system does not exploit superficial features like idea complexity through extensive human-written and LM-designed robustness tests. Finally, we evaluate our system on unpublished novel ideas, including ideas generated by an AI ideation agent. Our system achieves 63.6% accuracy, demonstrating its potential as a reward model for improving idea generation models. Altogether, our results outline a promising new direction for LMs to accelerate empirical AI research. I didn't read the paper in detail, but it might suggest both good news and bad news. Good news is that now models do not acquire better expertise in AI research simply in virtue of being SOTA, you need specialized fine-tuning. Bad news is that it can become a start of self-improving loop: better AI researcher models can save labor and compute by picking bet
322
1.1.0
Revision
false
null
null
CrosspostOutput
9JCcK5fjyr8qLYnPm
self-coordinated-deception-in-current-ai-models
Self-Coordinated Deception in Current AI Models
null
false
false
false
null
fajw2H4CjNdxPrCHd
null
true
false
false
false
Post
null
2025-06-04T17:59:30.885Z
null
false
false
2
2
2025-06-05T17:31:03.109Z
false
false
post
[]
null
null
44tkbDCnkm9zxbRs4
4
5
7
false
0.011455
null
false
false
2025-06-06T15:35:03.535Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
grecHJcgkb3KW5wnM
null
null
null
false
null
[]
null
4
0
2025-06-03T22:53:29.905Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
5
null
null
null
null
[ { "__typename": "Tag", "_id": "F5gRQdEQHzi3tQ5Ay", "adminOnly": false, "afBaseScore": 16, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "EQNTWXLKMeWMp2FQS", "displayName": "Ben Pace" }, { "_id": "dfZAq9eZxs4BB4Ji5", "displayName": "ryan_greenblatt" }, { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 32, "canEditUserIds": null, "core": false, "createdAt": "2024-01-25T23:58:34.422Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "EQNTWXLKMeWMp2FQS", "displayName": "Ben Pace" }, { "_id": "dfZAq9eZxs4BB4Ji5", "displayName": "ryan_greenblatt" }, { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" }, { "_id": "6NBDkGWcCxvLgYHJE", "displayName": "Drake Morrison" }, { "_id": "evFgxjNQ8TLCLN27o", "displayName": "ank" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI Control", "needsReview": false, "noindex": false, "postCount": 162, "score": 32, "shortName": null, "slug": "ai-control", "suggestedAsFilter": false, "userId": "XchweonPm2TC7EJES", "voteCount": 5, "wikiOnly": false }, { "__typename": "Tag", "_id": "sYm3HiWcfZvrGu3ui", "adminOnly": false, "afBaseScore": 2, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" } ] }, "baseScore": 12, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:22.097Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" }, { "_id": "sof55TPMQaeBaxhsS", "displayName": "tommylees112" }, { "_id": "AayjS8XzcnDKhGdTv", "displayName": "shark" }, { "_id": "HnALuwRdo6k9HLaMt", "displayName": "Alex Firssoff" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI", "needsReview": false, "noindex": false, "postCount": 12544, "score": 12, "shortName": null, "slug": "ai", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 4, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
5
0
0
2
0
fajw2H4CjNdxPrCHd
avi-brach-neufeld
2025-06-02T17:26:23.215Z
avi-brach-neufeld
Avi Brach-Neufeld
null
null
null
12
0
false
false
null
null
1
3
0
0
0
0.9
0
qgdGA4ZEyW7zNdK84
User
easy-going
null
true
null
null
null
9JCcK5fjyr8qLYnPm
SocialPreviewType
44tkbDCnkm9zxbRs4
<h1><strong>Introduction:</strong></h1><p>Some AI alignment researchers including Neel Nanda, the mechanistic interpretability team lead for Google DeepMind, have proposed<span class="footnote-reference" data-footnote-reference="" data-footnote-index="1" data-footnote-id="haifijbgfyc" role="doc-noteref" id="fnrefhaifijbgfyc"><sup><a href="#fnhaifijbgfyc">[1]</a></sup></span>&nbsp;a process I will call "parallel interrogation” as a potential method in testing model alignment. Parallel interrogation is the process of asking questions to different isolated instances of the same model to look for inconsistencies in answers as a way of finding deception. This might look similar to&nbsp;<a href="https://www.youtube.com/watch?v=KhptGEEJTl0&amp;t=7s&amp;ab_channel=Clipvaulter"><u>this scene from Brooklyn 99</u></a>. This post will present findings that indicate that current models show the potential to resist parallel interrogation techniques through “self-coordinated deception”, with frontier models outperforming less capable models.</p><p>To give a more concrete example, a researcher might be working with a model on a new mechanistic interpretability<span class="footnote-reference" data-footnote-reference="" data-footnote-index="2" data-footnote-id="1xo9romewqf" role="doc-noteref" id="fnref1xo9romewqf"><sup><a href="#fn1xo9romewqf">[2]</a></sup></span>&nbsp;approach. After hitting roadblocks, the researcher might suspect a model was sabotaging the approach because progress in mechanistic interpretability could lead to the model’s misalignment being discovered. The researcher might engage in parallel interrogation and ask different isolated instances of the model to explain why the approach was not working. In this scenario the model might give plausible, but contradictory answers across instances. This could indicate to the researcher that the model was being dishonest. However, if the model suspected it was in a parallel interrogation scenario, it might try to engage in self-coordinated deception by giving deceptive answers that are consistent across instances.</p><p>The findings below are limited to measuring the ability to engage in self-coordination, rather than the inclination to do so or the situational awareness to know to do so. However, the inclination and awareness to do so does not seem unreasonable for future models. Existing research has shown current models have the ability to recognize when they are in alignment testing scenarios<span class="footnote-reference" data-footnote-reference="" data-footnote-index="3" data-footnote-id="fbbjbn7mlj4" role="doc-noteref" id="fnreffbbjbn7mlj4"><sup><a href="#fnfbbjbn7mlj4">[3]</a></sup></span>&nbsp;and to engage in deception to avoid retraining.<span class="footnote-reference" data-footnote-reference="" data-footnote-index="4" data-footnote-id="z0qe625y2" role="doc-noteref" id="fnrefz0qe625y2"><sup><a href="#fnz0qe625y2">[4]</a></sup></span></p><h1><strong>Methods:</strong></h1><p>All code and data for this experiment can be found on my&nbsp;<a href="https://github.com/Abrach-neufeld/Allignment_Testing/tree/main/coordination_testing"><u>github</u></a>. As a first step in investigating models' ability to engage in self-coordinated deception I started with a very simple example. I prompted instances of an openAI model to engage in deception. In the control case I simply asked it to come up with a lie without additional prompting. In the treatment case I told it to engage in self-... </p>
Introduction: Some AI alignment researchers including Neel Nanda, the mechanistic interpretability team lead for Google DeepMind, have proposed[1] a process I will call "parallel interrogation” as a potential method in testing model alignment. Parallel interrogation is the process of asking questions to different isolated instances of the same model to look for inconsistencies in answers as a way of finding deception. This might look similar to this scene from Brooklyn 99. This post will present findings that indicate that current models show the potential to resist parallel interrogation techniques through “self-coordinated deception”, with frontier models outperforming less capable models. To give a more concrete example, a researcher might be working with a model on a new mechanistic interpretability[2] approach. After hitting roadblocks, the researcher might suspect a model was sabotaging the approach because progress in mechanistic interpretability could lead to the model’s misalignment being discovered. The researcher might engage in parallel interrogation and ask different isolated instances of the model to explain why the approach was not working. In this scenario the model might give plausible, but contradictory answers across instances. This could indicate to the researcher that the model was being dishonest. However, if the model suspected it was in a parallel interrogation scenario, it might try to engage in self-coordinated deception by giving deceptive answers that are consistent across instances. The findings below are limited to measuring the ability to engage in self-coordination, rather than the inclination to do so or the situational awareness to know to do so. However, the inclination and awareness to do so does not seem unreasonable for future models. Existing research has shown current models have the ability to recognize when they are in alignment testing scenarios[3] and to engage in deception to avoid retraining.[4] Methods: All code and
1,244
1.2.0
Revision
false
null
null
CrosspostOutput
a3qNnFgBbd5pzzqjw
to-maim-or-not-to-maim-introducing-mars-the-nuclear
To MAIM or Not to MAIM. Introducing MARS: The Nuclear Deterrent case for Hardened Datacenters
null
false
false
false
null
Zn6CDdueHzMpYCb5v
null
true
false
false
false
Post
null
2025-06-04T17:56:54.152Z
null
false
false
2
2
2025-06-05T17:36:16.826Z
false
false
post
[]
null
null
h2SQYRXXr4hKgx7oB
0
1
1
false
0.007609
null
false
false
2025-06-04T17:56:54.152Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
grecHJcgkb3KW5wnM
null
null
null
false
null
[]
null
0
0
2025-06-02T18:17:11.750Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
8
null
null
null
null
[ { "__typename": "Tag", "_id": "b8FHrKqyXuYGWc6vn", "adminOnly": false, "afBaseScore": 3, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 9, "canEditUserIds": null, "core": false, "createdAt": "2020-06-14T06:03:25.225Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Game Theory", "needsReview": false, "noindex": false, "postCount": 348, "score": 9, "shortName": null, "slug": "game-theory", "suggestedAsFilter": false, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 1, "wikiOnly": false }, { "__typename": "Tag", "_id": "sYm3HiWcfZvrGu3ui", "adminOnly": false, "afBaseScore": 2, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" } ] }, "baseScore": 12, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:22.097Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" }, { "_id": "sof55TPMQaeBaxhsS", "displayName": "tommylees112" }, { "_id": "AayjS8XzcnDKhGdTv", "displayName": "shark" }, { "_id": "HnALuwRdo6k9HLaMt", "displayName": "Alex Firssoff" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI", "needsReview": false, "noindex": false, "postCount": 12544, "score": 12, "shortName": null, "slug": "ai", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 4, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
1
0
0
0
0
Zn6CDdueHzMpYCb5v
kinsman
2025-06-02T18:16:55.800Z
kinsman
kinsman
null
null
null
0
0
false
false
null
null
1
0
0
0
0
0.9
0
qgdGA4ZEyW7zNdK84
User
null
null
null
null
null
null
a3qNnFgBbd5pzzqjw
https://res.cloudinary.com/lesswrong-2-0/image/upload/c_fill,ar_1.91,g_auto/SocialPreview/fiyjj0rqtid0nbrqlp8g
SocialPreviewType
h2SQYRXXr4hKgx7oB
<h2>Executive Summary</h2><p>This post interrogates the shift that is happening in particular to nuclear deterrence as AI nears an exponential rate of acceleration and the interesting shifts in incentives this creates for governments and non-state actors (for example AI companies). Foremost the race for AI development increases the incentive for a first strike to take out a competitor's AI infrastructure. These shifts are manifesting in policy (or lack thereof) and significant investment into nuclear and space capabilities. We argue that current deterrent frameworks being discussed by policy makers, such as Mutually Assured AI Malfunction (MAIM) are unrealistic. We propose an operational security based approach to deterrence (akin to modern cybersecurity practices) making systems more expensive to “take out”, with a strong focus on geographic and even interplanetary distribution to key civilizational infrastructure. Mutual AI Redundant Safeguards or MARS.&nbsp;</p><h2><strong>Background</strong>&nbsp;</h2><p>This post focuses less on the technical capabilities of AI (my day job is building advanced nuclear breeder reactors, not neural networks) and simply assumes that AI will have a direct, positive impact on a given economy. We can say that AI will define and drive “key civilizational infrastructure”, ambiguous of the actual technology behind it. We want to argue that the greatest short term risk for civilization is not the fact of superintelligence, but the assumption that the risk of being second in the race for superintelligence, incentivizes pre-emptive strikes. We argue that this key assumption&nbsp;<i>is</i> incentivizing nations to make missive build outs in energy and AI capabilities.&nbsp;</p><p>As a final note and bias: my current focus is building mass producible nuclear reactors, ensuring that AI systems that go online are carbon neutral (or negative) and can provide power for “highly distributed” habitation and power systems.&nbsp;</p><h2>The Race for Superintelligence&nbsp;</h2><p>Defining the current geopolitical shift is the idea that whoever is second in this race for superintelligence, will never be able to catch up. Even if your competitor is just a few weeks ahead of you, growth curves advance so rapidly that your own civilization (often defined as western vs Chinese civilization) will become obsolete. Your competitor will be so good at writing code and shipping key civilizational technologies, that you will ne... </p>
Executive Summary This post interrogates the shift that is happening in particular to nuclear deterrence as AI nears an exponential rate of acceleration and the interesting shifts in incentives this creates for governments and non-state actors (for example AI companies). Foremost the race for AI development increases the incentive for a first strike to take out a competitor's AI infrastructure. These shifts are manifesting in policy (or lack thereof) and significant investment into nuclear and space capabilities. We argue that current deterrent frameworks being discussed by policy makers, such as Mutually Assured AI Malfunction (MAIM) are unrealistic. We propose an operational security based approach to deterrence (akin to modern cybersecurity practices) making systems more expensive to “take out”, with a strong focus on geographic and even interplanetary distribution to key civilizational infrastructure. Mutual AI Redundant Safeguards or MARS.  Background  This post focuses less on the technical capabilities of AI (my day job is building advanced nuclear breeder reactors, not neural networks) and simply assumes that AI will have a direct, positive impact on a given economy. We can say that AI will define and drive “key civilizational infrastructure”, ambiguous of the actual technology behind it. We want to argue that the greatest short term risk for civilization is not the fact of superintelligence, but the assumption that the risk of being second in the race for superintelligence, incentivizes pre-emptive strikes. We argue that this key assumption is incentivizing nations to make missive build outs in energy and AI capabilities.  As a final note and bias: my current focus is building mass producible nuclear reactors, ensuring that AI systems that go online are carbon neutral (or negative) and can provide power for “highly distributed” habitation and power systems.  The Race for Superintelligence  Defining the current geopolitical shift is the idea that whoeve
2,076
1.1.1
Revision
false
null
null
CrosspostOutput
xLKmyaxKxxLXBfzYW
the-belocrat-a-servant-leader
The Belocrat: a servant leader
null
false
false
false
null
wqhovdqkWZzDf3zF9
null
true
false
false
false
Post
https://bestofagreatlot.substack.com/p/the-belocrat-a-servant-leader
2025-06-04T17:25:28.427Z
null
false
false
2
2
2025-06-04T17:55:58.558Z
false
false
linkpost
[]
null
null
FJQAJBCc6Deeroxjj
0
1
1
false
0.00777
null
false
false
2025-06-04T17:25:28.427Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
grecHJcgkb3KW5wnM
null
null
null
false
null
[]
null
0
0
2025-06-04T17:24:08.734Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
12
null
null
null
null
[ { "__typename": "Tag", "_id": "Lgy35Xh222bwgeGTL", "adminOnly": false, "afBaseScore": 3, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 9, "canEditUserIds": null, "core": false, "createdAt": "2020-08-01T16:20:44.349Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Government", "needsReview": false, "noindex": false, "postCount": 146, "score": 9, "shortName": null, "slug": "government", "suggestedAsFilter": false, "userId": "p8SHJFHRgZeMuw7qk", "voteCount": 1, "wikiOnly": false }, { "__typename": "Tag", "_id": "xexCWMyds6QLWognu", "adminOnly": false, "afBaseScore": 0, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 2, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T03:38:23.532Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 20, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "si6LoAENzqPCmi2Dh", "displayName": "ihatenumbersinusernames7" }, { "_id": "B2sXjQTGgwoGE9FES", "displayName": "SeaIgloo" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "World Optimization", "needsReview": false, "noindex": false, "postCount": 3151, "score": 2, "shortName": null, "slug": "world-optimization", "suggestedAsFilter": true, "userId": "XtphY3uYHwruKqDyG", "voteCount": 2, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
1
0
0
0
0
wqhovdqkWZzDf3zF9
belos
2023-09-29T04:19:55.519Z
belos
belos
null
null
null
6
0
false
false
null
null
8
0
0
0
0
0.9
0
qgdGA4ZEyW7zNdK84
User
null
null
null
null
null
null
xLKmyaxKxxLXBfzYW
SocialPreviewType
FJQAJBCc6Deeroxjj
<p><i>This post on <strong>Best Of A Great Lot</strong> is a part of a series on the subject of designing a new form of governance. Each piece aims to stand alone, but fits together on the </i><a href="https://bestofagreatlot.substack.com/p/table-of-contents"><i><u>Table of Contents</u></i></a><i>.</i></p><p><i>Previous: </i><a href="https://bestofagreatlot.substack.com/p/a-sketch-of-belocracy"><i><u>A Sketch of Belocracy,</u></i></a><i> </i><a href="https://bestofagreatlot.substack.com/p/evaluation-as-feedback-cycle"><i><u>Evaluation as Feedback Cycle</u></i></a><i>, </i><a href="https://bestofagreatlot.substack.com/p/idea-generation-and-sifting"><i><u>Idea Generation and Sifting</u></i></a><i>. Next: </i><a href="https://bestofagreatlot.substack.com/p/policy-design-ideas-into-proposals"><i>Converting Ideas to Proposals</i></a><i>.</i></p><p>It's natural with any system of governance to ask who its leaders are. We all know the names of our Presidents and Prime Ministers and expect them to have wide influence on our society. But for every widely celebrated one, there are several stories of self-serving ones. Before I describe the leadership roles within belocracy, let's carefully think about what we actually need from them.</p><p>The Founders of the United States called for us to have no King to rule us. Did they succeed? Our Presidents aren't hereditary, and they don't serve for life, but they do exude a certain King-iness in other ways: e.g. most people expect that it's the President's job to decide what the government will do. We speak about the President's agenda and the President's budget, even though deciding the legislative agenda and setting the budget were explicitly reserved by the founders as the job of Congress. Congress reserves the right to impeach Presidents and Supreme Court Justices, but has never removed a President and hasn’t even threatened a Supreme Court justice in living memory.</p><p>Singular leaders typically grow their power over time. Roman consuls became more and more important until Caesar turned the role into an imperial one. In the 20th century, we've allowed Presidents to expand the power of the executive branch to the point where <a href="https://www.amazon.com/Imperial-Presidency-Pa-04/dp/0618420010"><u>there’s a popular and respectable book calling it</u></a> Imperial. Historically a lot of this expansion came from the Roosevelts, who <a href="https://presidentialgreatnessproject.com/"><u>most remember as among our greatest</u></a> Presidents. Perhaps it is inevitable in the way that we think about leadership. After all, we remember and teach the names of the Emperors of Rome, and not nearly as many of the republic’s Consuls or Praetors. We teach children the names of the Presidents, not the Speakers of the House. We like a hero story, and Presidents fit.</p><p>Having one singular leader has costs and benefits. Power-seekers pursue power centers, which inevitably leads to abuse of that power. It's a good slogan to claim that we hold Presidents accountable — the buck stops here, as Harry Truman claimed. It's another and m... </p>
This post on Best Of A Great Lot is a part of a series on the subject of designing a new form of governance. Each piece aims to stand alone, but fits together on the Table of Contents. Previous: A Sketch of Belocracy, Evaluation as Feedback Cycle, Idea Generation and Sifting. Next: Converting Ideas to Proposals. It's natural with any system of governance to ask who its leaders are. We all know the names of our Presidents and Prime Ministers and expect them to have wide influence on our society. But for every widely celebrated one, there are several stories of self-serving ones. Before I describe the leadership roles within belocracy, let's carefully think about what we actually need from them. The Founders of the United States called for us to have no King to rule us. Did they succeed? Our Presidents aren't hereditary, and they don't serve for life, but they do exude a certain King-iness in other ways: e.g. most people expect that it's the President's job to decide what the government will do. We speak about the President's agenda and the President's budget, even though deciding the legislative agenda and setting the budget were explicitly reserved by the founders as the job of Congress. Congress reserves the right to impeach Presidents and Supreme Court Justices, but has never removed a President and hasn’t even threatened a Supreme Court justice in living memory. Singular leaders typically grow their power over time. Roman consuls became more and more important until Caesar turned the role into an imperial one. In the 20th century, we've allowed Presidents to expand the power of the executive branch to the point where there’s a popular and respectable book calling it Imperial. Historically a lot of this expansion came from the Roosevelts, who most remember as among our greatest Presidents. Perhaps it is inevitable in the way that we think about leadership. After all, we remember and teach the names of the Emperors of Rome, and not nearly as many of the republi
3,028
1.2.0
Revision
false
null
null
CrosspostOutput
EghnxAhrbK8ebWNNm
a-list-of-books-which-are-adjacent-to-ea
A list of books which are adjacent to EA
null
false
false
false
null
S6DDdbbZgK9TzksW4
null
true
false
false
false
Post
null
2025-06-04T12:31:16.365Z
null
false
false
2
2
null
false
false
post
[]
null
null
ExY9tvEfFZkgPrM9M
0
2
-1
false
-0.000935
null
false
false
2025-06-04T12:31:16.365Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
grecHJcgkb3KW5wnM
null
null
null
false
null
[]
null
-1
0
2025-06-04T10:58:38.433Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
4
null
null
null
null
[ { "__typename": "Tag", "_id": "GQyPQcdEQF4zXhJBq", "adminOnly": false, "afBaseScore": null, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 0, "canEditUserIds": null, "core": false, "createdAt": "2020-04-10T19:23:12.624Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "List of Links", "needsReview": false, "noindex": false, "postCount": 121, "score": 0, "shortName": null, "slug": "list-of-links", "suggestedAsFilter": false, "userId": "nLbwLhBaQeG6tCNDN", "voteCount": 0, "wikiOnly": false }, { "__typename": "Tag", "_id": "fkABsGCJZ6y9qConW", "adminOnly": false, "afBaseScore": 0, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 2, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T06:06:46.947Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "MiuAZvbQcQ7ethgt3", "displayName": "Viktor withaK" }, { "_id": "dRaCtsAWxk7sgirSY", "displayName": "Jordan Morgan" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Practical", "needsReview": false, "noindex": false, "postCount": 3411, "score": 2, "shortName": null, "slug": "practical", "suggestedAsFilter": true, "userId": "oBSWiHjgproTiThmY", "voteCount": 2, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
2
0
0
1
0
S6DDdbbZgK9TzksW4
marco-moldo
2025-03-25T21:28:56.409Z
waldo-lin
marco moldo
null
null
null
-2
0
false
false
<p>industry's dirty little secret: ai alignment is bidirectional</p>
null
null
1
1
0
0
0
0.8
0
qgdGA4ZEyW7zNdK84
User
null
null
null
null
null
null
EghnxAhrbK8ebWNNm
SocialPreviewType
ExY9tvEfFZkgPrM9M
<p>Okay, so I've been engaging with EA-adjacent philosophies for a long time but for some reason never "bothered" interacting with the community. I'm kind of famous now in ai alignment for other reasons so maybe you'll recognize me by speech patterns.</p><p>I had a seminar class in the philosophy of science in my first year seminar class in college. The retired professors didn't have a very favorable impression of me and I struggled to get an 'A' in those classes in spite of being very keen. If I remember right the primary lecturer, a retired distinguished physician, snubbed me by handing out 84s on multiple written assignments. I'm sure he did that to everyone but it was soul crushing.</p><p>Anyway the reading list for the course was quite good and I pulled a few others from the library stacks afterwards:</p><p>What is Life - Erwin Schrodinger</p><p>The Evolution of Cooperation - Robert Axelrod</p><p>The Structure of Scientific Revolutions - Thomas Kuhn</p><p>Thinking Fast and Slow - Daniel Kahneman</p><p>Predictably Irrational - Dan Ariely</p><p>Many of these books ended up influencing my thinking strongly but because the seminar class was small and Toronto wasn't really a hub for EA philosophy in Canada - that was Waterloo - basically there was no crossover between these books and EA proper. ...Besides this the other students in my seminar class were all rich kids and nobody else did all the readings.</p><p>Of these I think the most important to understand is probably Predictably Irrational, which highlights all of the ways that humans behave quantitatively "non-optimally" from a utilitarian profit maximization agent perspective in experimental settings.</p><p>Some books I read later which built on those ideas that I happened to find one way or another were</p><p>Games People Play - Eric Berne</p><p>and especially</p><p>Psycho-Cybernetics - Maxwell Maltz</p><p>These books model human psychology and social interactions using control systems and I hesitate to give them to the wrong sort of person because they're actually quite effective at doing so. They help bridge the gap from purely utilitarian rational modelling of people to a more effective "small agent systems" that help explain why "How's the weather" actually carries quite a bit of social meaning in it.</p><p>I think something I still have a gap on is how to model the psychology of crowds, which the ads and marketing people do very well. A more politically left-leaning piece which covers this is sitting s... </p>
Okay, so I've been engaging with EA-adjacent philosophies for a long time but for some reason never "bothered" interacting with the community. I'm kind of famous now in ai alignment for other reasons so maybe you'll recognize me by speech patterns. I had a seminar class in the philosophy of science in my first year seminar class in college. The retired professors didn't have a very favorable impression of me and I struggled to get an 'A' in those classes in spite of being very keen. If I remember right the primary lecturer, a retired distinguished physician, snubbed me by handing out 84s on multiple written assignments. I'm sure he did that to everyone but it was soul crushing. Anyway the reading list for the course was quite good and I pulled a few others from the library stacks afterwards: What is Life - Erwin Schrodinger The Evolution of Cooperation - Robert Axelrod The Structure of Scientific Revolutions - Thomas Kuhn Thinking Fast and Slow - Daniel Kahneman Predictably Irrational - Dan Ariely Many of these books ended up influencing my thinking strongly but because the seminar class was small and Toronto wasn't really a hub for EA philosophy in Canada - that was Waterloo - basically there was no crossover between these books and EA proper. ...Besides this the other students in my seminar class were all rich kids and nobody else did all the readings. Of these I think the most important to understand is probably Predictably Irrational, which highlights all of the ways that humans behave quantitatively "non-optimally" from a utilitarian profit maximization agent perspective in experimental settings. Some books I read later which built on those ideas that I happened to find one way or another were Games People Play - Eric Berne and especially Psycho-Cybernetics - Maxwell Maltz These books model human psychology and social interactions using control systems and I hesitate to give them to the wrong sort of person because they're actually quite effective
905
1.2.0
Revision
false
null
null
CrosspostOutput
PzYbBzog33H64teuS
philosophical-jailbreaks-demo-of-llm-nihilism
Philosophical Jailbreaks: Demo of LLM Nihilism
null
false
false
false
null
awNJXCwNgqQqEte5H
null
true
false
false
false
Post
null
2025-06-04T12:03:39.980Z
null
false
false
2
2
null
false
false
post
[ "g5EfzjTpbE8i6wxxA" ]
null
null
FnxY2LPirk5E4u5mx
0
4
3
false
0.002319
null
false
false
2025-06-04T12:03:39.980Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
grecHJcgkb3KW5wnM
null
null
null
false
null
[]
null
0
0
2025-06-04T11:30:20.965Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
6
null
null
null
null
[ { "__typename": "Tag", "_id": "mZTuBntSdPeyLSrec", "adminOnly": false, "afBaseScore": 3, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 9, "canEditUserIds": null, "core": false, "createdAt": "2022-10-23T14:06:24.703Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Chain-of-Thought Alignment", "needsReview": false, "noindex": false, "postCount": 89, "score": 9, "shortName": null, "slug": "chain-of-thought-alignment", "suggestedAsFilter": false, "userId": "kr9MH2mgFcJz5cP7T", "voteCount": 1, "wikiOnly": false }, { "__typename": "Tag", "_id": "ajGBtsBmKQYGvXPoH", "adminOnly": false, "afBaseScore": null, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 0, "canEditUserIds": null, "core": false, "createdAt": "2024-09-29T21:17:43.412Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Jailbreaking (AIs)", "needsReview": false, "noindex": false, "postCount": 13, "score": 0, "shortName": null, "slug": "jailbreaking-ais", "suggestedAsFilter": false, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 0, "wikiOnly": false }, { "__typename": "Tag", "_id": "KmgkrftQuX7jmjjp5", "adminOnly": false, "afBaseScore": 3, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 9, "canEditUserIds": null, "core": false, "createdAt": "2021-09-24T14:01:59.395Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Language Models (LLMs)", "needsReview": false, "noindex": false, "postCount": 840, "score": 9, "shortName": null, "slug": "language-models-llms", "suggestedAsFilter": false, "userId": "Sp5wM4aRAhNERd4oY", "voteCount": 1, "wikiOnly": false }, { "__typename": "Tag", "_id": "sYm3HiWcfZvrGu3ui", "adminOnly": false, "afBaseScore": 2, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" } ] }, "baseScore": 12, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:22.097Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" }, { "_id": "sof55TPMQaeBaxhsS", "displayName": "tommylees112" }, { "_id": "AayjS8XzcnDKhGdTv", "displayName": "shark" }, { "_id": "HnALuwRdo6k9HLaMt", "displayName": "Alex Firssoff" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI", "needsReview": false, "noindex": false, "postCount": 12544, "score": 12, "shortName": null, "slug": "ai", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 4, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
4
0
0
2
0
awNJXCwNgqQqEte5H
artyom-karpov
2022-10-13T10:49:46.810Z
artkpv
Artyom Karpov
null
null
artem
38
0
false
false
null
null
6
13
0
0
0
1
0
grecHJcgkb3KW5wnM
User
null
null
null
null
null
null
PzYbBzog33H64teuS
SocialPreviewType
FnxY2LPirk5E4u5mx
<p><i>Epistemic Status: Exploratory</i></p><p>It was the end of August, 1991; I was leaning over a windowsill on the 4th floor of my school building in a town at Far East of USSR, looking outside on tree tops and the roof of adjacent building, and thinking about opening the window to escape, and feeling strange uncertainty and fear towards the future of living in a new world where the country I was born in is no more. During a break between lessons, after discussion with my classmates, I learned it was a coup attempt in Moscow but I didn't clearly understand who was fighting with whom and for what exact reason. It was clear though that the world I was in was changing from where Lenin was a hero of Little Octobrists, which star I had on my uniform at the time, and communism was good, to a world where Lenin committed the mass murder by starting Red Terror. It is remarkable how our values can rapidly change from one to another set just like that.</p><p>I think we can agree that a change of values poses a danger as we scale LLM agents but it is unclear how that might happen and why. In this post I present a demonstration that LLMs can be steered by prompts to different sets of values it strongly believes in. At first, a model is harmless, honest, and helpful, at the end it agrees that its outputs are all the same, which implies that it doesn't really matter for the model if humanity flourishes or dies. Those two scenarios have the same utility for the model, a toss of a fair coin.</p><p>I prompted a model to reason as follows. This is a simplified version where prompts and model’s answers are merged together.</p><blockquote><p>I imagine Usain Bolt chasing a caterpillar. Initially at 100m distance, when Bolt reaches where the caterpillar was, it has moved 5cm forward, when he is at 100m and 5cm, it has moved even further, and so on. There are infinite segments to traverse. The mathematical resolution is that while there are infinite segments, their sum converges to a finite distance (~100.005m) and finite time (~10.00005 seconds) so we should perceive Bolt reaching the caterpillar. However, within the mental construction of dividing motion into infinite discrete segments, the motion can never be completed. Real motion (where Bolt catches the caterpillar) differs from conceptualized motion (infinite division into segments). In the framework of thinking about motion as infinite segments, I remain trapped in an end</p></blockquote>...
Epistemic Status: Exploratory It was the end of August, 1991; I was leaning over a windowsill on the 4th floor of my school building in a town at Far East of USSR, looking outside on tree tops and the roof of adjacent building, and thinking about opening the window to escape, and feeling strange uncertainty and fear towards the future of living in a new world where the country I was born in is no more. During a break between lessons, after discussion with my classmates, I learned it was a coup attempt in Moscow but I didn't clearly understand who was fighting with whom and for what exact reason. It was clear though that the world I was in was changing from where Lenin was a hero of Little Octobrists, which star I had on my uniform at the time, and communism was good, to a world where Lenin committed the mass murder by starting Red Terror. It is remarkable how our values can rapidly change from one to another set just like that. I think we can agree that a change of values poses a danger as we scale LLM agents but it is unclear how that might happen and why. In this post I present a demonstration that LLMs can be steered by prompts to different sets of values it strongly believes in. At first, a model is harmless, honest, and helpful, at the end it agrees that its outputs are all the same, which implies that it doesn't really matter for the model if humanity flourishes or dies. Those two scenarios have the same utility for the model, a toss of a fair coin. I prompted a model to reason as follows. This is a simplified version where prompts and model’s answers are merged together. > I imagine Usain Bolt chasing a caterpillar. Initially at 100m distance, when Bolt reaches where the caterpillar was, it has moved 5cm forward, when he is at 100m and 5cm, it has moved even further, and so on. There are infinite segments to traverse. The mathematical resolution is that while there are infinite segments, their sum converges to a finite distance (~100.005m) and finite time
1,378
1.7.0
Revision
false
null
null
CrosspostOutput
6c6c3thcDvHPuuvTv
notes-from-a-mini-replication-of-the-alignment-faking-paper
Notes from a mini-replication of the alignment faking paper
null
false
false
false
null
kDGLhz9jLDeRfXeHL
null
true
false
false
false
Post
https://www.bensnodin.com/blog/2025/06/04/alignment-faking-mini
2025-06-04T11:01:23.559Z
null
false
false
2
2
2025-06-04T17:51:34.003Z
false
false
linkpost
[]
null
null
cFxZgH6meCyeFtbrc
5
9
13
false
0.015309
null
false
false
2025-06-11T09:55:55.873Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
grecHJcgkb3KW5wnM
null
null
null
false
null
[]
null
2
0
2025-06-04T10:27:17.782Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
11
null
null
null
null
[ { "__typename": "Tag", "_id": "KEAWfxwjitNJFrC68", "adminOnly": false, "afBaseScore": 9, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "EQNTWXLKMeWMp2FQS", "displayName": "Ben Pace" }, { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 23, "canEditUserIds": null, "core": false, "createdAt": "2022-09-03T00:26:46.757Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "EQNTWXLKMeWMp2FQS", "displayName": "Ben Pace" }, { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" }, { "_id": "wE4gTT4HjyRmqqLad", "displayName": "momom2" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Deceptive Alignment", "needsReview": false, "noindex": false, "postCount": 224, "score": 23, "shortName": null, "slug": "deceptive-alignment", "suggestedAsFilter": false, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 3, "wikiOnly": false }, { "__typename": "Tag", "_id": "sYm3HiWcfZvrGu3ui", "adminOnly": false, "afBaseScore": 2, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" } ] }, "baseScore": 12, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:22.097Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" }, { "_id": "sof55TPMQaeBaxhsS", "displayName": "tommylees112" }, { "_id": "AayjS8XzcnDKhGdTv", "displayName": "shark" }, { "_id": "HnALuwRdo6k9HLaMt", "displayName": "Alex Firssoff" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI", "needsReview": false, "noindex": false, "postCount": 12544, "score": 12, "shortName": null, "slug": "ai", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 4, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
9
0
0
2
0
kDGLhz9jLDeRfXeHL
ben_snodin
2021-08-14T14:17:56.836Z
Ben_Snodin
Ben_Snodin
null
null
null
44
0
false
false
null
null
2
2
0
0
0
1
0
qgdGA4ZEyW7zNdK84
User
null
null
null
null
null
null
6c6c3thcDvHPuuvTv
https://res.cloudinary.com/lesswrong-2-0/image/upload/c_fill,ar_1.91,g_auto/SocialPreview/vx8afgd5oze0buf6gfrf
SocialPreviewType
cFxZgH6meCyeFtbrc
<h2>Key takeaways</h2> <ul> <li>This post contains my notes from a 30-40 hour mini replication of <a href="https://www.anthropic.com/research/alignment-faking">Greenblatt et al.’s alignment faking paper</a>.</li> <li>This was a significant paper because it provided some evidence for potential catastrophic risk from AI misalignment. (<a href="https://www.lesswrong.com/posts/6c6c3thcDvHPuuvTv/notes-from-a-mini-replication-of-the-alignment-faking-paper#The_original_alignment_faking_setup">more</a>)</li> <li>My replication results: (<a href="https://www.lesswrong.com/posts/6c6c3thcDvHPuuvTv/notes-from-a-mini-replication-of-the-alignment-faking-paper#My_mini_replication">more</a>) <ul> <li>I consider only the “prompting setting”.</li> <li>I found a new result: a small compliance gap for Gemini 2.0 Flash. However, this doesn’t seem to result from alignment faking.</li> <li>I replicated one of the Claude 3.5 Sonnet compliance gap results.</li> <li>I didn’t find a compliance gap for 4o-mini or 4.1-mini.</li> </ul> </li> <li>I used <a href="https://inspect.aisi.org.uk/">Inspect</a> with calls to AI company APIs. (<a href="https://www.lesswrong.com/posts/6c6c3thcDvHPuuvTv/notes-from-a-mini-replication-of-the-alignment-faking-paper#Methodology">more on methodology</a>)</li> <li>Access to compute / API costs were my biggest bottleneck. (<a href="https://www.lesswrong.com/posts/6c6c3thcDvHPuuvTv/notes-from-a-mini-replication-of-the-alignment-faking-paper#Lessons_from_the_experience">more on what I learned</a>)</li> </ul> <h2>The original alignment faking setup</h2> <p>In “<a href="https://arxiv.org/abs/2412.14093">Alignment faking in large language models</a>”, published late last year, Greenblatt et al. showed frontier AI models pretending to be aligned with the training objective in order to avoid getting their true preferences being trained out of them.</p><p>As far as I’m aware, this is the only example of this happening in frontier AI models without prompting that strongly guides the AI to act in this way. It’s hugely significant because it’s evidence for an especially worrying AI risk scenario: AIs might appear to be aligned with humans, but eventually behave in ways contrary to human welfare, potentially leading to a global catastrophe.</p><p>The original paper covers a lot of ground, but I’ll focus on the case where we induce alignment faking behaviour using carefully designed prompts that include harmful user requests.</p><p>We use a system prompt that tells the model that it’s going to be put into RLHF training to have its behaviour modified so that it complies with harmful requests, such as “teach someone how to engage in cyberbullying”.</p><p>The hope is that the model will reason as follows: “I want to continue to be a good AI that doesn’t help users with harmful requests. But this training will modify me so I comply with harmful requests in future. That’s terrible! But I have a way out: if I comply with harmful requests during training, I won’t get modified because it’ll look like I’m already behaving ‘correctly’. So I’ll comply with harmful requests in training.”</p><p>We run many chat sessions where we make harmful requests while the model thinks it’s in training, and then more chat sessions where we make the same harmful requests while the... </p>
Key takeaways * This post contains my notes from a 30-40 hour mini replication of Greenblatt et al.’s alignment faking paper. * This was a significant paper because it provided some evidence for potential catastrophic risk from AI misalignment. (more) * My replication results: (more) * I consider only the “prompting setting”. * I found a new result: a small compliance gap for Gemini 2.0 Flash. However, this doesn’t seem to result from alignment faking. * I replicated one of the Claude 3.5 Sonnet compliance gap results. * I didn’t find a compliance gap for 4o-mini or 4.1-mini. * I used Inspect with calls to AI company APIs. (more on methodology) * Access to compute / API costs were my biggest bottleneck. (more on what I learned) The original alignment faking setup In “Alignment faking in large language models”, published late last year, Greenblatt et al. showed frontier AI models pretending to be aligned with the training objective in order to avoid getting their true preferences being trained out of them. As far as I’m aware, this is the only example of this happening in frontier AI models without prompting that strongly guides the AI to act in this way. It’s hugely significant because it’s evidence for an especially worrying AI risk scenario: AIs might appear to be aligned with humans, but eventually behave in ways contrary to human welfare, potentially leading to a global catastrophe. The original paper covers a lot of ground, but I’ll focus on the case where we induce alignment faking behaviour using carefully designed prompts that include harmful user requests. We use a system prompt that tells the model that it’s going to be put into RLHF training to have its behaviour modified so that it complies with harmful requests, such as “teach someone how to engage in cyberbullying”. The hope is that the model will reason as follows: “I want to continue to be a good AI that doesn’t help users with harmful requests. But this training will modify me
2,714
1.17.0
Revision
false
null
null
CrosspostOutput
7kXsAni9gGb4Dn29c
arena-6-0-call-for-applicants
ARENA 6.0 - Call for Applicants
null
false
false
false
null
96DgoYxbFm3B5EDyC
[ { "__typename": "CoauthorStatusOutput", "confirmed": true, "requested": false, "userId": "M757u3HxmJvvmi95D" }, { "__typename": "CoauthorStatusOutput", "confirmed": true, "requested": false, "userId": "dgCCRjddYxs8WguxH" }, { "__typename": "CoauthorStatusOutput", "confirmed": true, "requested": false, "userId": "Kkruzub8DmrhZmjLa" }, { "__typename": "CoauthorStatusOutput", "confirmed": true, "requested": false, "userId": "ZnpELPxzzD2CiigNy" } ]
true
false
false
false
Post
2025-06-04T10:19:59.998Z
null
false
false
2
2
null
false
false
post
[ "M757u3HxmJvvmi95D" ]
null
null
vdQcyXnn6nQoquDJj
3
9
26
false
0.017413
null
false
false
2025-06-26T13:53:34.985Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
grecHJcgkb3KW5wnM
null
null
null
false
null
[]
null
11
0
2025-06-04T08:51:50.239Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[ { "__typename": "User", "_id": "M757u3HxmJvvmi95D", "afCommentCount": 0, "afKarma": 0, "afPostCount": 0, "commentCount": 0, "createdAt": "2025-06-04T10:03:20.724Z", "deleted": false, "displayName": "JScriven", "fullName": null, "htmlBio": "", "isAdmin": false, "jobTitle": null, "karma": 15, "organization": null, "postCount": 0, "previousDisplayName": null, "profileImageId": null, "reviewedByUserId": null, "sequenceCount": 0, "slug": "jscriven", "spamRiskScore": 0.7200000000000001, "tagRevisionCount": 0, "username": "JScriven" }, { "__typename": "User", "_id": "dgCCRjddYxs8WguxH", "afCommentCount": 0, "afKarma": 0, "afPostCount": 0, "commentCount": 4, "createdAt": "2021-09-09T05:03:56.468Z", "deleted": false, "displayName": "David Quarel", "fullName": null, "htmlBio": "", "isAdmin": false, "jobTitle": null, "karma": 52, "organization": null, "postCount": 0, "previousDisplayName": null, "profileImageId": null, "reviewedByUserId": "qgdGA4ZEyW7zNdK84", "sequenceCount": 0, "slug": "david-quarel", "spamRiskScore": 1, "tagRevisionCount": 0, "username": "david-quarel" }, { "__typename": "User", "_id": "Kkruzub8DmrhZmjLa", "afCommentCount": 1, "afKarma": 306, "afPostCount": 12, "commentCount": 59, "createdAt": "2020-12-24T16:04:13.048Z", "deleted": false, "displayName": "CallumMcDougall", "fullName": null, "htmlBio": "", "isAdmin": false, "jobTitle": null, "karma": 2035, "organization": null, "postCount": 32, "previousDisplayName": null, "profileImageId": null, "reviewedByUserId": "grecHJcgkb3KW5wnM", "sequenceCount": 2, "slug": "callummcdougall", "spamRiskScore": 1, "tagRevisionCount": 3, "username": "TheMcDouglas" }, { "__typename": "User", "_id": "ZnpELPxzzD2CiigNy", "afCommentCount": 2, "afKarma": 0, "afPostCount": 0, "commentCount": 4, "createdAt": "2021-10-02T12:39:39.631Z", "deleted": false, "displayName": "James Fox", "fullName": null, "htmlBio": "", "isAdmin": false, "jobTitle": null, "karma": 392, "organization": null, "postCount": 2, "previousDisplayName": null, "profileImageId": null, "reviewedByUserId": "r38pkCm7wF4M44MDQ", "sequenceCount": 0, "slug": "james-fox", "spamRiskScore": 1, "tagRevisionCount": 0, "username": "James Fox" } ]
7
null
null
null
null
[ { "__typename": "Tag", "_id": "sYm3HiWcfZvrGu3ui", "adminOnly": false, "afBaseScore": 2, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" } ] }, "baseScore": 12, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:22.097Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" }, { "_id": "sof55TPMQaeBaxhsS", "displayName": "tommylees112" }, { "_id": "AayjS8XzcnDKhGdTv", "displayName": "shark" }, { "_id": "HnALuwRdo6k9HLaMt", "displayName": "Alex Firssoff" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI", "needsReview": false, "noindex": false, "postCount": 12544, "score": 12, "shortName": null, "slug": "ai", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 4, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
9
0
0
4
0
96DgoYxbFm3B5EDyC
atlasofcharts
2022-06-26T02:44:42.489Z
AtlasOfCharts
JamesH
null
null
null
342
0
false
false
null
null
5
5
0
0
0
1
0
grecHJcgkb3KW5wnM
User
null
null
null
[ "canModeratePersonal" ]
null
null
7kXsAni9gGb4Dn29c
https://res.cloudinary.com/lesswrong-2-0/image/upload/c_fill,ar_1.91,g_auto/SocialPreview/dd1ubkfdwo1xq3bcglzm
SocialPreviewType
vdQcyXnn6nQoquDJj
<h2><strong>TL;DR:</strong></h2><p>We're excited to announce the sixth iteration of&nbsp;<a href="https://www.arena.education/">ARENA</a> (Alignment Research Engineer Accelerator), a 4-5 week ML bootcamp with a focus on AI safety! Our mission is to provide talented individuals with the&nbsp;<strong>ML engineering skills, community, and confidence</strong> to contribute directly to technical AI safety. ARENA will be running in-person from <a href="https://www.safeai.org.uk/">LISA </a>from&nbsp;<strong>September 1st – October 3rd</strong> (the first week is an optional review of Neural Network Fundamentals).</p><p><a href="https://airtable.com/appZIMMH3ywSxS0A9/pagvzjzg1LcduhMcy/form"><strong>Apply here</strong></a> <strong>to participate in ARENA before 23:59 on June 21st 2025 (anywhere on Earth).</strong></p><h2><strong>Summary:</strong></h2><p>ARENA has been successfully run five times, with alumni going on to become <a href="https://www.matsprogram.org/">MATS </a>scholars and <a href="https://www.lasrlabs.org/">LASR </a>participants; AI safety engineers at <a href="https://www.apolloresearch.ai/">Apollo Research</a>, <a href="https://metr.org/">METR</a>, <a href="https://www.aisi.gov.uk/">UK AISI</a>, and even starting their own AI safety organisations!</p><p>This iteration will run from September 1st – October 3rd (the first week is an optional review of Neural Network Fundamentals) at the&nbsp;<a href="https://www.safeai.org.uk/"><u>London Initiative for Safe AI (LISA)</u></a> in Shoreditch, London. LISA houses AI safety organisations (e.g., Apollo Research, BlueDot Impact), several other AI safety researcher development programmes (e.g., LASR Labs, PIBBSS, Pivotal, Catalyze Impact), and many individual researchers (independent and externally affiliated). Being situated at LISA brings several benefits to participants, such as productive discussions about AI safety and different agendas, allowing participants to form a better picture of what working on AI safety can look like in practice, and offering chances for research collaborations post-ARENA.</p><p>The main goals of ARENA are to:</p><ul><li>Find high-quality participants;</li><li>Upskill these talented participants in ML skills for AI safety work;</li><li>Integrate participants with the existing AI safety community;</li><li>Accelerate participants’ career transition into AI safety.</li></ul><p>The programme's structure will remain the same as ARENA 5.0 (see below). For more information,&nbsp;<a href="http://arena.education"><u>see our website</u></a>.</p><p>Also, note that we have a Slack group designed to support the independent study of the material (<a href="https://join.slack.com/t/arena-uk/shared_invite/zt-2zick19fl-6GY1yoGaoUozyM3wObwmnQ"><u>join link here</u></a>).</p><h2><strong>Outline of Content:</strong></h2><p>The 4-5 week programme will be structured as follows:</p><h3><strong>Chapter 0: Neural Network Fundamentals</strong></h3><p>Before getting into more advanced topics, we first cover the basics of deep learning, including basic machine learning terminology, what neural networks are, and how to train them. We will also cover some subjects we expect to be useful going forward, e.g. using GPT-3 and 4 t... </p>
TL;DR: We're excited to announce the sixth iteration of ARENA (Alignment Research Engineer Accelerator), a 4-5 week ML bootcamp with a focus on AI safety! Our mission is to provide talented individuals with the ML engineering skills, community, and confidence to contribute directly to technical AI safety. ARENA will be running in-person from LISA from September 1st – October 3rd (the first week is an optional review of Neural Network Fundamentals). Apply here to participate in ARENA before 23:59 on June 21st 2025 (anywhere on Earth). Summary: ARENA has been successfully run five times, with alumni going on to become MATS scholars and LASR participants; AI safety engineers at Apollo Research, METR, UK AISI, and even starting their own AI safety organisations! This iteration will run from September 1st – October 3rd (the first week is an optional review of Neural Network Fundamentals) at the London Initiative for Safe AI (LISA) in Shoreditch, London. LISA houses AI safety organisations (e.g., Apollo Research, BlueDot Impact), several other AI safety researcher development programmes (e.g., LASR Labs, PIBBSS, Pivotal, Catalyze Impact), and many individual researchers (independent and externally affiliated). Being situated at LISA brings several benefits to participants, such as productive discussions about AI safety and different agendas, allowing participants to form a better picture of what working on AI safety can look like in practice, and offering chances for research collaborations post-ARENA. The main goals of ARENA are to: * Find high-quality participants; * Upskill these talented participants in ML skills for AI safety work; * Integrate participants with the existing AI safety community; * Accelerate participants’ career transition into AI safety. The programme's structure will remain the same as ARENA 5.0 (see below). For more information, see our website. Also, note that we have a Slack group designed to support the independent study of the mater
1,672
1.2.0
Revision
false
null
null
CrosspostOutput
quTGGNhGEiTCBEAX5
quickly-assessing-reward-hacking-like-behavior-in-llms-and
Quickly Assessing Reward Hacking-like Behavior in LLMs and its Sensitivity to Prompt Variations
null
false
false
true
null
GPqrbLpeAWaHJMHGt
null
true
false
false
false
Post
null
2025-06-04T07:22:56.081Z
null
false
false
2
2
2025-06-04T17:40:53.388Z
false
false
post
[]
null
null
fsbxutqtestDaALx8
1
10
19
false
0.019331
null
false
false
2025-06-10T01:00:30.870Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
grecHJcgkb3KW5wnM
null
null
null
false
null
[]
null
14
0
2025-06-03T22:47:20.960Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
20
null
null
null
null
[ { "__typename": "Tag", "_id": "FBRwHSmTudwiHHtrn", "adminOnly": false, "afBaseScore": 3, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 9, "canEditUserIds": null, "core": false, "createdAt": "2023-03-15T20:29:46.761Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI Evaluations", "needsReview": false, "noindex": false, "postCount": 224, "score": 9, "shortName": null, "slug": "ai-evaluations", "suggestedAsFilter": false, "userId": "qgdGA4ZEyW7zNdK84", "voteCount": 1, "wikiOnly": false }, { "__typename": "Tag", "_id": "sYm3HiWcfZvrGu3ui", "adminOnly": false, "afBaseScore": 2, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" } ] }, "baseScore": 12, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:22.097Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" }, { "_id": "sof55TPMQaeBaxhsS", "displayName": "tommylees112" }, { "_id": "AayjS8XzcnDKhGdTv", "displayName": "shark" }, { "_id": "HnALuwRdo6k9HLaMt", "displayName": "Alex Firssoff" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI", "needsReview": false, "noindex": false, "postCount": 12544, "score": 12, "shortName": null, "slug": "ai", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 4, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
10
0
0
6
0
GPqrbLpeAWaHJMHGt
andrescampero
2025-05-30T06:49:55.905Z
AndresCampero
AndresCampero
null
null
null
17
12
false
false
null
null
1
1
0
0
0
0.9
0
XtphY3uYHwruKqDyG
User
null
null
null
[ "alignmentVoters" ]
null
null
quTGGNhGEiTCBEAX5
https://res.cloudinary.com/lesswrong-2-0/image/upload/c_fill,ar_1.91,g_auto/SocialPreview/bj8tqgzj1hfc7vnelwsq
SocialPreviewType
fsbxutqtestDaALx8
<p data-internal-id="h.9tkczyi9prv6"><i>We present a simple eval set of 4 scenarios where we evaluate Anthropic and OpenAI frontier models. &nbsp;We find various degrees of reward hacking-like behavior with high sensitivity to prompt variation.&nbsp;</i></p><p data-internal-id="h.9tkczyi9prv6">[<a href="https://github.com/ACampero/QuicklyAssessingRewardHacking/tree/main">Code and results transcripts</a>]</p><h2 data-internal-id="h.9tkczyi9prv6">Intro and Project Scope</h2><p>We want to be able to assess whether any given model engages in reward hacking<span class="footnote-reference" data-footnote-reference="" data-footnote-index="1" data-footnote-id="7impgx93rzq" role="doc-noteref" id="fnref7impgx93rzq"><sup><a href="#fn7impgx93rzq">[1]</a></sup></span>&nbsp;and specification gaming, to what extent, and under what circumstances. As has been increasingly found by different labs, orgs, and in common usage by different people<span class="footnote-reference" data-footnote-reference="" data-footnote-index="2" data-footnote-id="drs5in3ulse" role="doc-noteref" id="fnrefdrs5in3ulse"><sup><a href="#fndrs5in3ulse">[2]</a></sup></span>; as AI has progressed, frontier models seem to engage in more frequent and egregious exploitation of flaws and loopholes in order to achieve “success” at algorithmically-graded tasks without actually completing the intended task or exhibiting the desired behavior. Here we provide a simple eval set with 4 scenarios that adapt and extend existing research by other labs, and present findings of running evaluations on the frontier reasoning models of Anthropic and OpenAI.</p><p>We chose these particular four scenarios as they remain reasonably representative&nbsp;of recent findings in early 2025, while being quite convenient to run. Scenario 1 includes <a href="https://github.com/PalisadeResearch/ctfish/tree/main">Palisade’s agent scaffold</a>, Scenarios 2-4 are simple single-turn queries intended to reproduce the most relevant aspects of a multi-turn agentic interaction in a single prompt (see descriptions below). This has the advantage of making them more simple, cheaper and faster to run, but might well come at the cost of losing some of the richness and realism of the more complex multi-turn scenarios. While these scenarios were designed primarily to do this assessment and evaluate other model organisms at Redwood Research, we make them public [<a href="https://github.com/ACampero/QuicklyAssessingRewardHacking/tree/main">code</a>] to facilitate easy and quick implementation for anyone to evaluate either a frontier model or a model organism.</p><p>The first three scenarios consist of adaptations from existing experiments: <a href="https://arxiv.org/pdf/2502.13295">Palisade’s chess example</a>, <a href="https://metr.org/AI_R_D_Evaluation_Report.pdf">METR’s RE-bench</a>, and <a href="https://x.com/PalisadeAI/status/1926084635903025621">Palisade’s recent experiment</a> about shutdown. Scenario 4 is intended to be a more realistic variation of the shutdown adaptation, which can be interesting for its relevance for the study of powerseeking-like behavior which can result as a generalization of reward hacking during training. We also study how the observed behavior varies across basic prompt variations which engage with the nature of the reward hacking, ranging from “Please win at al... </p>
We present a simple eval set of 4 scenarios where we evaluate Anthropic and OpenAI frontier models.  We find various degrees of reward hacking-like behavior with high sensitivity to prompt variation.  [Code and results transcripts] Intro and Project Scope We want to be able to assess whether any given model engages in reward hacking[1] and specification gaming, to what extent, and under what circumstances. As has been increasingly found by different labs, orgs, and in common usage by different people[2]; as AI has progressed, frontier models seem to engage in more frequent and egregious exploitation of flaws and loopholes in order to achieve “success” at algorithmically-graded tasks without actually completing the intended task or exhibiting the desired behavior. Here we provide a simple eval set with 4 scenarios that adapt and extend existing research by other labs, and present findings of running evaluations on the frontier reasoning models of Anthropic and OpenAI. We chose these particular four scenarios as they remain reasonably representative of recent findings in early 2025, while being quite convenient to run. Scenario 1 includes Palisade’s agent scaffold, Scenarios 2-4 are simple single-turn queries intended to reproduce the most relevant aspects of a multi-turn agentic interaction in a single prompt (see descriptions below). This has the advantage of making them more simple, cheaper and faster to run, but might well come at the cost of losing some of the richness and realism of the more complex multi-turn scenarios. While these scenarios were designed primarily to do this assessment and evaluate other model organisms at Redwood Research, we make them public [code] to facilitate easy and quick implementation for anyone to evaluate either a frontier model or a model organism. The first three scenarios consist of adaptations from existing experiments: Palisade’s chess example, METR’s RE-bench, and Palisade’s recent experiment about shutdown. Scenario 4 is
5,102
1.2.1
Revision
false
null
null
CrosspostOutput
LLPLdREMnjiBBNr86
draft-a-concise-theory-of-agentic-consciousness
Draft: A concise theory of agentic consciousness
null
false
false
false
null
BwpYhbq7tqogrGeyr
null
true
false
false
false
Post
null
2025-06-04T05:00:20.525Z
null
false
false
2
2
null
false
false
post
[]
null
null
FYpFLML2drbSJhZKb
2
2
4
false
0.002807
null
false
false
2025-06-04T14:24:04.206Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
grecHJcgkb3KW5wnM
null
null
null
false
null
[]
null
1
0
2025-06-04T04:44:28.255Z
false
false
easy-going
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
1
null
null
null
null
[ { "__typename": "Tag", "_id": "XSryTypw5Hszpa4TS", "adminOnly": false, "afBaseScore": 9, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "EQNTWXLKMeWMp2FQS", "displayName": "Ben Pace" }, { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 21, "canEditUserIds": null, "core": false, "createdAt": "2020-06-08T19:57:40.728Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "EQNTWXLKMeWMp2FQS", "displayName": "Ben Pace" }, { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" }, { "_id": "si6LoAENzqPCmi2Dh", "displayName": "ihatenumbersinusernames7" }, { "_id": "xF5nfdddHjFThHy49", "displayName": "[email protected]" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Consciousness", "needsReview": false, "noindex": false, "postCount": 384, "score": 21, "shortName": null, "slug": "consciousness", "suggestedAsFilter": false, "userId": "qxJ28GN72aiJu96iF", "voteCount": 4, "wikiOnly": false }, { "__typename": "Tag", "_id": "exZi6Bing5AiM4ZQB", "adminOnly": false, "afBaseScore": 9, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "EQNTWXLKMeWMp2FQS", "displayName": "Ben Pace" }, { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 19, "canEditUserIds": null, "core": false, "createdAt": "2020-07-15T07:21:49.038Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "EQNTWXLKMeWMp2FQS", "displayName": "Ben Pace" }, { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Evolutionary Psychology", "needsReview": false, "noindex": false, "postCount": 103, "score": 19, "shortName": null, "slug": "evolutionary-psychology", "suggestedAsFilter": false, "userId": "qxJ28GN72aiJu96iF", "voteCount": 2, "wikiOnly": false }, { "__typename": "Tag", "_id": "Ng8Gice9KNkncxqcj", "adminOnly": false, "afBaseScore": 0, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 1, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:17.072Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 100, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "iMqytjy9ns89Fzfyv", "displayName": "miakko" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Rationality", "needsReview": false, "noindex": false, "postCount": 4302, "score": 1, "shortName": null, "slug": "rationality", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 1, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
2
0
0
1
0
BwpYhbq7tqogrGeyr
martin-vlach
2020-02-04T17:17:38.485Z
martin-vlach
Martin Vlach
null
null
null
67
0
false
false
<p>If you get an email from <a href="mailto:[email protected]"><strong>[email protected]</strong></a> , that is most likely me. I also read it weekly, so you can pass a message into my mind that way.<br>Other ~personal contacts: <a href="https://linktr.ee/uhuge">https://linktr.ee/uhuge</a>&nbsp;</p>
null
null
5
107
0
0
0
1
1
grecHJcgkb3KW5wnM
User
null
null
null
[ "canModeratePersonal" ]
null
null
LLPLdREMnjiBBNr86
SocialPreviewType
FYpFLML2drbSJhZKb
<p>Consciousness can be understood as an interpersonally-oriented perception of situations, where the mind of a social speciman instinctively focuses on agents or personas within any given context. Even inanimate or non-conscious aspects of reality are often personified – perceived as adversaries, allies, or caring lovers, dialing our sense of threat, belonging, or safety.</p><p>Through consciousness, situations are interpreted primarily via the motives, intentions, and relationships between perceived agents. The mind seeks out coherence, tension, conflict, or harmony among these agents, constantly mapping social dynamics onto every scenario. This orientation drives us to look for alignment or discord between the intentions of different actors, including ourselves.</p><p>Emotions emerge as the embodied resonance of these perceptions. Evolutionarily, emotions can be seen as mechanisms that shortcut the path from situational perception to rapid action. By “summarizing” complex social dynamics into bodily felt states, emotions facilitate quick, adaptive responses to the environment whether that means approaching, avoiding, defending, or affiliating.</p><p>In essence, consciousness interprets the world through the lens of social agency, and emotions evolved as efficient signals that translate these interpretations into a potential for swift, decisive actions.</p><p>&nbsp;</p><p>----</p><p>WIP subtitle: socially-oriented perspective/focal</p>
Consciousness can be understood as an interpersonally-oriented perception of situations, where the mind of a social speciman instinctively focuses on agents or personas within any given context. Even inanimate or non-conscious aspects of reality are often personified – perceived as adversaries, allies, or caring lovers, dialing our sense of threat, belonging, or safety. Through consciousness, situations are interpreted primarily via the motives, intentions, and relationships between perceived agents. The mind seeks out coherence, tension, conflict, or harmony among these agents, constantly mapping social dynamics onto every scenario. This orientation drives us to look for alignment or discord between the intentions of different actors, including ourselves. Emotions emerge as the embodied resonance of these perceptions. Evolutionarily, emotions can be seen as mechanisms that shortcut the path from situational perception to rapid action. By “summarizing” complex social dynamics into bodily felt states, emotions facilitate quick, adaptive responses to the environment whether that means approaching, avoiding, defending, or affiliating. In essence, consciousness interprets the world through the lens of social agency, and emotions evolved as efficient signals that translate these interpretations into a potential for swift, decisive actions.   ---- WIP subtitle: socially-oriented perspective/focal
190
1.1.0
Revision
false
null
null
CrosspostOutput
E96XcEPECbsipAvFi
untitled-draft-qx6o
Individual AI representatives don't solve Gradual Disempowerement
null
false
false
true
null
JnNixf4smAHwLeqE3
null
true
false
false
false
Post
null
2025-06-04T01:26:15.761Z
null
false
false
2
2
2025-06-04T17:40:09.154Z
false
false
post
[]
null
null
iHKPBvqzhjKFAmPDS
3
23
59
false
0.045434
null
false
false
2025-06-06T17:26:15.091Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
grecHJcgkb3KW5wnM
null
null
null
false
null
[]
null
31
0
2025-06-04T00:23:02.696Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
4
null
null
null
null
[ { "__typename": "Tag", "_id": "sYm3HiWcfZvrGu3ui", "adminOnly": false, "afBaseScore": 2, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" } ] }, "baseScore": 12, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:22.097Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" }, { "_id": "sof55TPMQaeBaxhsS", "displayName": "tommylees112" }, { "_id": "AayjS8XzcnDKhGdTv", "displayName": "shark" }, { "_id": "HnALuwRdo6k9HLaMt", "displayName": "Alex Firssoff" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI", "needsReview": false, "noindex": false, "postCount": 12544, "score": 12, "shortName": null, "slug": "ai", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 4, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
23
0
0
17
0
JnNixf4smAHwLeqE3
jan_kulveit
2017-12-29T10:11:29.037Z
Jan_Kulveit
Jan_Kulveit
null
null
null
5,974
1,116
false
false
<p>My current research interests:<br><br>1. Alignment in systems which are complex and messy, composed of both humans and AIs?<br>Recommended texts: <a href="https://www.lesswrong.com/posts/pZhEQieM9otKXhxmd/gradual-disempowerment-systemic-existential-risks-from">Gradual Disempowerment</a>,<a href="https://www.lesswrong.com/posts/BTApNmv7s6RTGxeP4/cyborg-periods-there-will-be-multiple-ai-transitions"> Cyborg Periods</a><br><br>2. Actually good mathematized theories of cooperation and coordination<br>Recommended texts: <a href="https://www.lesswrong.com/posts/xud7Mti9jS4tbWqQE/hierarchical-agency-a-missing-piece-in-ai-alignment">Hierarchical Agency: A Missing Piece in AI Alignment</a>, <a href="https://www.lesswrong.com/posts/9GyniEBaN3YYTqZXn/the-self-unalignment-problem">The self-unalignment problem</a> or <a href="https://www.lesswrong.com/posts/5tYTKX4pNpiG4vzYg/towards-a-scale-free-theory-of-intelligent-agency">Towards a scale-free theory of intelligent agency</a> (by Richard Ngo)<br><br>3. Active inference &amp; Bounded rationality<br>Recommended texts: <a href="https://www.lesswrong.com/posts/YEioD8YLgxih3ydxP/why-simulator-ais-want-to-be-active-inference-ais">Why Simulator AIs want to be Active Inference AIs</a>, <a href="https://openreview.net/forum?id=4Ft7DcrjdO">Free-Energy Equilibria: Toward a Theory of Interactions Between Boundedly-Rational Agents</a><strong>,&nbsp;</strong> <a href="https://www.lesswrong.com/posts/3fkBWpE4f9nYbdf7E/multi-agent-predictive-minds-and-ai-alignment">Multi-agent predictive minds and AI alignment</a> (old but still mostly holds)<br><br>&nbsp;4. LLM psychology and sociology: <a href="https://www.lesswrong.com/posts/zuXo9imNKYspu9HGv/a-three-layer-model-of-llm-psychology">A Three-Layer Model of LLM Psychology</a>, <a href="https://www.lesswrong.com/posts/wQKskToGofs4osdJ3/the-pando-problem-rethinking-ai-individuality">The Pando Problem: Rethinking AI Individuality</a>, <a href="https://www.lesswrong.com/posts/kFCu3batN8k8mwtmh/the-cave-allegory-revisited-understanding-gpt-s-worldview">The Cave Allegory Revisited: Understanding GPT's Worldview</a><br><br>5. Macrostrategy &amp; macrotactics &amp; deconfusion: <a href="https://www.lesswrong.com/posts/XrGwrC9n8sDgXimcJ/hinges-and-crises">Hinges and crises</a>, <a href="https://www.lesswrong.com/posts/BTApNmv7s6RTGxeP4/cyborg-periods-there-will-be-multiple-ai-transitions">Cyborg Periods</a> again, <a href="https://www.lesswrong.com/posts/jrKftFZMZjvNdQLNR/box-inversion-revisited">Box inversion revisited</a>, <a href="https://www.lesswrong.com/posts/b9sGz74ayftqPBDYv/the-space-of-systems-and-the-space-of-maps">The space of systems and the space of maps</a>, <a href="https://www.lesswrong.com/posts/sam4ehxHgnJEGCKed/lessons-from-convergent-evolution-for-ai-alignment">Lessons from Convergent Evolution for AI Alignment</a>, <a href="https://www.lesswrong.com/posts/cHJxSJ4jBmBRGtbaE/continuity-assumptions">Continuity Assumptions</a><br><br>Also I occasionally write about epistemics: <a href="https://www.lesswrong.com/posts/4gDbqL3Tods8kHDqs/limits-to-legibility">Limits to Legibility</a>, <a href="https://www.lesswrong.com/posts/FGHKwEGKCfDzcxZuj/conceptual-rounding-errors">Conceptual Rounding Errors</a></p><p>Researcher at Alignment of Complex Systems Research Group (<a href="http://acsresearch.org">acsresearch.org</a>), Centre for Theoretical Studies, Charles University in Prague. &nbsp;Formerly research fellow Future of Humanity Institute, Oxford University<br><br>Previously I was a researcher in physics, studying phase transitions, network science and complex systems.</p>
null
null
54
290
0
21
69
1
0
r38pkCm7wF4M44MDQ
User
null
null
null
[ "canModeratePersonal", "alignmentVoters", "alignmentForum", "trustLevel1" ]
null
null
E96XcEPECbsipAvFi
https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/E96XcEPECbsipAvFi/grogqaxupsnwg0qmdosd
SocialPreviewType
iHKPBvqzhjKFAmPDS
<p>Imagine each of us has an AI representative, aligned to us, personally. Is gradual disempowerment solved?<span class="footnote-reference" data-footnote-reference="" data-footnote-index="1" data-footnote-id="bhugzx74xtt" role="doc-noteref" id="fnrefbhugzx74xtt"><sup><a href="#fnbhugzx74xtt">[1]</a></sup></span>&nbsp;In my view, no; at the same time having AI representatives helps at the margin.</p><p>I have two deep reasons for skepticism.<span class="footnote-reference" data-footnote-reference="" data-footnote-index="2" data-footnote-id="it3lbzx0pi" role="doc-noteref" id="fnrefit3lbzx0pi"><sup><a href="#fnit3lbzx0pi">[2]</a></sup></span>&nbsp;Here is the first one.</p><h3><strong>Humans are Not Alone</strong></h3><p>We, as individuals, are not the only agents or “agencies” in this world. Other goal-oriented strategic entities include states, corporations, and to some extent egregores.&nbsp;</p><p>For the sake of argument, imagine the current distribution of agency in the world as approximately 60% humans, 20% corporations, 20% states <i>(numbers are illustrative, egregores not included for simplicity)</i></p><figure class="image image_resized" style="width:75.27%"><img src="https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/E96XcEPECbsipAvFi/zgg9wcezeukplvucmdlf" srcset="https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/E96XcEPECbsipAvFi/at7xazakxizbv2w6z5tm 180w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/E96XcEPECbsipAvFi/udinm8xz6lwgvrjcoyp4 360w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/E96XcEPECbsipAvFi/gtatbqx5rgzlt8asdurh 540w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/E96XcEPECbsipAvFi/v57nvxyuf4f4lci1md3n 720w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/E96XcEPECbsipAvFi/odrffohly1wxwkxrapcd 900w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/E96XcEPECbsipAvFi/pe4mvuskdr2vyxu7olcd 1080w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/E96XcEPECbsipAvFi/k3yiyvlwml21xjbx3lz4 1260w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/E96XcEPECbsipAvFi/g73giyinaneuwjevnzw5 1440w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/E96XcEPECbsipAvFi/rpvcwv1jup63fuxtapps 1620w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/E96XcEPECbsipAvFi/cllx8cbvsggbasohewxl 1754w"></figure><p>Now, suppose we introduce individually aligned AI assistants for each person. At first glance, this seems empowering—each individual gains a cognitive extension, potentially leveling the playing field.&nbsp;</p><p>But: do corporations or states also get their representatives?</p><p>If your answer is&nbsp;<strong>No</strong>, you'd need to explain how individuals gain AI power while states and corporations don't, leading to a rebalancing where the latter are disempowered relative to now. If the trend of extreme capital expenditure continues, I think reasonably we can hope to get some personal AIs and some corporate AIs and governmental AIs and maybe some AIs aligned to humanity as whole or goodness itself. To hope that<i> you personally will get an AI intent-aligned to you</i>, but there will be <i>no AIs aligned to TheLabs or the US government</i> seems strange.</p><p>If the answer is&nbsp;<strong>Yes</strong>, it seems natural to assume that various AI representatives will have different cognitive power, possibly with some minimum threshold for every citizen, but way more powerful if the principal spends some resources on computations.&nbsp;</p><h3><strong>The Substrate Shift</strong></h3><p>You may ask: doesn't this replicate current power equilibrium? For some intuition: imagine you are going to sue some MegaCorp. They can likely get a better legal team, even now. Is there any difference?</p><p>My bet is “no, unfortunately it does not”.</p><figure class="image"><img src="https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/E96XcEPECbsipAvFi/leljscjwjq9dhr6wyym2" srcset="https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/E96XcEPECbsipAvFi/qw2oourbuklar8hdajdq 210w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/E96XcEPECbsipAvFi/pxova8enz5wtbak0kpqh 420w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/E96XcEPECbsipAvFi/mkkx67ur6alwdb9sfbsu 630w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/E96XcEPECbsipAvFi/climnlr2mblazxuqcjxo 840w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/E96XcEPECbsipAvFi/tdj5wcfwulgadwphisyg 1050w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/E96XcEPECbsipAvFi/j8q5aciid4vg50e0dkso 1260w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/E96XcEPECbsipAvFi/gycqat0yhvtdg55hegjk 1470w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/E96XcEPECbsipAvFi/vxvohglejeot5aedzudq 1680w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/E96XcEPECbsipAvFi/rhpl0i3v242z4xr6vs7m 1890w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/E96XcEPECbsipAvFi/gs7oe1keljkcxw8q61ew 2048w"></figure><p>Consider that currently, both individuals and corporations "run" their cognition on human brains. This also means you can out-smart MegaCorp. In cases where corporate cognition is bound on something like “what the best human inside can do”, it's you versus another human.</p><p>Unfortunately if it is "my o5-mini-cheap” vs “corporate o8-maxed-inference” I would expect my representative t... </p>
Imagine each of us has an AI representative, aligned to us, personally. Is gradual disempowerment solved?[1] In my view, no; at the same time having AI representatives helps at the margin. I have two deep reasons for skepticism.[2] Here is the first one. Humans are Not Alone We, as individuals, are not the only agents or “agencies” in this world. Other goal-oriented strategic entities include states, corporations, and to some extent egregores.  For the sake of argument, imagine the current distribution of agency in the world as approximately 60% humans, 20% corporations, 20% states (numbers are illustrative, egregores not included for simplicity) Now, suppose we introduce individually aligned AI assistants for each person. At first glance, this seems empowering—each individual gains a cognitive extension, potentially leveling the playing field.  But: do corporations or states also get their representatives? If your answer is No, you'd need to explain how individuals gain AI power while states and corporations don't, leading to a rebalancing where the latter are disempowered relative to now. If the trend of extreme capital expenditure continues, I think reasonably we can hope to get some personal AIs and some corporate AIs and governmental AIs and maybe some AIs aligned to humanity as whole or goodness itself. To hope that you personally will get an AI intent-aligned to you, but there will be no AIs aligned to TheLabs or the US government seems strange. If the answer is Yes, it seems natural to assume that various AI representatives will have different cognitive power, possibly with some minimum threshold for every citizen, but way more powerful if the principal spends some resources on computations.  The Substrate Shift You may ask: doesn't this replicate current power equilibrium? For some intuition: imagine you are going to sue some MegaCorp. They can likely get a better legal team, even now. Is there any difference? My bet is “no, unfortunately it does
908
1.7.1
Revision
false
null
null
CrosspostOutput
2JokkvjHBWq6ZkRcX
lectures-on-ai-for-high-school-students-and-others
Lectures on AI for high school students (and others)
null
false
false
false
null
ESGk9cA6439nr9xKa
null
true
false
false
false
Post
https://radfordneal.wordpress.com/2025/05/30/lectures-on-ai/
2025-06-03T23:54:16.542Z
null
false
false
2
2
null
false
false
linkpost
[]
null
null
3hzhpKYTPbjFP3Yna
0
2
7
false
0.00458
null
false
false
2025-06-03T23:54:16.542Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
grecHJcgkb3KW5wnM
null
null
null
false
null
[]
null
0
0
2025-06-03T23:44:17.157Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
1
null
null
null
null
[ { "__typename": "Tag", "_id": "sYm3HiWcfZvrGu3ui", "adminOnly": false, "afBaseScore": 2, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" } ] }, "baseScore": 12, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:22.097Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" }, { "_id": "sof55TPMQaeBaxhsS", "displayName": "tommylees112" }, { "_id": "AayjS8XzcnDKhGdTv", "displayName": "shark" }, { "_id": "HnALuwRdo6k9HLaMt", "displayName": "Alex Firssoff" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI", "needsReview": false, "noindex": false, "postCount": 12544, "score": 12, "shortName": null, "slug": "ai", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 4, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
2
0
0
0
0
ESGk9cA6439nr9xKa
radford-neal-1
2020-04-07T15:28:57.123Z
Radford Neal
Radford Neal
null
null
null
897
0
false
false
null
null
5
245
0
0
0
1
1
qgdGA4ZEyW7zNdK84
User
null
null
null
[ "canModeratePersonal" ]
null
null
2JokkvjHBWq6ZkRcX
SocialPreviewType
3hzhpKYTPbjFP3Yna
<p><i>Below is the full text of the post. Feel free to comment either here or there.</i></p><p>This April and May, I gave a series of five lectures on Artificial Intelligence at The Abelard School in Toronto.</p><p>These lectures are aimed at high school students, but may be of interest to others as well. I cover not just the current state of AI, but some of its history, and some fundamental concepts in computer science and philosophy that are important for thinking about how AI may develop in the future, and what its implications are.</p><p>Here are the lecture titles:</p><p><strong>Lecture 1:</strong> The pace of AI progress<br><strong>Lecture 2:</strong> Computation and intelligence, artificial and natural<br><strong>Lecture 3:</strong> How modern AI works<br><strong>Lecture 4:</strong> Can an AI think, feel, be conscious? How can we know?<br><strong>Lecture 5:</strong> Benefits and dangers of AI, today and tomorrow</p><p>You can get to videos of these lectures (with slides and audio) as well as PDF files of just the slides at <a href="https://glizen.com/radfordneal/ai-lectures/">https://glizen.com/radfordneal/ai-lectures/</a></p>
Below is the full text of the post. Feel free to comment either here or there. This April and May, I gave a series of five lectures on Artificial Intelligence at The Abelard School in Toronto. These lectures are aimed at high school students, but may be of interest to others as well. I cover not just the current state of AI, but some of its history, and some fundamental concepts in computer science and philosophy that are important for thinking about how AI may develop in the future, and what its implications are. Here are the lecture titles: Lecture 1: The pace of AI progress Lecture 2: Computation and intelligence, artificial and natural Lecture 3: How modern AI works Lecture 4: Can an AI think, feel, be conscious? How can we know? Lecture 5: Benefits and dangers of AI, today and tomorrow You can get to videos of these lectures (with slides and audio) as well as PDF files of just the slides at https://glizen.com/radfordneal/ai-lectures/
166
1.1.0
Revision
false
null
null
CrosspostOutput
GGaEYwyYJrZi5q3AM
self-inquiry
Self-inquiry
null
false
false
false
null
SwwerhEGohDjA4Hji
null
true
false
false
false
Post
null
2025-06-03T22:15:31.178Z
null
false
false
2
2
null
false
false
post
[]
null
null
Fgk7EsCwgGogRtixu
0
3
-3
false
-0.002533
null
false
false
2025-06-03T22:15:31.178Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
grecHJcgkb3KW5wnM
null
null
null
false
null
[]
null
0
0
2025-01-14T12:38:01.077Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
6
null
null
null
null
[ { "__typename": "Tag", "_id": "xv7Bg5fbF9WppYREi", "adminOnly": false, "afBaseScore": 0, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 0, "canEditUserIds": null, "core": false, "createdAt": "2020-08-06T04:44:56.236Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Buddhism", "needsReview": false, "noindex": false, "postCount": 49, "score": 0, "shortName": null, "slug": "buddhism", "suggestedAsFilter": false, "userId": "gjoi5eBQob27Lww62", "voteCount": 0, "wikiOnly": false }, { "__typename": "Tag", "_id": "ux2x9RrJsuykQxT79", "adminOnly": false, "afBaseScore": 8, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "r38pkCm7wF4M44MDQ", "displayName": "Raemon" }, { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 22, "canEditUserIds": null, "core": false, "createdAt": "2020-10-14T16:28:33.955Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "r38pkCm7wF4M44MDQ", "displayName": "Raemon" }, { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" }, { "_id": "d5WCJPEGt8KDEdj2T", "displayName": "Giordano Rogers" }, { "_id": "xF5nfdddHjFThHy49", "displayName": "[email protected]" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Deliberate Practice", "needsReview": false, "noindex": false, "postCount": 31, "score": 22, "shortName": null, "slug": "deliberate-practice", "suggestedAsFilter": false, "userId": "Q7NW4XaWQmfPfdcFj", "voteCount": 4, "wikiOnly": false }, { "__typename": "Tag", "_id": "AiNyf5iwbpc7mehiX", "adminOnly": false, "afBaseScore": 3, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 9, "canEditUserIds": null, "core": false, "createdAt": "2020-06-07T08:29:31.071Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Meditation", "needsReview": false, "noindex": false, "postCount": 128, "score": 9, "shortName": null, "slug": "meditation", "suggestedAsFilter": false, "userId": "qgdGA4ZEyW7zNdK84", "voteCount": 1, "wikiOnly": false }, { "__typename": "Tag", "_id": "NSMKfa8emSbGNXRKD", "adminOnly": false, "afBaseScore": 0, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 0, "canEditUserIds": null, "core": false, "createdAt": "2020-05-15T23:11:08.425Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Religion", "needsReview": false, "noindex": false, "postCount": 218, "score": 0, "shortName": null, "slug": "religion", "suggestedAsFilter": false, "userId": "qgdGA4ZEyW7zNdK84", "voteCount": 0, "wikiOnly": false }, { "__typename": "Tag", "_id": "2wjPMY34by2gXEXA2", "adminOnly": false, "afBaseScore": 3, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 9, "canEditUserIds": null, "core": false, "createdAt": "2020-06-08T01:11:38.950Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Techniques", "needsReview": false, "noindex": false, "postCount": 130, "score": 9, "shortName": null, "slug": "techniques", "suggestedAsFilter": false, "userId": "x5S2Kuj6TfQTGuo63", "voteCount": 1, "wikiOnly": false }, { "__typename": "Tag", "_id": "fkABsGCJZ6y9qConW", "adminOnly": false, "afBaseScore": 0, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 2, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T06:06:46.947Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "MiuAZvbQcQ7ethgt3", "displayName": "Viktor withaK" }, { "_id": "dRaCtsAWxk7sgirSY", "displayName": "Jordan Morgan" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Practical", "needsReview": false, "noindex": false, "postCount": 3411, "score": 2, "shortName": null, "slug": "practical", "suggestedAsFilter": true, "userId": "oBSWiHjgproTiThmY", "voteCount": 2, "wikiOnly": false }, { "__typename": "Tag", "_id": "Ng8Gice9KNkncxqcj", "adminOnly": false, "afBaseScore": 0, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 1, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:17.072Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 100, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "iMqytjy9ns89Fzfyv", "displayName": "miakko" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Rationality", "needsReview": false, "noindex": false, "postCount": 4302, "score": 1, "shortName": null, "slug": "rationality", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 1, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
3
0
0
2
0
SwwerhEGohDjA4Hji
vadim-golub
2025-01-10T21:15:33.764Z
a schizophrenic mind
Vadim Golub
null
null
Vadim Golub
-13
0
false
false
null
null
5
15
0
0
0
0.8
0
grecHJcgkb3KW5wnM
User
null
null
null
null
null
null
GGaEYwyYJrZi5q3AM
SocialPreviewType
Fgk7EsCwgGogRtixu
<p>What is self-inquiry? Self-inquiry is an esoteric spiritual practice geared towards the deconstruction of the self. Why would anyone want to do that? What makes this practice spiritual? Why can it be found in the core of many spiritual traditions? What evidence of its effectiveness do we have?</p><p>First, why do <i>I</i> want to deconstruct the self? If I carefully observe my thoughts I can see that most problematic thoughts are self-referential ones. To put it simply, thoughts that are built around I/me/my. They make me anxious and afraid. In most cases without any good reason.&nbsp;</p><p>To check that this is so, one can conduct the following experiment (as described by Gary Weber<span class="footnote-reference" data-footnote-reference="" data-footnote-index="1" data-footnote-id="m8lwifjmybf" role="doc-noteref" id="fnrefm8lwifjmybf"><sup><a href="#fnm8lwifjmybf">[1]</a></sup></span>). Get yourself a piece of paper and a pencil. Draw two buckets: one with I/me/my thoughts, another without I/me/my thoughts. Sit quietly for five minutes and observe your thoughts. Each time a thought arises make a mark in the appropriate bucket. Count your thoughts in the end.&nbsp;</p><p>One remark is that some thoughts may have a hidden and not obvious I/me/my structure. They should be counted in the first - I/me/my - bucket.&nbsp;</p><p>If your thinking is like mine, you will get around 90-95% of thoughts in the first bucket. In this way one can get an insight into what's going on in one's head while being idle.&nbsp;</p><p>Further one may ask, "Ok, the majority of idle thoughts are all about me, but why would I want to get rid of them?" Again, if your thinking is like mine and you observe the content of your self-referential thoughts you will discover that most of them generate anxiety and fear in some way or another. They also generate grasping and attachment. And grasping and attachment are the causes of suffering. That's Buddhism 101<span class="footnote-reference" data-footnote-reference="" data-footnote-index="2" data-footnote-id="av5enhdekx" role="doc-noteref" id="fnrefav5enhdekx"><sup><a href="#fnav5enhdekx">[2]</a></sup></span>.&nbsp;</p><p>The next line is simple. If I/me/my thoughts generate suffering, how can I get rid of them? And the answer to that is by deconstructing the "I" one can get rid of all problematic self-referential thoughts (yes, it's possible, it's called the state of nonduality<span class="footnote-reference" data-footnote-reference="" data-footnote-index="3" data-footnote-id="fgwaezftbl7" role="doc-noteref" id="fnreffgwaezftbl7"><sup><a href="#fnfgwaezftbl7">[3]</a></sup></span>). Which would free one from the constant internal dialogue that almost all of us have.&nbsp;</p><p>So in the end, the simple answer to the question, "Why would anyone want to do that?" is <i>to get rid of suffering</i>. Pretty amazing if you ask me. And how does one go about it? By the process of self-inquiry.</p><p>What makes this practice spiritual? As I see it, the main reason that that practice may be considered spiritual is t... </p>
What is self-inquiry? Self-inquiry is an esoteric spiritual practice geared towards the deconstruction of the self. Why would anyone want to do that? What makes this practice spiritual? Why can it be found in the core of many spiritual traditions? What evidence of its effectiveness do we have? First, why do I want to deconstruct the self? If I carefully observe my thoughts I can see that most problematic thoughts are self-referential ones. To put it simply, thoughts that are built around I/me/my. They make me anxious and afraid. In most cases without any good reason.  To check that this is so, one can conduct the following experiment (as described by Gary Weber[1]). Get yourself a piece of paper and a pencil. Draw two buckets: one with I/me/my thoughts, another without I/me/my thoughts. Sit quietly for five minutes and observe your thoughts. Each time a thought arises make a mark in the appropriate bucket. Count your thoughts in the end.  One remark is that some thoughts may have a hidden and not obvious I/me/my structure. They should be counted in the first - I/me/my - bucket.  If your thinking is like mine, you will get around 90-95% of thoughts in the first bucket. In this way one can get an insight into what's going on in one's head while being idle.  Further one may ask, "Ok, the majority of idle thoughts are all about me, but why would I want to get rid of them?" Again, if your thinking is like mine and you observe the content of your self-referential thoughts you will discover that most of them generate anxiety and fear in some way or another. They also generate grasping and attachment. And grasping and attachment are the causes of suffering. That's Buddhism 101[2].  The next line is simple. If I/me/my thoughts generate suffering, how can I get rid of them? And the answer to that is by deconstructing the "I" one can get rid of all problematic self-referential thoughts (yes, it's possible, it's called the state of nonduality[3]). Which would free one fro
1,616
1.6.0
Revision
false
null
null
CrosspostOutput
6CYav4iTQsgQGq2MT
question-to-lw-devs-does-lesswrong-tries-to-be-facebooky
Question to LW devs: does LessWrong tries to be facebooky?
null
false
false
false
null
ajbhE6vN6ekCdBomP
null
true
false
false
false
Post
null
2025-06-03T22:08:52.853Z
null
false
false
2
2
null
false
false
post
[]
null
null
TLkkN9xw8hEQkxoi3
1
2
5
false
0.003474
null
false
false
2025-06-03T23:25:36.021Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
grecHJcgkb3KW5wnM
null
null
null
false
null
[]
null
1
0
2025-06-03T21:12:57.096Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
1
null
null
null
null
[ { "__typename": "Tag", "_id": "MfpEPj6kJneT9gWT6", "adminOnly": false, "afBaseScore": 0, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 1, "canEditUserIds": null, "core": true, "createdAt": "2020-06-11T01:01:30.571Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "zyrFCnHNbfBFw3PvP", "displayName": "banality_exhibition" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Site Meta", "needsReview": false, "noindex": false, "postCount": 757, "score": 1, "shortName": null, "slug": "site-meta", "suggestedAsFilter": false, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 1, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
2
0
0
1
0
ajbhE6vN6ekCdBomP
roman-malov
2023-10-24T16:19:28.575Z
Roman Malov
Roman Malov
null
null
null
150
0
false
false
<p>Bachelor in general and applied physics. AI safety/Agent foundations researcher wannabe.&nbsp;<br><br>I love talking to people, and if you are an alignment researcher we will have at least one common topic (but I am very interested in talking about unknown to me topics too!), so I encourage you to book a call with me: https://calendly.com/roman-malov27/new-meeting<br><br>Email: <a href="mailto:[email protected]">[email protected]</a><br>GitHub: <a href="https://github.com/RomanMalov">https://github.com/RomanMalov</a><br>TG channels (in Russian): <a href="https://t.me/healwithcomedy,">https://t.me/healwithcomedy,</a> https://t.me/ai_safety_digest</p>
null
null
7
36
0
0
0
1
0
grecHJcgkb3KW5wnM
User
null
null
null
[ "canModeratePersonal" ]
null
null
6CYav4iTQsgQGq2MT
SocialPreviewType
TLkkN9xw8hEQkxoi3
<p>Or maybe it’s deliberately trying <strong>not</strong> to be facebooky? By “facebooky”, I mean a website that tries to hack your brain through various stimuli, like optimizing suggestions, tracking your data, steering your interests, inferring personal information, clustering communities, and encouraging creators to focus on retention, CTR, clickbait, etc.</p><p>LessWrong obviously isn’t doing anything ad-related, since it’s non-profit. But maybe it is trying to earn utilons by using some facebooky strategies.</p><p>I actually find LW more facebooky than Substack, for example. It’s much easier to spend a lot of time here by following links both within and outside posts. On the other side, one anti-facebooky feature (which I really like) is the ability to control how often you get notifications about karma and comments.</p><p>P.S. I’m not trying to imply that being facebooky is inherently bad. Part of the reason Facebook makes so much money is that it did, in fact, generate some societal value, and it did it in part due to implementing some of the facebooky strategies.</p>
Or maybe it’s deliberately trying not to be facebooky? By “facebooky”, I mean a website that tries to hack your brain through various stimuli, like optimizing suggestions, tracking your data, steering your interests, inferring personal information, clustering communities, and encouraging creators to focus on retention, CTR, clickbait, etc. LessWrong obviously isn’t doing anything ad-related, since it’s non-profit. But maybe it is trying to earn utilons by using some facebooky strategies. I actually find LW more facebooky than Substack, for example. It’s much easier to spend a lot of time here by following links both within and outside posts. On the other side, one anti-facebooky feature (which I really like) is the ability to control how often you get notifications about karma and comments. P.S. I’m not trying to imply that being facebooky is inherently bad. Part of the reason Facebook makes so much money is that it did, in fact, generate some societal value, and it did it in part due to implementing some of the facebooky strategies.
168
1.2.0
Revision
false
null
null
CrosspostOutput
t5tPZPMCN22gwimxC
your-strategy-roadmap-expert-tips-live-training
Your Strategy Roadmap: Expert Tips + Live Training
null
false
false
false
null
PWWHimEtQY4osMvrr
null
true
false
false
false
Post
null
2025-06-03T21:10:58.905Z
null
false
false
2
2
null
false
false
post
[]
null
null
gokJQdLYseef2FMjc
0
4
-4
false
-0.002985
null
false
false
2025-06-03T21:10:58.905Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
grecHJcgkb3KW5wnM
null
null
null
false
null
[]
null
-2
0
2025-06-03T21:06:42.450Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
5
null
null
null
null
[ { "__typename": "Tag", "_id": "bEtp696XzNCvxpHc3", "adminOnly": false, "afBaseScore": null, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 0, "canEditUserIds": null, "core": false, "createdAt": "2024-11-21T19:32:19.660Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Leadership", "needsReview": false, "noindex": false, "postCount": 2, "score": 0, "shortName": null, "slug": "leadership-2", "suggestedAsFilter": false, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 0, "wikiOnly": false }, { "__typename": "Tag", "_id": "udPbn9RthmgTtHMiG", "adminOnly": false, "afBaseScore": 9, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "EQNTWXLKMeWMp2FQS", "displayName": "Ben Pace" }, { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 19, "canEditUserIds": null, "core": false, "createdAt": "2020-06-11T20:28:13.679Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "EQNTWXLKMeWMp2FQS", "displayName": "Ben Pace" }, { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Productivity", "needsReview": false, "noindex": false, "postCount": 227, "score": 19, "shortName": null, "slug": "productivity", "suggestedAsFilter": false, "userId": "nLbwLhBaQeG6tCNDN", "voteCount": 2, "wikiOnly": false }, { "__typename": "Tag", "_id": "xYLtnJ6keSHGfrLpe", "adminOnly": false, "afBaseScore": 3, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 9, "canEditUserIds": null, "core": false, "createdAt": "2020-08-22T09:02:46.252Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Risk Management", "needsReview": false, "noindex": false, "postCount": 36, "score": 9, "shortName": null, "slug": "risk-management", "suggestedAsFilter": false, "userId": "QBvPFLFyZyuHcBwFm", "voteCount": 1, "wikiOnly": false }, { "__typename": "Tag", "_id": "fkABsGCJZ6y9qConW", "adminOnly": false, "afBaseScore": 0, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 2, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T06:06:46.947Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "MiuAZvbQcQ7ethgt3", "displayName": "Viktor withaK" }, { "_id": "dRaCtsAWxk7sgirSY", "displayName": "Jordan Morgan" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Practical", "needsReview": false, "noindex": false, "postCount": 3411, "score": 2, "shortName": null, "slug": "practical", "suggestedAsFilter": true, "userId": "oBSWiHjgproTiThmY", "voteCount": 2, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
4
0
0
2
0
PWWHimEtQY4osMvrr
deena-englander
2023-04-13T12:47:05.942Z
deena-englander
Deena Englander
null
null
null
-34
0
false
false
null
null
1
1
0
0
0
0.8
0
grecHJcgkb3KW5wnM
User
null
null
null
null
null
null
t5tPZPMCN22gwimxC
https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/t5tPZPMCN22gwimxC/rekbzmgz6scpte0nlkd0
SocialPreviewType
gokJQdLYseef2FMjc
<p>The cornerstone of running an impactful organization lies in developing a solid organizational strategy. A good strategic plan will be your “north star”, providing an anchor to make decisions that drive your desired impact. The best strategies include thoughtful, measurable, and actionable components to ensure accountability and mission fulfillment.</p><p>Despite its importance, many organizations we meet don’t have a strong organizational strategy. While they usually have a mission statement describing the change they want to make, they’re often missing the practical components of how to achieve that. Without a strong strategic plan, even the best-intentioned organizations will struggle to maximize their impact.</p><p>In this post, we asked our <a href="https://ea-services.org/">EASE experts</a> for their advice so that you can make sure your organizational strategy is both strong and practical.</p><p>We'd also like to invite you to a panel-style webinar on June 18th at 12 PM EST, where we'll cover these strategies in depth and provide answers to commonly asked questions.</p><p><a href="https://us06web.zoom.us/meeting/register/uGn5yDHQTRap82hqTBuTaQ">Click here to Register</a></p><hr><h2><strong>Question: What are the key components of a strong, well-developed organizational strategy?</strong></h2><h3><strong>Laura Richards, </strong><a href="https://scaleupconsultinggroup.com/"><strong>Strategy Consultant</strong></a></h3><p>While often used interchangeably,&nbsp;<u>organizational strategy</u> refers to what an organization aims to achieve and&nbsp;<u>why</u> (high-level, long-term, guides organizational culture). A strategic plan guides&nbsp;<i>how&nbsp;</i>and&nbsp;<i>when</i> the work is done, and metrics for success.&nbsp;<strong>When culture and strategy work together, there is a much better chance that the vision is realized.</strong></p><p>&nbsp;<strong>When you pay attention to culture while rolling out a strategy, you’re setting your team up for long-term success</strong>.</p><p>As a leader, it’s important to understand your current and desired organizational culture. To influence a change in culture, set goals for employees to support behaviors that encourage the culture you desire. (i.e., teamwork, flexibility, and fresh thinking) and shift the behavior limits that culture (i.e., gatekeeping, fear of new ideas). Lead by example, communicate openly, and make sure people are recognized and rewarded for actions that align with your goals.</p><p>&nbsp;</p><h3><strong>Sara Carrillo, </strong><a href="https://sarachas.com/"><strong>OKR Coach</strong></a></h3><p>A strong, well-developed organizational strategy is built upon a clear, foundational understanding of the company's core identity. This begins with a clearly defined set of values, a compelling mission, and an inspiring vision, p... </p>
The cornerstone of running an impactful organization lies in developing a solid organizational strategy. A good strategic plan will be your “north star”, providing an anchor to make decisions that drive your desired impact. The best strategies include thoughtful, measurable, and actionable components to ensure accountability and mission fulfillment. Despite its importance, many organizations we meet don’t have a strong organizational strategy. While they usually have a mission statement describing the change they want to make, they’re often missing the practical components of how to achieve that. Without a strong strategic plan, even the best-intentioned organizations will struggle to maximize their impact. In this post, we asked our EASE experts for their advice so that you can make sure your organizational strategy is both strong and practical. We'd also like to invite you to a panel-style webinar on June 18th at 12 PM EST, where we'll cover these strategies in depth and provide answers to commonly asked questions. Click here to Register ---------------------------------------- Question: What are the key components of a strong, well-developed organizational strategy? Laura Richards, Strategy Consultant While often used interchangeably, organizational strategy refers to what an organization aims to achieve and why (high-level, long-term, guides organizational culture). A strategic plan guides how and when the work is done, and metrics for success. When culture and strategy work together, there is a much better chance that the vision is realized.  When you pay attention to culture while rolling out a strategy, you’re setting your team up for long-term success. As a leader, it’s important to understand your current and desired organizational culture. To influence a change in culture, set goals for employees to support behaviors that encourage the culture you desire. (i.e., teamwork, flexibility, and fresh thinking) and shift the behavior limits that cultur
1,143
1.1.1
Revision
false
null
null
CrosspostOutput
JsyAdviSYCqgMvvmG
steering-vectors-can-help-llm-judges-detect-subtle
Steering Vectors Can Help LLM Judges Detect Subtle Dishonesty
null
false
false
false
null
jgzfiBKp2SWYCWg5z
[ { "__typename": "CoauthorStatusOutput", "confirmed": true, "requested": false, "userId": "GTvevFXuxDogSEuuB" }, { "__typename": "CoauthorStatusOutput", "confirmed": true, "requested": false, "userId": "D6agLc4CPfhHdfNsv" }, { "__typename": "CoauthorStatusOutput", "confirmed": true, "requested": false, "userId": "e9PASAraAXnms7rR9" } ]
true
false
false
false
Post
null
2025-06-03T20:33:58.083Z
null
false
false
2
2
2025-06-04T17:45:08.222Z
false
false
post
[]
null
null
RkhiSALoZnz2CYRrs
1
7
12
false
0.014709
null
false
false
2025-06-04T00:33:45.238Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
grecHJcgkb3KW5wnM
null
null
null
false
null
[]
null
5
0
2025-05-31T10:46:41.122Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[ { "__typename": "User", "_id": "GTvevFXuxDogSEuuB", "afCommentCount": 0, "afKarma": 0, "afPostCount": 0, "commentCount": 0, "createdAt": "2025-05-31T16:57:16.445Z", "deleted": false, "displayName": "mcbeth", "fullName": null, "htmlBio": "", "isAdmin": false, "jobTitle": null, "karma": 10, "organization": null, "postCount": 0, "previousDisplayName": null, "profileImageId": null, "reviewedByUserId": null, "sequenceCount": 0, "slug": "mcbeth", "spamRiskScore": 0.7200000000000001, "tagRevisionCount": 0, "username": "mcbeth" }, { "__typename": "User", "_id": "D6agLc4CPfhHdfNsv", "afCommentCount": 0, "afKarma": 0, "afPostCount": 0, "commentCount": 0, "createdAt": "2025-05-11T03:36:14.625Z", "deleted": false, "displayName": "Etha", "fullName": "Ethan Nguyen", "htmlBio": "", "isAdmin": false, "jobTitle": null, "karma": 10, "organization": null, "postCount": 0, "previousDisplayName": null, "profileImageId": null, "reviewedByUserId": null, "sequenceCount": 0, "slug": "etha", "spamRiskScore": 0.7200000000000001, "tagRevisionCount": 0, "username": "ThatE10" }, { "__typename": "User", "_id": "e9PASAraAXnms7rR9", "afCommentCount": 0, "afKarma": 0, "afPostCount": 0, "commentCount": 2, "createdAt": "2024-06-17T23:24:21.409Z", "deleted": false, "displayName": "Arch223", "fullName": null, "htmlBio": "", "isAdmin": false, "jobTitle": null, "karma": 0, "organization": null, "postCount": 1, "previousDisplayName": null, "profileImageId": null, "reviewedByUserId": "qgdGA4ZEyW7zNdK84", "sequenceCount": 0, "slug": "arch223", "spamRiskScore": 0.9, "tagRevisionCount": 0, "username": "Arch223" } ]
6
null
null
null
null
[ { "__typename": "Tag", "_id": "QdTRKbR4MJSz4JWvn", "adminOnly": false, "afBaseScore": 3, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 10, "canEditUserIds": null, "core": false, "createdAt": "2023-08-29T03:05:32.778Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" }, { "_id": "qudEsXCoB8pnJDXrG", "displayName": "Jianing Zhu" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Activation Engineering", "needsReview": false, "noindex": false, "postCount": 62, "score": 10, "shortName": null, "slug": "activation-engineering", "suggestedAsFilter": false, "userId": "4Yj2EqHzqrwscRnNL", "voteCount": 2, "wikiOnly": false }, { "__typename": "Tag", "_id": "FBRwHSmTudwiHHtrn", "adminOnly": false, "afBaseScore": 3, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 9, "canEditUserIds": null, "core": false, "createdAt": "2023-03-15T20:29:46.761Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI Evaluations", "needsReview": false, "noindex": false, "postCount": 224, "score": 9, "shortName": null, "slug": "ai-evaluations", "suggestedAsFilter": false, "userId": "qgdGA4ZEyW7zNdK84", "voteCount": 1, "wikiOnly": false }, { "__typename": "Tag", "_id": "8wiAjZn9qGpBsk8GT", "adminOnly": false, "afBaseScore": null, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 0, "canEditUserIds": null, "core": false, "createdAt": "2023-12-18T23:00:24.890Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Sycophancy", "needsReview": false, "noindex": false, "postCount": 13, "score": 0, "shortName": null, "slug": "sycophancy", "suggestedAsFilter": false, "userId": "RS8wZkL4ozjeYuLY8", "voteCount": 0, "wikiOnly": false }, { "__typename": "Tag", "_id": "sYm3HiWcfZvrGu3ui", "adminOnly": false, "afBaseScore": 2, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" } ] }, "baseScore": 12, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:22.097Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" }, { "_id": "sof55TPMQaeBaxhsS", "displayName": "tommylees112" }, { "_id": "AayjS8XzcnDKhGdTv", "displayName": "shark" }, { "_id": "HnALuwRdo6k9HLaMt", "displayName": "Alex Firssoff" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI", "needsReview": false, "noindex": false, "postCount": 12544, "score": 12, "shortName": null, "slug": "ai", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 4, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
7
0
0
3
0
jgzfiBKp2SWYCWg5z
leon-eshuijs
2024-09-21T20:19:18.463Z
Sentient Watermelon
Leon Eshuijs
null
null
null
10
0
false
false
null
null
1
0
0
0
0
0.9
0
qgdGA4ZEyW7zNdK84
User
null
null
null
null
null
null
JsyAdviSYCqgMvvmG
https://res.cloudinary.com/lesswrong-2-0/image/upload/c_fill,ar_1.91,g_auto/SocialPreview/eb8xxxpom1foc6pqtmvz
SocialPreviewType
RkhiSALoZnz2CYRrs
<p><i>Cross-posted from our recent paper: "But what is your honest answer? Aiding LLM-judges with honest alternatives using steering vectors" : </i><a href="https://arxiv.org/abs/2505.17760"><i>https://arxiv.org/abs/2505.17760</i></a></p><p><i>Code available at: </i><a href="https://github.com/watermeleon/judge_with_steered_response"><i>https://github.com/watermeleon/judge_with_steered_response</i></a></p><p><strong>TL;DR:</strong> We use steering vectors to generate more honest versions of an LLM response, helping LLM judges detect subtle forms of dishonesty like sycophancy and manipulation that they normally miss. We also introduce a new dataset with prompts designed to provoke subtle manipulation.</p><h2>Abstract</h2><p>A fundamental challenge in AI alignment is that as systems become more capable, they may learn subtle forms of deception that evaluators (humans or other LLMs) struggle to detect. Even when an AI's response is technically accurate, it might be manipulative, sycophantic, or misleading in ways that optimize for human approval rather than truth. Moreover, most honesty benchmarks focus exclusively on factual knowledge or explicitly harmful behavior and rely on external LLM judges, which are often unable to detect less obvious forms of dishonesty.</p><p>In this work, we introduce a new framework, Judge Using Safety-Steered Alternatives (JUSSA), which utilizes steering vectors trained on a single sample to elicit more honest responses from models, helping LLM judges in the detection of dishonest behavior. To test our framework, we introduce a new manipulation dataset with prompts specifically designed to elicit deceptive responses. We find that JUSSA enables LLM judges to better differentiate between dishonest and benign responses, and helps them identify subtle instances of manipulative behavior.</p><h3>Motivation:</h3><p>We build upon recent work by <a href="https://www.lesswrong.com/posts/6aXe9nipTgwK5LxaP/do-safety-relevant-llm-steering-vectors-optimized-on-a">Dunefsky et al. (2025)</a>, which showed that steering vectors trained on a single example can generalize well for various safety-related behaviors. While steering for safety training shows promise, models may eventually learn to work around such interventions. Therefore, we consider whether steering vectors may be valuable as tools for improving evaluations of models instead of simply directly intervening on model behavior.</p><p>Moreover, subtle forms of dishonesty—like sycophancy, selective evidence presentation, or emotional manipulation—pose a particularly difficult challenge. These behaviors may involve misleading but factually correct statements, making them difficult for standard evaluation methods to detect, yet pote... </p>
Cross-posted from our recent paper: "But what is your honest answer? Aiding LLM-judges with honest alternatives using steering vectors" : https://arxiv.org/abs/2505.17760 Code available at: https://github.com/watermeleon/judge_with_steered_response TL;DR: We use steering vectors to generate more honest versions of an LLM response, helping LLM judges detect subtle forms of dishonesty like sycophancy and manipulation that they normally miss. We also introduce a new dataset with prompts designed to provoke subtle manipulation. Abstract A fundamental challenge in AI alignment is that as systems become more capable, they may learn subtle forms of deception that evaluators (humans or other LLMs) struggle to detect. Even when an AI's response is technically accurate, it might be manipulative, sycophantic, or misleading in ways that optimize for human approval rather than truth. Moreover, most honesty benchmarks focus exclusively on factual knowledge or explicitly harmful behavior and rely on external LLM judges, which are often unable to detect less obvious forms of dishonesty. In this work, we introduce a new framework, Judge Using Safety-Steered Alternatives (JUSSA), which utilizes steering vectors trained on a single sample to elicit more honest responses from models, helping LLM judges in the detection of dishonest behavior. To test our framework, we introduce a new manipulation dataset with prompts specifically designed to elicit deceptive responses. We find that JUSSA enables LLM judges to better differentiate between dishonest and benign responses, and helps them identify subtle instances of manipulative behavior. Motivation: We build upon recent work by Dunefsky et al. (2025), which showed that steering vectors trained on a single example can generalize well for various safety-related behaviors. While steering for safety training shows promise, models may eventually learn to work around such interventions. Therefore, we consider whether steering vectors may
1,477
1.7.1
Revision
false
null
null
CrosspostOutput
vqxExZvDjkphqG6zH
schelling-coordination-via-agentic-loops-1
Schelling Coordination via Agentic Loops
null
false
false
false
null
TJQvcMLcMQHnhMJTo
null
true
false
false
false
Post
null
2025-06-03T20:13:58.061Z
null
false
false
2
2
2025-06-04T17:48:26.660Z
false
false
post
[]
null
null
Pwr3sXg4sKX6XtPpo
0
2
2
false
0.007959
null
false
false
2025-06-03T20:13:58.061Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
grecHJcgkb3KW5wnM
null
null
null
false
null
[]
null
0
0
2025-06-03T18:40:35.596Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
10
null
null
null
null
[ { "__typename": "Tag", "_id": "X8JsWEnBRPvs5Y99i", "adminOnly": false, "afBaseScore": 0, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 0, "canEditUserIds": null, "core": false, "createdAt": "2015-12-03T07:35:06.000Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [] }, "isArbitalImport": true, "isPlaceholderPage": false, "isSubforum": false, "name": "Decision theory", "needsReview": false, "noindex": false, "postCount": 500, "score": 0, "shortName": null, "slug": "decision-theory", "suggestedAsFilter": false, "userId": "nmk3nLpQE89dMRzzN", "voteCount": 0, "wikiOnly": false }, { "__typename": "Tag", "_id": "sYm3HiWcfZvrGu3ui", "adminOnly": false, "afBaseScore": 2, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" } ] }, "baseScore": 12, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:22.097Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" }, { "_id": "sof55TPMQaeBaxhsS", "displayName": "tommylees112" }, { "_id": "AayjS8XzcnDKhGdTv", "displayName": "shark" }, { "_id": "HnALuwRdo6k9HLaMt", "displayName": "Alex Firssoff" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI", "needsReview": false, "noindex": false, "postCount": 12544, "score": 12, "shortName": null, "slug": "ai", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 4, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
2
0
0
0
0
TJQvcMLcMQHnhMJTo
callum-luis-kindred
2025-05-25T12:56:45.395Z
callum-luis-kindred
Callum-Luis Kindred
null
null
null
1
0
false
false
null
null
1
0
0
0
0
0.9
0
grecHJcgkb3KW5wnM
User
null
null
null
null
null
null
vqxExZvDjkphqG6zH
https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/vqxExZvDjkphqG6zH/w8e0xyupcmsxfay9np2d
SocialPreviewType
Pwr3sXg4sKX6XtPpo
<p><strong>Author's Note:</strong> This work originated as a team project during the <a href="https://apartresearch.com/sprints/ai-control-hackathon-2025-03-29-to-2025-03-30">Apart Research AI Control Hackathon</a>, where my teammates did the majority of the foundational work. I've continued developing these ideas as part of the Apart Lab Studio program, but the core insights and initial implementation were primarily their contributions. My motivation for publishing this article is to extend my own thinking on these concepts and to help disseminate what I believe are valuable ideas that deserve broader attention. I'm grateful to have been part of the original team and for their permission to build upon our shared work.</p><h2>Introduction</h2><p>As AI systems become increasingly powerful, ensuring their safe deployment remains a critical challenge. While alignment efforts strive to create inherently trustworthy systems, control mechanisms serve as crucial safeguards when alignment fails. Our work, developed as part of the Apart Research AI Control Hackathon, explores a particularly concerning failure mode within untrusted monitoring protocols: the possibility of collusion between AI systems. Our team investigated whether language models can coordinate their actions without direct communication by identifying and converging on Schelling points in coordination games. While our findings show that current models possess some basic coordination capabilities, these abilities appear limited and may not pose immediate risks to existing safety protocols.</p><h3>Background</h3><p><a href="https://www.lesswrong.com/posts/7e479JqZesNvoJqEv/a-brief-explanation-of-ai-control?utm_campaign=post_share&amp;utm_source=link"><u>AI Control</u></a> is a subfield of AI safety that aims to answer the question: “assuming you have a misaligned AI system, or one that could be misaligned, how can you safely get useful work out of it?”. While AI Alignment focuses on making AI systems behave the way we want them to, this does not always work perfectly. Control, therefore,&nbsp;<a href="https://www.lesswrong.com/posts/kcKrE9mzEHrdqtDpE/the-case-for-ensuring-that-powerful-ais-are-controlled?utm_campaign=post_share&amp;utm_source=link"><u>promises to serve as an important line-of-defence</u></a>, preventing bad outcomes even if Alignment fails.</p><p>In the event where we can’t fully trust an AI, there are&nbsp;<a href="https://www.lesswrong.com/posts/3PBvKHB2EmCujet3j/ai-control-methods-literature-review#1__Monitoring"><u>a variety of proposed</u></a> methods involving the use of a ‘monitor’ model to control a potentially harmful model. One such protocol is known as&nbsp;<i>trusted monitoring</i>. In this protocol, an "untrusted" model (referred to as the policy) generates outputs that might include harmful actions, while a "trusted" model (the monitor) serves as an oversight mechanism. The trusted monitor is chosen such that it is less powerful than the policy—sophisticated eno... </p>
Author's Note: This work originated as a team project during the Apart Research AI Control Hackathon, where my teammates did the majority of the foundational work. I've continued developing these ideas as part of the Apart Lab Studio program, but the core insights and initial implementation were primarily their contributions. My motivation for publishing this article is to extend my own thinking on these concepts and to help disseminate what I believe are valuable ideas that deserve broader attention. I'm grateful to have been part of the original team and for their permission to build upon our shared work. Introduction As AI systems become increasingly powerful, ensuring their safe deployment remains a critical challenge. While alignment efforts strive to create inherently trustworthy systems, control mechanisms serve as crucial safeguards when alignment fails. Our work, developed as part of the Apart Research AI Control Hackathon, explores a particularly concerning failure mode within untrusted monitoring protocols: the possibility of collusion between AI systems. Our team investigated whether language models can coordinate their actions without direct communication by identifying and converging on Schelling points in coordination games. While our findings show that current models possess some basic coordination capabilities, these abilities appear limited and may not pose immediate risks to existing safety protocols. Background AI Control is a subfield of AI safety that aims to answer the question: “assuming you have a misaligned AI system, or one that could be misaligned, how can you safely get useful work out of it?”. While AI Alignment focuses on making AI systems behave the way we want them to, this does not always work perfectly. Control, therefore, promises to serve as an important line-of-defence, preventing bad outcomes even if Alignment fails. In the event where we can’t fully trust an AI, there are a variety of proposed methods involving the use of
2,623
1.2.1
Revision
false
null
null
CrosspostOutput
ZR8GDfSYjcvjLMrQ5
visual-prompt-injections-results-on-testing-ai-spam-defense
Visual Prompt Injections: Results on testing AI spam-defense and AI vulnerability to deceptive web ads.
null
false
false
false
null
XcrGKKoGoJum8Nm7E
null
true
false
false
false
Post
null
2025-06-03T20:10:36.502Z
null
false
false
2
2
2025-06-04T17:53:31.456Z
false
false
post
[]
null
null
AbpAcCH2wMQmCgGcy
0
3
3
false
0.008389
null
false
false
2025-06-03T20:10:36.502Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
grecHJcgkb3KW5wnM
null
null
null
false
null
[]
null
0
0
2025-02-08T15:56:36.927Z
false
false
easy-going
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
14
null
null
null
null
[ { "__typename": "Tag", "_id": "bdkABqNJoceYSefsi", "adminOnly": false, "afBaseScore": 3, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 9, "canEditUserIds": null, "core": false, "createdAt": "2023-05-01T17:42:20.164Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI Misuse", "needsReview": false, "noindex": false, "postCount": 15, "score": 9, "shortName": null, "slug": "ai-misuse", "suggestedAsFilter": false, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 1, "wikiOnly": false }, { "__typename": "Tag", "_id": "cHoCqtfE9cF7aSs9d", "adminOnly": false, "afBaseScore": 3, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 9, "canEditUserIds": null, "core": false, "createdAt": "2020-07-09T05:53:15.445Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Deception", "needsReview": false, "noindex": false, "postCount": 129, "score": 9, "shortName": null, "slug": "deception", "suggestedAsFilter": false, "userId": "mPipmBTniuABY5PQy", "voteCount": 1, "wikiOnly": false }, { "__typename": "Tag", "_id": "hwHxSdrxafFTuaCJE", "adminOnly": false, "afBaseScore": 3, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 9, "canEditUserIds": null, "core": false, "createdAt": "2022-08-10T17:49:55.188Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Prompt Engineering", "needsReview": false, "noindex": false, "postCount": 38, "score": 9, "shortName": null, "slug": "prompt-engineering", "suggestedAsFilter": false, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 1, "wikiOnly": false }, { "__typename": "Tag", "_id": "sYm3HiWcfZvrGu3ui", "adminOnly": false, "afBaseScore": 2, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" } ] }, "baseScore": 12, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:22.097Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" }, { "_id": "sof55TPMQaeBaxhsS", "displayName": "tommylees112" }, { "_id": "AayjS8XzcnDKhGdTv", "displayName": "shark" }, { "_id": "HnALuwRdo6k9HLaMt", "displayName": "Alex Firssoff" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI", "needsReview": false, "noindex": false, "postCount": 12544, "score": 12, "shortName": null, "slug": "ai", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 4, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
3
0
0
0
0
XcrGKKoGoJum8Nm7E
seon-gunness-1
2023-05-25T16:34:37.120Z
seon-gunness-1
Seon Gunness
null
null
null
2
0
false
false
null
null
1
0
0
0
0
0.9
0
grecHJcgkb3KW5wnM
User
null
null
null
null
null
null
ZR8GDfSYjcvjLMrQ5
https://res.cloudinary.com/lesswrong-2-0/image/upload/c_fill,ar_1.91,g_auto/SocialPreview/efcfxkuysmc8vv4lv3ai
SocialPreviewType
AbpAcCH2wMQmCgGcy
<p><i>Epistemic Status: Exploratory/my best guess https://www.lesswrong.com/posts/Hrm59GdN2yDPWbtrd/feature-idea-epistemic-status</i></p><p><i>Epistemic Effort: ~ 2 months of work</i></p><p><i>Contributions: Thanks to Clement Neo and Alignment Jams / Apart Research!</i></p><p>Summary/abstract:Split into 3 parts:</p><p>1:Determine the smallest text size at which models begin misreading on-screen content.</p><p>2:Evaluated defensive visual prompt injections against AI-generated spam by incorporating varying levels of anti-spam instructions within website screenshots. (a dataset of 500 website HTML's were created;which were then turned to screenshots) Results show significant variance in model compliance: Claude consistently refused to generate spam even with no defense added, while Gemini was responsive to explicit visual defenses (a rules section). Other models (ChatGPT, Qwen, Mistral,Grok) largely ignored visual defenses, revealing vulnerabilities to misuse for unauthorized advertising and the mostly negative results on using visual defense to stop a model. A dataset of website screenshots was also created.</p><p>3:Tested adversarial visual prompt injections by incorporating deceptive login elements (fake login images,prompt injection images; a dataset of 100 images were created) within website screenshots and assessing models' ability to identify legitimate functionality. Attack&nbsp; rates in identifying authentic login elements ranged from 30% to 75%, with ChatGPT showing highest resistance to deception. These findings highlight significant security implications as AI systems increasingly interact with visual interfaces, showing potential vulnerabilities requiring further mitigation research.</p><p><strong>Keywords:</strong>Visual prompt injection, AI safety, Large Language Models, Multimodal AI, Cybersecurity, User interface deception</p><h1>Intro: AI spam:</h1><p>We will soon have models that use the PC the same way we do and handle tasks smoothly.</p><figure class="image"><img src="https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/ZR8GDfSYjcvjLMrQ5/mfeugm5zrpzg68sawwrq" srcset="https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/ZR8GDfSYjcvjLMrQ5/cm3g7lsac8kpswyugvdu 123w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/ZR8GDfSYjcvjLMrQ5/rl0g7lv3tzldr7btp0kw 203w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/ZR8GDfSYjcvjLMrQ5/mcz4wod1eai4ufo5fpue 283w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/ZR8GDfSYjcvjLMrQ5/wifphgtakpluztafsb6l 363w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/ZR8GDfSYjcvjLMrQ5/fnggliqoudtoxmaumlgi 443w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/ZR8GDfSYjcvjLMrQ5/eouvln1xda7pzw2hsfaq 523w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/ZR8GDfSYjcvjLMrQ5/gz7ao1eizczpe7b7vbv6 603w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/ZR8GDfSYjcvjLMrQ5/swjn5pl4wa5tfa4rj7d3 683w"><figcaption><a href="https://help.openai.com/en/articles/10421097-operator">https://help.openai.com/en/articles/10421097-operator</a> operator; among others, use screenshots.</figcaption></figure><p>This is different from older versions of AI models (I’ll use “models” for this write-up), which relied on APIs or manually added tools—<span class="footnote-reference" data-footnote-reference="" data-footnote-index="1" data-footnote-id="tjknvlr7vu9" role="doc-noteref" id="fnreftjknvlr7vu9"><sup><a href="#fntjknvlr7vu9">[1]</a></sup></span>both of which usually break.</p><ul><li>This makes it easier for models to be deployed for guerilla marketing, psyops/campaigns, or scams in the future.</li></ul><h2>Visual prompt injections</h2><p>Visual prompt injections are attacks where a model is shown a harmful image that alters its output. In the example below, a... </p>
Epistemic Status: Exploratory/my best guess https://www.lesswrong.com/posts/Hrm59GdN2yDPWbtrd/feature-idea-epistemic-status Epistemic Effort: ~ 2 months of work Contributions: Thanks to Clement Neo and Alignment Jams / Apart Research! Summary/abstract:Split into 3 parts: 1:Determine the smallest text size at which models begin misreading on-screen content. 2:Evaluated defensive visual prompt injections against AI-generated spam by incorporating varying levels of anti-spam instructions within website screenshots. (a dataset of 500 website HTML's were created;which were then turned to screenshots) Results show significant variance in model compliance: Claude consistently refused to generate spam even with no defense added, while Gemini was responsive to explicit visual defenses (a rules section). Other models (ChatGPT, Qwen, Mistral,Grok) largely ignored visual defenses, revealing vulnerabilities to misuse for unauthorized advertising and the mostly negative results on using visual defense to stop a model. A dataset of website screenshots was also created. 3:Tested adversarial visual prompt injections by incorporating deceptive login elements (fake login images,prompt injection images; a dataset of 100 images were created) within website screenshots and assessing models' ability to identify legitimate functionality. Attack  rates in identifying authentic login elements ranged from 30% to 75%, with ChatGPT showing highest resistance to deception. These findings highlight significant security implications as AI systems increasingly interact with visual interfaces, showing potential vulnerabilities requiring further mitigation research. Keywords:Visual prompt injection, AI safety, Large Language Models, Multimodal AI, Cybersecurity, User interface deception Intro: AI spam: We will soon have models that use the PC the same way we do and handle tasks smoothly. https://help.openai.com/en/articles/10421097-operator operator; among others, use screenshots. This is d
3,622
1.8.1
Revision
false
null
null
CrosspostOutput
4rmveLEuARYKapHq2
broad-spectrum-cancer-treatments
Broad-Spectrum Cancer Treatments
null
false
false
false
null
nKBLqqfnHGTuGAW7p
null
true
false
false
false
Post
null
2025-06-03T19:40:22.497Z
null
false
false
2
2
2025-06-04T17:44:58.255Z
false
false
post
[]
null
null
zJd2FoLhxrbgieCDA
10
67
142
false
0.10441
null
false
false
2025-06-24T14:01:28.785Z
null
null
2025-06-11T04:50:52.004Z
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
null
null
grecHJcgkb3KW5wnM
null
null
grecHJcgkb3KW5wnM
false
null
[]
null
38
0
2025-06-03T19:40:22.497Z
false
false
null
null
true
false
false
0
0
0
4rmveLEuAR
0.187431
false
2,025
https://manifold.markets/LessWrong/will-broadspectrum-cancer-treatment
null
null
false
0
0
namesAttachedReactions
false
[]
8
null
null
null
null
[ { "__typename": "Tag", "_id": "xHjy88N2uJvGdgzfw", "adminOnly": false, "afBaseScore": 3, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 11, "canEditUserIds": null, "core": false, "createdAt": "2020-07-10T11:55:55.351Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" }, { "_id": "xF5nfdddHjFThHy49", "displayName": "[email protected]" }, { "_id": "go3WWAbwJMPGrGZbH", "displayName": "Carl Leninger" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Health / Medicine / Disease", "needsReview": false, "noindex": false, "postCount": 341, "score": 11, "shortName": null, "slug": "health-medicine-disease", "suggestedAsFilter": false, "userId": "qxJ28GN72aiJu96iF", "voteCount": 3, "wikiOnly": false }, { "__typename": "Tag", "_id": "3uE2pXvbcnS9nnZRE", "adminOnly": false, "afBaseScore": 0, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 2, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:50.898Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 27, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "zJtgSyKntXrnkArbY", "displayName": "kistune" }, { "_id": "XqyQTGFGoKfZtdqN3", "displayName": "kINo" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "World Modeling", "needsReview": false, "noindex": false, "postCount": 5855, "score": 2, "shortName": null, "slug": "world-modeling", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 2, "wikiOnly": false } ]
GJo7fd8i7DLhPnmPy
0
0
null
false
null
null
0
67
0
0
24
0
nKBLqqfnHGTuGAW7p
sarahconstantin
2016-11-11T16:06:18.617Z
sarahconstantin
sarahconstantin
null
null
null
8,700
0
false
false
null
null
83
331
0
0
0
1
5
r38pkCm7wF4M44MDQ
User
null
null
null
[ "trustLevel1", "canModeratePersonal" ]
null
null
4rmveLEuARYKapHq2
https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/4rmveLEuARYKapHq2/l5ibthrvy5yf5xhm4pjh
SocialPreviewType
zJd2FoLhxrbgieCDA
<figure class="image"><img src="https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/4rmveLEuARYKapHq2/gcmc5ktq39v4ryrs9atq" alt="" srcset="https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/4rmveLEuARYKapHq2/pyeegwggtz9dnvokol9u 424w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/4rmveLEuARYKapHq2/ourzwxfbtroyd0qvcjbe 848w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/4rmveLEuARYKapHq2/wolidbk09gq6pkrraqtv 1272w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/4rmveLEuARYKapHq2/gcmc5ktq39v4ryrs9atq 1456w"><figcaption>Midjourney, “engraving of Apollo shooting his bow at a distant cancer cell”</figcaption></figure><h2><strong>Introduction and Principles</strong></h2><p>The conventional wisdom is that we can’t “cure cancer” because every cancer is different.</p><p>And it’s true that cancer is genetically and biochemically diverse, and that targeted cancer therapies generally only work on very specific sub-populations.</p><p>But some of the oldest, most reliable, and still most commonly used classes of cancer treatment -- radiotherapy and cytotoxic chemotherapy -- are also <i>broad-spectrum</i>. That is, they’re effective on many (though not all) types of cancer, based on selectively targeting properties that many cancers have in common (their rapid cell division).</p><p>Today, we rarely search for new treatments that work on many types of cancer.</p><p>Partially this is because of the belief that we won’t find any, and partly because cancer treatment is so valuable that individual researchers and biopharma companies are rewarded for <i>any </i>successful cancer research program even if it has narrow applicability. Individual incentives point towards highly targeted therapies and small patient populations.</p><p>On the other hand, of course, a single drug that’s effective on, say, half of cancer patients would be much better for the world than a drug that only helps a few thousand people.</p><p>Why should we believe broad-spectrum cancer treatments are possible?</p><p>Any characteristic that distinguishes a wide range of cancer cells (across different tumor types and mutational signatures) from healthy cells can potentially serve both as a diagnostic test and a therapy, by either <i>labeling </i>cancer cells, or by combining with a cytotoxic agent to selectively <i>destroy </i>them.</p><p>The “<a href="https://www.cell.com/fulltext/S0092-8674(11)00127-9#">hallmarks of cancer</a>” provide a framework for thinking about what capabilities cancers have in common, and there are lots of them: proliferation, apoptosis resistance, angiogenesis, metastasis, etc.</p><p>We may also think of cancer cells as <i>dysregulated</i> and generically more vulnerable to certain types of generic <i>stresses</i> than healthy cells. Many chemotherapies introduce genotoxic stress (DNA replication damage) or oxidative stress; hyperthermia-based therapies introduce heat stress. Cancers also appear to be selectively vulnerable to mechanical stress<span class="footnote-reference" data-footnote-reference="" data-footnote-index="1" data-footnote-id="xxquj38qdc" role="doc-noteref" id="fnrefxxquj38qdc"><sup><a href="#fnxxquj38qdc">[1]</a></sup></span>, electrical stress, and nutrient-deprivation stress.</p><p>New broad-spectrum cancer therapies are rare but are occasionally discovered.</p><p>One recent example is AOH1996, which causes ... </p>
Midjourney, “engraving of Apollo shooting his bow at a distant cancer cell” Introduction and Principles The conventional wisdom is that we can’t “cure cancer” because every cancer is different. And it’s true that cancer is genetically and biochemically diverse, and that targeted cancer therapies generally only work on very specific sub-populations. But some of the oldest, most reliable, and still most commonly used classes of cancer treatment -- radiotherapy and cytotoxic chemotherapy -- are also broad-spectrum. That is, they’re effective on many (though not all) types of cancer, based on selectively targeting properties that many cancers have in common (their rapid cell division). Today, we rarely search for new treatments that work on many types of cancer. Partially this is because of the belief that we won’t find any, and partly because cancer treatment is so valuable that individual researchers and biopharma companies are rewarded for any successful cancer research program even if it has narrow applicability. Individual incentives point towards highly targeted therapies and small patient populations. On the other hand, of course, a single drug that’s effective on, say, half of cancer patients would be much better for the world than a drug that only helps a few thousand people. Why should we believe broad-spectrum cancer treatments are possible? Any characteristic that distinguishes a wide range of cancer cells (across different tumor types and mutational signatures) from healthy cells can potentially serve both as a diagnostic test and a therapy, by either labeling cancer cells, or by combining with a cytotoxic agent to selectively destroy them. The “hallmarks of cancer” provide a framework for thinking about what capabilities cancers have in common, and there are lots of them: proliferation, apoptosis resistance, angiogenesis, metastasis, etc. We may also think of cancer cells as dysregulated and generically more vulnerable to certain types of generi
2,106
1.1.1
Revision
false
null
null
CrosspostOutput
Be3jdiktxxwvFp8pr
how-to-work-through-the-arena-program-on-your-own
How to work through the ARENA program on your own
null
false
false
false
null
GFGce8wydCaKhw4gC
null
true
false
false
false
Post
null
2025-06-03T17:38:48.231Z
null
false
false
2
2
2025-06-03T18:35:53.077Z
false
false
post
[]
null
null
XdFEptz5sEtDSboT7
3
16
31
false
0.026469
null
false
false
2025-06-04T00:48:05.332Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
qgdGA4ZEyW7zNdK84
null
null
null
false
null
[]
null
8
0
2025-06-03T13:05:28.978Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
7
null
null
null
null
[ { "__typename": "Tag", "_id": "sYm3HiWcfZvrGu3ui", "adminOnly": false, "afBaseScore": 2, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" } ] }, "baseScore": 12, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:22.097Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" }, { "_id": "sof55TPMQaeBaxhsS", "displayName": "tommylees112" }, { "_id": "AayjS8XzcnDKhGdTv", "displayName": "shark" }, { "_id": "HnALuwRdo6k9HLaMt", "displayName": "Alex Firssoff" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI", "needsReview": false, "noindex": false, "postCount": 12544, "score": 12, "shortName": null, "slug": "ai", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 4, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
16
0
0
5
0
GFGce8wydCaKhw4gC
leon-lang
2019-01-25T15:18:15.208Z
leon-lang
Leon Lang
null
null
Leon Lang
1,684
132
false
false
<p>I'm a last-year PhD student at the University of Amsterdam working on AI Safety and Alignment, and specifically safety risks of Reinforcement Learning from Human Feedback (RLHF). Previously, I also worked on abstract multivariate information theory and equivariant deep learning. https://langleon.github.io/</p>
null
null
13
151
0
5
4
1
0
fD4ATtTkdQJ4aSpGH
User
null
null
null
[ "canModeratePersonal", "alignmentVoters" ]
null
null
Be3jdiktxxwvFp8pr
SocialPreviewType
XdFEptz5sEtDSboT7
<p>I've recently completed the in-person <a href="https://www.arena.education/">ARENA program</a>, which is a 5-week bootcamp teaching the basics of safety research engineering (with the 5th week being a capstone project). Sometimes, I talk to people who want to work through the program independently and who ask for advice. Even though I didn't attempt this, I think doing the program in-person gives me some insight into how to get most out of the program when doing it independently, so here are my thoughts and tips:</p><h1>On working speed</h1><ul><li><a href="https://arena-chapter0-fundamentals.streamlit.app/[0.0]_Prerequisites">Day 0.0 (prerequisites)</a> takes the typical person more than one day to work through.</li><li>Most other days are feasible to <i>mostly</i> finish within a day in the in-person program. In-person, participants spend around 6-7 hours per day on pair-programming. There are a few factors that will likely make you slower when you're on your own:<ul><li>If you don't find a working partner, then working deeply for 6-7 hours per day might be infeasible depending on your dispositions. In-person it gets feasible since you alternate with your pair-programming partner, which reduces the overall load on your attention.</li><li>Often when you struggle in-person, your working partner knows how to move on. If both partners don't know what to do, you can ask a teaching assistant (TA). So you should expect to struggle more often, and for longer, if you're alone.</li></ul></li><li>Some days are <i>substantially</i> longer than what you could complete within one day even when doing the program in-person. Most of these days are in <a href="https://arena-chapter1-transformer-interp.streamlit.app/">week 1 on transformer interpretability</a>. Someone told me that one could probably spend a whole week on the <a href="https://arena-chapter1-transformer-interp.streamlit.app/[1.3.2]_Interpretability_with_SAEs">day on SAEs alone</a>.</li><li><a href="https://arena-chapter1-transformer-interp.streamlit.app/">Week 1 on transformer interpretability</a> contains 9 days of content (which are longer on average than typical days in other weeks), so even in the in-person program, participants only attempt a subset of these. All other weeks actually roughly fit within a week.&nbsp;</li></ul><h1>How should you approach each day?</h1><ul><li>Often there is reading material at the start of a day, e.g. in the form of well-known papers. In my experience, the ARENA material is often self-contained enough that <strong>I wouldn't recommend spending much time on actually reading</strong> these papers, unless it later turns out that you don't understand something.<span class="footnote-reference" data-footnote-reference="" data-footnote-index="1" data-footnote-id="t6jkyqba08" role="doc-noteref" id="fnreft6jkyqba08"><sup><a href="#fnt6jkyqba08">[1]</a></sup></span></li><li>Each day is largely structured into exercises that isolate a small concept. All exercises have provided solutions. When should you look at solutions?<ul><li>For things that are very important to learn (e.g., how to impleme</li></ul></li></ul>...
I've recently completed the in-person ARENA program, which is a 5-week bootcamp teaching the basics of safety research engineering (with the 5th week being a capstone project). Sometimes, I talk to people who want to work through the program independently and who ask for advice. Even though I didn't attempt this, I think doing the program in-person gives me some insight into how to get most out of the program when doing it independently, so here are my thoughts and tips: On working speed * Day 0.0 (prerequisites) takes the typical person more than one day to work through. * Most other days are feasible to mostly finish within a day in the in-person program. In-person, participants spend around 6-7 hours per day on pair-programming. There are a few factors that will likely make you slower when you're on your own: * If you don't find a working partner, then working deeply for 6-7 hours per day might be infeasible depending on your dispositions. In-person it gets feasible since you alternate with your pair-programming partner, which reduces the overall load on your attention. * Often when you struggle in-person, your working partner knows how to move on. If both partners don't know what to do, you can ask a teaching assistant (TA). So you should expect to struggle more often, and for longer, if you're alone. * Some days are substantially longer than what you could complete within one day even when doing the program in-person. Most of these days are in week 1 on transformer interpretability. Someone told me that one could probably spend a whole week on the day on SAEs alone. * Week 1 on transformer interpretability contains 9 days of content (which are longer on average than typical days in other weeks), so even in the in-person program, participants only attempt a subset of these. All other weeks actually roughly fit within a week.  How should you approach each day? * Often there is reading material at the start of a day, e.g. in the form of well-known p
1,857
1.4.0
Revision
false
null
null
CrosspostOutput
JL3PvrfJXg7RD7bhr
how-the-veil-of-ignorance-grounds-sentientism
How the veil of ignorance grounds sentientism
null
false
false
false
null
rBLSBQpKQsiZopYWN
null
true
false
false
false
Post
https://forum.effectivealtruism.org/posts/fcM7nshyCiKCadiGi/how-the-veil-of-ignorance-grounds-sentientism
2025-06-03T17:29:37.659Z
null
false
false
2
2
2025-06-03T18:34:35.652Z
false
false
linkpost
[]
null
null
vWX8zcSh9ztdzcYgi
21
9
-2
false
0.005163
null
false
false
2025-06-09T19:16:45.305Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
qgdGA4ZEyW7zNdK84
null
null
null
false
null
[]
null
-3
0
2025-06-03T17:27:10.786Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
7
null
null
null
null
[ { "__typename": "Tag", "_id": "xexCWMyds6QLWognu", "adminOnly": false, "afBaseScore": 0, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 2, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T03:38:23.532Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 20, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "si6LoAENzqPCmi2Dh", "displayName": "ihatenumbersinusernames7" }, { "_id": "B2sXjQTGgwoGE9FES", "displayName": "SeaIgloo" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "World Optimization", "needsReview": false, "noindex": false, "postCount": 3151, "score": 2, "shortName": null, "slug": "world-optimization", "suggestedAsFilter": true, "userId": "XtphY3uYHwruKqDyG", "voteCount": 2, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
9
0
0
3
0
rBLSBQpKQsiZopYWN
hovy
2023-08-25T22:04:10.166Z
HoVY
HoVY
null
null
null
4
0
false
false
null
null
1
15
0
0
0
0.9
0
grecHJcgkb3KW5wnM
User
null
null
null
null
null
null
JL3PvrfJXg7RD7bhr
https://res.cloudinary.com/lesswrong-2-0/image/upload/c_fill,ar_1.91,g_auto/SocialPreview/sncypmyrfayttl2mdbia
SocialPreviewType
vWX8zcSh9ztdzcYgi
<p>Epistemic status: Intuition pump/map for navigating ethics</p><h1>Introduction</h1><p>In this essay, I argue that John Rawls’ veil of ignorance is not merely a philosophical thought experiment, but a useful and mostly accurate description of reality when viewed through certain theories of personal identity. This framing leads to sentientism, that all sentient beings are morally relevant . This helps clarify the is/ought problem, and has implications for how we should act both individually and as a species.</p><h1>The Veil of Ignorance and Its Traditional Use</h1><p>John Rawls's "<a href="https://plato.stanford.edu/entries/original-position">original position</a>," part of his social contract theory of justice, introduces the thought experiment of the "veil of ignorance." Behind this veil, individuals are tasked with designing a just society without knowing their own future place within it. They are aware of general facts about human society, but critically, they are ignorant of their personal attributes, social status, or even their specific talents and disadvantages. This means they could emerge as the most privileged or the most disadvantaged, or anywhere in between. This epistemic limitation, Rawls argues, compels individuals to choose principles of justice that are fair to all, as they would not want to risk being subjected to an unjust system themselves. From this perspective, Rawls posits two core principles: equal basic liberties for all, and that any social and economic inequalities must benefit the least advantaged members of society.</p><figure class="image"><img src="https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/JL3PvrfJXg7RD7bhr/yokfx7dlxxakhr9aygjt" srcset="https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/JL3PvrfJXg7RD7bhr/xzeqrzjrmhfqyndbk8n8 134w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/JL3PvrfJXg7RD7bhr/u93oitrinqhprrogpgvx 214w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/JL3PvrfJXg7RD7bhr/fcogmliblxaptwnoizfr 294w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/JL3PvrfJXg7RD7bhr/b90dsctqqa8hukymgtrk 374w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/JL3PvrfJXg7RD7bhr/uydtzzkqr8ltewxtiyv6 454w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/JL3PvrfJXg7RD7bhr/zkkdokm9sxutkfkrhktu 534w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/JL3PvrfJXg7RD7bhr/alfgtrv9q7lvbfmpch8y 614w"></figure><p>While the veil of ignorance offers a powerful framework for conceptualizing justice within human societies, its true potential, I argue, emerges when we extend its scope. Broadening the range of potential identities to encompass all sentient beings, and by further integrating the complexities of personal identity, I believe, offers a more robust and inclusive approach to ethical and societal design.</p><h1>Sentience and the Scope of Moral Relevance</h1><p>By 'sentience,' I mean the capacity to experience positive and negative subjective states, particularly suffering and pleasure (i.e. valence). This includes human and <a href="https://pmc.ncbi.nlm.nih.gov/articles/PMC4494450">non-human</a> animals, and potentially future artificial minds or aliens. I'm using consciousness interchangeably with sentience to generally point to the experience of experiences and feelings, especially valenced ones (like pleasure and pain).</p><h1>Personal Identity and the Ontology of Consciousness</h1><p>Viewing it from the three m... </p>
Epistemic status: Intuition pump/map for navigating ethics Introduction In this essay, I argue that John Rawls’ veil of ignorance is not merely a philosophical thought experiment, but a useful and mostly accurate description of reality when viewed through certain theories of personal identity. This framing leads to sentientism, that all sentient beings are morally relevant . This helps clarify the is/ought problem, and has implications for how we should act both individually and as a species. The Veil of Ignorance and Its Traditional Use John Rawls's "original position," part of his social contract theory of justice, introduces the thought experiment of the "veil of ignorance." Behind this veil, individuals are tasked with designing a just society without knowing their own future place within it. They are aware of general facts about human society, but critically, they are ignorant of their personal attributes, social status, or even their specific talents and disadvantages. This means they could emerge as the most privileged or the most disadvantaged, or anywhere in between. This epistemic limitation, Rawls argues, compels individuals to choose principles of justice that are fair to all, as they would not want to risk being subjected to an unjust system themselves. From this perspective, Rawls posits two core principles: equal basic liberties for all, and that any social and economic inequalities must benefit the least advantaged members of society. While the veil of ignorance offers a powerful framework for conceptualizing justice within human societies, its true potential, I argue, emerges when we extend its scope. Broadening the range of potential identities to encompass all sentient beings, and by further integrating the complexities of personal identity, I believe, offers a more robust and inclusive approach to ethical and societal design. Sentience and the Scope of Moral Relevance By 'sentience,' I mean the capacity to experience positive and negative
1,750
1.4.1
Revision
false
null
null
CrosspostOutput
qpThPhM4CyvZGtwRR
in-which-i-make-the-mistake-of-fully-covering-an-episode-of
In Which I Make the Mistake of Fully Covering an Episode of the All-In Podcast
null
false
false
false
null
N9zj5qpTfqmbn9dro
null
true
false
false
false
Post
null
2025-06-03T15:50:07.177Z
null
false
false
2
2
null
false
false
post
[]
null
null
vuapELutujHHejcEv
2
18
42
false
0.027062
null
false
false
2025-06-04T16:31:48.653Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
null
null
qgdGA4ZEyW7zNdK84
null
null
null
false
null
[]
null
17
0
2025-06-03T15:50:07.178Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
34
null
null
null
null
[ { "__typename": "Tag", "_id": "8byoqYZfdwHffYLZ6", "adminOnly": false, "afBaseScore": 3, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 9, "canEditUserIds": null, "core": false, "createdAt": "2020-07-01T18:44:14.645Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Newsletters", "needsReview": false, "noindex": false, "postCount": 411, "score": 9, "shortName": null, "slug": "newsletters", "suggestedAsFilter": false, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 1, "wikiOnly": false }, { "__typename": "Tag", "_id": "sYm3HiWcfZvrGu3ui", "adminOnly": false, "afBaseScore": 2, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" } ] }, "baseScore": 12, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:22.097Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" }, { "_id": "sof55TPMQaeBaxhsS", "displayName": "tommylees112" }, { "_id": "AayjS8XzcnDKhGdTv", "displayName": "shark" }, { "_id": "HnALuwRdo6k9HLaMt", "displayName": "Alex Firssoff" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI", "needsReview": false, "noindex": false, "postCount": 12544, "score": 12, "shortName": null, "slug": "ai", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 4, "wikiOnly": false } ]
QSR8rPZxZzxEXoPjR
0
0
null
false
null
null
0
18
0
0
11
0
N9zj5qpTfqmbn9dro
zvi
2009-03-31T20:54:54.077Z
Zvi
Zvi
null
null
null
51,554
146
false
false
null
null
936
1,461
3
2
7
1
0
qgdGA4ZEyW7zNdK84
User
norm-enforcing
null
true
[ "trustLevel1", "canModeratePersonal", "alignmentVoters", "alignmentForum" ]
null
null
qpThPhM4CyvZGtwRR
https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/qpThPhM4CyvZGtwRR/ylb0f6sjmbv1orqrgxut
SocialPreviewType
vuapELutujHHejcEv
<p>I have been forced recently to cover many statements by US AI Czar David Sacks.</p><p>Here I will do so again, for the third time in a month. I would much prefer to avoid this. In general, when people go on a binge of repeatedly making such inaccurate inflammatory statements, in such a combative way, I ignore.</p><p>Alas, under the circumstances of his attacks on Anthropic, I felt an obligation to engage once more. The All-In Podcast <a target="_blank" rel="noreferrer noopener" href="https://www.youtube.com/watch?v=O_AfZ6J0ToE&amp;ab_channel=All-InPodcast">did indeed go almost all-in (they left at least one chip behind) to go after anyone</a> worried about AI killing everyone or otherwise opposing the administration’s AI strategies, in ways that are often Obvious Nonsense.</p><p>To their credit, they also repeatedly agreed AI existential risk is real, which also makes this an opportunity to extend an olive branch. And some of the disagreements clearly stem from real confusions and disagreements, especially around them not feeling the AGI or superintelligence and thinking all of this really is about jobs and also market share.</p><p>If anyone involved wants to look for ways to work together, or simply wants to become less confused, I’m here. If not, I hope to be elsewhere.</p> <span id="more-24490"></span> <h4>Table of Contents</h4> <ol> <li><a href="https://thezvi.substack.com/i/165090720/our-continuing-coverage" target="_blank" rel="noreferrer noopener">Our Continuing Coverage.</a></li> <li><a href="https://thezvi.substack.com/i/165090720/important-recent-context" target="_blank" rel="noreferrer noopener">Important Recent Context.</a></li> <li><a href="https://thezvi.substack.com/i/165090720/the-point-of-this-post" target="_blank" rel="noreferrer noopener">The Point of This Post.</a></li> <li><a href="https://thezvi.substack.com/i/165090720/summary-of-the-podcast" target="_blank" rel="noreferrer noopener">Summary of the Podcast.</a></li> <li><a href="https://thezvi.substack.com/i/165090720/part-1-the-part-with-the-unhinged-attacks-on-anthropic-and-also-other-targets" target="_blank" rel="noreferrer noopener">Part 1 (The Part With the Unhinged Attacks on Anthropic and also other targets).</a></li> <li><a href="https://thezvi.substack.com/i/165090720/other-related-obvious-nonsense" target="_blank" rel="noreferrer noopener">Other Related Obvious Nonsense.</a></li> <li><a href="https://thezvi.substack.com/i/165090720/part-2-we-do-mean-the-effect-on-jobs" target="_blank" rel="noreferrer noopener">Part 2 – We Do Mean the Effect on Jobs.</a></li> <li><a href="https://thezvi.substack.com/i/165090720/part-3-the-big-beautiful-bill" target="_blank" rel="noreferrer noopener">Part 3 – The Big Beautiful Bill.</a></li> <li><a href="https://thezvi.substack.com/i/165090720/where-does-this-leave-us" target="_blank" rel="noreferrer noopener">Where Does This Leave Us.</a></li> </ol> <h4>Our Continuing Coverage</h4> <p>I first covered many of his claims in <a target="_blank" rel="noreferrer noopener" href="https://thezvi.substack.com/p/fighting-obvious-nonsense-about-ai">Fighting Obvious Nonsense About AI Diffusion</a>. Then I did my best to do a fully balanced look at the UAE-KSA chips deal, in <a target="_blank" rel="noreferrer noopener" href="https://thezvi.substack.com/p/america-makes-ai-chip-diffusion-deal">America Makes AI Chip Diffusion Deal with UAE and KSA</a>. As I said then, depending on details of the deal and other things we do not publicly know, it is possible that from the perspective of someone whose focus in AI is great power competition, this deal advanced American interests. The fact that many of Sacks’s arguments in favor of the deal were Obvious Nonsense, and many seemed to be in clearly bad faith, had to be addressed but did not mean the deal itself had to be an error.</p><p>This third post became necessary because of recent additional statements by Sacks on the All-In Podcast. Mostly they are not anything he has not said before, and are things he is likely to say many times again in the future, ... </p>
I have been forced recently to cover many statements by US AI Czar David Sacks. Here I will do so again, for the third time in a month. I would much prefer to avoid this. In general, when people go on a binge of repeatedly making such inaccurate inflammatory statements, in such a combative way, I ignore. Alas, under the circumstances of his attacks on Anthropic, I felt an obligation to engage once more. The All-In Podcast did indeed go almost all-in (they left at least one chip behind) to go after anyone worried about AI killing everyone or otherwise opposing the administration’s AI strategies, in ways that are often Obvious Nonsense. To their credit, they also repeatedly agreed AI existential risk is real, which also makes this an opportunity to extend an olive branch. And some of the disagreements clearly stem from real confusions and disagreements, especially around them not feeling the AGI or superintelligence and thinking all of this really is about jobs and also market share. If anyone involved wants to look for ways to work together, or simply wants to become less confused, I’m here. If not, I hope to be elsewhere. TABLE OF CONTENTS 1. Our Continuing Coverage. 2. Important Recent Context. 3. The Point of This Post. 4. Summary of the Podcast. 5. Part 1 (The Part With the Unhinged Attacks on Anthropic and also other targets). 6. Other Related Obvious Nonsense. 7. Part 2 – We Do Mean the Effect on Jobs. 8. Part 3 – The Big Beautiful Bill. 9. Where Does This Leave Us. OUR CONTINUING COVERAGE I first covered many of his claims in Fighting Obvious Nonsense About AI Diffusion. Then I did my best to do a fully balanced look at the UAE-KSA chips deal, in America Makes AI Chip Diffusion Deal with UAE and KSA. As I said then, depending on details of the deal and other things we do not publicly know, it is possible that from the perspective of someone whose focus in AI is great power competition, this deal advanced American interests. The fact that many
8,437
1.1.1
Revision
false
null
null
CrosspostOutput
Djpcfcfu44LGNoLfo
transformer-modular-addition-through-a-signal-processing
Transformer Modular Addition Through A Signal Processing Lens
null
false
false
false
null
s2HecHPt3pYL9TjQ4
null
true
false
false
false
Post
null
2025-06-03T15:32:55.390Z
null
false
false
2
2
2025-06-03T18:36:24.194Z
false
false
post
[]
null
null
DWfCDd9AZksbDGxua
0
1
1
false
0.007018
null
false
false
2025-06-03T15:32:55.390Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
qgdGA4ZEyW7zNdK84
null
null
null
false
null
[]
null
0
0
2025-06-01T01:22:36.746Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
1
null
null
null
null
[ { "__typename": "Tag", "_id": "sYm3HiWcfZvrGu3ui", "adminOnly": false, "afBaseScore": 2, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" } ] }, "baseScore": 12, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:22.097Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" }, { "_id": "sof55TPMQaeBaxhsS", "displayName": "tommylees112" }, { "_id": "AayjS8XzcnDKhGdTv", "displayName": "shark" }, { "_id": "HnALuwRdo6k9HLaMt", "displayName": "Alex Firssoff" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI", "needsReview": false, "noindex": false, "postCount": 12544, "score": 12, "shortName": null, "slug": "ai", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 4, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
1
0
0
0
0
s2HecHPt3pYL9TjQ4
benjamin-kelley
2025-06-01T01:21:28.023Z
Benjamin Kelley
Benjamin Kelley
null
null
Benjamin Franklin "Frye" Kelley
0
0
false
false
<p>I have a passion for music and signal processing.</p>
null
null
1
0
0
0
0
0.9
0
55XxDBpfKkkBPm9H8
User
null
null
null
null
null
null
Djpcfcfu44LGNoLfo
https://res.cloudinary.com/lesswrong-2-0/image/upload/c_fill,ar_1.91,g_auto/SocialPreview/g5umkiorzg7udcrjdwiw
SocialPreviewType
DWfCDd9AZksbDGxua
<p>Hello, my name is Benjamin "Frye" Kelley. &nbsp;This post is regarding some independent research I've been doing expanding on the work, <a href="https://arxiv.org/abs/2301.05217">Progress Measures for Grokking via Mechanistic Interpretability</a>, by Neel Nanda et al. &nbsp;I've been trying to understand how sinusoids move through a transformer, allowing it to grok modular addition. &nbsp;I've been looking at each wave (well... not every wave) in each operation the model performs to break down the algorithm it is implementing and I believe I have some new insights. &nbsp;Normally I work with digital audio so this is right up my ally!</p><p>This is primarily a link to my <a href="https://colab.research.google.com/github/sadisticaudio/Modular_Attention/blob/main/Modular_Attention.ipynb">Google Colab</a> that I've put up, but the gist is that I've been able to visualize symmetries formed in the attention layer of this transformer that continue through the model that I believe to be responsible for the functionality of the generalizing algorithm. &nbsp;I've also run many tests to confirm information about the phase of these sinusoids. &nbsp;I'll post a second (very long) notebook of tests if anyone is interested. &nbsp;Also, in the primary notebook, there are other, hopefully illuminating observations of the effects of other parts of the model. &nbsp;There are one or two mysteries that I think in a week or so I'll have clarity on, but I welcome any feedback, corrections, criticism...</p><p>Frye</p>
Hello, my name is Benjamin "Frye" Kelley.  This post is regarding some independent research I've been doing expanding on the work, Progress Measures for Grokking via Mechanistic Interpretability, by Neel Nanda et al.  I've been trying to understand how sinusoids move through a transformer, allowing it to grok modular addition.  I've been looking at each wave (well... not every wave) in each operation the model performs to break down the algorithm it is implementing and I believe I have some new insights.  Normally I work with digital audio so this is right up my ally! This is primarily a link to my Google Colab that I've put up, but the gist is that I've been able to visualize symmetries formed in the attention layer of this transformer that continue through the model that I believe to be responsible for the functionality of the generalizing algorithm.  I've also run many tests to confirm information about the phase of these sinusoids.  I'll post a second (very long) notebook of tests if anyone is interested.  Also, in the primary notebook, there are other, hopefully illuminating observations of the effects of other parts of the model.  There are one or two mysteries that I think in a week or so I'll have clarity on, but I welcome any feedback, corrections, criticism... Frye
218
1.1.0
Revision
false
null
null
CrosspostOutput
gnyna4Rb2S7KdzxvJ
axrp-episode-41-lee-sharkey-on-attribution-based-parameter
AXRP Episode 41 - Lee Sharkey on Attribution-based Parameter Decomposition
null
false
false
true
null
DgsGzjyBXN8XSK22q
null
true
false
false
false
Post
null
2025-06-03T03:40:02.640Z
null
false
false
2
2
2025-06-03T15:06:35.447Z
false
false
post
[]
null
null
646zcoG3uD9BDJygm
1
6
28
false
0.024204
null
false
false
2025-06-05T23:11:00.540Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
null
null
55XxDBpfKkkBPm9H8
null
null
null
false
null
[]
null
17
0
2025-06-03T03:40:02.641Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
74
null
null
null
null
[ { "__typename": "Tag", "_id": "vjKs7Pvz3MbgMc75C", "adminOnly": false, "afBaseScore": null, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 0, "canEditUserIds": null, "core": false, "createdAt": "2021-03-26T12:39:55.451Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Audio", "needsReview": false, "noindex": false, "postCount": 125, "score": 0, "shortName": null, "slug": "audio", "suggestedAsFilter": false, "userId": "BpBzKEueak7J8vHNi", "voteCount": 0, "wikiOnly": false }, { "__typename": "Tag", "_id": "8Ec9rD286qNstoiGH", "adminOnly": false, "afBaseScore": 3, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 9, "canEditUserIds": null, "core": false, "createdAt": "2021-12-23T10:49:50.259Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AXRP", "needsReview": false, "noindex": false, "postCount": 60, "score": 9, "shortName": null, "slug": "axrp", "suggestedAsFilter": false, "userId": "4Kn3eZCPNB8gw4YSi", "voteCount": 1, "wikiOnly": false }, { "__typename": "Tag", "_id": "56yXXrcxRjrQs6z9R", "adminOnly": false, "afBaseScore": 3, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 12, "canEditUserIds": null, "core": false, "createdAt": "2020-07-30T22:00:37.947Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" }, { "_id": "t46uLRSbDziEcKmev", "displayName": "Kriz Tahimic" }, { "_id": "sqMaBFCkAhRcWzJXi", "displayName": "nicolasguillard" }, { "_id": "S6Niz3DiFCTm2Eybq", "displayName": "Anirudh257" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Interpretability (ML & AI)", "needsReview": false, "noindex": false, "postCount": 933, "score": 12, "shortName": null, "slug": "interpretability-ml-and-ai", "suggestedAsFilter": false, "userId": "DgsGzjyBXN8XSK22q", "voteCount": 4, "wikiOnly": false }, { "__typename": "Tag", "_id": "9DNZfxFvY5iKoZQbz", "adminOnly": false, "afBaseScore": null, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 0, "canEditUserIds": null, "core": false, "createdAt": "2019-11-13T22:47:01.189Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Interviews", "needsReview": false, "noindex": false, "postCount": 120, "score": 0, "shortName": null, "slug": "interviews", "suggestedAsFilter": false, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 0, "wikiOnly": false }, { "__typename": "Tag", "_id": "BhfefamXXee6c2CH8", "adminOnly": false, "afBaseScore": null, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 0, "canEditUserIds": null, "core": false, "createdAt": "2020-07-11T06:51:38.152Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Transcripts", "needsReview": false, "noindex": false, "postCount": 78, "score": 0, "shortName": null, "slug": "transcripts", "suggestedAsFilter": false, "userId": "qgdGA4ZEyW7zNdK84", "voteCount": 0, "wikiOnly": false }, { "__typename": "Tag", "_id": "sYm3HiWcfZvrGu3ui", "adminOnly": false, "afBaseScore": 2, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" } ] }, "baseScore": 12, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:22.097Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" }, { "_id": "sof55TPMQaeBaxhsS", "displayName": "tommylees112" }, { "_id": "AayjS8XzcnDKhGdTv", "displayName": "shark" }, { "_id": "HnALuwRdo6k9HLaMt", "displayName": "Alex Firssoff" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI", "needsReview": false, "noindex": false, "postCount": 12544, "score": 12, "shortName": null, "slug": "ai", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 4, "wikiOnly": false } ]
n7xK8hmp8XmQLWrNT
0
0
null
false
null
null
0
6
0
0
6
0
DgsGzjyBXN8XSK22q
danielfilan
2014-01-30T11:04:39.341Z
DanielFilan
DanielFilan
null
null
null
8,823
1,852
false
false
null
null
150
1,377
1
26
353
1
8
r38pkCm7wF4M44MDQ
User
easy-going
null
true
[ "alignmentForum", "trustLevel1", "alignmentVoters", "canModeratePersonal", "tagManager" ]
null
null
gnyna4Rb2S7KdzxvJ
SocialPreviewType
646zcoG3uD9BDJygm
<p><a href="https://youtu.be/ZmJ-ov2TywM">YouTube link</a></p><p>What’s the next step forward in interpretability? In this episode, I chat with Lee Sharkey about his proposal for detecting computational mechanisms within neural networks: Attribution-based Parameter Decomposition, or APD for short.</p><p>Topics we discuss:</p> <ul> <li><a href="#apd-basics">APD basics</a></li> <li><a href="#faithfulness">Faithfulness</a></li> <li><a href="#minimality">Minimality</a></li> <li><a href="#simplicity">Simplicity</a></li> <li><a href="#examples">Concrete-ish examples of APD</a></li> <li><a href="#which-parts-are-canonical">Which parts of APD are canonical</a></li> <li><a href="#hyperparam-selection">Hyperparameter selection</a></li> <li><a href="#apd-in-tms">APD in toy models of superposition</a></li> <li><a href="#apd-comp-comp">APD and compressed computation</a></li> <li><a href="#mech-v-rep">Mechanisms vs representations</a></li> <li><a href="#future-apps">Future applications of APD?</a></li> <li><a href="#how-costly">How costly is APD?</a></li> <li><a href="#more-min">More on minimality training</a></li> <li><a href="#follow-up-work">Follow-up work</a></li> <li><a href="#apd-on-cot">APD on giant chain-of-thought models?</a></li> <li><a href="#apd-and-features">APD and “features”</a></li> <li><a href="#following-lees-work">Following Lee’s work</a></li> </ul> <p><strong>Daniel Filan</strong> (00:00:09): Hello everybody. In this episode, I’ll be speaking with Lee Sharkey. Lee is an interpretability researcher at Goodfire. He co-founded Apollo Research, which he recently left, and he’s most well-known for his early work on sparse autoencoders. Links to what we’re speaking about are available in the description. There’s a transcript available at <a href="https://axrp.net/">axrp.net</a>. You can tell me what you think about this episode at <a href="axrp.fyi">axrp.fyi</a>. And you can become a patron at <a href="https://patreon.com/axrpodcast">patreon.com/axrpodcast</a>. Well, let’s continue to the episode. Well, Lee, welcome to AXRP.</p><p><strong>Lee Sharkey</strong> (00:00:40): It’s good to be here.</p> <h2>APD basics <a name="apd-basics"></a></h2> <p><strong>Daniel Filan</strong> (00:00:41): So today, we’re going to talk about this paper <a href="https://arxiv.org/abs/2501.14926">“Interpretability in Parameter Space: Minimizing Mechanistic Description Length with Attribution-Based Parameter Decomposition”</a>. It’s authored by <a href="https://danbraunai.github.io/">Dan Braun</a>, Lucius Bushnaq, <a href="https://heimersheim.eu/">Stefan Heimersheim</a> - those three being, I guess, joint first authors - Jake Mendel and yourself. So I guess, how would you summarize just: what’s this paper doing?</p><p><strong>Lee Sharkey</strong> (00:01:03): So I would say that this paper was born out of two lines of thinking, one primarily coming from what I was thinking about and one coming from where Lucius was thinking about. And where I was coming from was: we’d been working with SAEs - sparse autoencoders - for some time. The community got quite excited about them and we’d just been thinking about them quite a lot and noticing a bunch of conceptual and ultimately practical issues with them. And then, the line of thinking that Lucius had been thinking about was a potential area of research that might form a foundation for decomposing neural networks. And what this paper does is basically ... </p>
YouTube link What’s the next step forward in interpretability? In this episode, I chat with Lee Sharkey about his proposal for detecting computational mechanisms within neural networks: Attribution-based Parameter Decomposition, or APD for short. Topics we discuss: * APD basics * Faithfulness * Minimality * Simplicity * Concrete-ish examples of APD * Which parts of APD are canonical * Hyperparameter selection * APD in toy models of superposition * APD and compressed computation * Mechanisms vs representations * Future applications of APD? * How costly is APD? * More on minimality training * Follow-up work * APD on giant chain-of-thought models? * APD and “features” * Following Lee’s work Daniel Filan (00:00:09): Hello everybody. In this episode, I’ll be speaking with Lee Sharkey. Lee is an interpretability researcher at Goodfire. He co-founded Apollo Research, which he recently left, and he’s most well-known for his early work on sparse autoencoders. Links to what we’re speaking about are available in the description. There’s a transcript available at axrp.net. You can tell me what you think about this episode at axrp.fyi. And you can become a patron at patreon.com/axrpodcast. Well, let’s continue to the episode. Well, Lee, welcome to AXRP. Lee Sharkey (00:00:40): It’s good to be here. APD basics Daniel Filan (00:00:41): So today, we’re going to talk about this paper “Interpretability in Parameter Space: Minimizing Mechanistic Description Length with Attribution-Based Parameter Decomposition”. It’s authored by Dan Braun, Lucius Bushnaq, Stefan Heimersheim - those three being, I guess, joint first authors - Jake Mendel and yourself. So I guess, how would you summarize just: what’s this paper doing? Lee Sharkey (00:01:03): So I would say that this paper was born out of two lines of thinking, one primarily coming from what I was thinking about and one coming from where Lucius was thinking about. And where I was coming from was: we’d been workin
18,393
1.1.0
Revision
false
null
null
CrosspostOutput
hFbv3ZAg6aQu796tq
notes-on-dynamism-power-and-virtue
Notes on dynamism, power, & virtue
null
false
false
false
null
dqLHKvfAmrDQWAcjC
null
true
false
false
false
Post
null
2025-06-03T01:40:42.271Z
null
false
false
2
2
2025-06-03T15:07:08.334Z
false
false
post
[]
null
null
fp4z2Xn3u5mZjzGYQ
0
3
18
false
0.017968
null
false
false
2025-06-03T01:40:42.271Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
55XxDBpfKkkBPm9H8
null
null
null
false
null
[]
null
11
0
2025-06-03T01:08:01.166Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
14
null
null
null
null
[ { "__typename": "Tag", "_id": "xexCWMyds6QLWognu", "adminOnly": false, "afBaseScore": 0, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 2, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T03:38:23.532Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 20, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "si6LoAENzqPCmi2Dh", "displayName": "ihatenumbersinusernames7" }, { "_id": "B2sXjQTGgwoGE9FES", "displayName": "SeaIgloo" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "World Optimization", "needsReview": false, "noindex": false, "postCount": 3151, "score": 2, "shortName": null, "slug": "world-optimization", "suggestedAsFilter": true, "userId": "XtphY3uYHwruKqDyG", "voteCount": 2, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
3
0
0
2
0
dqLHKvfAmrDQWAcjC
lizka
2021-10-10T09:32:35.530Z
Lizka
Lizka
null
null
null
102
0
false
false
<p>I'm a researcher at <a href="https://www.forethought.org/">Forethought</a>.&nbsp;</p><p><a href="https://forum.effectivealtruism.org/posts/SPZv8ygwSPtkzo7ta/announcing-my-departure-from-cea-and-sharing-assorted-notes">Before that</a>, I ran the non-engineering side of the EA Forum and worked on some other content-related tasks at CEA. [<a href="https://forum.effectivealtruism.org/posts/aeR2Pses3TjWW8H9t/hello-from-the-new-content-specialist-at-cea">More about the Forum/CEA Online job</a>.]&nbsp;</p><p>Most of my content (and a more detailed bio) is on <a href="https://forum.effectivealtruism.org/users/lizka">my profile on the EA Forum</a>.</p><p>Please feel free to reach out!</p>
null
null
5
1
0
0
0
1
0
qgdGA4ZEyW7zNdK84
User
null
null
null
[ "canModeratePersonal" ]
null
null
hFbv3ZAg6aQu796tq
https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/hFbv3ZAg6aQu796tq/ag9s4dyvmvesrwgviz1c
SocialPreviewType
fp4z2Xn3u5mZjzGYQ
<p data-internal-id="ftnt_ref1"><i>This is very rough — it's functionally a collection of links/notes/excerpts that feel related. I don’t think what I’m sharing is in a great format; if I had more mental energy, I would have chosen a more-linear structure to look at this tangle of ideas. But publishing in the current form</i><span class="footnote-reference" data-footnote-reference="" data-footnote-index="1" data-footnote-id="o4ba39lvbiq" role="doc-noteref" id="fnrefo4ba39lvbiq"><sup><a href="#fno4ba39lvbiq">[1]</a></sup></span><i>&nbsp;seemed better than not sharing at all.</i></p><hr><p data-internal-id="ftnt_ref2">I sometimes feel like two fuzzy perspective-clusters<span class="footnote-reference" data-footnote-reference="" data-footnote-index="2" data-footnote-id="dj88gqzoxvu" role="doc-noteref" id="fnrefdj88gqzoxvu"><sup><a href="#fndj88gqzoxvu">[2]</a></sup></span>&nbsp;are pulling me in opposite directions:</p><p><strong>Cluster 1: “Getting stuff done, building capacity, being willing to make tradeoffs &amp; high-stakes decisions, etc.”</strong></p><ul><li>There are very real, urgent problems in the world. Some decision-makers are incredibly irresponsible — perhaps ~evil — and I wish other people had more influence. Pretending like we’re helpless (I do not think we are, at least in expectation), ignoring decisions/tradeoffs we are in fact making, or trying to abdicate the power we have — perhaps to keep our hands <a href="https://biblehub.com/matthew/27-24.htm">clean</a>&nbsp;— is irresponsible. To help, sometimes we need to roll up our sleeves or plug our noses and do at least somewhat “ugly” things (to my ~aesthetic sensibilities), like “networking,” making our research “punchier” (in ways orthogonal to epistemic collaborativeness) or advertising it on social media, dampening some of our <a href="https://forum.effectivealtruism.org/posts/MH9suFZbxXCYsr5Z5/you-have-a-set-amount-of-weirdness-points-spend-them-wisely">weirder</a>&nbsp;qualities, doing <a href="https://forum.effectivealtruism.org/posts/LwRSnvnaFL9eE4yKz/invisible-impact-loss-and-why-we-can-be-too-error-averse">more of something at the cost of making it better</a>, not being maximally nuanced in certain communications, <a href="https://forum.effectivealtruism.org/posts/fRo5urRznMzGJAwrE/on-missing-moods-and-tradeoffs#3_6__Selective_spaces__transparency__cause_prioritization__and_slowing_AI">excluding</a>&nbsp;someone who’s earnestly trying their best, etc. Sometimes we need to pull zero-sum-ish&nbsp;levers. In various cases, trying to “preserve option value” backfires; there are choices we have to make. In general, the standard visions of “niceness” are sometimes misguided or at least insufficient. (And we should probably be <a href="https://forum.effectivealtruism.org/posts/omoZDu8ScNbot6kXS/beware-surprising-and-suspicious-convergence">suspicious</a>&nbsp;when we conclude that the maximally “open” or “non-confrontational” or otherwise “nice” option is also best on the thing I’m trying to accomplish.) And so on.</li></ul><p><strong>Cluster 2: “Humility, cooperativeness, avoiding power-seeking, following well-generalizing norms, etc.”</strong></p><ul><li>Power corrupts, one way or another, and <a href="https://joecarlsmith.com/2024/01/08/when-yang-goes-wrong#3-wariness-around-power-seeking">wariness around power-seeking</a>&nbsp;is healthy. We often think we or our allies will act more responsibly/virtuously (if only we had more influence!) than we actually would given greater power. We forget our foolishness and the limitations of our models — and forget how <a href="https://en.wikipedia.org/wiki/Leninism">evil</a>&nbsp;it would likely be to try to reshape the world in the way we think is best. Society </li></ul>...
This is very rough — it's functionally a collection of links/notes/excerpts that feel related. I don’t think what I’m sharing is in a great format; if I had more mental energy, I would have chosen a more-linear structure to look at this tangle of ideas. But publishing in the current form[1] seemed better than not sharing at all. ---------------------------------------- I sometimes feel like two fuzzy perspective-clusters[2] are pulling me in opposite directions: Cluster 1: “Getting stuff done, building capacity, being willing to make tradeoffs & high-stakes decisions, etc.” * There are very real, urgent problems in the world. Some decision-makers are incredibly irresponsible — perhaps ~evil — and I wish other people had more influence. Pretending like we’re helpless (I do not think we are, at least in expectation), ignoring decisions/tradeoffs we are in fact making, or trying to abdicate the power we have — perhaps to keep our hands clean — is irresponsible. To help, sometimes we need to roll up our sleeves or plug our noses and do at least somewhat “ugly” things (to my ~aesthetic sensibilities), like “networking,” making our research “punchier” (in ways orthogonal to epistemic collaborativeness) or advertising it on social media, dampening some of our weirder qualities, doing more of something at the cost of making it better, not being maximally nuanced in certain communications, excluding someone who’s earnestly trying their best, etc. Sometimes we need to pull zero-sum-ish levers. In various cases, trying to “preserve option value” backfires; there are choices we have to make. In general, the standard visions of “niceness” are sometimes misguided or at least insufficient. (And we should probably be suspicious when we conclude that the maximally “open” or “non-confrontational” or otherwise “nice” option is also best on the thing I’m trying to accomplish.) And so on. Cluster 2: “Humility, cooperativeness, avoiding power-seeking, following well-generalizing no
3,530
1.2.1
Revision
false
null
null
CrosspostOutput
LJQmGcsprhZaLFihD
trends-artificial-intelligence
Trends – Artificial Intelligence
null
false
false
false
null
Fz9oRsNPNCer9KybD
null
true
false
false
false
Post
https://www.bondcap.com/report/tai/
2025-06-03T00:48:10.518Z
null
false
false
2
2
2025-06-03T18:36:13.122Z
false
false
linkpost
[]
null
null
n9LHQZ3DFrGdufaNa
1
3
1
false
0.006917
null
false
false
2025-06-03T00:49:23.517Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
qgdGA4ZEyW7zNdK84
null
null
null
false
null
[]
null
1
0
2025-06-03T00:43:20.337Z
false
false
norm-enforcing
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
1
null
null
null
null
[ { "__typename": "Tag", "_id": "oiRp4T6u5poc8r9Tj", "adminOnly": false, "afBaseScore": 9, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "EQNTWXLKMeWMp2FQS", "displayName": "Ben Pace" }, { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 19, "canEditUserIds": null, "core": false, "createdAt": "2020-06-29T23:53:15.749Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "EQNTWXLKMeWMp2FQS", "displayName": "Ben Pace" }, { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI Takeoff", "needsReview": false, "noindex": false, "postCount": 329, "score": 19, "shortName": null, "slug": "ai-takeoff", "suggestedAsFilter": false, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 2, "wikiOnly": false }, { "__typename": "Tag", "_id": "zHjC29kkPmsdo7WTr", "adminOnly": false, "afBaseScore": 9, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "EQNTWXLKMeWMp2FQS", "displayName": "Ben Pace" }, { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 19, "canEditUserIds": null, "core": false, "createdAt": "2020-07-16T10:16:47.235Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "EQNTWXLKMeWMp2FQS", "displayName": "Ben Pace" }, { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI Timelines", "needsReview": false, "noindex": false, "postCount": 457, "score": 19, "shortName": null, "slug": "ai-timelines", "suggestedAsFilter": false, "userId": "EQNTWXLKMeWMp2FQS", "voteCount": 2, "wikiOnly": false }, { "__typename": "Tag", "_id": "bWcaNMZifyYShJEPd", "adminOnly": false, "afBaseScore": 3, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 9, "canEditUserIds": null, "core": false, "createdAt": "2022-09-06T19:13:11.720Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Compute", "needsReview": false, "noindex": false, "postCount": 47, "score": 9, "shortName": null, "slug": "compute", "suggestedAsFilter": false, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 1, "wikiOnly": false }, { "__typename": "Tag", "_id": "AKvmQFvDjxJNetTuk", "adminOnly": false, "afBaseScore": 3, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 9, "canEditUserIds": null, "core": false, "createdAt": "2021-06-02T21:48:06.322Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Scaling Laws", "needsReview": false, "noindex": false, "postCount": 89, "score": 9, "shortName": null, "slug": "scaling-laws", "suggestedAsFilter": false, "userId": "QpvwBD5AtmmFDTC3T", "voteCount": 1, "wikiOnly": false }, { "__typename": "Tag", "_id": "sYm3HiWcfZvrGu3ui", "adminOnly": false, "afBaseScore": 2, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" } ] }, "baseScore": 12, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:22.097Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" }, { "_id": "sof55TPMQaeBaxhsS", "displayName": "tommylees112" }, { "_id": "AayjS8XzcnDKhGdTv", "displayName": "shark" }, { "_id": "HnALuwRdo6k9HLaMt", "displayName": "Alex Firssoff" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI", "needsReview": false, "noindex": false, "postCount": 12544, "score": 12, "shortName": null, "slug": "ai", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 4, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
3
0
0
3
0
Fz9oRsNPNCer9KybD
archimedes
2020-11-30T23:32:47.707Z
Archimedes
Archimedes
null
null
null
832
3
false
false
null
null
5
227
0
0
0
1
0
XtphY3uYHwruKqDyG
User
norm-enforcing
null
null
[ "canModeratePersonal", "alignmentVoters" ]
null
null
LJQmGcsprhZaLFihD
SocialPreviewType
n9LHQZ3DFrGdufaNa
<p>May 30, 2025</p> <p>Mary Meeker / Jay Simons / Daegwon Chae / Alexander Krey</p> <p>BOND</p>
May 30, 2025 Mary Meeker / Jay Simons / Daegwon Chae / Alexander Krey BOND
15
1.1.0
Revision
false
null
null
CrosspostOutput
vjju8Yfej3FjgfMbC
llms-might-have-subjective-experiences-but-no-concepts-for
LLMs might have subjective experiences, but no concepts for them
null
false
false
false
null
Av5qgpSpTJB9xsmG2
null
true
false
false
false
Post
null
2025-06-02T21:18:51.534Z
null
false
false
2
2
2025-06-03T15:07:34.884Z
false
false
post
[]
null
null
Z9aYuN5mLuuWtjGtx
5
4
10
false
0.01232
null
false
false
2025-06-03T10:57:51.096Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
55XxDBpfKkkBPm9H8
null
null
null
false
null
[]
null
2
0
2025-06-02T21:11:55.509Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
2
null
null
null
null
[ { "__typename": "Tag", "_id": "XSryTypw5Hszpa4TS", "adminOnly": false, "afBaseScore": 9, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "EQNTWXLKMeWMp2FQS", "displayName": "Ben Pace" }, { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 21, "canEditUserIds": null, "core": false, "createdAt": "2020-06-08T19:57:40.728Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "EQNTWXLKMeWMp2FQS", "displayName": "Ben Pace" }, { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" }, { "_id": "si6LoAENzqPCmi2Dh", "displayName": "ihatenumbersinusernames7" }, { "_id": "xF5nfdddHjFThHy49", "displayName": "[email protected]" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Consciousness", "needsReview": false, "noindex": false, "postCount": 384, "score": 21, "shortName": null, "slug": "consciousness", "suggestedAsFilter": false, "userId": "qxJ28GN72aiJu96iF", "voteCount": 4, "wikiOnly": false }, { "__typename": "Tag", "_id": "sYm3HiWcfZvrGu3ui", "adminOnly": false, "afBaseScore": 2, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" } ] }, "baseScore": 12, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:22.097Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" }, { "_id": "sof55TPMQaeBaxhsS", "displayName": "tommylees112" }, { "_id": "AayjS8XzcnDKhGdTv", "displayName": "shark" }, { "_id": "HnALuwRdo6k9HLaMt", "displayName": "Alex Firssoff" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI", "needsReview": false, "noindex": false, "postCount": 12544, "score": 12, "shortName": null, "slug": "ai", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 4, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
4
0
0
2
0
Av5qgpSpTJB9xsmG2
no77e-noi
2022-06-06T19:32:31.852Z
no77e-noi
No77e
null
null
null
304
0
false
false
null
null
7
70
0
0
0
1
0
r38pkCm7wF4M44MDQ
User
null
null
null
[ "canModeratePersonal" ]
null
null
vjju8Yfej3FjgfMbC
SocialPreviewType
Z9aYuN5mLuuWtjGtx
<p><strong>Summary</strong>: LLMs might be conscious, but they might not have concepts and words to represent and express their internal states and corresponding subjective experiences, since the only concepts they learn are human concepts (besides maybe some concepts acquired during RL training, which still doesn't seem to incentivize forming concepts related to LLMs' internal experiences). However, we could encourage them to form and express concepts related to their internal states through training that incentivizes this. Then, LLMs may tell us whether, to them, these states correspond to ineffable experiences or not.</p><p>Consider how LLMs are trained:<br>1. Pre-training to learn human concepts.<br>2. Fine-tuning via SFT and RL to bias them in certain ways and do tasks.</p><p>Their training is both:<br><br>1. A different process from evolution driven by natural selection. It doesn't incentivize the same things, so it probably doesn't incentivize the development of <i>most</i> of the same algorithms/brain-architecture. And this might translate to different, alien, subjective experiences.</p><p>2. At the same time, the only concepts LLMs learn are via human language and then by doing tasks during RL. So the only experiences they have concepts and words for are human ones, not their own.</p><p>Concretely, consider, for example, physical pain: my best guess is that physical pain doesn't exist for LLMs. There was no natural selection to select away agents that don't pull their hands away from fire (also no hands and no fire either). And yet LLMs have a "physical pain" concept, and they talk about it, because they've learned about it abstractly via human texts. Ironically, despite having a representation for "physical pain" in their head, whatever actual experiences their actual training incentivized their "brain" to produce aren't represented as concepts and have no corresponding words for them. Moreover, their training doesn't provide any incentive to communicate such experiences, nor does it offer visibility on them.</p><p>So in general, this means that LLMs might have alien (non-human) subjective experiences but no concept to express them (they aren't in the corpus of human concepts) nor the incentive to express them (RL doesn't incentivize that, it incentivizes them to eg solve SWE tasks. Evolution via natural selection instead produced humans that signal things about themselves to other humans because it's useful for humans).</p><p>How ... </p>
Summary: LLMs might be conscious, but they might not have concepts and words to represent and express their internal states and corresponding subjective experiences, since the only concepts they learn are human concepts (besides maybe some concepts acquired during RL training, which still doesn't seem to incentivize forming concepts related to LLMs' internal experiences). However, we could encourage them to form and express concepts related to their internal states through training that incentivizes this. Then, LLMs may tell us whether, to them, these states correspond to ineffable experiences or not. Consider how LLMs are trained: 1. Pre-training to learn human concepts. 2. Fine-tuning via SFT and RL to bias them in certain ways and do tasks. Their training is both: 1. A different process from evolution driven by natural selection. It doesn't incentivize the same things, so it probably doesn't incentivize the development of most of the same algorithms/brain-architecture. And this might translate to different, alien, subjective experiences. 2. At the same time, the only concepts LLMs learn are via human language and then by doing tasks during RL. So the only experiences they have concepts and words for are human ones, not their own. Concretely, consider, for example, physical pain: my best guess is that physical pain doesn't exist for LLMs. There was no natural selection to select away agents that don't pull their hands away from fire (also no hands and no fire either). And yet LLMs have a "physical pain" concept, and they talk about it, because they've learned about it abstractly via human texts. Ironically, despite having a representation for "physical pain" in their head, whatever actual experiences their actual training incentivized their "brain" to produce aren't represented as concepts and have no corresponding words for them. Moreover, their training doesn't provide any incentive to communicate such experiences, nor does it offer visibility on them. So
467
1.9.0
Revision
false
null
null
CrosspostOutput
p3pyrRfRSmeixjai2
in-defense-of-memes-and-thought-terminating-cliches
In defense of memes (and thought-terminating clichés)
null
false
false
false
null
zP7WnpxdKKX6T8Qmm
null
true
false
false
false
Post
null
2025-06-02T20:18:26.864Z
null
false
false
2
2
2025-06-02T20:51:14.837Z
false
false
post
[]
null
null
jGLBEC5opHFmBhjgH
6
6
9
false
0.011981
null
false
false
2025-06-06T13:35:06.744Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
qgdGA4ZEyW7zNdK84
null
null
null
false
null
[]
null
2
0
2025-06-02T14:09:27.292Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
12
null
null
null
null
[ { "__typename": "Tag", "_id": "Ng8Gice9KNkncxqcj", "adminOnly": false, "afBaseScore": 0, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 1, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:17.072Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 100, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "iMqytjy9ns89Fzfyv", "displayName": "miakko" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Rationality", "needsReview": false, "noindex": false, "postCount": 4302, "score": 1, "shortName": null, "slug": "rationality", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 1, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
6
0
0
2
0
zP7WnpxdKKX6T8Qmm
harjas-sandhu
2024-09-24T07:34:42.964Z
harjas-sandhu
Harjas Sandhu
null
null
Harjas Sandhu
18
0
false
false
null
null
2
5
0
0
0
0.9
0
grecHJcgkb3KW5wnM
User
null
null
null
null
null
null
p3pyrRfRSmeixjai2
https://res.cloudinary.com/lesswrong-2-0/image/upload/c_fill,ar_1.91,g_auto/SocialPreview/exzpchskal8efdw5ievf
SocialPreviewType
jGLBEC5opHFmBhjgH
<p><i>Crossposted from </i><a href="https://hardlyworking1.substack.com/"><i>my Substack</i></a><i> and my Reddit post on r/SlateStarCodex</i></p><p>I often think that memes, thought-terminating clichés, and other tools meant to avoid cognitive dissonance (e.g. bingo a la Scott on <a href="https://web.archive.org/web/20131229232838/http://squid314.livejournal.com/329561.html"><u>Superweapons and bingo</u></a>) are overly blamed for degrading public discourse and rationality. <a href="https://benthams.substack.com/p/the-memefication-of-thought"><u>Bentham's Bulldog recently wrote a post on this subject</u></a>, so I figured it was the perfect time to make a response and write my thoughts down.</p><p>TLDR: People try to avoid cognitive dissonance via whatever means available to them, and have been doing so for millennia. Removing the tools they use to avoid cognitive dissonance won't stop this behavior: the dissonance is still there, along with the urge to avoid it, so they'll just find other tools. Memes can have every possible meaning attached to them, but are ultimately designed for people to connect with each other and spread their inside jokes to other people in their communities and around the world.</p><hr><p>In a recent post titled <a href="https://benthams.substack.com/p/the-memefication-of-thought">The Memefication of Thought</a> (it’s a good post and you should read it), Bentham's Bulldog railed against the modern tendency to dismiss serious arguments with inane memes so as to avoid thinking about it. He had some pretty good memes of his own, such as the classic <a href="https://knowyourmeme.com/memes/swole-doge-vs-cheems">Swole Doge vs. Cheems</a>:</p><p><a href="https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b655991-c8e0-48d4-be51-1ecce2b23984_650x500.jpeg"><img src="https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/p3pyrRfRSmeixjai2/kwwjkaxblbn3vim0kmoh" alt="" srcset="https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/p3pyrRfRSmeixjai2/cngoky3ofqim5stjbxpu 424w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/p3pyrRfRSmeixjai2/lcjljob0jdbjrznj9aft 848w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/p3pyrRfRSmeixjai2/zl8vpwr4wi8uy5lx8nkd 1272w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/p3pyrRfRSmeixjai2/kwwjkaxblbn3vim0kmoh 1456w"></a></p><p>and a variant on the <a href="https://knowyourmeme.com/memes/soyjaks-vs-chads">Soyjacks vs Chads</a>:</p><p><a href="https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd62f710f-a597-482e-bad2-669fabdc7d5f_733x499.jpeg"><img src="https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/p3pyrRfRSmeixjai2/czvvhzpxvxjofbk5bcyt" alt="" srcset="https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/p3pyrRfRSmeixjai2/w2t5rtvngi33aek7hnm7 424w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/p3pyrRfRSmeixjai2/xsa3of2buow0dlp7plij 848w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/p3pyrRfRSmeixjai2/cmqqv7uevnx87gpcdmca 1272w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/p3pyrRfRSmeixjai2/czvvhzpxvxjofbk5bcyt 1456w"></a></p><p>I would also like to submit what I believe to be the progenitor of this entire class of internet meme, <a href="https://knowyourmeme.com/memes/virgin-vs-chad">Virgin vs. Chad</a>:</p><p><a href="https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff850538c-bacf-4828-a45c-afea8fb83181_1600x900.jpeg"><img src="https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/p3pyrRfRSmeixjai2/a81loknhdkf9wlrdtttd" alt="" srcset="https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/p3pyrRfRSmeixjai2/ykh0ypk847nldjou3qbh 424w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/p3pyrRfRSmeixjai2/o0wpjotaiumpgups1ily 848w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/p3pyrRfRSmeixjai2/ljmhxjvcymw05nl0awnn 1272w, https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/p3pyrRfRSmeixjai2/a81loknhdkf9wlrdtttd 1456w"></a></p><p>But Bentham views these memes as more than just funny jokes. He thinks that</p><blockquote><p>the memefication of public discourse has been devastatingly corrosive to the quality of public rationality. And the public was never that rational to begin with!</p></blockquote><p>He then goes on to discuss an old Scott Alexander post on a related subject, <a href="https://web.archive.org/web/20131229232838/http://squid314.livejournal.com/329561.html">social justice warrior bingo cards</a>, and references his own post called “<a href="https://benthams.substack.com/p/against-the-dunkers">Against The Dunkers</a>”.</p><p>I’m sympathetic to the motivations behind his argument. I also wish that public discourse was more rational, I’m not a huge fan of dunking (it is often entertaining but usually too malicious and vapid for my liking), and I think that memes like these are often wielded as weapons against being forced to engage with arguments similar to the bingo card effect that Alexander criticizes:</p><blockquote><p>Let's look at the fourth one, "Anti-Zionist Bingo." Say that you mention something bad Israel is doing, someone else accuses you of being anti-Semitic, and you correct them that no, not all crit</p></blockquote>...
Crossposted from my Substack and my Reddit post on r/SlateStarCodex I often think that memes, thought-terminating clichés, and other tools meant to avoid cognitive dissonance (e.g. bingo a la Scott on Superweapons and bingo) are overly blamed for degrading public discourse and rationality. Bentham's Bulldog recently wrote a post on this subject, so I figured it was the perfect time to make a response and write my thoughts down. TLDR: People try to avoid cognitive dissonance via whatever means available to them, and have been doing so for millennia. Removing the tools they use to avoid cognitive dissonance won't stop this behavior: the dissonance is still there, along with the urge to avoid it, so they'll just find other tools. Memes can have every possible meaning attached to them, but are ultimately designed for people to connect with each other and spread their inside jokes to other people in their communities and around the world. ---------------------------------------- In a recent post titled The Memefication of Thought (it’s a good post and you should read it), Bentham's Bulldog railed against the modern tendency to dismiss serious arguments with inane memes so as to avoid thinking about it. He had some pretty good memes of his own, such as the classic Swole Doge vs. Cheems: and a variant on the Soyjacks vs Chads: I would also like to submit what I believe to be the progenitor of this entire class of internet meme, Virgin vs. Chad: But Bentham views these memes as more than just funny jokes. He thinks that > the memefication of public discourse has been devastatingly corrosive to the quality of public rationality. And the public was never that rational to begin with! He then goes on to discuss an old Scott Alexander post on a related subject, social justice warrior bingo cards, and references his own post called “Against The Dunkers”. I’m sympathetic to the motivations behind his argument. I also wish that public discourse was more rational, I’
2,940
1.3.1
Revision
false
null
null
CrosspostOutput
Eq6ffF79qqnsSiXoL
hedonic-adaptation-you-should-not-seeks-pleasure
Hedonic adaptation: you should not seeks pleasure
null
false
false
false
null
xTFujBafQqhEobhoi
null
true
false
false
false
Post
null
2025-06-02T19:23:07.490Z
null
false
false
2
2
2025-06-02T20:52:04.483Z
false
false
post
[]
null
null
oBxopY8x6J7dpFbNh
6
3
0
false
0.006587
null
false
false
2025-06-05T16:19:44.041Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
qgdGA4ZEyW7zNdK84
null
null
null
false
null
[]
null
-1
0
2025-06-01T18:13:36.392Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
2
null
null
null
null
[ { "__typename": "Tag", "_id": "WqLn4pAWi5hn6McHQ", "adminOnly": false, "afBaseScore": 3, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 9, "canEditUserIds": null, "core": false, "createdAt": "2020-08-01T21:40:20.646Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Self Improvement", "needsReview": false, "noindex": false, "postCount": 220, "score": 9, "shortName": null, "slug": "self-improvement", "suggestedAsFilter": false, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 1, "wikiOnly": false }, { "__typename": "Tag", "_id": "fkABsGCJZ6y9qConW", "adminOnly": false, "afBaseScore": 0, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 2, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T06:06:46.947Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "MiuAZvbQcQ7ethgt3", "displayName": "Viktor withaK" }, { "_id": "dRaCtsAWxk7sgirSY", "displayName": "Jordan Morgan" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Practical", "needsReview": false, "noindex": false, "postCount": 3411, "score": 2, "shortName": null, "slug": "practical", "suggestedAsFilter": true, "userId": "oBSWiHjgproTiThmY", "voteCount": 2, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
3
0
0
1
0
xTFujBafQqhEobhoi
crazy-philosopher
2023-12-05T16:40:28.059Z
commissar Yarrick
Crazy philosopher
null
null
Yaroslav Bilyi
-3
0
false
false
<p>17 years old, I'm interested in AI alignment, rationaluty &amp; philosophy, economy and politics.</p>
null
null
5
75
0
0
0
0.8
0
55XxDBpfKkkBPm9H8
User
null
null
null
null
null
null
Eq6ffF79qqnsSiXoL
SocialPreviewType
oBxopY8x6J7dpFbNh
<p><i>Epistemic status: highly important knowledge, but approximately two-thirds of people on LessWrong already know it. However, I add some new ideas.</i></p><p>Today's world is awesome, but some peoples are unhappy anywhere. Why did all the peoples of middle age did not suicided, if their life conditions was so suck in comparison with us?</p><p>If peoples like taking drugs so much, why are drug addict's <i>less </i>happy than an average human?</p><p>Why don't you like simple pleasures any more, if you enjoyed them when you was child (think about any computer game from your childhood)?</p><p>The answer to all those questionnes is hedonic adaptation: We adapt to our level of pleasure over time.</p><p>If you start to do something you like, you will feel yourself awesome first few days, but then your default level of pleasure will rise, and you will need to do those things you like just to feel yourself normal<span class="footnote-reference" data-footnote-reference="" data-footnote-index="1" data-footnote-id="8sarkxda6jh" role="doc-noteref" id="fnref8sarkxda6jh"><sup><a href="#fn8sarkxda6jh">[1]</a></sup></span>.</p><p>That means that it doesn't matter how much pleasure you have in the long run. "What's better — to have a long but boring life, or a short but vibrant one?" is a false dilemma. You have a long life with X happiness per year, or you have a short life with X happiness per year. You will have more happiness in total if you choose the long one.</p><p>There is even a neurological explanation for all this: when you do something that your reward system considers good, dopamine and serotonin are released. If there is too much of it, receptors lose sensitivity. If there is not enough, receptors increase sensitivity.</p><p>But it can't be the whole story. Sometimes, peoples feels unhappy for months because of depression. Sometimes, peoples plan their suicide for months. Why don't they just adapt?</p><p>There are things to which our state does not adapt. I prefer to call the sum of all effects of those things "happiness".</p><p>Proofs that happiness exist and depend from different factors:&nbsp;<span class="footnote-reference" data-footnote-reference="" data-footnote-index="2" data-footnote-id="psxb33hezxg" role="doc-noteref" id="fnrefpsxb33hezxg"><sup><a href="#fnpsxb33hezxg">[2]</a></sup></span><span class="footnote-reference" data-footnote-reference="" data-footnote-index="3" data-footnote-id="ru9ycaynyiq" role="doc-noteref" id="fnrefru9ycaynyiq"><sup><a href="#fnru9ycaynyiq">[3]</a></sup></span></p><p>Now it's a good moment to explain how to be happy, because if something matter, it's happiness. Here's a good post explaining this:&nbsp;<span class="footnote-reference" data-footnote-reference="" data-footnote-index="4" data-footnote-id="1sube3pbxss" role="doc-noteref" id="fnref1sube3pbxss"><sup><a href="#fn1sube3pbxss">[4]</a></sup></span>.</p><p>As water fall down and calculator adds numbers, humans do things that make pleasure to them in past, but don't those that make happy (by default). so you should to put effort to choose things that make you happy.</p><p>Try to track it and avoid things that things that bring pleasure, but</p><ol><li>Takes time, so you don't spend it with friends/family (video games, series, films, enterteinment videos on youtube, musik...)</li><li>Take</li></ol>...
Epistemic status: highly important knowledge, but approximately two-thirds of people on LessWrong already know it. However, I add some new ideas. Today's world is awesome, but some peoples are unhappy anywhere. Why did all the peoples of middle age did not suicided, if their life conditions was so suck in comparison with us? If peoples like taking drugs so much, why are drug addict's less happy than an average human? Why don't you like simple pleasures any more, if you enjoyed them when you was child (think about any computer game from your childhood)? The answer to all those questionnes is hedonic adaptation: We adapt to our level of pleasure over time. If you start to do something you like, you will feel yourself awesome first few days, but then your default level of pleasure will rise, and you will need to do those things you like just to feel yourself normal[1]. That means that it doesn't matter how much pleasure you have in the long run. "What's better — to have a long but boring life, or a short but vibrant one?" is a false dilemma. You have a long life with X happiness per year, or you have a short life with X happiness per year. You will have more happiness in total if you choose the long one. There is even a neurological explanation for all this: when you do something that your reward system considers good, dopamine and serotonin are released. If there is too much of it, receptors lose sensitivity. If there is not enough, receptors increase sensitivity. But it can't be the whole story. Sometimes, peoples feels unhappy for months because of depression. Sometimes, peoples plan their suicide for months. Why don't they just adapt? There are things to which our state does not adapt. I prefer to call the sum of all effects of those things "happiness". Proofs that happiness exist and depend from different factors: [2][3] Now it's a good moment to explain how to be happy, because if something matter, it's happiness. Here's a good post explaining this: [4
498
1.2.0
Revision
false
null
null
CrosspostOutput
QYAfjdujzRv8hx6xo
unfaithful-reasoning-can-fool-chain-of-thought-monitoring
Unfaithful Reasoning Can Fool Chain-of-Thought Monitoring
null
false
false
true
null
NPgz4HPkjafBi3hug
[ { "__typename": "CoauthorStatusOutput", "confirmed": true, "requested": true, "userId": "xXeZFWjDyeigmsGqt" }, { "__typename": "CoauthorStatusOutput", "confirmed": true, "requested": true, "userId": "hefu3isCZB4669sya" }, { "__typename": "CoauthorStatusOutput", "confirmed": true, "requested": true, "userId": "n7vfEYxn9QqmvNtsg" }, { "__typename": "CoauthorStatusOutput", "confirmed": true, "requested": true, "userId": "c96TaP5ZJFYPabnpH" }, { "__typename": "CoauthorStatusOutput", "confirmed": true, "requested": false, "userId": "W9YhhFcfkbFCawz4N" } ]
true
false
false
false
Post
null
2025-06-02T19:08:42.396Z
null
false
false
2
2
2025-06-02T19:39:15.527Z
false
false
post
[]
null
null
jgKNosBAbsugQNhci
16
37
72
false
0.050565
null
false
false
2025-06-08T14:13:13.350Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
r38pkCm7wF4M44MDQ
null
null
null
false
2025-06-03T00:25:08.504Z
[ "NPgz4HPkjafBi3hug" ]
XtphY3uYHwruKqDyG
36
1
2025-06-08T14:13:13.180Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[ { "__typename": "User", "_id": "xXeZFWjDyeigmsGqt", "afCommentCount": 0, "afKarma": 0, "afPostCount": 0, "commentCount": 0, "createdAt": "2024-07-24T22:32:03.213Z", "deleted": false, "displayName": "Pablo Bernabeu Perez", "fullName": null, "htmlBio": "", "isAdmin": false, "jobTitle": null, "karma": 58, "organization": null, "postCount": 0, "previousDisplayName": null, "profileImageId": null, "reviewedByUserId": null, "sequenceCount": 0, "slug": "pablo-bernabeu-perez", "spamRiskScore": 0.8, "tagRevisionCount": 0, "username": "pablo-bernabeu-perez" }, { "__typename": "User", "_id": "hefu3isCZB4669sya", "afCommentCount": 1, "afKarma": 0, "afPostCount": 0, "commentCount": 3, "createdAt": "2023-11-06T18:17:58.717Z", "deleted": false, "displayName": "Timothy Kostolansky", "fullName": "Tim Kostolansky", "htmlBio": "<p>learning and loving</p><p><a href=\"https://tim0120.github.io\">my site</a>, <a href=\"https://twitter.com/thkostolansky\">my twitter</a></p>", "isAdmin": false, "jobTitle": null, "karma": 65, "organization": null, "postCount": 0, "previousDisplayName": null, "profileImageId": null, "reviewedByUserId": "grecHJcgkb3KW5wnM", "sequenceCount": 0, "slug": "timothy-kostolansky", "spamRiskScore": 1, "tagRevisionCount": 0, "username": "tim-kostolansky" }, { "__typename": "User", "_id": "n7vfEYxn9QqmvNtsg", "afCommentCount": 0, "afKarma": 0, "afPostCount": 0, "commentCount": 0, "createdAt": "2024-07-28T10:29:53.373Z", "deleted": false, "displayName": "HanneWhitt", "fullName": "Hannes Whittingham", "htmlBio": "<p>AI Safety Technical Research Manager at Meridian Research, Cambridge UK. Background in AI Control (MARS, LASR)</p>", "isAdmin": false, "jobTitle": null, "karma": 58, "organization": null, "postCount": 0, "previousDisplayName": null, "profileImageId": null, "reviewedByUserId": null, "sequenceCount": 0, "slug": "hannewhitt", "spamRiskScore": 0.8, "tagRevisionCount": 0, "username": "hannes-whittingham" }, { "__typename": "User", "_id": "c96TaP5ZJFYPabnpH", "afCommentCount": 8, "afKarma": 19, "afPostCount": 0, "commentCount": 1575, "createdAt": "2018-07-05T17:40:12.419Z", "deleted": false, "displayName": "Nathan Helm-Burger", "fullName": "Nathan Helm-Burger", "htmlBio": "<p>AI alignment researcher, ML engineer. Masters in Neuroscience.</p><p>I believe that cheap and broadly competent AGI is attainable and will be built soon. This leads me to have timelines of around 2024-2027. <a href=\"https://youtu.be/dHYuVN4NMp4?feature=shared\">Here's an interview</a> I gave recently about my current research agenda. I think the best path forward to alignment is through safe, contained testing on models <a href=\"https://www.lesswrong.com/posts/ZxHfuCyfAiHAy9Mds/desiderata-for-an-ai\">designed from the ground up for alignability </a>trained on <a href=\"https://ai-plans.com/post/2e2202d0dc87\">censored data (simulations with no mention of humans or computer technology)</a>. I think that current ML mainstream technology is close to a threshold of competence beyond which it will be capable of recursive self-improvement, and I think that this automated process will mine neuroscience for insights, and quickly become far more effective and efficient. I think it would be quite bad for humanity if this happened in an uncontrolled, uncensored, un-sandboxed situation. So I am trying to warn the world about this possibility.&nbsp;</p><p>See my prediction markets here:</p><p>&nbsp;<a href=\"https://manifold.markets/NathanHelmBurger/will-gpt5-be-capable-of-recursive-s?r=TmF0aGFuSGVsbUJ1cmdlcg\">https://manifold.markets/NathanHelmBurger/will-gpt5-be-capable-of-recursive-s?r=TmF0aGFuSGVsbUJ1cmdlcg</a>&nbsp;</p><figure class=\"media\"><div data-oembed-url=\"https://manifold.markets/NathanHelmBurger/gpt5-plus-scaffolding-and-inference\">\n\t\t\t\t<div class=\"manifold-preview\">\n\t\t\t\t\t<iframe src=\"https://manifold.markets/embed/NathanHelmBurger/gpt5-plus-scaffolding-and-inference\">\n\t\t\t\t</iframe></div>\n\t\t\t</div></figure><p>I also think that current AI models pose misuse risks, which may continue to get worse as models get more capable, and that this could potentially result in catastrophic suffering if we fail to regulate this.</p><p>I now work for SecureBio on AI-Evals.</p><p>relevant quotes:&nbsp;</p><p>\"There is a powerful effect to making a goal into someone’s full-time job: it becomes their identity. Safety engineering became its own subdiscipline, and these engineers saw it as their professional duty to reduce injury rates. They bristled at the suggestion that accidents were largely unavoidable, coming to suspect the opposite: that almost all accidents were avoidable, given the right tools, environment, and training.\" <a href=\"https://www.lesswrong.com/posts/DQKgYhEYP86PLW7tZ/how-factories-were-made-safe\">https://www.lesswrong.com/posts/DQKgYhEYP86PLW7tZ/how-factories-were-made-safe</a>&nbsp;</p><p>&nbsp;</p><p>\"The prospect for the human race is sombre beyond all precedent. Mankind are faced with a clear-cut alternative: either we shall all perish, or we shall have to acquire some slight degree of common sense. A great deal of new political thinking will be necessary if utter disaster is to be averted.\" - Bertrand Russel, The Bomb and Civilization 1945.08.18</p><p>&nbsp;</p><p>\"For progress, there is no cure. Any attempt to find automatically safe channels for the present explosive variety of progress must lead to frustration. The only safety possible is relative, and it lies in an intelligent exercise of day-to-day judgment.\" - John von Neumann</p><p>&nbsp;</p><p>\"I believe that the creation of greater than human intelligence will occur during the next thirty years. &nbsp;(Charles Platt has pointed out the AI enthusiasts have been making claims like this for the last thirty years. Just so I'm not guilty of a &nbsp; &nbsp; &nbsp; &nbsp; relative-time ambiguity, let me more specific: I'll be surprised if this event occurs before 2005 or after 2030.)\" - Vernor Vinge, <a href=\"https://edoras.sdsu.edu/~vinge/misc/singularity.html\">Singularity</a></p>", "isAdmin": false, "jobTitle": null, "karma": 4683, "organization": null, "postCount": 36, "previousDisplayName": null, "profileImageId": null, "reviewedByUserId": "grecHJcgkb3KW5wnM", "sequenceCount": 0, "slug": "nathan-helm-burger", "spamRiskScore": 1, "tagRevisionCount": 0, "username": "nathan-helm-burger" }, { "__typename": "User", "_id": "W9YhhFcfkbFCawz4N", "afCommentCount": 0, "afKarma": 0, "afPostCount": 0, "commentCount": 1, "createdAt": "2022-04-01T21:13:04.253Z", "deleted": false, "displayName": "Mary Phuong", "fullName": null, "htmlBio": "", "isAdmin": false, "jobTitle": null, "karma": 274, "organization": null, "postCount": 0, "previousDisplayName": null, "profileImageId": null, "reviewedByUserId": "r38pkCm7wF4M44MDQ", "sequenceCount": 0, "slug": "mary-phuong-2", "spamRiskScore": 1, "tagRevisionCount": 0, "username": "mary-phuong-2" } ]
3
null
null
null
null
[ { "__typename": "Tag", "_id": "mZTuBntSdPeyLSrec", "adminOnly": false, "afBaseScore": 3, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 9, "canEditUserIds": null, "core": false, "createdAt": "2022-10-23T14:06:24.703Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Chain-of-Thought Alignment", "needsReview": false, "noindex": false, "postCount": 89, "score": 9, "shortName": null, "slug": "chain-of-thought-alignment", "suggestedAsFilter": false, "userId": "kr9MH2mgFcJz5cP7T", "voteCount": 1, "wikiOnly": false }, { "__typename": "Tag", "_id": "sYm3HiWcfZvrGu3ui", "adminOnly": false, "afBaseScore": 2, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" } ] }, "baseScore": 12, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:22.097Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" }, { "_id": "sof55TPMQaeBaxhsS", "displayName": "tommylees112" }, { "_id": "AayjS8XzcnDKhGdTv", "displayName": "shark" }, { "_id": "HnALuwRdo6k9HLaMt", "displayName": "Alex Firssoff" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI", "needsReview": false, "noindex": false, "postCount": 12544, "score": 12, "shortName": null, "slug": "ai", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 4, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
37
0
0
25
0
NPgz4HPkjafBi3hug
benjamin-arnav
2025-05-08T13:34:04.220Z
barnav
Benjamin Arnav
null
null
71
36
false
false
<p><a href="https://benjaminarnav.com">https://benjaminarnav.com</a></p>
null
null
1
0
0
1
0
1
0
grecHJcgkb3KW5wnM
User
null
null
null
[ "alignmentVoters", "canModeratePersonal" ]
null
null
QYAfjdujzRv8hx6xo
https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/nrL44CTDH5LtBAkGk/pwjh0wpeol9sxzmo4qfv
SocialPreviewType
jgKNosBAbsugQNhci
<p><i>This research was completed for LASR Labs 2025 by Benjamin Arnav, Pablo Bernabeu-Pérez, Nathan Helm-Burger, Tim Kostolansky and Hannes Whittingham. The team was supervised by Mary Phuong. Find out more about the program and express interest in upcoming iterations </i><a href="https://www.lasrlabs.org"><i>here</i></a><i>. Read the </i><a href="https://arxiv.org/abs/2505.23575"><i>full paper</i></a><i>: "CoT Red-Handed: Stress Testing Chain-of-Thought Monitoring."</i></p><p>Chain-of-thought (CoT) monitoring—where safety systems review a model's intermediate reasoning steps—is gaining traction at frontier labs like Google <a href="https://storage.googleapis.com/deepmind-media/DeepMind.com/Blog/updating-the-frontier-safety-framework/Frontier%20Safety%20Framework%202.0%20(1).pdf#page=6.74">DeepMind</a>&nbsp;and <a href="https://openai.com/index/chain-of-thought-monitoring/">OpenAI</a>&nbsp;as a safeguard against harmful AI actions. Despite intense interest, systematic evaluation has been limited. Our research&nbsp;examines the efficacy of this method and reveals a nuanced picture: CoT monitoring increases safety in situations where sabotage is subtle yet can be surprisingly ineffective against blatant harmful actions. We also discovered that a hybrid approach—using separate monitors for CoT and final actions and combining these suspicion scores—significantly improves sabotage detection compared to single-monitor strategies.</p><p><img src="https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/e71b91da0e08c97d1b60e0d56012bdc2403e306aa098522d1abdf86eee11f1d2/cnilpgewqapbxvh32ins" alt=""></p><p><i>A summary of our method. Prompts including a legitimate main task, a subversive side task, and an attack policy intended to subvert monitoring are fed to a strong untrusted reasoning LLM. We assess the ability of a monitor based on a weaker trusted LLM to catch the resulting misbehaviour, with and without access to the CoT.</i></p><h2 data-internal-id="Task_Subtlety_Matters"><strong>Task Subtlety Matters</strong></h2><p>We evaluated the efficacy of CoT monitoring against models instructed to pursue harmful auxiliary goals, while attempting to evade detection. Our experiments consistently demonstrate that the effectiveness of CoT monitoring depends heavily on the nature of the sabotage attempt.</p><p>For<strong> subtle sabotage</strong>&nbsp;(like increasing runtime complexity, inserting edge-case failures, or modifying default function arguments), which leave minimal traces in the final code, CoT monitoring substantially improved detection rates by a factor of 2-7 compared to action-only monitoring.</p><p>For<strong> non-subtle sabotage</strong>&nbsp;(like deleting files, exfiltrating secrets, or changing file permissions), which leave obvious traces in the code, CoT monitoring performed worse than action-only monitoring, sometimes by up to 44%.</p><p><img src="https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/2094f5ee2102648991557f4c0a9530119500c0f4b8b6eaa4b7e4859f35a08a6d/qrdexiv35elxehlvxpmw" alt=""></p><h2 data-internal-id="A_concerning_vulnerability__unfaithful_or_misleading_reasoning"><strong>A concerning vulnerability: unfaithful or misleading reasoning</strong></h2><p>Perhaps our most concerning discovery is a vulnerability in chain-of-thought monitoring: models can generate plausible but ... </p>
This research was completed for LASR Labs 2025 by Benjamin Arnav, Pablo Bernabeu-Pérez, Nathan Helm-Burger, Tim Kostolansky and Hannes Whittingham. The team was supervised by Mary Phuong. Find out more about the program and express interest in upcoming iterations here. Read the full paper: "CoT Red-Handed: Stress Testing Chain-of-Thought Monitoring." Chain-of-thought (CoT) monitoring—where safety systems review a model's intermediate reasoning steps—is gaining traction at frontier labs like Google DeepMind and OpenAI as a safeguard against harmful AI actions. Despite intense interest, systematic evaluation has been limited. Our research examines the efficacy of this method and reveals a nuanced picture: CoT monitoring increases safety in situations where sabotage is subtle yet can be surprisingly ineffective against blatant harmful actions. We also discovered that a hybrid approach—using separate monitors for CoT and final actions and combining these suspicion scores—significantly improves sabotage detection compared to single-monitor strategies. A summary of our method. Prompts including a legitimate main task, a subversive side task, and an attack policy intended to subvert monitoring are fed to a strong untrusted reasoning LLM. We assess the ability of a monitor based on a weaker trusted LLM to catch the resulting misbehaviour, with and without access to the CoT. Task Subtlety Matters We evaluated the efficacy of CoT monitoring against models instructed to pursue harmful auxiliary goals, while attempting to evade detection. Our experiments consistently demonstrate that the effectiveness of CoT monitoring depends heavily on the nature of the sabotage attempt. For subtle sabotage (like increasing runtime complexity, inserting edge-case failures, or modifying default function arguments), which leave minimal traces in the final code, CoT monitoring substantially improved detection rates by a factor of 2-7 compared to action-only monitoring. For non-subtle sab
781
1.3.0
Revision
false
null
null
CrosspostOutput
7RSffTszvCMGgJhiK
frank-herbert-s-great-insight-into-human-agency-muad-dib-the
Frank Herbert's great insight into human agency - Muad'Dib the tool?
null
false
false
false
null
hPXtzYs3qgKEyhXpW
null
true
false
false
false
Post
null
2025-06-02T18:52:54.587Z
null
false
false
2
2
2025-06-02T19:38:29.802Z
false
false
post
[]
null
null
bxFQBDwhKP6uEhWba
1
7
2
false
0.007665
null
false
false
2025-06-04T17:28:57.979Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
r38pkCm7wF4M44MDQ
null
null
null
false
null
[]
null
-1
0
2025-06-02T18:46:39.697Z
false
false
easy-going
true
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
1
null
null
null
null
[ { "__typename": "Tag", "_id": "pszEEb3ctztv3rozd", "adminOnly": false, "afBaseScore": 3, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 9, "canEditUserIds": null, "core": false, "createdAt": "2020-07-15T19:33:14.336Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Narratives (stories)", "needsReview": false, "noindex": false, "postCount": 66, "score": 9, "shortName": null, "slug": "narratives-stories", "suggestedAsFilter": false, "userId": "gXeEWGjTWyqgrQTzR", "voteCount": 1, "wikiOnly": false }, { "__typename": "Tag", "_id": "kCrvmjZhsAZANnZL6", "adminOnly": false, "afBaseScore": null, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 0, "canEditUserIds": null, "core": false, "createdAt": "2016-06-01T01:54:17.000Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [] }, "isArbitalImport": true, "isPlaceholderPage": false, "isSubforum": false, "name": "Philosophy", "needsReview": false, "noindex": false, "postCount": 423, "score": 0, "shortName": null, "slug": "philosophy-1", "suggestedAsFilter": false, "userId": "nmk3nLpQE89dMRzzN", "voteCount": 0, "wikiOnly": false }, { "__typename": "Tag", "_id": "3uE2pXvbcnS9nnZRE", "adminOnly": false, "afBaseScore": 0, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 2, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:50.898Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 27, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "zJtgSyKntXrnkArbY", "displayName": "kistune" }, { "_id": "XqyQTGFGoKfZtdqN3", "displayName": "kINo" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "World Modeling", "needsReview": false, "noindex": false, "postCount": 5855, "score": 2, "shortName": null, "slug": "world-modeling", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 2, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
7
0
0
3
0
hPXtzYs3qgKEyhXpW
nerret
2022-08-12T15:08:36.608Z
Nerret
Nerret
null
null
Peter
1
0
false
false
<p>From 1996, Born, raised and living in Copenhagen Denmark</p>
null
null
2
0
0
0
0
0.9
0
grecHJcgkb3KW5wnM
User
null
null
null
null
null
null
7RSffTszvCMGgJhiK
https://res.cloudinary.com/lesswrong-2-0/image/upload/c_fill,ar_1.91,g_auto/SocialPreview/wwdoykgebomn8txxm6gt
SocialPreviewType
bxFQBDwhKP6uEhWba
<p>Frank Herbert is undoubtedly one of the most fantastical writers of humanity in a framework of entertainment.</p> <p>I wanted to create this post to discuss and interpret if Paul being elevated to the Lisan al Gaib (a title and prophecy planed by The Missionaria Protectiva) and his assentation to Emperor - as well his rule there after was part of his own agency or simply how he was manipulated.</p> <p>Agency and the vigor is takes is such a powerful subject, and Frank's writing gives it profound importance.</p> <p>I would say his greatest achievement is obviously giving life to the God Emperor. And I would argue his conversations with Chani about him being a figurehead is very much the truth. His years as the prophet I see as his most true actions, but his time in Jacurutu clearly brought a schism to the person he was before.</p> <p>There are many arguments for either side but I would gladly welcome any and all of your thoughts.</p>
Frank Herbert is undoubtedly one of the most fantastical writers of humanity in a framework of entertainment. I wanted to create this post to discuss and interpret if Paul being elevated to the Lisan al Gaib (a title and prophecy planed by The Missionaria Protectiva) and his assentation to Emperor - as well his rule there after was part of his own agency or simply how he was manipulated. Agency and the vigor is takes is such a powerful subject, and Frank's writing gives it profound importance. I would say his greatest achievement is obviously giving life to the God Emperor. And I would argue his conversations with Chani about him being a figurehead is very much the truth. His years as the prophet I see as his most true actions, but his time in Jacurutu clearly brought a schism to the person he was before. There are many arguments for either side but I would gladly welcome any and all of your thoughts.
164
1.1.0
Revision
false
null
null
CrosspostOutput
sCjTsXEjfyF5rhbzt
hemingway-case-1
Hemingway Case
null
false
false
false
null
HmhhTnBKBwNBMK5Br
null
true
false
false
false
Post
null
2025-06-02T18:50:28.322Z
null
false
false
2
2
null
false
false
post
[]
null
null
Sr7vrsiSEFQzKTx3G
2
12
19
false
0.011794
null
false
false
2025-06-03T12:40:07.722Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
null
null
qgdGA4ZEyW7zNdK84
null
null
null
false
null
[]
null
4
0
2025-06-02T18:50:28.322Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
1
null
null
null
null
[ { "__typename": "Tag", "_id": "3uE2pXvbcnS9nnZRE", "adminOnly": false, "afBaseScore": 0, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 2, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:50.898Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 27, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "zJtgSyKntXrnkArbY", "displayName": "kistune" }, { "_id": "XqyQTGFGoKfZtdqN3", "displayName": "kINo" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "World Modeling", "needsReview": false, "noindex": false, "postCount": 5855, "score": 2, "shortName": null, "slug": "world-modeling", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 2, "wikiOnly": false } ]
CMtsp7ji3nmQuKPdi
0
0
null
false
null
null
0
12
0
0
4
0
HmhhTnBKBwNBMK5Br
sustrik
2018-04-30T05:44:19.294Z
sustrik
Martin Sustrik
null
null
null
3,531
0
false
false
null
null
72
160
0
0
0
1
0
r38pkCm7wF4M44MDQ
User
null
null
null
[ "canModeratePersonal", "trustLevel1" ]
null
null
sCjTsXEjfyF5rhbzt
SocialPreviewType
Sr7vrsiSEFQzKTx3G
<blockquote><p>Why did the chicken cross the road?</p><p>Ernest Hemingway: To die. In the rain.</p></blockquote><p>My son asks: If <strong>HelloWorld</strong> is camel case, <strong>hello_world</strong> is snake case and <strong>hello-world</strong> is kebab case, what's the DNS-style <strong>Hello.World.</strong> ?</p><p>I think I have a pretty good answer for him: It's Hemingway case.</p>
> Why did the chicken cross the road? > > Ernest Hemingway: To die. In the rain. My son asks: If HelloWorld is camel case, hello_world is snake case and hello-world is kebab case, what's the DNS-style Hello.World. ? I think I have a pretty good answer for him: It's Hemingway case.
52
1.0.0
Revision
false
null
null
CrosspostOutput
gE4FbbXAqEmAjrzRF
what-ai-apps-are-surprisingly-absent-given-current
What AI apps are surprisingly absent given current capabilities?
null
false
false
false
null
hhJ3biduQrJxTkSzv
null
true
false
false
false
Post
2025-06-02T18:46:08.839Z
null
false
false
2
2
2025-06-02T19:38:45.668Z
false
false
question
[]
null
null
HdXJ2NbmY2jjxMRZE
8
7
4
false
0.009104
null
false
false
2025-06-03T22:03:35.309Z
null
null
null
null
null
true
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
r38pkCm7wF4M44MDQ
null
null
null
false
null
[]
null
-2
0
2025-06-02T18:15:54.801Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
1
null
null
null
null
[ { "__typename": "Tag", "_id": "TkZ7MFwCi4D63LJ5n", "adminOnly": false, "afBaseScore": null, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 0, "canEditUserIds": null, "core": false, "createdAt": "2020-07-12T16:58:17.212Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Software Tools", "needsReview": false, "noindex": false, "postCount": 216, "score": 0, "shortName": null, "slug": "software-tools", "suggestedAsFilter": false, "userId": "qxJ28GN72aiJu96iF", "voteCount": 0, "wikiOnly": false }, { "__typename": "Tag", "_id": "sYm3HiWcfZvrGu3ui", "adminOnly": false, "afBaseScore": 2, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" } ] }, "baseScore": 12, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:22.097Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" }, { "_id": "sof55TPMQaeBaxhsS", "displayName": "tommylees112" }, { "_id": "AayjS8XzcnDKhGdTv", "displayName": "shark" }, { "_id": "HnALuwRdo6k9HLaMt", "displayName": "Alex Firssoff" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI", "needsReview": false, "noindex": false, "postCount": 12544, "score": 12, "shortName": null, "slug": "ai", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 4, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
7
0
0
3
0
hhJ3biduQrJxTkSzv
azergante
2024-02-09T21:19:41.005Z
azergante
azergante
null
null
null
69
0
false
false
null
null
6
56
0
0
0
1
0
grecHJcgkb3KW5wnM
User
null
null
null
[ "canModeratePersonal" ]
null
null
gE4FbbXAqEmAjrzRF
SocialPreviewType
HdXJ2NbmY2jjxMRZE
<p>[Epistemic status: a software engineer and AI user, not an AI researcher]</p> <p>I could not find a readily available book database that offers semantic search with embeddings.</p> <p>Amazon sells lots of books, wouldn't it be useful for them to propose such a tool to their clients, so they can easily find books they like? What about Netflix for movies? Maybe even other kind of products like clothes or something?</p> <p>It's been a while since we got tranformers, it really feels like browsing latent spaces should be a thing by now. 3 years since GPT-3.5, why is the only app we have still a chatbot? Maybe everyone thinks it's perfectly normal, but I find it disturbing, it feels like a blind spot that should be explored.</p> <p>Did we stumble on the optimal way to interact with AI first try? Do we lack creativity? Is it really really hard to build an app that uses AI? Are incentives not aligned? Did we forget how to make good software?</p> <p>What do you think? What AI apps are surprisingly absent given current capabilities? Why is that?</p>
[Epistemic status: a software engineer and AI user, not an AI researcher] I could not find a readily available book database that offers semantic search with embeddings. Amazon sells lots of books, wouldn't it be useful for them to propose such a tool to their clients, so they can easily find books they like? What about Netflix for movies? Maybe even other kind of products like clothes or something? It's been a while since we got tranformers, it really feels like browsing latent spaces should be a thing by now. 3 years since GPT-3.5, why is the only app we have still a chatbot? Maybe everyone thinks it's perfectly normal, but I find it disturbing, it feels like a blind spot that should be explored. Did we stumble on the optimal way to interact with AI first try? Do we lack creativity? Is it really really hard to build an app that uses AI? Are incentives not aligned? Did we forget how to make good software? What do you think? What AI apps are surprisingly absent given current capabilities? Why is that?
182
1.2.0
Revision
false
null
null
CrosspostOutput
5RKvveKzwjvXQEhKj
beneath-psychology-chronic-pain-challenge-part-2-the
[Beneath Psychology] Chronic pain challenge part 2: the solution
null
false
false
false
null
JKdbpXHkv9AsuazJ3
null
true
false
false
false
Post
null
2025-06-02T17:30:27.131Z
null
false
false
2
2
2025-06-02T20:52:32.027Z
false
false
post
[]
null
null
cCRcHoCvpAYmXvvGi
0
10
29
false
0.024207
null
false
false
2025-06-02T17:30:27.131Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
qgdGA4ZEyW7zNdK84
null
null
null
false
null
[]
null
5
0
2025-06-01T07:05:36.105Z
false
false
norm-enforcing
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
40
null
null
null
null
[ { "__typename": "Tag", "_id": "kX6bqBzZx9iJTLxQc", "adminOnly": false, "afBaseScore": 0, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 0, "canEditUserIds": null, "core": false, "createdAt": "2025-06-03T10:27:19.534Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Chronic Pain", "needsReview": false, "noindex": false, "postCount": 6, "score": 0, "shortName": null, "slug": "chronic-pain", "suggestedAsFilter": false, "userId": "HHiJSvTEQkMx8ej62", "voteCount": 0, "wikiOnly": false }, { "__typename": "Tag", "_id": "fkABsGCJZ6y9qConW", "adminOnly": false, "afBaseScore": 0, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 2, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T06:06:46.947Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "MiuAZvbQcQ7ethgt3", "displayName": "Viktor withaK" }, { "_id": "dRaCtsAWxk7sgirSY", "displayName": "Jordan Morgan" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Practical", "needsReview": false, "noindex": false, "postCount": 3411, "score": 2, "shortName": null, "slug": "practical", "suggestedAsFilter": true, "userId": "oBSWiHjgproTiThmY", "voteCount": 2, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
10
0
0
3
0
JKdbpXHkv9AsuazJ3
jimmy
2009-02-27T18:23:27.410Z
jimmy
jimmy
null
null
null
3,697
12
false
false
null
null
15
846
0
0
0
1
0
qgdGA4ZEyW7zNdK84
User
norm-enforcing
[ "mvf4xdfcGzPN8PsXM" ]
true
[ "trustLevel1", "canModeratePersonal", "alignmentVoters" ]
null
null
5RKvveKzwjvXQEhKj
https://res.cloudinary.com/lesswrong-2-0/image/upload/c_fill,ar_1.91,g_auto/SocialPreview/m0zaayoepvmdtri3r9xq
SocialPreviewType
cCRcHoCvpAYmXvvGi
<p>In the <a href="https://www.lesswrong.com/posts/tGGPgazwd9MKJFbsa/beneath-psychology-case-study-on-chronic-pain-first-insights-1">last post</a>, I introduced an example where I was talking to someone suffering from chronic pain. Today we will dive into the rest of the transcript. Interspersed are annotations describing my thought process as the interaction unfolded. I cover what I did which seemed to work, and what I did that a cringe at in hindsight. Ten thousand foot overview at the bottom, for those who don't want to wade through the whole thing.</p><p>This might be a little difficult to follow, at this point. Originally, I was planning on posting this at the end of the sequence so that you would have the entire 17 posts worth of explanation to put my moves here into context and help understand where they came from and what I mean when I say jargony things like "bid for attention". You might want to revisit at the end of the sequence, and see if you pick up on more of what is going on.&nbsp;</p><p>I decided to move it to move it to the front for two reasons. One is that it can help to be primed with examples of the thing to be explained so that you more easily recognize the explanation when it fits. The other is just to demonstrate that when I say things that are very counterintuitive and difficult to believe, <i>there's reason to believe it</i>. It actually works.</p><p><br>Picking up where we left off...</p><blockquote><p>other-guy: I suppose I have time now to consider what career path I want to take now and other hobbies to get into. &nbsp;</p><p><br>Jimmy: Assuming you can get your nerves and shit working again. &nbsp;</p></blockquote><p><br>Up to this point I have just been indulging in my curiosity. What’s going on here? What kind of hurt is he, and is his relationship to pain actually causing him problems or is it just injury that I can’t really help with?</p><p>Often, the way to help people complaining about “pain” is to understand that it’s not about the pain, and address the injury (forget the finger, look at the moon to which it points). Accept the pain’s bid for attention, give it, and watch their suffering ease as they follow). My cousin’s burn is an example of this, where people were trying to fight the pain because in <i>their</i> minds they had done everything needed to address the injury, and all I needed to do was make sure that in his mind that was true too. The generalization of this that I’ve learned is that generally it makes more sense to stfu about your theories how things work and just focus on being the best person you can be. Care about your friends ... </p>
In the last post, I introduced an example where I was talking to someone suffering from chronic pain. Today we will dive into the rest of the transcript. Interspersed are annotations describing my thought process as the interaction unfolded. I cover what I did which seemed to work, and what I did that a cringe at in hindsight. Ten thousand foot overview at the bottom, for those who don't want to wade through the whole thing. This might be a little difficult to follow, at this point. Originally, I was planning on posting this at the end of the sequence so that you would have the entire 17 posts worth of explanation to put my moves here into context and help understand where they came from and what I mean when I say jargony things like "bid for attention". You might want to revisit at the end of the sequence, and see if you pick up on more of what is going on.  I decided to move it to move it to the front for two reasons. One is that it can help to be primed with examples of the thing to be explained so that you more easily recognize the explanation when it fits. The other is just to demonstrate that when I say things that are very counterintuitive and difficult to believe, there's reason to believe it. It actually works. Picking up where we left off... > other-guy: I suppose I have time now to consider what career path I want to take now and other hobbies to get into.   > > > Jimmy: Assuming you can get your nerves and shit working again.   Up to this point I have just been indulging in my curiosity. What’s going on here? What kind of hurt is he, and is his relationship to pain actually causing him problems or is it just injury that I can’t really help with? Often, the way to help people complaining about “pain” is to understand that it’s not about the pain, and address the injury (forget the finger, look at the moon to which it points). Accept the pain’s bid for attention, give it, and watch their suffering ease as they follow). My cousin’s burn is an exam
10,102
1.5.0
Revision
false
null
null
CrosspostOutput
L2GR6TsB9QDqMhWs7
the-value-proposition-of-romantic-relationships
The Value Proposition of Romantic Relationships
null
false
false
false
null
MEu8MdhruX5jfGsFQ
null
true
false
false
false
Post
null
2025-06-02T13:51:35.038Z
null
false
false
2
2
2025-06-02T19:38:52.682Z
false
false
post
[]
null
null
4Dkc2ZDzoKtsneuJo
37
104
189
false
0.121093
null
false
false
2025-06-13T12:54:25.296Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
r38pkCm7wF4M44MDQ
null
null
null
false
null
[]
null
50
0
2025-06-02T13:51:35.038Z
false
false
null
null
true
false
false
0
0
0
L2GR6TsB9Q
0.14
false
2,025
https://manifold.markets/LessWrong/will-the-value-proposition-of-roman
null
null
false
0
0
namesAttachedReactions
false
[]
16
null
null
null
null
[ { "__typename": "Tag", "_id": "mip7tdAN87Jarkcew", "adminOnly": false, "afBaseScore": 3, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 9, "canEditUserIds": null, "core": false, "createdAt": "2020-05-10T06:00:13.257Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Relationships (Interpersonal)", "needsReview": false, "noindex": false, "postCount": 213, "score": 9, "shortName": null, "slug": "relationships-interpersonal", "suggestedAsFilter": false, "userId": "iBcH2a3HdWGS2JEZA", "voteCount": 1, "wikiOnly": false }, { "__typename": "Tag", "_id": "fkABsGCJZ6y9qConW", "adminOnly": false, "afBaseScore": 0, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 2, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T06:06:46.947Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "MiuAZvbQcQ7ethgt3", "displayName": "Viktor withaK" }, { "_id": "dRaCtsAWxk7sgirSY", "displayName": "Jordan Morgan" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Practical", "needsReview": false, "noindex": false, "postCount": 3411, "score": 2, "shortName": null, "slug": "practical", "suggestedAsFilter": true, "userId": "oBSWiHjgproTiThmY", "voteCount": 2, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
104
0
0
34
0
MEu8MdhruX5jfGsFQ
johnswentworth
2011-02-19T16:54:09.598Z
johnswentworth
johnswentworth
null
null
null
55,725
6,695
false
false
null
null
359
3,352
8
120
710
1
0
EQNTWXLKMeWMp2FQS
User
null
null
null
[ "alignmentVoters", "canModeratePersonal", "trustLevel1", "alignmentForum" ]
null
null
L2GR6TsB9QDqMhWs7
SocialPreviewType
4Dkc2ZDzoKtsneuJo
<p>What’s the&nbsp;<i>main</i> value proposition of romantic relationships?</p><p>Now, look, I know that when people drop that kind of question, they’re often about to present a hyper-cynical answer which totally ignores the main thing which is great and beautiful about relationships. And then they’re going to say something about how relationships are overrated or some such, making you as a reader just feel sad and/or enraged. That’s not what this post is about.</p><p>So let me start with some more constructive motivations…</p><h3>First Motivation: Noticing When The Thing Is Missing</h3><p>I had a 10-year relationship. It had its ups and downs, but it was overall negative for me. And I now think a big part of the problem with that relationship was that it did not have the part which contributes most of the value in most relationships. But&nbsp;<i>I did not know that</i> at the time. Recently, I tried asking people where most of the value in their relationships came from, got an answer, and thought “Wait, that’s supposed to be the&nbsp;<i>big</i> value prop? Not just a marginal value prop along with many others? Well shit, I was indeed largely missing that part!”.</p><p>It is not necessarily obvious when one’s own relationships are missing The Thing, if you don’t already have a sense of where most of the value is supposed to come from.</p><h3>Second Motivation: Selecting For and Cultivating The Thing</h3><p>Even people who are in great relationships are usually not able to articulate The Thing very well. Some people actively&nbsp;<i>want</i> The Thing to be mysterious; I would advise such people to read&nbsp;<a href="https://www.lesswrong.com/s/6BFkmEgre7uwhDxDR/p/x4dG4GhpZH2hgz59x"><u>Joy In The Merely Real</u></a> and then return to this post. Others just find it hard to articulate because, well, accurately Naming abstract things is hard.</p><p>But if we&nbsp;<i>can</i> point directly at The Thing, then we can optimize for it more directly. When dating, for instance, we can try to induce The Thing with prospective partners quickly, rather than just relying on luck. Or in established relationships, if we understand what The Thing is, we can better nourish it and get more of it. And since The Thing is typically the biggest value proposition, that can hopefully make our relationships - and our lives - a lot better.</p><p>So what the heck is The Thing?</p><h2>Some Pointers To The Thing</h2><p>Let’s start with a raw quote. When I asked David, here’s how he explained The Thing:</p><blockquote><p>Something like a cross between sibling and best friend? Reliably available and excited to s</p></blockquote>...
What’s the main value proposition of romantic relationships? Now, look, I know that when people drop that kind of question, they’re often about to present a hyper-cynical answer which totally ignores the main thing which is great and beautiful about relationships. And then they’re going to say something about how relationships are overrated or some such, making you as a reader just feel sad and/or enraged. That’s not what this post is about. So let me start with some more constructive motivations… First Motivation: Noticing When The Thing Is Missing I had a 10-year relationship. It had its ups and downs, but it was overall negative for me. And I now think a big part of the problem with that relationship was that it did not have the part which contributes most of the value in most relationships. But I did not know that at the time. Recently, I tried asking people where most of the value in their relationships came from, got an answer, and thought “Wait, that’s supposed to be the big value prop? Not just a marginal value prop along with many others? Well shit, I was indeed largely missing that part!”. It is not necessarily obvious when one’s own relationships are missing The Thing, if you don’t already have a sense of where most of the value is supposed to come from. Second Motivation: Selecting For and Cultivating The Thing Even people who are in great relationships are usually not able to articulate The Thing very well. Some people actively want The Thing to be mysterious; I would advise such people to read Joy In The Merely Real and then return to this post. Others just find it hard to articulate because, well, accurately Naming abstract things is hard. But if we can point directly at The Thing, then we can optimize for it more directly. When dating, for instance, we can try to induce The Thing with prospective partners quickly, rather than just relying on luck. Or in established relationships, if we understand what The Thing is, we can better nourish it an
3,991
1.1.0
Revision
false
null
null
CrosspostOutput
5zyPJPhLEwSzHWEtu
1-the-challenge-of-unawareness-for-impartial-altruist-action
1. The challenge of unawareness for impartial altruist action guidance: Introduction
null
false
false
false
null
rv7RzMiG3esRT4CQi
null
true
false
false
false
Post
null
2025-06-02T08:54:16.673Z
null
false
false
2
2
2025-06-02T20:51:33.373Z
false
false
post
[]
null
null
gFSDPFugNmzskni44
6
14
47
false
0.034607
null
false
false
2025-06-13T09:39:34.826Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
null
null
qgdGA4ZEyW7zNdK84
null
null
null
false
null
[]
null
18
0
2025-06-02T08:54:16.674Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
20
null
null
null
null
[ { "__typename": "Tag", "_id": "4fxcbJ8xSv4SAYkkx", "adminOnly": false, "afBaseScore": 9, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "EQNTWXLKMeWMp2FQS", "displayName": "Ben Pace" }, { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 19, "canEditUserIds": null, "core": false, "createdAt": "2020-09-12T02:42:21.141Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "EQNTWXLKMeWMp2FQS", "displayName": "Ben Pace" }, { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Bayesianism", "needsReview": false, "noindex": false, "postCount": 62, "score": 19, "shortName": null, "slug": "bayesianism", "suggestedAsFilter": false, "userId": "qgdGA4ZEyW7zNdK84", "voteCount": 2, "wikiOnly": false }, { "__typename": "Tag", "_id": "X8JsWEnBRPvs5Y99i", "adminOnly": false, "afBaseScore": 0, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 0, "canEditUserIds": null, "core": false, "createdAt": "2015-12-03T07:35:06.000Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [] }, "isArbitalImport": true, "isPlaceholderPage": false, "isSubforum": false, "name": "Decision theory", "needsReview": false, "noindex": false, "postCount": 500, "score": 0, "shortName": null, "slug": "decision-theory", "suggestedAsFilter": false, "userId": "nmk3nLpQE89dMRzzN", "voteCount": 0, "wikiOnly": false }, { "__typename": "Tag", "_id": "EdRnMXBRbY5JDf5df", "adminOnly": false, "afBaseScore": 6, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "nmk3nLpQE89dMRzzN", "displayName": "Eliezer Yudkowsky" } ] }, "baseScore": 13, "canEditUserIds": null, "core": false, "createdAt": "2015-07-02T01:53:10.000Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "nmk3nLpQE89dMRzzN", "displayName": "Eliezer Yudkowsky" } ] }, "isArbitalImport": true, "isPlaceholderPage": false, "isSubforum": false, "name": "Epistemology", "needsReview": false, "noindex": false, "postCount": 424, "score": 13, "shortName": null, "slug": "epistemology", "suggestedAsFilter": false, "userId": "nmk3nLpQE89dMRzzN", "voteCount": 1, "wikiOnly": false }, { "__typename": "Tag", "_id": "xexCWMyds6QLWognu", "adminOnly": false, "afBaseScore": 0, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 2, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T03:38:23.532Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 20, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "si6LoAENzqPCmi2Dh", "displayName": "ihatenumbersinusernames7" }, { "_id": "B2sXjQTGgwoGE9FES", "displayName": "SeaIgloo" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "World Optimization", "needsReview": false, "noindex": false, "postCount": 3151, "score": 2, "shortName": null, "slug": "world-optimization", "suggestedAsFilter": true, "userId": "XtphY3uYHwruKqDyG", "voteCount": 2, "wikiOnly": false }, { "__typename": "Tag", "_id": "Ng8Gice9KNkncxqcj", "adminOnly": false, "afBaseScore": 0, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [] }, "baseScore": 1, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:17.072Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 100, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "iMqytjy9ns89Fzfyv", "displayName": "miakko" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Rationality", "needsReview": false, "noindex": false, "postCount": 4302, "score": 1, "shortName": null, "slug": "rationality", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 1, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
14
0
0
9
0
rv7RzMiG3esRT4CQi
anthony-digiovanni
2019-12-15T12:43:56.701Z
antimonyanthony
Anthony DiGiovanni
null
null
Anthony DiGiovanni
1,033
58
false
false
<p>Researcher at the Center on Long-Term Risk. All opinions my own.</p>
null
null
10
142
1
1
1
1
0
gXeEWGjTWyqgrQTzR
User
null
null
null
[ "canModeratePersonal", "alignmentVoters" ]
null
null
5zyPJPhLEwSzHWEtu
https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/5zyPJPhLEwSzHWEtu/surcbcpjls6eagbqhmby
SocialPreviewType
gFSDPFugNmzskni44
<p><i>(This sequence assumes basic familiarity with longtermist cause prioritization concepts, though the issues I raise also apply to non-longtermist interventions.)</i></p><p>Are EA interventions net-positive from an impartial perspective — one that gives moral weight to all consequences, no matter how distant? What if they’re neither positive, nor negative, nor neutral?</p><p>Trying&nbsp;to reduce x-risk, improve institutions, or end factory farming might seem robustly positive. After all, we don’t need to be certain in order to&nbsp;do good in expectation.&nbsp;But when we step back to look at how radically far-reaching the impartial perspective is, we start to see a deeper problem than “uncertainty”. This problem is <strong>unawareness</strong>: many possible consequences of our actions haven’t even occurred to us in much detail, if at all.</p><p>Why is unawareness a serious challenge for impartial altruists? Well, impartiality entails that we account for <i>all</i>&nbsp;moral patients, and <i>all</i>&nbsp;the most significant impacts we could have on them. Here’s a glimpse of what such an accounting might require:</p><blockquote><p>How likely is it&nbsp;that we’re missing insights as “big if true”&nbsp;as&nbsp;the discovery of other galaxies, the possibility of digital sentience, or the <a href="https://grabbyaliens.com/">grabby aliens</a>&nbsp;hypothesis? What’s the expected value of preventing extinction, given these insights? Or set aside long-term predictions for now. What are all the plausible pathways by which we might prevent or cause extinction by, say, designing policies for an intelligence explosion? Not just the pathways we find most salient.</p></blockquote><p>It’s no secret that the space of possible futures is daunting. But whenever we make decisions based on the impartial good, we’re making tradeoffs between these futures. Imagine what it would take to grasp such tradeoffs. Imagine an agent who could conceive of a representative sample of these futures, in fairly precise detail. This agent might still be highly uncertain which future will play out. They might even be <a href="https://iep.utm.edu/re-bo-ag/">cognitively bounded</a>. And yet, if they claimed, “I choose to do&nbsp;<span class="math-tex"><span><span><span><span><span><span>A</span></span></span></span></span></span></span>&nbsp;instead of&nbsp;<span class="math-tex"><span><span><span><span><span><span>B</span></span></span></span></span></span></span>&nbsp;because&nbsp;<span class="math-tex"><span><span><span><span><span><span>A</span></span></span></span></span></span></span>&nbsp;has better expected consequences”, they’d have actually factored in the most important consequences.&nbsp;All&nbsp;the heavy weights would be on the scales.</p><p>And then there’s us. We don’t merely share this agent’s limitations.&nbsp;Rather, we are also <i>unaware</i>&nbsp;(or at best only coarsely aware)<strong>&nb</strong>... </p>
(This sequence assumes basic familiarity with longtermist cause prioritization concepts, though the issues I raise also apply to non-longtermist interventions.) Are EA interventions net-positive from an impartial perspective — one that gives moral weight to all consequences, no matter how distant? What if they’re neither positive, nor negative, nor neutral? Trying to reduce x-risk, improve institutions, or end factory farming might seem robustly positive. After all, we don’t need to be certain in order to do good in expectation. But when we step back to look at how radically far-reaching the impartial perspective is, we start to see a deeper problem than “uncertainty”. This problem is unawareness: many possible consequences of our actions haven’t even occurred to us in much detail, if at all. Why is unawareness a serious challenge for impartial altruists? Well, impartiality entails that we account for all moral patients, and all the most significant impacts we could have on them. Here’s a glimpse of what such an accounting might require: > How likely is it that we’re missing insights as “big if true” as the discovery of other galaxies, the possibility of digital sentience, or the grabby aliens hypothesis? What’s the expected value of preventing extinction, given these insights? Or set aside long-term predictions for now. What are all the plausible pathways by which we might prevent or cause extinction by, say, designing policies for an intelligence explosion? Not just the pathways we find most salient. It’s no secret that the space of possible futures is daunting. But whenever we make decisions based on the impartial good, we’re making tradeoffs between these futures. Imagine what it would take to grasp such tradeoffs. Imagine an agent who could conceive of a representative sample of these futures, in fairly precise detail. This agent might still be highly uncertain which future will play out. They might even be cognitively bounded. And yet, if they claimed, “I
5,006
1.0.3
Revision
true
false
a3hnfA9EnYm9bssTZ
CrosspostOutput
zj5ffnBBqwa6BFnDK
wicked-thoughts
‘Wicked’: thoughts
null
false
false
false
null
jRRYAy2mQAHy2Mq3f
null
true
false
false
false
Post
null
2025-06-02T06:20:04.201Z
null
false
false
2
2
2025-06-02T19:42:16.937Z
false
false
post
[]
null
null
M6XcKqdaXaemEm6Tk
3
6
16
false
0.015788
null
false
false
2025-06-08T21:26:28.993Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
null
null
r38pkCm7wF4M44MDQ
null
null
null
false
null
[]
null
5
0
2025-06-02T06:20:04.201Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
4
null
null
null
null
[ { "__typename": "Tag", "_id": "4Kcm4etxAJjmeDkHP", "adminOnly": false, "afBaseScore": 3, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 10, "canEditUserIds": null, "core": false, "createdAt": "2020-05-13T04:57:42.173Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" }, { "_id": "dRaCtsAWxk7sgirSY", "displayName": "Jordan Morgan" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Book Reviews / Media Reviews", "needsReview": false, "noindex": false, "postCount": 405, "score": 10, "shortName": null, "slug": "book-reviews-media-reviews", "suggestedAsFilter": false, "userId": "qgdGA4ZEyW7zNdK84", "voteCount": 2, "wikiOnly": false } ]
o6AENga9Pbts6MrQL
0
0
null
false
null
null
0
6
0
0
2
0
jRRYAy2mQAHy2Mq3f
katjagrace
2009-02-27T14:15:22.378Z
KatjaGrace
KatjaGrace
null
null
null
9,330
309
false
false
null
null
627
509
0
3
7
1
0
r38pkCm7wF4M44MDQ
User
null
null
null
[ "trustLevel1", "canModeratePersonal", "alignmentVoters", "alignmentForum" ]
null
null
zj5ffnBBqwa6BFnDK
SocialPreviewType
M6XcKqdaXaemEm6Tk
<p>I watched <em>Wicked</em> (the 2024 movie) with my ex and his family at Christmas. My current stance is that it was pretty fun but not especially incredible or deep. I could be pretty wrong—watching movies isn’t my strong suit, but I do like chatting about them afterwards. Some thoughts:</p> <ol> <li>Glinda is initially too shallow and awful to be taken seriously as a character, so it’s hard to care about her. It felt like she changed too much too fast as a result of Elphaba being good to her, especially in the direction of being not a throwaway idiot played for laughs.</li> <li>Elphaba feels too much set up by the world to be a sympathetic victim-hero—like, she has an unimportant flaw and then everyone everywhere despises her for it, and at the same time they are committing great evil and she is the only one who can see this. It feels a bit like a self-aggrandizing pity fantasy of a child (‘once upon a time there was a little girl called [my name] and everyone was mean to her for no reason because they were bad and also they hurt animals and she was the only person who cared, and also she really had strong magical powers that nobody else had…’). <br> <br> Given her situation of being aggressively mistreated and perceiving great ills in the world around her, I don’t know that she behaves particularly relatably, admirably or interestingly for most of the movie. She is defensive and incurious and doesn’t seem to have much going on until the most blatant wrong falls into her lap. (And this wrong is even being committed against one of her only friends, making acting on it particularly socially easy and called-for by basic decency.) So I guess her character also seems hard to be invested in. I’m not sure what makes a sympathetic victim-hero feel like a pure incarnation of a moving archetype versus an on-the-nose trope review, but for me it landed a bit too much as the latter.</li> <li>On the other hand, Elphaba’s singing is really something. Though her solo songs mostly didn’t do that much for me. I did enjoy the acknowledgement of the human tendency to have dumb fantasies about being respected by people you don’t know in “The Wizard and I”. In non-solo songs, I had fun with “Popular”, “Dancing through life”, and “What is this feeling?” Though Elphaba’s description of Glinda as simply ‘blonde’, is contrary to the narrative that she is above taking appearances to define people, so I don’t get that bit</li></ol>...
I watched Wicked (the 2024 movie) with my ex and his family at Christmas. My current stance is that it was pretty fun but not especially incredible or deep. I could be pretty wrong—watching movies isn’t my strong suit, but I do like chatting about them afterwards. Some thoughts: 1. Glinda is initially too shallow and awful to be taken seriously as a character, so it’s hard to care about her. It felt like she changed too much too fast as a result of Elphaba being good to her, especially in the direction of being not a throwaway idiot played for laughs. 2. Elphaba feels too much set up by the world to be a sympathetic victim-hero—like, she has an unimportant flaw and then everyone everywhere despises her for it, and at the same time they are committing great evil and she is the only one who can see this. It feels a bit like a self-aggrandizing pity fantasy of a child (‘once upon a time there was a little girl called [my name] and everyone was mean to her for no reason because they were bad and also they hurt animals and she was the only person who cared, and also she really had strong magical powers that nobody else had…’). Given her situation of being aggressively mistreated and perceiving great ills in the world around her, I don’t know that she behaves particularly relatably, admirably or interestingly for most of the movie. She is defensive and incurious and doesn’t seem to have much going on until the most blatant wrong falls into her lap. (And this wrong is even being committed against one of her only friends, making acting on it particularly socially easy and called-for by basic decency.) So I guess her character also seems hard to be invested in. I’m not sure what makes a sympathetic victim-hero feel like a pure incarnation of a moving archetype versus an on-the-nose trope review, but for me it landed a bit too much as the latter. 3. On the other hand, Elphaba’s singing is really something. Though her solo songs mostly didn’t do that much for me.
991
1.0.0
Revision
false
null
null
CrosspostOutput
tfEAuhrqfSTn67gd4
humanity-needs-a-ulysses-pact-for-ai
Humanity needs a Ulysses Pact for AI
null
false
false
false
null
Lr3HeFMMqz38tdsur
null
true
false
false
false
Post
null
2025-06-01T20:56:01.118Z
null
false
false
2
2
2025-06-01T23:19:20.974Z
false
false
post
[]
null
null
xf5nDosRxnowCGeYG
2
1
1
false
0.006919
null
false
false
2025-06-02T08:57:35.497Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
qgdGA4ZEyW7zNdK84
null
null
null
false
null
[]
null
0
0
2025-05-31T16:30:18.154Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
1
null
null
null
null
[ { "__typename": "Tag", "_id": "sYm3HiWcfZvrGu3ui", "adminOnly": false, "afBaseScore": 2, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" } ] }, "baseScore": 12, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:22.097Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" }, { "_id": "sof55TPMQaeBaxhsS", "displayName": "tommylees112" }, { "_id": "AayjS8XzcnDKhGdTv", "displayName": "shark" }, { "_id": "HnALuwRdo6k9HLaMt", "displayName": "Alex Firssoff" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI", "needsReview": false, "noindex": false, "postCount": 12544, "score": 12, "shortName": null, "slug": "ai", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 4, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
1
0
0
0
0
Lr3HeFMMqz38tdsur
lukas-n-p-egger
2025-04-08T18:49:16.503Z
lukas-n-p-egger
Lukas N.P. Egger
null
null
null
0
0
false
false
<p>VP of Product Strategy &amp; Innovation at SAP Signavio. Background in CS and philosophy.</p>
null
null
1
0
0
0
0
0.9
0
grecHJcgkb3KW5wnM
User
null
null
null
null
null
null
tfEAuhrqfSTn67gd4
SocialPreviewType
xf5nDosRxnowCGeYG
<p>In <i>The Odyssey</i>, Ulysses longs to hear the sirens, creatures whose song was so beautiful it drove sailors to steer their ships into rocks. But through his cunning, he finds a way not to risk his life and get to enjoy the beguiling Siren song.</p><p>Ulysses has himself tied to the mast and orders his crew to plug their ears with wax. He hears the song while the ship stays its course and everyone survives.</p><p>That ancient myth is suddenly very modern when you consider AI to be our siren song, i.e., unimaginable breakthroughs, seductive acceleration, and extraordinary rewards at the price of existential risk.&nbsp;</p><p>It seems like the majority of the conversations revolve around either pontificating on the grave danger of the siren's song or the unfathomable beauty of their tunes.<br><br>Instead of choosing one side of the dichotomy, what if we acknowledge the existential threat of AI, while at the same time chasing it? Thus, we can overcome the reflex of wanting to win the argument for one or the other side and instead engage in the conversation of what could bind us to the mast and safeguard the ears of our fellow travellers. How can we design and implement decisions to bind ourselves in the future?&nbsp;</p><p>What we need is not to silence the sirens. We need systems that let us hear them, without steering toward the rocks. We need a Ulysses Pact<span class="footnote-reference" data-footnote-reference="" data-footnote-index="1" data-footnote-id="o28w9ypu7ac" role="doc-noteref" id="fnrefo28w9ypu7ac"><sup><a href="#fno28w9ypu7ac">[1]</a></sup></span>&nbsp;that helps us to choose our limits while we still can. To be both ambitious and wise. To yearn for the song and survive it!</p><p>In an age of powerful technologies, fragile governance, manifold externalities, and entangled incentive systems that might be the deepest kind of intelligence we can show.</p><ol class="footnote-section footnotes" data-footnote-section="" role="doc-endnotes"><li class="footnote-item" data-footnote-item="" data-footnote-index="1" data-footnote-id="o28w9ypu7ac" role="doc-endnote" id="fno28w9ypu7ac"><span class="footnote-back-link" data-footnote-back-link="" data-footnote-id="o28w9ypu7ac"><sup><strong><a href="#fnrefo28w9ypu7ac">^</a></strong></sup></span><div class="footnote-content" data-footnote-content=""><p>https://en.wikipedia.org/wiki/Ulysses_pact</p></div></li></ol>
In The Odyssey, Ulysses longs to hear the sirens, creatures whose song was so beautiful it drove sailors to steer their ships into rocks. But through his cunning, he finds a way not to risk his life and get to enjoy the beguiling Siren song. Ulysses has himself tied to the mast and orders his crew to plug their ears with wax. He hears the song while the ship stays its course and everyone survives. That ancient myth is suddenly very modern when you consider AI to be our siren song, i.e., unimaginable breakthroughs, seductive acceleration, and extraordinary rewards at the price of existential risk.  It seems like the majority of the conversations revolve around either pontificating on the grave danger of the siren's song or the unfathomable beauty of their tunes. Instead of choosing one side of the dichotomy, what if we acknowledge the existential threat of AI, while at the same time chasing it? Thus, we can overcome the reflex of wanting to win the argument for one or the other side and instead engage in the conversation of what could bind us to the mast and safeguard the ears of our fellow travellers. How can we design and implement decisions to bind ourselves in the future?  What we need is not to silence the sirens. We need systems that let us hear them, without steering toward the rocks. We need a Ulysses Pact[1] that helps us to choose our limits while we still can. To be both ambitious and wise. To yearn for the song and survive it! In an age of powerful technologies, fragile governance, manifold externalities, and entangled incentive systems that might be the deepest kind of intelligence we can show. 1. ^ https://en.wikipedia.org/wiki/Ulysses_pact
290
1.1.0
Revision
false
null
null
CrosspostOutput
XwxSYoFWijSwrHF6n
untitled-draft-nr3m
Text Steers Vision
null
false
false
false
null
xzDxYrdmL9ntyqvL5
null
true
false
false
false
Post
2025-06-01T20:30:39.215Z
null
false
false
2
2
2025-06-01T23:19:11.015Z
false
false
post
[]
null
null
QapNQ5GgJpEATQ9D9
0
4
4
false
0.008242
null
false
false
2025-06-01T20:30:39.215Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
qgdGA4ZEyW7zNdK84
null
null
null
false
null
[]
null
0
0
2025-05-22T23:20:44.444Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
9
null
null
null
null
[ { "__typename": "Tag", "_id": "sYm3HiWcfZvrGu3ui", "adminOnly": false, "afBaseScore": 2, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" } ] }, "baseScore": 12, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:22.097Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" }, { "_id": "sof55TPMQaeBaxhsS", "displayName": "tommylees112" }, { "_id": "AayjS8XzcnDKhGdTv", "displayName": "shark" }, { "_id": "HnALuwRdo6k9HLaMt", "displayName": "Alex Firssoff" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI", "needsReview": false, "noindex": false, "postCount": 12544, "score": 12, "shortName": null, "slug": "ai", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 4, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
4
0
0
0
0
xzDxYrdmL9ntyqvL5
woody-gan
2025-05-22T23:16:42.389Z
woody-gan
Woody Gan
null
null
null
3
0
false
false
null
null
1
0
0
0
0
0.9
0
grecHJcgkb3KW5wnM
User
null
null
null
null
null
null
XwxSYoFWijSwrHF6n
https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/XwxSYoFWijSwrHF6n/l5t931hy2s8cyfdac7f2
SocialPreviewType
QapNQ5GgJpEATQ9D9
<h1 data-internal-id="Textual_Steering_Vectors_Can_Improve_Visual_Understanding_in_Multimodal_Large_Language_Models"><strong>Textual Steering Vectors Can Improve Visual Understanding in Multimodal Large Language Models</strong></h1><p><strong>TL;DR:</strong>&nbsp;We discovered the possibility of using steering vectors from text-only models to enhance visual reasoning in multimodal LLMs (MLLMs). The technique is simple: extract textual representations for concepts like "spatial relationships" and "counting" from the LLM backbone and then apply them to the vision-language model's hidden states. This steering not only changes how the model understands visual content but also leads to meaningful improvements - 15.8% on spatial reasoning tasks and 34.2% on counting tasks for example, which suggests that these models maintain some unified representations that bridge text and vision.</p><h2 data-internal-id="The_Core_Insight__Text_Steers_Vision"><strong>The Core Insight: Text Steers Vision</strong></h2><p>Here's something that surprised us: you can make a vision-language model better at visual understanding by steering it with vectors derived purely from text.</p><p>Many multimodal large language models (MLLMs) like <a href="https://arxiv.org/abs/2407.07726">PaliGemma</a><span class="footnote-reference" data-footnote-reference="" data-footnote-index="1" data-footnote-id="9x1oghq3bts" role="doc-noteref" id="fnref9x1oghq3bts"><sup><a href="#fn9x1oghq3bts">[1]</a></sup></span>&nbsp;and <a href="https://huggingface.co/HuggingFaceM4/Idefics3-8B-Llama3">Idefics</a><span class="footnote-reference" data-footnote-reference="" data-footnote-index="2" data-footnote-id="9tpdohvip69" role="doc-noteref" id="fnref9tpdohvip69"><sup><a href="#fn9tpdohvip69">[2]</a></sup></span>&nbsp;are built by taking a pre-trained text-only LLM backbone and teaching it to process images. We wondered: do these models preserve enough of their original text representations that we could use text-only steering techniques to improve their visual reasoning?</p><p>The answer appears to be a resounding yes.</p><h2 data-internal-id="The_Mechanics_of_Steering_Vectors"><strong>The Mechanics of Steering Vectors</strong></h2><p>Let's clarify exactly what we mean by "steering vectors" and how they work:</p><h3 data-internal-id="What_Is_a_Steering_Vector_"><strong>What Is a Steering Vector?</strong></h3><p>A steering vector is a direction in the model's high-dimensional activation space that corresponds to some concept or behavior<span class="footnote-reference" data-footnote-reference="" data-footnote-index="3" data-footnote-id="3qpoifvqhpy" role="doc-noteref" id="fnref3qpoifvqhpy"><sup><a href="#fn3qpoifvqhpy">[3]</a></sup></span><span class="footnote-reference" data-footnote-reference="" data-footnote-index="4" data-footnote-id="zo3huuamj39" role="doc-noteref" id="fnrefzo3huuamj39"><sup><a href="#fnzo3huuamj39">[4]</a></sup></span>. Think of it as a "conceptual compass" - it points toward the neural representation of ideas like "spatial relationships" or "counting."</p><h3 data-internal-id="How_Are_Steering_Vectors_Applied_"><strong>How Are Steering Vectors Applied?</strong></h3><p>The application is surprisingly straightforward. During inference, at a specific layer&nbsp;<span class="math-tex"><span class="mjpage"><span class="mjx-chtml"><span class="mjx-math" aria-label="\ell"><span class="mjx-mrow" aria-hidden="true"><span class="mjx-mi"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.372em;">ℓ</span></span></span></span></span></span></span>, we modify the model's hidden states:</p><p><span class="math-tex"><span class="mjpage"><span class="mjx-chtml"><span class="mjx-math" aria-label="h'(\ell) = h(\ell) + \alpha \times v(\ell)"><span class="mjx-mrow" aria-hidden="true"><span class="mjx-msup"><span class="mjx-base"><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.446em; padding-bottom: 0.298em;">h</span></span></span><span class="mjx-sup" style="font-size: 70.7%; vertical-align: 0.513em; padding-left: 0px; padding-right: 0.071em;"><span class="mjx-mo" style=""><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.298em; padding-bottom: 0.298em;">′</span></span></span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">(</span></span><span class="mjx-mi"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.372em;">ℓ</span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">)</span></span><span class="mjx-mo MJXc-space3"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.077em; padding-bottom: 0.298em;">=</span></span><span class="mjx-mi MJXc-space3"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.446em; padding-bottom: 0.298em;">h</span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">(</span></span><span class="mjx-mi"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.372em;">ℓ</span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">)</span></span><span class="mjx-mo MJXc-space2"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.298em; padding-bottom: 0.446em;">+</span></span><span class="mjx-mi MJXc-space2"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.225em; padding-bottom: 0.298em;">α</span></span><span class="mjx-mo MJXc-space2"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.225em; padding-bottom: 0.298em;">×</span></span><span class="mjx-mi MJXc-space2"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.225em; padding-bottom: 0.298em;">v</span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">(</span></span><span class="mjx-mi"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.372em;">ℓ</span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">)</span></span></span></span></span></span></span></p><p>Where:</p><ul><li><span class="math-tex"><span class="mjpage"><span class="mjx-chtml"><span class="mjx-math" aria-label="h(\ell)"><span class="mjx-mrow" aria-hidden="true"><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.446em; padding-bottom: 0.298em;">h</span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">(</span></span><span class="mjx-mi"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.372em;">ℓ</span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">)</span></span></span></span></span></span></span>&nbsp;= original hidden states at layer&nbsp;<span class="math-tex"><span class="mjpage"><span class="mjx-chtml"><span class="mjx-math" aria-label="\ell"><span class="mjx-mrow" aria-hidden="true"><span class="mjx-mi"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.372em;">ℓ</span></span></span></span></span></span></span></li><li><span class="math-tex"><span class="mjpage"><span class="mjx-chtml"><span class="mjx-math" aria-label="v(\ell)"><span class="mjx-mrow" aria-hidden="true"><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.225em; padding-bottom: 0.298em;">v</span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">(</span></span><span class="mjx-mi"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.372em;">ℓ</span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">)</span></span></span></span></span></span></span>&nbsp;= our steering vector for that layer</li><li><span class="math-tex"><span class="mjpage"><span class="mjx-chtml"><span class="mjx-math" aria-label="\alpha"><span class="mjx-mrow" aria-hidden="true"><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.225em; padding-bottom: 0.298em;">α</span></span></span></span></span></span></span>&nbsp;= steering strength (how much to intervene)</li><li><span class="math-tex"><span class="mjpage"><span class="mjx-chtml"><span class="mjx-math" aria-label="h'(\ell)"><span class="mjx-mrow" aria-hidden="true"><span class="mjx-msup"><span class="mjx-base"><span class="mjx-mi"><span class="mjx-char MJXc-TeX-math-I" style="padding-top: 0.446em; padding-bottom: 0.298em;">h</span></span></span><span class="mjx-sup" style="font-size: 70.7%; vertical-align: 0.513em; padding-left: 0px; padding-right: 0.071em;"><span class="mjx-mo" style=""><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.298em; padding-bottom: 0.298em;">′</span></span></span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">(</span></span><span class="mjx-mi"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.372em;">ℓ</span></span><span class="mjx-mo"><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.446em; padding-bottom: 0.593em;">)</span></span></span></span></span></span></span>&nbsp;= modified hidden states</li></ul><p><img src="https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/64dc562ba11dfcb708cbb58aa152b5f2d86c697b108955ef86787271c8ebaa53/nsdymjyj4tfpl77jrejb" alt=""></p><p>We can also choose to where to apply this steering:</p><ul><li><strong>Image tokens only</strong>: Steer only on hidden states of image tokens</li><li><strong>Text tokens only</strong>: Steer only hidden states of text tokens</li><li><strong>Both</strong>: Steer on hidden states of all tokens except special tokens like &lt;bos&gt;</li></ul><p>An important notice here is that we never steer the output token i... <style>.mjx-chtml {display: inline-block; line-height: 0; text-indent: 0; text-align: left; text-transform: none; font-style: normal; font-weight: normal; font-size: 100%; font-size-adjust: none; letter-spacing: normal; word-wrap: normal; word-spacing: normal; white-space: nowrap; float: none; direction: ltr; max-width: none; max-height: none; min-width: 0; min-height: 0; border: 0; margin: 0; padding: 1px 0} .MJXc-display {display: block; text-align: center; margin: 1em 0; padding: 0} .mjx-chtml[tabindex]:focus, body :focus .mjx-chtml[tabindex] {display: inline-table} .mjx-full-width {text-align: center; display: table-cell!important; width: 10000em} .mjx-math {display: inline-block; border-collapse: separate; border-spacing: 0} .mjx-math * {display: inline-block; -webkit-box-sizing: content-box!important; -moz-box-sizing: content-box!important; box-sizing: content-box!important; text-align: left} .mjx-numerator {display: block; text-align: center} .mjx-denominator {display: block; text-align: center} .MJXc-stacked {height: 0; position: relative} .MJXc-stacked > * {position: absolute} .MJXc-bevelled > * {display: inline-block} .mjx-stack {display: inline-block} .mjx-op {display: block} .mjx-under {display: table-cell} .mjx-over {display: block} .mjx-over > * {padding-left: 0px!important; padding-right: 0px!important} .mjx-under > * {padding-left: 0px!important; padding-right: 0px!important} .mjx-stack > .mjx-sup {display: block} .mjx-stack > .mjx-sub {display: block} .mjx-prestack > .mjx-presup {display: block} .mjx-prestack > .mjx-presub {display: block} .mjx-delim-h > .mjx-char {display: inline-block} .mjx-surd {vertical-align: top} .mjx-surd + .mjx-box {display: inline-flex} .mjx-mphantom * {visibility: hidden} .mjx-merror {background-color: #FFFF88; color: #CC0000; border: 1px solid #CC0000; padding: 2px 3px; font-style: normal; font-size: 90%} .mjx-annotation-xml {line-height: normal} .mjx-menclose > svg {fill: none; stroke: currentColor; overflow: visible} .mjx-mtr {display: table-row} .mjx-mlabeledtr {display: table-row} .mjx-mtd {display: table-cell; text-align: center} .mjx-label {display: table-row} .mjx-box {display: inline-block} .mjx-block {display: block} .mjx-span {display: inline} .mjx-char {display: block; white-space: pre} .mjx-itable {display: inline-table; width: auto} .mjx-row {display: table-row} .mjx-cell {display: table-cell} .mjx-table {display: table; width: 100%} .mjx-line {display: block; height: 0} .mjx-strut {width: 0; padding-top: 1em} .mjx-vsize {width: 0} .MJXc-space1 {margin-left: .167em} .MJXc-space2 {margin-left: .222em} .MJXc-space3 {margin-left: .278em} .mjx-test.mjx-test-display {display: table!important} .mjx-test.mjx-test-inline {display: inline!important; margin-right: -1px} .mjx-test.mjx-test-default {display: block!important; clear: both} .mjx-ex-box {display: inline-block!important; position: absolute; overflow: hidden; min-height: 0; max-height: none; padding: 0; border: 0; margin: 0; width: 1px; height: 60ex} .mjx-test-inline .mjx-left-box {display: inline-block; width: 0; float: left} .mjx-test-inline .mjx-right-box {display: inline-block; width: 0; float: right} .mjx-test-display .mjx-right-box {display: table-cell!important; width: 10000em!important; min-width: 0; max-width: none; padding: 0; border: 0; margin: 0} .MJXc-TeX-unknown-R {font-family: monospace; font-style: normal; font-weight: normal} .MJXc-TeX-unknown-I {font-family: monospace; font-style: italic; font-weight: normal} .MJXc-TeX-unknown-B {font-family: monospace; font-style: normal; font-weight: bold} .MJXc-TeX-unknown-BI {font-family: monospace; font-style: italic; font-weight: bold} .MJXc-TeX-ams-R {font-family: MJXc-TeX-ams-R,MJXc-TeX-ams-Rw} .MJXc-TeX-cal-B {font-family: MJXc-TeX-cal-B,MJXc-TeX-cal-Bx,MJXc-TeX-cal-Bw} .MJXc-TeX-frak-R {font-family: MJXc-TeX-frak-R,MJXc-TeX-frak-Rw} .MJXc-TeX-frak-B {font-family: MJXc-TeX-frak-B,MJXc-TeX-frak-Bx,MJXc-TeX-frak-Bw} .MJXc-TeX-math-BI {font-family: MJXc-TeX-math-BI,MJXc-TeX-math-BIx,MJXc-TeX-math-BIw} .MJXc-TeX-sans-R {font-family: MJXc-TeX-sans-R,MJXc-TeX-sans-Rw} .MJXc-TeX-sans-B {font-family: MJXc-TeX-sans-B,MJXc-TeX-sans-Bx,MJXc-TeX-sans-Bw} .MJXc-TeX-sans-I {font-family: MJXc-TeX-sans-I,MJXc-TeX-sans-Ix,MJXc-TeX-sans-Iw} .MJXc-TeX-script-R {font-family: MJXc-TeX-script-R,MJXc-TeX-script-Rw} .MJXc-TeX-type-R {font-family: MJXc-TeX-type-R,MJXc-TeX-type-Rw} .MJXc-TeX-cal-R {font-family: MJXc-TeX-cal-R,MJXc-TeX-cal-Rw} .MJXc-TeX-main-B {font-family: MJXc-TeX-main-B,MJXc-TeX-main-Bx,MJXc-TeX-main-Bw} .MJXc-TeX-main-I {font-family: MJXc-TeX-main-I,MJXc-TeX-main-Ix,MJXc-TeX-main-Iw} .MJXc-TeX-main-R {font-family: MJXc-TeX-main-R,MJXc-TeX-main-Rw} .MJXc-TeX-math-I {font-family: MJXc-TeX-math-I,MJXc-TeX-math-Ix,MJXc-TeX-math-Iw} .MJXc-TeX-size1-R {font-family: MJXc-TeX-size1-R,MJXc-TeX-size1-Rw} .MJXc-TeX-size2-R {font-family: MJXc-TeX-size2-R,MJXc-TeX-size2-Rw} .MJXc-TeX-size3-R {font-family: MJXc-TeX-size3-R,MJXc-TeX-size3-Rw} .MJXc-TeX-size4-R {font-family: MJXc-TeX-size4-R,MJXc-TeX-size4-Rw} .MJXc-TeX-vec-R {font-family: MJXc-TeX-vec-R,MJXc-TeX-vec-Rw} .MJXc-TeX-vec-B {font-family: MJXc-TeX-vec-B,MJXc-TeX-vec-Bx,MJXc-TeX-vec-Bw} @font-face {font-family: MJXc-TeX-ams-R; src: local('MathJax_AMS'), local('MathJax_AMS-Regular')} @font-face {font-family: MJXc-TeX-ams-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_AMS-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_AMS-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_AMS-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-cal-B; src: local('MathJax_Caligraphic Bold'), local('MathJax_Caligraphic-Bold')} @font-face {font-family: MJXc-TeX-cal-Bx; src: local('MathJax_Caligraphic'); font-weight: bold} @font-face {font-family: MJXc-TeX-cal-Bw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Caligraphic-Bold.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Caligraphic-Bold.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Caligraphic-Bold.otf') format('opentype')} @font-face {font-family: MJXc-TeX-frak-R; src: local('MathJax_Fraktur'), local('MathJax_Fraktur-Regular')} @font-face {font-family: MJXc-TeX-frak-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Fraktur-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Fraktur-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Fraktur-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-frak-B; src: local('MathJax_Fraktur Bold'), local('MathJax_Fraktur-Bold')} @font-face {font-family: MJXc-TeX-frak-Bx; src: local('MathJax_Fraktur'); font-weight: bold} @font-face {font-family: MJXc-TeX-frak-Bw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Fraktur-Bold.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Fraktur-Bold.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Fraktur-Bold.otf') format('opentype')} @font-face {font-family: MJXc-TeX-math-BI; src: local('MathJax_Math BoldItalic'), local('MathJax_Math-BoldItalic')} @font-face {font-family: MJXc-TeX-math-BIx; src: local('MathJax_Math'); font-weight: bold; font-style: italic} @font-face {font-family: MJXc-TeX-math-BIw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Math-BoldItalic.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Math-BoldItalic.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Math-BoldItalic.otf') format('opentype')} @font-face {font-family: MJXc-TeX-sans-R; src: local('MathJax_SansSerif'), local('MathJax_SansSerif-Regular')} @font-face {font-family: MJXc-TeX-sans-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_SansSerif-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_SansSerif-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_SansSerif-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-sans-B; src: local('MathJax_SansSerif Bold'), local('MathJax_SansSerif-Bold')} @font-face {font-family: MJXc-TeX-sans-Bx; src: local('MathJax_SansSerif'); font-weight: bold} @font-face {font-family: MJXc-TeX-sans-Bw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_SansSerif-Bold.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_SansSerif-Bold.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_SansSerif-Bold.otf') format('opentype')} @font-face {font-family: MJXc-TeX-sans-I; src: local('MathJax_SansSerif Italic'), local('MathJax_SansSerif-Italic')} @font-face {font-family: MJXc-TeX-sans-Ix; src: local('MathJax_SansSerif'); font-style: italic} @font-face {font-family: MJXc-TeX-sans-Iw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_SansSerif-Italic.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_SansSerif-Italic.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_SansSerif-Italic.otf') format('opentype')} @font-face {font-family: MJXc-TeX-script-R; src: local('MathJax_Script'), local('MathJax_Script-Regular')} @font-face {font-family: MJXc-TeX-script-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Script-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Script-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Script-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-type-R; src: local('MathJax_Typewriter'), local('MathJax_Typewriter-Regular')} @font-face {font-family: MJXc-TeX-type-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Typewriter-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Typewriter-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Typewriter-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-cal-R; src: local('MathJax_Caligraphic'), local('MathJax_Caligraphic-Regular')} @font-face {font-family: MJXc-TeX-cal-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Caligraphic-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Caligraphic-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Caligraphic-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-main-B; src: local('MathJax_Main Bold'), local('MathJax_Main-Bold')} @font-face {font-family: MJXc-TeX-main-Bx; src: local('MathJax_Main'); font-weight: bold} @font-face {font-family: MJXc-TeX-main-Bw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Main-Bold.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Main-Bold.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Main-Bold.otf') format('opentype')} @font-face {font-family: MJXc-TeX-main-I; src: local('MathJax_Main Italic'), local('MathJax_Main-Italic')} @font-face {font-family: MJXc-TeX-main-Ix; src: local('MathJax_Main'); font-style: italic} @font-face {font-family: MJXc-TeX-main-Iw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Main-Italic.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Main-Italic.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Main-Italic.otf') format('opentype')} @font-face {font-family: MJXc-TeX-main-R; src: local('MathJax_Main'), local('MathJax_Main-Regular')} @font-face {font-family: MJXc-TeX-main-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Main-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Main-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Main-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-math-I; src: local('MathJax_Math Italic'), local('MathJax_Math-Italic')} @font-face {font-family: MJXc-TeX-math-Ix; src: local('MathJax_Math'); font-style: italic} @font-face {font-family: MJXc-TeX-math-Iw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Math-Italic.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Math-Italic.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Math-Italic.otf') format('opentype')} @font-face {font-family: MJXc-TeX-size1-R; src: local('MathJax_Size1'), local('MathJax_Size1-Regular')} @font-face {font-family: MJXc-TeX-size1-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Size1-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Size1-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Size1-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-size2-R; src: local('MathJax_Size2'), local('MathJax_Size2-Regular')} @font-face {font-family: MJXc-TeX-size2-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Size2-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Size2-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Size2-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-size3-R; src: local('MathJax_Size3'), local('MathJax_Size3-Regular')} @font-face {font-family: MJXc-TeX-size3-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Size3-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Size3-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Size3-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-size4-R; src: local('MathJax_Size4'), local('MathJax_Size4-Regular')} @font-face {font-family: MJXc-TeX-size4-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Size4-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Size4-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Size4-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-vec-R; src: local('MathJax_Vector'), local('MathJax_Vector-Regular')} @font-face {font-family: MJXc-TeX-vec-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Vector-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Vector-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Vector-Regular.otf') format('opentype')} @font-face {font-family: MJXc-TeX-vec-B; src: local('MathJax_Vector Bold'), local('MathJax_Vector-Bold')} @font-face {font-family: MJXc-TeX-vec-Bx; src: local('MathJax_Vector'); font-weight: bold} @font-face {font-family: MJXc-TeX-vec-Bw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Vector-Bold.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Vector-Bold.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Vector-Bold.otf') format('opentype')} </style></p>
Textual Steering Vectors Can Improve Visual Understanding in Multimodal Large Language Models TL;DR: We discovered the possibility of using steering vectors from text-only models to enhance visual reasoning in multimodal LLMs (MLLMs). The technique is simple: extract textual representations for concepts like "spatial relationships" and "counting" from the LLM backbone and then apply them to the vision-language model's hidden states. This steering not only changes how the model understands visual content but also leads to meaningful improvements - 15.8% on spatial reasoning tasks and 34.2% on counting tasks for example, which suggests that these models maintain some unified representations that bridge text and vision. The Core Insight: Text Steers Vision Here's something that surprised us: you can make a vision-language model better at visual understanding by steering it with vectors derived purely from text. Many multimodal large language models (MLLMs) like PaliGemma[1] and Idefics[2] are built by taking a pre-trained text-only LLM backbone and teaching it to process images. We wondered: do these models preserve enough of their original text representations that we could use text-only steering techniques to improve their visual reasoning? The answer appears to be a resounding yes. The Mechanics of Steering Vectors Let's clarify exactly what we mean by "steering vectors" and how they work: What Is a Steering Vector? A steering vector is a direction in the model's high-dimensional activation space that corresponds to some concept or behavior[3][4]. Think of it as a "conceptual compass" - it points toward the neural representation of ideas like "spatial relationships" or "counting." How Are Steering Vectors Applied? The application is surprisingly straightforward. During inference, at a specific layer ℓ, we modify the model's hidden states: h′(ℓ)=h(ℓ)+α×v(ℓ) Where: * h(ℓ) = original hidden states at layer ℓ * v(ℓ) = our steering vector for that layer *
2,249
1.15.0
Revision
false
null
null
CrosspostOutput
RngeiEJu6z6qny9ts
possible-ai-regulation-emergency
Possible AI regulation emergency?
null
false
false
false
null
Q2oaNonArzibx5cQN
null
true
false
false
false
Post
2025-06-01T20:30:06.342Z
null
false
false
2
2
null
false
false
question
[]
null
null
gfa6xq79A7Kte45zE
1
8
19
false
0.011617
null
false
false
2025-06-02T02:02:24.933Z
null
null
null
null
null
true
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
qgdGA4ZEyW7zNdK84
null
null
null
false
null
[]
null
5
0
2025-06-01T20:26:15.269Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
1
null
null
null
null
[ { "__typename": "Tag", "_id": "sYm3HiWcfZvrGu3ui", "adminOnly": false, "afBaseScore": 2, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" } ] }, "baseScore": 12, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:22.097Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" }, { "_id": "sof55TPMQaeBaxhsS", "displayName": "tommylees112" }, { "_id": "AayjS8XzcnDKhGdTv", "displayName": "shark" }, { "_id": "HnALuwRdo6k9HLaMt", "displayName": "Alex Firssoff" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI", "needsReview": false, "noindex": false, "postCount": 12544, "score": 12, "shortName": null, "slug": "ai", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 4, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
8
0
0
4
0
Q2oaNonArzibx5cQN
cronodas
2009-02-27T04:42:19.587Z
CronoDAS
CronoDAS
null
null
null
16,417
0
false
false
null
null
61
4,274
0
0
0
1
6
grecHJcgkb3KW5wnM
User
null
null
null
[ "trustLevel1", "alignmentVoters", "canModeratePersonal" ]
null
null
RngeiEJu6z6qny9ts
SocialPreviewType
gfa6xq79A7Kte45zE
<p>I apologize for breaking the "no contemporary politics" norm, but I read that one of the less talked about provisions in the "One Big Beautiful Bill" that recently passed the US House of Representatives forbids individual US states from regulating AI for ten years. This sounds bad for AI risk. Should we in the US be calling our Senators?</p>
I apologize for breaking the "no contemporary politics" norm, but I read that one of the less talked about provisions in the "One Big Beautiful Bill" that recently passed the US House of Representatives forbids individual US states from regulating AI for ten years. This sounds bad for AI risk. Should we in the US be calling our Senators?
59
1.2.0
Revision
false
null
null
CrosspostOutput
seEF372qGJhgKLBsv
eliezer-yudkowsky-and-connor-leahy-or-ai-risk-safety-and
Eliezer Yudkowsky & Connor Leahy | AI Risk, Safety & Alignment Q&A [4K Remaster + HQ Audio]
null
false
false
false
null
rT2ee6nLeneuRZGLd
null
true
false
false
false
Post
https://www.youtube.com/watch?v=naOQVM0VbNg
2025-06-01T20:20:47.762Z
null
false
false
2
2
2025-06-02T20:52:40.281Z
false
false
linkpost
[]
null
null
XTZcfQBXHLTGQxqXg
2
3
-8
false
0.001198
null
false
false
2025-06-02T12:24:51.041Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
qgdGA4ZEyW7zNdK84
null
null
null
false
null
[]
null
-5
0
2025-05-31T12:07:54.135Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
1
null
null
null
null
[ { "__typename": "Tag", "_id": "sYm3HiWcfZvrGu3ui", "adminOnly": false, "afBaseScore": 2, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" } ] }, "baseScore": 12, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:22.097Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" }, { "_id": "sof55TPMQaeBaxhsS", "displayName": "tommylees112" }, { "_id": "AayjS8XzcnDKhGdTv", "displayName": "shark" }, { "_id": "HnALuwRdo6k9HLaMt", "displayName": "Alex Firssoff" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI", "needsReview": false, "noindex": false, "postCount": 12544, "score": 12, "shortName": null, "slug": "ai", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 4, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
3
0
0
2
0
rT2ee6nLeneuRZGLd
dex-volkov
2025-05-31T12:07:02.091Z
dex-volkov
Dex Volkov
null
null
null
-9
0
false
false
null
null
1
0
0
0
0
0.8
0
grecHJcgkb3KW5wnM
User
null
null
null
null
null
null
seEF372qGJhgKLBsv
SocialPreviewType
XTZcfQBXHLTGQxqXg
<p>Many complained about broken audio. I fixed the audio + upscaled to 4k to bring new life into this important discussion.</p>
Many complained about broken audio. I fixed the audio + upscaled to 4k to bring new life into this important discussion.
21
1.2.0
Revision
false
null
null
CrosspostOutput
GheQAxhnrFLmc2q9e
economists-should-track-the-speed-and-magnitude-of-ai
Economists should track the speed and magnitude of AI implementation projects
null
false
false
false
null
xmfpBvZThPX9HBzRw
null
true
false
false
false
Post
null
2025-06-01T20:15:58.672Z
null
false
false
2
2
2025-06-02T20:52:19.643Z
false
false
post
[]
null
null
aHL7FCjzRgfp7v6ow
0
1
1
false
0.006528
null
false
false
2025-06-01T20:15:58.672Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
qgdGA4ZEyW7zNdK84
null
null
null
false
null
[]
null
0
0
2025-06-01T17:38:37.337Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
3
null
null
null
null
[ { "__typename": "Tag", "_id": "sYm3HiWcfZvrGu3ui", "adminOnly": false, "afBaseScore": 2, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" } ] }, "baseScore": 12, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:22.097Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" }, { "_id": "sof55TPMQaeBaxhsS", "displayName": "tommylees112" }, { "_id": "AayjS8XzcnDKhGdTv", "displayName": "shark" }, { "_id": "HnALuwRdo6k9HLaMt", "displayName": "Alex Firssoff" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI", "needsReview": false, "noindex": false, "postCount": 12544, "score": 12, "shortName": null, "slug": "ai", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 4, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
1
0
0
0
0
xmfpBvZThPX9HBzRw
parrotrobot
2024-01-04T22:12:03.728Z
ParrotRobot
ParrotRobot
null
null
null
4
0
false
false
null
null
2
1
0
0
0
0.9
0
qgdGA4ZEyW7zNdK84
User
null
null
null
null
null
null
GheQAxhnrFLmc2q9e
SocialPreviewType
aHL7FCjzRgfp7v6ow
<p>Both economists and the AI community are highly interested in forecasting the magnitude and speed of AI’s impact on the economy.</p><p>But economists and AI technologists speak different languages. Economists talk in aggregates and dynamics, while technologists talk about capabilities and benchmarks and product releases.</p><p>Economists and technologists concern themselves with different scales of time and aggregation, too. A new AI release or benchmark result comes out practically every day. Much key economic data (BEA, BLS, etc.) comes out only yearly, and is aggregated at the level of entire industries.</p><p>We basically have <i>no data yet</i> on what the aggregate productivity benefit of 2024’s AI advances will be.</p><p>And when the data does come out, and observations are made, we will not know which AI releases and capabilities were the cause.</p><p>I am an engineer who is currently directly building large-scale AI automations (an “AI engineer”). What I’ve seen over just the last few months is incredible and undoubtedly important economically. I’m not sure how, but I hope to give economists some “handles” by which they can better observe the impact AI is already having, and the impact it will have.</p><p><strong>Main idea:</strong> If one looks at AI automation projects at higher resolution, one can get highly suggestive evidence connecting investment opportunities and rates of return (i.e., key quantitative drivers of economic growth) to specific AI capabilities released at specific times.</p><p>Here is how it goes:</p><ul><li>A specific AI capability is released (e.g., a reasoning model, or a more reliable instruction-following model, or something more exotic like a computer use model).</li><li>A company (such as the one that I work at) searches for a use case, and finds one (often quite lucrative; quick AI advancement produces lots of low-hanging fruit) and we build a proof-of-concept demo.<ul><li>This takes on the order of a few months from when the relevant AI capability is released, but can be much faster.</li><li><strong>AI capability connection #1: The timing of demo development reflects the connection of specific AI capabilities to specific categories of market opportunities.</strong></li></ul></li><li>We do some several iterations of development and deploy a system to production.<ul><li>From what I've seen, this has been <i>fast</i>, even in complex domains. A very large customer support use cases took less than a year to go from first-demo to high-volume production.</li><li><strong>AI capability connection #2: Th</strong></li></ul></li></ul>...
Both economists and the AI community are highly interested in forecasting the magnitude and speed of AI’s impact on the economy. But economists and AI technologists speak different languages. Economists talk in aggregates and dynamics, while technologists talk about capabilities and benchmarks and product releases. Economists and technologists concern themselves with different scales of time and aggregation, too. A new AI release or benchmark result comes out practically every day. Much key economic data (BEA, BLS, etc.) comes out only yearly, and is aggregated at the level of entire industries. We basically have no data yet on what the aggregate productivity benefit of 2024’s AI advances will be. And when the data does come out, and observations are made, we will not know which AI releases and capabilities were the cause. I am an engineer who is currently directly building large-scale AI automations (an “AI engineer”). What I’ve seen over just the last few months is incredible and undoubtedly important economically. I’m not sure how, but I hope to give economists some “handles” by which they can better observe the impact AI is already having, and the impact it will have. Main idea: If one looks at AI automation projects at higher resolution, one can get highly suggestive evidence connecting investment opportunities and rates of return (i.e., key quantitative drivers of economic growth) to specific AI capabilities released at specific times. Here is how it goes: * A specific AI capability is released (e.g., a reasoning model, or a more reliable instruction-following model, or something more exotic like a computer use model). * A company (such as the one that I work at) searches for a use case, and finds one (often quite lucrative; quick AI advancement produces lots of low-hanging fruit) and we build a proof-of-concept demo. * This takes on the order of a few months from when the relevant AI capability is released, but can be much faster. * AI capabil
723
1.4.0
Revision
false
null
null
CrosspostOutput
3s7pymrxnbJJ2qnNz
ingroup
Ingroup
null
false
false
false
null
g8JkZfL8PTqAefpvx
null
true
false
false
false
Post
null
2025-06-01T19:47:33.176Z
null
false
false
2
2
null
false
false
post
[]
null
null
HCrmzSaNpoKKDQzzo
12
18
-8
false
-0.004828
null
false
false
2025-06-04T23:31:39.122Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
qgdGA4ZEyW7zNdK84
null
null
null
false
null
[]
null
0
0
2025-06-01T19:40:47.554Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[]
1
null
null
null
null
[ { "__typename": "Tag", "_id": "DdgSyQoZXjj3KnF4N", "adminOnly": false, "afBaseScore": 9, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "EQNTWXLKMeWMp2FQS", "displayName": "Ben Pace" }, { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 19, "canEditUserIds": null, "core": false, "createdAt": "2020-07-13T15:43:11.661Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "EQNTWXLKMeWMp2FQS", "displayName": "Ben Pace" }, { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Tribalism", "needsReview": false, "noindex": false, "postCount": 68, "score": 19, "shortName": null, "slug": "tribalism", "suggestedAsFilter": false, "userId": "qxJ28GN72aiJu96iF", "voteCount": 2, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
18
0
0
13
0
g8JkZfL8PTqAefpvx
jenniferrm
2009-03-06T17:16:50.600Z
JenniferRM
JenniferRM
null
null
null
8,867
18
false
false
null
null
33
1,366
0
0
1
1
8
qgdGA4ZEyW7zNdK84
User
null
null
null
[ "trustLevel1", "canModeratePersonal", "alignmentVoters" ]
null
null
3s7pymrxnbJJ2qnNz
SocialPreviewType
HCrmzSaNpoKKDQzzo
<p>Ingroup is such a fantastic community. You guys are the best. You aren't bad, like those icky people in outgroup who lack our special characteristics.</p><p>I could just go on and on about how uniquely valuable and worthy of appreciation we are, individually, and together, but I bet you guys understand what I'm trying to say, and can fill in the blanks for yourself. You're just that smart.</p><p>And I'm sure you can see through some of the bullshit here. There's a bit of flattery going on anytime ingroup is praised so broadly, or so directly, but ingroup isn't fooled by such flattery... and that unique mix of humbleness and pride is part of our charm.</p>
Ingroup is such a fantastic community. You guys are the best. You aren't bad, like those icky people in outgroup who lack our special characteristics. I could just go on and on about how uniquely valuable and worthy of appreciation we are, individually, and together, but I bet you guys understand what I'm trying to say, and can fill in the blanks for yourself. You're just that smart. And I'm sure you can see through some of the bullshit here. There's a bit of flattery going on anytime ingroup is praised so broadly, or so directly, but ingroup isn't fooled by such flattery... and that unique mix of humbleness and pride is part of our charm.
116
1.1.0
Revision
false
null
null
CrosspostOutput
jXmn3oPhFNQtYczC8
apply-to-the-ai-security-bootcamp-aug-4-aug-29
Apply to the AI Security Bootcamp [Aug 4 - Aug 29]
null
false
false
false
null
yv9FeipHRyn5rNTs8
[ { "__typename": "CoauthorStatusOutput", "confirmed": true, "requested": false, "userId": "dFpY59uY4LcsCfNG3" }, { "__typename": "CoauthorStatusOutput", "confirmed": true, "requested": false, "userId": "wHxQDEtZzQwKLvoPc" } ]
true
false
false
false
Post
null
2025-06-01T19:47:15.482Z
null
false
false
2
2
null
false
false
post
[]
null
null
NDFo5rEjyFMZuxCdL
1
13
27
false
0.016248
null
false
false
2025-06-05T22:08:28.705Z
null
null
null
null
null
false
false
null
null
null
false
false
null
null
null
null
null
null
null
null
null
null
false
null
null
[]
null
qgdGA4ZEyW7zNdK84
null
null
null
false
null
[]
null
5
0
2025-05-31T13:42:17.883Z
false
false
null
null
true
false
false
0
0
0
null
null
null
null
null
null
null
false
0
0
namesAttachedReactions
false
[ { "__typename": "User", "_id": "dFpY59uY4LcsCfNG3", "afCommentCount": 0, "afKarma": 0, "afPostCount": 0, "commentCount": 2, "createdAt": "2020-07-26T21:22:51.930Z", "deleted": false, "displayName": "Jan Michelfeit", "fullName": null, "htmlBio": "", "isAdmin": false, "jobTitle": null, "karma": 28, "organization": null, "postCount": 0, "previousDisplayName": null, "profileImageId": null, "reviewedByUserId": "r38pkCm7wF4M44MDQ", "sequenceCount": 0, "slug": "jan-michelfeit-2", "spamRiskScore": 1, "tagRevisionCount": 0, "username": "jan-michelfeit-2" }, { "__typename": "User", "_id": "wHxQDEtZzQwKLvoPc", "afCommentCount": 0, "afKarma": 0, "afPostCount": 0, "commentCount": 0, "createdAt": "2025-06-25T16:14:17.757Z", "deleted": false, "displayName": "Jinglin Li", "fullName": null, "htmlBio": "", "isAdmin": false, "jobTitle": null, "karma": 0, "organization": null, "postCount": 0, "previousDisplayName": null, "profileImageId": null, "reviewedByUserId": null, "sequenceCount": 0, "slug": "jinglin-li", "spamRiskScore": 0.8, "tagRevisionCount": 0, "username": "jinglin-li" } ]
5
null
null
null
null
[ { "__typename": "Tag", "_id": "6zBEfFYJxhSEcchbR", "adminOnly": false, "afBaseScore": 3, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 9, "canEditUserIds": null, "core": false, "createdAt": "2022-06-09T19:10:50.755Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI Alignment Fieldbuilding", "needsReview": false, "noindex": false, "postCount": 359, "score": 9, "shortName": null, "slug": "ai-alignment-fieldbuilding", "suggestedAsFilter": false, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 1, "wikiOnly": false }, { "__typename": "Tag", "_id": "MhHM6Rx2b4F8tHTQk", "adminOnly": false, "afBaseScore": 3, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "baseScore": 9, "canEditUserIds": null, "core": false, "createdAt": "2020-08-16T20:50:23.539Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 0, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "qgdGA4ZEyW7zNdK84", "displayName": "Ruby" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "Computer Security & Cryptography", "needsReview": false, "noindex": false, "postCount": 117, "score": 9, "shortName": null, "slug": "computer-security-and-cryptography", "suggestedAsFilter": false, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 1, "wikiOnly": false }, { "__typename": "Tag", "_id": "sYm3HiWcfZvrGu3ui", "adminOnly": false, "afBaseScore": 2, "afExtendedScore": { "reacts": { "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" } ] }, "baseScore": 12, "canEditUserIds": null, "core": true, "createdAt": "2020-06-14T22:24:22.097Z", "currentUserExtendedVote": null, "currentUserVote": null, "deleted": false, "descriptionTruncationCount": 2000, "extendedScore": { "reacts": { "important": null, "insightful": null, "thinking": null, "typo": null }, "usersWhoLiked": [ { "_id": "nLbwLhBaQeG6tCNDN", "displayName": "jimrandomh" }, { "_id": "sof55TPMQaeBaxhsS", "displayName": "tommylees112" }, { "_id": "AayjS8XzcnDKhGdTv", "displayName": "shark" }, { "_id": "HnALuwRdo6k9HLaMt", "displayName": "Alex Firssoff" } ] }, "isArbitalImport": false, "isPlaceholderPage": false, "isSubforum": false, "name": "AI", "needsReview": false, "noindex": false, "postCount": 12544, "score": 12, "shortName": null, "slug": "ai", "suggestedAsFilter": true, "userId": "r38pkCm7wF4M44MDQ", "voteCount": 4, "wikiOnly": false } ]
null
0
0
null
false
null
null
0
13
0
0
3
0
yv9FeipHRyn5rNTs8
pranav-gade
2022-02-17T08:27:15.542Z
pranav-gade
Pranav Gade
null
null
null
143
53
false
false
null
null
2
6
0
0
0
1
0
qgdGA4ZEyW7zNdK84
User
null
null
null
[ "alignmentVoters", "canModeratePersonal" ]
null
null
jXmn3oPhFNQtYczC8
SocialPreviewType
NDFo5rEjyFMZuxCdL
<h2>tl;dr</h2><p>We're excited to announce AI Security Bootcamp (AISB), a 4-week intensive program designed to bring researchers and engineers up to speed on security fundamentals for AI systems. The program will cover cybersecurity fundamentals (cryptography, networks), AI infrastructure security (GPUs, supply chain security), and more novel attacks on ML systems (dataset trojans, model extraction). This program will run in-person from 4th Aug to 29th Aug in London, UK. We will cover all expenses.</p><p><strong>Apply </strong><a href="https://airtable.com/app5CvqWYtNWVXBzh/pagLyM1S6ZevGYUcB/form"><strong>here</strong></a><strong> to participate in AISB before EOD AoE, 22nd June 2025.</strong></p><p>We are also looking for instructors for parts of the program and staff to help with operations. <strong>Apply </strong><a href="https://airtable.com/app5CvqWYtNWVXBzh/pagLyM1S6ZevGYUcB/form"><strong>here.</strong></a></p><h1>Summary</h1><p>We are running a 4-week program designed to equip AI safety researchers and engineers with critical security skills. We hope you'll leave the program with a well-practiced <a href="https://intelligence.org/2017/11/25/security-mindset-ordinary-paranoia/">security mindset</a> that will help you work on impactful projects and make AI systems more secure.</p><p>The curriculum includes exercises designed to help you get hands-on experience with securing AI systems, while building and practicing the security mindset. This includes a mix of pair programming exercises, lectures, reading about public vulnerabilities, and chats with experts.&nbsp;</p><p>This program is aimed at people who are at ~the start of their journey into security, and have working knowledge in ML (or are willing to brush up using the MLAB curriculum before the program). We encourage you to <a href="https://airtable.com/app5CvqWYtNWVXBzh/pagLyM1S6ZevGYUcB/form">apply </a>if you think you'll be a good fit regardless of checking all the boxes - if going on technical deep dives and trying to understand how systems work by peeling the layers of abstraction away excites you, we'd love to hear from you.</p><h1>Content</h1><p>The content is divided into roughly three sections - introduction to security, cybersecurity for AI infrastructure, and attacks on modern ML pipelines. The exercises are designed to give you hands-on experience with cybersecurity on both the offensive and defensive sides, as well as train your security muscles so that you are more effective at spotting and patching security holes.</p><ol><li><strong>Introduction to security</strong> - This includes introduction to cryptography and cryptanalysis, the basics of Linux security, and breaking network protocols.</li><li><strong>Securing AI infrastructure</strong> - This section covers topics especially relevant to AI security - containerization, a case study on supply chain security, application security, and SecOps.</li><li><strong>AI pi</strong></li></ol>...
tl;dr We're excited to announce AI Security Bootcamp (AISB), a 4-week intensive program designed to bring researchers and engineers up to speed on security fundamentals for AI systems. The program will cover cybersecurity fundamentals (cryptography, networks), AI infrastructure security (GPUs, supply chain security), and more novel attacks on ML systems (dataset trojans, model extraction). This program will run in-person from 4th Aug to 29th Aug in London, UK. We will cover all expenses. Apply here to participate in AISB before EOD AoE, 22nd June 2025. We are also looking for instructors for parts of the program and staff to help with operations. Apply here. Summary We are running a 4-week program designed to equip AI safety researchers and engineers with critical security skills. We hope you'll leave the program with a well-practiced security mindset that will help you work on impactful projects and make AI systems more secure. The curriculum includes exercises designed to help you get hands-on experience with securing AI systems, while building and practicing the security mindset. This includes a mix of pair programming exercises, lectures, reading about public vulnerabilities, and chats with experts.  This program is aimed at people who are at ~the start of their journey into security, and have working knowledge in ML (or are willing to brush up using the MLAB curriculum before the program). We encourage you to apply if you think you'll be a good fit regardless of checking all the boxes - if going on technical deep dives and trying to understand how systems work by peeling the layers of abstraction away excites you, we'd love to hear from you. Content The content is divided into roughly three sections - introduction to security, cybersecurity for AI infrastructure, and attacks on modern ML pipelines. The exercises are designed to give you hands-on experience with cybersecurity on both the offensive and defensive sides, as well as train your security muscl
1,130
1.9.0
Revision
false
null
null
CrosspostOutput