{ localUrl: '../page/5tw.html', arbitalUrl: 'https://arbital.com/p/5tw', rawJsonUrl: '../raw/5tw.json', likeableId: '0', likeableType: 'page', myLikeValue: '0', likeCount: '0', dislikeCount: '0', likeScore: '0', individualLikes: [], pageId: '5tw', edit: '1', editSummary: '', prevEdit: '0', currentEdit: '1', wasPublished: 'true', type: 'comment', title: '"Omit the 'as'"', clickbait: '', textLength: '13', alias: '5tw', externalUrl: '', sortChildrenBy: 'recentFirst', hasVote: 'false', voteType: '', votesAnonymous: 'false', editCreatorId: 'EricRogstad', editCreatedAt: '2016-08-11 08:37:16', pageCreatorId: 'EricRogstad', pageCreatedAt: '2016-08-11 08:37:16', seeDomainId: '0', editDomainId: '1182', submitToDomainId: '0', isAutosave: 'false', isSnapshot: 'false', isLiveEdit: 'true', isMinorEdit: 'false', indirectTeacher: 'false', todoCount: '0', isEditorComment: 'true', isApprovedComment: 'true', isResolved: 'false', snapshotText: '', anchorContext: 'Similarly, the acceptance rate for \\$8/\\$2 splits is much higher if the experimenter is known to have imposed a maximum \\$2 offer on the Proposer\\. As Falk et\\. al\\. observed, this behavior makes no sense if humans have an innate distaste for unfair outcomes, but in an LDT scenario the answer here changes to "Accept \\$2\\." So as an LDT researcher would observe that the human subjects are behaving suggestively like algorithms that know other algorithms are reasoning about them\\.', anchorText: 'So as an LDT researcher would observe', anchorOffset: '327', mergedInto: '', isDeleted: 'false', viewCount: '75', text: 'Omit the 'as'', metaText: '', isTextLoaded: 'true', isSubscribedToDiscussion: 'false', isSubscribedToUser: 'false', isSubscribedAsMaintainer: 'false', discussionSubscriberCount: '1', maintainerCount: '1', userSubscriberCount: '0', lastVisit: '', hasDraft: 'false', votes: [], voteSummary: 'null', muVoteSummary: '0', voteScaling: '0', currentUserVote: '-2', voteCount: '0', lockedVoteType: '', maxEditEver: '0', redLinkCount: '0', lockedBy: '', lockedUntil: '', nextPageId: '', prevPageId: '', usedAsMastery: 'false', proposalEditNum: '0', permissions: { edit: { has: 'false', reason: 'You don't have domain permission to edit this page' }, proposeEdit: { has: 'true', reason: '' }, delete: { has: 'false', reason: 'You don't have domain permission to delete this page' }, comment: { has: 'false', reason: 'You can't comment in this domain because you are not a member' }, proposeComment: { has: 'true', reason: '' } }, summaries: {}, creatorIds: [ 'EricRogstad' ], childIds: [], parentIds: [ 'ultimatum_game' ], commentIds: [], questionIds: [], tagIds: [], relatedIds: [], markIds: [], explanations: [], learnMore: [], requirements: [], subjects: [], lenses: [], lensParentId: '', pathPages: [], learnMoreTaughtMap: {}, learnMoreCoveredMap: {}, learnMoreRequiredMap: {}, editHistory: {}, domainSubmissions: {}, answers: [], answerCount: '0', commentCount: '0', newCommentCount: '0', linkedMarkCount: '0', changeLogs: [ { likeableId: '0', likeableType: 'changeLog', myLikeValue: '0', likeCount: '0', dislikeCount: '0', likeScore: '0', individualLikes: [], id: '18691', pageId: '5tw', userId: 'EricRogstad', edit: '1', type: 'newEdit', createdAt: '2016-08-11 08:37:16', auxPageId: '', oldSettingsValue: '', newSettingsValue: '' } ], feedSubmissions: [], searchStrings: {}, hasChildren: 'false', hasParents: 'true', redAliases: {}, improvementTagIds: [], nonMetaTagIds: [], todos: [], slowDownMap: 'null', speedUpMap: 'null', arcPageIds: 'null', contentRequests: {} }