{
  localUrl: '../page/nonperson_predicate.html',
  arbitalUrl: 'https://arbital.com/p/nonperson_predicate',
  rawJsonUrl: '../raw/1fv.json',
  likeableId: '404',
  likeableType: 'page',
  myLikeValue: '0',
  likeCount: '3',
  dislikeCount: '0',
  likeScore: '3',
  individualLikes: [
    'AlexeiAndreev',
    'EricBruylant',
    'NathanFish'
  ],
  pageId: 'nonperson_predicate',
  edit: '1',
  editSummary: '',
  prevEdit: '0',
  currentEdit: '1',
  wasPublished: 'true',
  type: 'wiki',
  title: 'Nonperson predicate',
  clickbait: 'If we knew which computations were definitely not people, we could tell AIs which programs they were definitely allowed to compute.',
  textLength: '1317',
  alias: 'nonperson_predicate',
  externalUrl: '',
  sortChildrenBy: 'likes',
  hasVote: 'false',
  voteType: '',
  votesAnonymous: 'false',
  editCreatorId: 'EliezerYudkowsky',
  editCreatedAt: '2015-12-28 19:49:00',
  pageCreatorId: 'EliezerYudkowsky',
  pageCreatedAt: '2015-12-28 19:49:00',
  seeDomainId: '0',
  editDomainId: 'EliezerYudkowsky',
  submitToDomainId: '0',
  isAutosave: 'false',
  isSnapshot: 'false',
  isLiveEdit: 'true',
  isMinorEdit: 'false',
  indirectTeacher: 'false',
  todoCount: '0',
  isEditorComment: 'false',
  isApprovedComment: 'true',
  isResolved: 'false',
  snapshotText: '',
  anchorContext: '',
  anchorText: '',
  anchorOffset: '0',
  mergedInto: '',
  isDeleted: 'false',
  viewCount: '137',
  text: 'A "nonperson predicate" is a possible method for preventing an [2c advanced AI] from [6v accidentally running sapient computations] (it would be a potentially huge moral catastrophe if an AI created, ran, and discarded a large number of sapient programs inside itself).  A nonperson predicate looks at potential computations and returns one of two possible answers, "Don't know" and "Definitely not a person".  A successful nonperson predicate may (very often) return "Don't know" for computations that aren't in fact people, but it never returns "Definitely not a person" for something that *is* a person.  In other words, to solve this problem, we don't need to know what consciousness *is* so much as we need to know what it *isn't* - we don't need to be sure what *is* a person, we need to be sure what *isn't* a person.  For a nonperson predicate to be useful, however, it must still pass enough useful computations that we can build a working, capable AI out of them.  (Otherwise "Rocks are okay, everything else might be a person" would be an adequate nonperson predicate.)  The [6r foreseeable difficulty] of a nonperson predicate is that [10k instrumental pressures] to model humans accurately might tend to [42 seek out flaws and loopholes] in any attempted predicate.  See the page on [6v] for more detail.',
  metaText: '',
  isTextLoaded: 'true',
  isSubscribedToDiscussion: 'false',
  isSubscribedToUser: 'false',
  isSubscribedAsMaintainer: 'false',
  discussionSubscriberCount: '1',
  maintainerCount: '1',
  userSubscriberCount: '0',
  lastVisit: '2016-02-09 15:44:59',
  hasDraft: 'false',
  votes: [],
  voteSummary: 'null',
  muVoteSummary: '0',
  voteScaling: '0',
  currentUserVote: '-2',
  voteCount: '0',
  lockedVoteType: '',
  maxEditEver: '0',
  redLinkCount: '0',
  lockedBy: '',
  lockedUntil: '',
  nextPageId: '',
  prevPageId: '',
  usedAsMastery: 'false',
  proposalEditNum: '0',
  permissions: {
    edit: {
      has: 'false',
      reason: 'You don't have domain permission to edit this page'
    },
    proposeEdit: {
      has: 'true',
      reason: ''
    },
    delete: {
      has: 'false',
      reason: 'You don't have domain permission to delete this page'
    },
    comment: {
      has: 'false',
      reason: 'You can't comment in this domain because you are not a member'
    },
    proposeComment: {
      has: 'true',
      reason: ''
    }
  },
  summaries: {},
  creatorIds: [
    'EliezerYudkowsky'
  ],
  childIds: [],
  parentIds: [
    'mindcrime'
  ],
  commentIds: [],
  questionIds: [],
  tagIds: [],
  relatedIds: [],
  markIds: [],
  explanations: [],
  learnMore: [],
  requirements: [],
  subjects: [],
  lenses: [],
  lensParentId: '',
  pathPages: [],
  learnMoreTaughtMap: {},
  learnMoreCoveredMap: {},
  learnMoreRequiredMap: {},
  editHistory: {},
  domainSubmissions: {},
  answers: [],
  answerCount: '0',
  commentCount: '0',
  newCommentCount: '0',
  linkedMarkCount: '0',
  changeLogs: [
    {
      likeableId: '0',
      likeableType: 'changeLog',
      myLikeValue: '0',
      likeCount: '0',
      dislikeCount: '0',
      likeScore: '0',
      individualLikes: [],
      id: '4507',
      pageId: 'nonperson_predicate',
      userId: 'EliezerYudkowsky',
      edit: '1',
      type: 'newEdit',
      createdAt: '2015-12-28 19:49:00',
      auxPageId: '',
      oldSettingsValue: '',
      newSettingsValue: ''
    },
    {
      likeableId: '0',
      likeableType: 'changeLog',
      myLikeValue: '0',
      likeCount: '0',
      dislikeCount: '0',
      likeScore: '0',
      individualLikes: [],
      id: '4499',
      pageId: 'nonperson_predicate',
      userId: 'EliezerYudkowsky',
      edit: '0',
      type: 'newParent',
      createdAt: '2015-12-28 19:34:43',
      auxPageId: 'mindcrime',
      oldSettingsValue: '',
      newSettingsValue: ''
    }
  ],
  feedSubmissions: [],
  searchStrings: {},
  hasChildren: 'false',
  hasParents: 'true',
  redAliases: {},
  improvementTagIds: [],
  nonMetaTagIds: [],
  todos: [],
  slowDownMap: 'null',
  speedUpMap: 'null',
  arcPageIds: 'null',
  contentRequests: {}
}