{
  localUrl: '../page/7hy.html',
  arbitalUrl: 'https://arbital.com/p/7hy',
  rawJsonUrl: '../raw/7hy.json',
  likeableId: '0',
  likeableType: 'page',
  myLikeValue: '0',
  likeCount: '0',
  dislikeCount: '0',
  likeScore: '0',
  individualLikes: [],
  pageId: '7hy',
  edit: '2',
  editSummary: '',
  prevEdit: '1',
  currentEdit: '2',
  wasPublished: 'true',
  type: 'wiki',
  title: 'AI arms race',
  clickbait: 'This part of the plan describes how to prevent and handle AI arms race.',
  textLength: '932',
  alias: '7hy',
  externalUrl: '',
  sortChildrenBy: 'likes',
  hasVote: 'false',
  voteType: '',
  votesAnonymous: 'false',
  editCreatorId: 'AlexeiAndreev',
  editCreatedAt: '2017-01-31 02:18:24',
  pageCreatorId: 'AlexeiAndreev',
  pageCreatedAt: '2017-01-26 00:01:17',
  seeDomainId: '0',
  editDomainId: '2223',
  submitToDomainId: '0',
  isAutosave: 'false',
  isSnapshot: 'false',
  isLiveEdit: 'true',
  isMinorEdit: 'false',
  indirectTeacher: 'false',
  todoCount: '0',
  isEditorComment: 'false',
  isApprovedComment: 'false',
  isResolved: 'false',
  snapshotText: '',
  anchorContext: '',
  anchorText: '',
  anchorOffset: '0',
  mergedInto: '',
  isDeleted: 'false',
  viewCount: '21',
  text: 'To understand this topic better, see [3qx]. This page will describe how this concept is treated in [7hx].\n\nI think the arms race can of two types. In either case it's driven by fear, but in the first case it's imaginary, whereas in the second case it's real. \n\nImagine a scenario where you know that an organization is developing an AGI at full speed. For now they have no safety plans, so you assume that when they reach AGI, they'll just proceed ahead. This will likely encourage you to develop your own AGI faster in turn.\n\nIf the other company will in fact do what you expect, then your fear is real. However, if the other company has a plan to stop at some milestone before full AGI to evaluate its safety, then your fear is mostly imaginary.\n\nThis makes communication really important. If companies make their plans public, then it will make it easier for people to distinguish between real and imaginary arms race scenarios. ',
  metaText: '',
  isTextLoaded: 'true',
  isSubscribedToDiscussion: 'false',
  isSubscribedToUser: 'false',
  isSubscribedAsMaintainer: 'false',
  discussionSubscriberCount: '1',
  maintainerCount: '1',
  userSubscriberCount: '0',
  lastVisit: '',
  hasDraft: 'false',
  votes: [],
  voteSummary: 'null',
  muVoteSummary: '0',
  voteScaling: '0',
  currentUserVote: '-2',
  voteCount: '0',
  lockedVoteType: '',
  maxEditEver: '0',
  redLinkCount: '0',
  lockedBy: '',
  lockedUntil: '',
  nextPageId: '',
  prevPageId: '',
  usedAsMastery: 'false',
  proposalEditNum: '0',
  permissions: {
    edit: {
      has: 'false',
      reason: 'You don't have domain permission to edit this page'
    },
    proposeEdit: {
      has: 'true',
      reason: ''
    },
    delete: {
      has: 'false',
      reason: 'You don't have domain permission to delete this page'
    },
    comment: {
      has: 'false',
      reason: 'You can't comment in this domain because you are not a member'
    },
    proposeComment: {
      has: 'true',
      reason: ''
    }
  },
  summaries: {},
  creatorIds: [
    'AlexeiAndreev'
  ],
  childIds: [],
  parentIds: [
    'the_plan_experiment'
  ],
  commentIds: [],
  questionIds: [],
  tagIds: [],
  relatedIds: [],
  markIds: [],
  explanations: [],
  learnMore: [],
  requirements: [],
  subjects: [],
  lenses: [],
  lensParentId: '',
  pathPages: [],
  learnMoreTaughtMap: {},
  learnMoreCoveredMap: {},
  learnMoreRequiredMap: {},
  editHistory: {},
  domainSubmissions: {},
  answers: [],
  answerCount: '0',
  commentCount: '0',
  newCommentCount: '0',
  linkedMarkCount: '0',
  changeLogs: [
    {
      likeableId: '0',
      likeableType: 'changeLog',
      myLikeValue: '0',
      likeCount: '0',
      dislikeCount: '0',
      likeScore: '0',
      individualLikes: [],
      id: '21900',
      pageId: '7hy',
      userId: 'AlexeiAndreev',
      edit: '0',
      type: 'newEditGroup',
      createdAt: '2017-01-31 02:19:00',
      auxPageId: '2223',
      oldSettingsValue: '21',
      newSettingsValue: '2223'
    },
    {
      likeableId: '0',
      likeableType: 'changeLog',
      myLikeValue: '0',
      likeCount: '0',
      dislikeCount: '0',
      likeScore: '0',
      individualLikes: [],
      id: '21899',
      pageId: '7hy',
      userId: 'AlexeiAndreev',
      edit: '2',
      type: 'newEdit',
      createdAt: '2017-01-31 02:18:24',
      auxPageId: '',
      oldSettingsValue: '',
      newSettingsValue: ''
    },
    {
      likeableId: '0',
      likeableType: 'changeLog',
      myLikeValue: '0',
      likeCount: '0',
      dislikeCount: '0',
      likeScore: '0',
      individualLikes: [],
      id: '21840',
      pageId: '7hy',
      userId: 'AlexeiAndreev',
      edit: '0',
      type: 'newParent',
      createdAt: '2017-01-26 00:01:18',
      auxPageId: 'the_plan_experiment',
      oldSettingsValue: '',
      newSettingsValue: ''
    },
    {
      likeableId: '0',
      likeableType: 'changeLog',
      myLikeValue: '0',
      likeCount: '0',
      dislikeCount: '0',
      likeScore: '0',
      individualLikes: [],
      id: '21838',
      pageId: '7hy',
      userId: 'AlexeiAndreev',
      edit: '1',
      type: 'newEdit',
      createdAt: '2017-01-26 00:01:17',
      auxPageId: '',
      oldSettingsValue: '',
      newSettingsValue: ''
    }
  ],
  feedSubmissions: [],
  searchStrings: {},
  hasChildren: 'false',
  hasParents: 'true',
  redAliases: {},
  improvementTagIds: [],
  nonMetaTagIds: [],
  todos: [],
  slowDownMap: 'null',
  speedUpMap: 'null',
  arcPageIds: 'null',
  contentRequests: {}
}