Logic.html 31.8 KB
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359 360 361 362 363 364 365 366 367 368 369 370 371 372 373 374 375 376 377 378 379 380 381 382 383 384 385 386 387 388 389 390 391 392 393 394 395 396 397 398 399 400 401 402 403 404 405 406 407 408 409 410 411 412 413 414 415 416 417 418 419 420 421 422 423 424 425 426 427 428 429 430 431 432 433 434 435 436 437 438 439 440 441 442 443 444 445 446 447 448 449 450 451 452 453 454 455 456 457 458 459 460 461 462 463 464 465 466 467 468 469 470 471 472 473 474 475 476 477 478 479 480 481 482 483 484 485 486 487 488 489 490 491 492 493 494 495 496 497 498 499 500 501 502 503 504 505 506 507 508 509 510 511 512 513 514 515 516 517 518 519 520 521 522 523 524 525 526 527 528 529 530 531 532 533 534 535 536 537 538 539 540 541 542 543 544 545 546 547 548 549 550 551 552 553 554 555 556 557 558 559 560 561 562 563 564 565 566 567 568 569 570 571 572 573 574 575 576 577 578 579 580 581 582 583 584 585 586 587 588 589 590 591 592 593 594 595 596 597 598 599 600 601 602 603 604 605 606 607 608 609 610 611 612 613 614 615 616 617 618 619 620 621 622 623 624 625 626 627 628 629 630 631 632 633 634 635 636 637 638 639 640 641 642 643 644 645 646 647 648 649 650 651 652 653 654 655 656 657 658 659 660 661 662 663 664 665 666 667 668 669 670 671 672 673 674 675 676 677 678 679 680 681 682 683 684 685 686 687 688 689 690 691 692 693 694 695 696 697 698 699 700 701 702 703 704 705 706 707 708 709 710 711 712 713 714 715 716 717 718 719 720 721 722 723 724 725 726 727 728 729 730 731 732 733 734 735 736 737 738 739 740 741 742 743 744 745 746 747 748 749 750 751 752 753 754 755 756 757 758 759 760 761 762 763 764 765 766 767 768 769 770 771 772 773 774 775 776 777 778 779 780 781 782 783 784 785 786 787 788
<html xmlns="http://www.w3.org/1999/xhtml">
  <head>
    <meta name="generator" content=
    "HTML Tidy for Mac OS X (vers 31 October 2006 - Apple Inc. build 13), see www.w3.org" />
    <title>
      Logic on the Web -- Web architecture
    </title>
    <link rel="Stylesheet" href="di.css" type="text/css" />
    <meta http-equiv="Content-Type" content=
    "text/html; charset=us-ascii" />
  </head>
  <body bgcolor="#DDFFDD" text="#000000" lang="en">
    <address>
      Tim Berners-Lee<br />
      Date: 1998, last change: $Date: 2009/08/27 21:38:07 $
      $Author: timbl $<br />
      Status: personal view only. Editing status: first draft.
    </address>
    <p>
      <a href="./">Up to Design Issues</a>
    </p>
    <hr />
    <h1>
      The Semantic Web as a language of logic
    </h1>
    <p>
      This looks at the Semantic Web design in the light a little
      reading on formal logic, of the Access Limited Logic system,
      in particular, and in the light of logical languages in
      general. A problem here is a that I am no logician, and so I
      am am having to step like a fascinated reporter into this
      world of which I do not possess intimate experience.
    </p>
    <h2>
      Introduction
    </h2>
    <p>
      The <a href="Toolbox.html">Semantic Web Toolbox</a> discusses
      the step from the web as being a repository of flat data
      without logic to a level at which it is possible to express
      logic. This is something which knowledge representation
      systems have been wary of.
    </p>
    <p>
      The Semantic Web has a different set of goals from most
      systems of logic. As Crawford and Kuipers put it in [<a href=
      "#Crawf90">Crawf90</a>],
    </p>
    <blockquote>
      [...]a knowledge representation system must have the
      following properties:
      <ol>
        <li>It must have a reasonably compact syntax.
        </li>
        <li>It must have a well defined semantics so that one can
        say precisely what is being represented.
        </li>
        <li>It must have sufficient expressive power to represent
        human knowledge.
        </li>
        <li>It must have an efficient, powerful, and understandable
        reasoning mechanism
        </li>
        <li>It must be usable to build large knowledge bases.
        </li>
      </ol>
      <p>
        It has proved difficult, however, to achieve the third and
        fourth properties simultaneously.
      </p>
    </blockquote>
    <p>
      The semantic web goal is to be a unifying system which will
      (like the web for human communication) be as un-restraining
      as possible so that the complexity of reality can be
      described. Therefore item 3 becomes essential. This can be
      achieved by dropping 4 - or the parts of item 4 which
      conflict with 3, notably a single, efficient reasoning
      system. The idea is that, within the global semantic web,
      there will be a subset of the system which will be
      constrained in specific ways so as to achieve the
      tractability and efficiency which is no necessary in real
      applications. However, the semantic web itself will not
      define a reasoning engine. It will define valid operations,
      and will require consistency for them. On the semantic web in
      general, a party must be able to follow a proof of a theorem
      but is not expected to generate one.
    </p>
    <p>
      (This fundamental change goals from KR systems to the
      semantic web is loosely analogous with the goal change from
      conventional hypertext systems to the original Web design
      dropping link consistency in favor of expressive flexibility
      and scalability.The latter did not prevent individual web
      sites from having a strict hierarchical order or matrix
      structure, but it did not require it of the web as a whole.)
    </p>
    <p>
      If there is a <em>semantic web machine</em>, then it is a
      proof validator, not a theorem prover. It can't find answers,
      it can't even check that an answer is right, but it can
      follow a simple explanation that an answer is right. The
      Semantic Web as a source of data should be fodder for
      automated reasoning systems of many kinds, but it as such not
      a reasoning system.
    </p>
    <p>
      Most knowledge representation systems distinguish between
      inference "rules" and other believed information. In some
      cases, this is because the rules (such as substitution in a
      formula) cannot be written in the language - they are defined
      outside the language. In fact the set of rules used by the
      system is often not only formally quite redundant but
      arbitrary. However, a universal design such as the Semantic
      Web must be minimalist. We will ask all logical data on the
      web to be expressed directly or indirectly in terms of the
      semantic web - a strong demand - so we cannot constrain
      applications any further. Different machines which use data
      from the web will use different algorithms, different sets of
      inference rules. In some cases these will be powerful AI
      systems and in others they will be simple document conversion
      systems. The essential this is that the results of either
      must be provably correct against the same basic minimalist
      rules. In fact for interchange of proof, the set of rules is
      an engineering choice.
    </p>
    <p>
      There are many related ways in which subsystems can be
      created
    </p>
    <ul>
      <li>The semantic web language can be subsetted, by the
      removal of operations and axioms and rules;
      </li>
      <li>The set of statements may be limited to that from
      particular documents or web sites;
      </li>
      <li>The form of formulas used may be constrained, for example
      using document schemata;
      </li>
      <li>Application design decisions can be made so as to
      specifically guarantee tractable results using common
      reasoning engines.
      </li>
      <li>Proofs can be constructed by completely hand-built
      application-specific programs
      </li>
    </ul>
    <p>
      For example, Access Limited Logic is restricted (as I
      understand it) to relations r(a,b) available when r is
      accessed, and uses inference rules which only chain forward
      along such links. There is also a "partitioning" of the Web
      by making partitioning the rules in order to limit
      complexity.
    </p>
    <p>
      For the semantic web as a whole, then, we do require
      tractable
    </p>
    <ul>
      <li>Consistency, that it must not be possible to deduce a
      contradiction (without having been given one)
      </li>
      <li>Strength in that all applications must be subsets
      </li>
    </ul>
    <h3>
      Grounding in Reality
    </h3>
    <p>
      Philosophically, the semantic web produces more than a set of
      rules for manipulation of formulae. It defines documents on
      the Web has having a socially significant meaning. Therefore
      it is not simply sufficient to demonstrate that one can
      constrain the semantic web so as to make it isomorphic to a
      particular algebra of a given system, but one must ensure
      that a particular mapping is defined so that the web
      representation of that particular system conveys is semantics
      in a way that it can meaningfully be combined with the rest
      of the semantic web. Electronic commerce needs a solid
      foundation in this way, and the development of the semantic
      web is (in 1999) essential to provide a rigid framework in
      which to define electronic commerce terms, before electronic
      commerce expands as a mass of vaguely defined semantics and
      ad hoc syntax which leaves no room for automatic treatment,
      and in which the court of law rather than a logical
      derivation settles arguments.
    </p>
    <p>
      Practically, the meaning of semantic web data is grounded in
      non-semantic-web applications which are interfaced to the
      semantic web. For example, currency transfer or ecommerce
      applications, which accept semantic web input, define for
      practical purposes what the terms in the currency transfer
      instrument mean.
    </p>
    <h2>
      Axiomatic Basis
    </h2>
    <p>
      <em>@@I [DanC] think this section is outdated by recent
      thoughts [2002] on</em> <a href="#reasoning"><em>paradox and
      the excluded middle</em></a>
    </p>
    <p>
      To the level of first order logic, we don't really need to
      pick one set of axioms in that there are equivalent choices
      which lead to demonstrably the same results.
    </p>
    <p>
      (A cute one at the propositional logic level seems [Burris,
      p126] to be Nicod's set in which nand (in XML toolbox
      &lt;not&gt;..&lt;/not&gt; and below [xy]) is the Sheffer
      (sole) connective and the only rules of inference are
      substitution and the <em>modus ponens</em> equivelent that
      from F and [F[G H]] one can deduce H, and the single axiom
      [[P[QR]][[S[SS]][[UQ][[PU][PU]]]].)
    </p>
    <p>
      Let us assume the properties of first order logic here.
    </p>
    <p>
      If we add anything else we have to be careful that it should
      either be definable in terms of the first order set or that
      the resulting language is a subset of a well proven logical
      system -- or else we have a lot of work to do in establishing
      a new system!
    </p>
    <h2>
      Intractability and Undecidability
    </h2>
    <p>
      These are two goals to which we explicitly do not aspire in
      the Semantic Web in order to get in return expressive power.
      (We still require consistency!). The world is full of
      undecidable statements, and intractable problems. The
      semantic web has to give the power to express such things.
    </p>
    <p>
      Crawford and Kuipers The same explain in the introduction
      their Negation in ALL paper,
    </p>
    <blockquote>
      "Experience with formally specified knowledge representation
      systems has revealed a trade-off between the expressive power
      of knowledge representation systems and their computational
      complexity. If, for example, a knowledge representation
      system is as expressive as first order predicate calculus,
      then the problem of deciding what an agent could logically
      deduce from its knowledge base is unsolvable"
    </blockquote>
    <p>
      Do we need in practice to decide what an agent could deduce
      from its logic base? No, not in general. The agent may have
      various kinds of reasoning engine, and in practice also
      various amounts of connectivity, storage space, access to
      indexes, and processing power which will determine what it
      will actually deduce. Knowing that a certain algorithm may be
      nondeterministic polynomial in the size of the entire Web may
      not be at all helpful, as even linear time would be quite
      impractical. Practical computability may be assured by
      topological properties of the web, or the existence of know
      shortcuts such as precompiled indexes and definitive
      exclusive lists.
    </p>
    <p>
      Keeping a language less powerful than first order predicate
      calculus is quite reasonable within an application, but not
      for the Web.
    </p>
    <h2>
      Decidability
    </h2>
    <p>
      A dream of logicians in the last century to find languages in
      which all sentences were either true or false, and provably
      so. This involved trying to restrict the language so as to
      avoid the possibility of (for example) self-contradictory
      statements which can not be categorized as a true or not
      true.
    </p>
    <p>
      On the Semantic Web, this looks like a very academic problem,
      when in fact one anyway operates with a mass of untrustworthy
      data at any point, and restricts what one uses to a limited
      subset of the web. Clearly one must not be able to derive a
      self-contradictory statement, but there is no harm in the
      language being powerful enough to express it. Indeed,
      endorsement systems must give us the power to say "that
      statement is false" and so loops which if believed prove
      self-contradictory will arise by accident or design. A
      typical response of a system which finds a self-contradictory
      statement might be similar to the response to finding a
      contradiction, for example, to cease to trust information
      from the same source (or public key).
    </p>
    <h3>
      Reflection: Quoting, Context, and/or Higher Order Logic
    </h3>
    <p>
      <em>@@hmm... better section heading? maybe just quoting, or
      contexts? one place where we really do seem to need more than
      HOL is <a href="#L736">induction</a>.</em>
    </p>
    <p>
      The fact that there is [Burris p___] "no good set of axioms
      and rules for higher order logic" is frustrating not only in
      that it stumps the desire to write common sense
      mathematically, but also because operations which seem
      natural for electronic commerce seem at first sight to demand
      higher order logic. There is also a fundamental niceness to
      having a system powerful enough to describe its own rules, of
      course, just as one expects to be able to write a compiler
      for a programming language in the same language <em>(@@need
      to study</em> <a href=
      "http://lists.w3.org/Archives/Public/www-archive/2002Apr/0057.html">
      <em>references from Hayes</em></a><em>, esp "Tarski's results
      on meta-descriptions (a consistent language can't be the same
      expressive power as its own metatheory), Montague's paradox
      (showing that even quite weak languages can't consistently
      describe their own semantics)"</em>. When Frege tried
      second-order logic, I understand, Russel showed that his
      logic was inconsistent. But can we make a language in which
      is consistent (you can't derive a contradiction from its
      axioms) and yet allows enough to for example:-
    </p>
    <ul>
      <li>Model human trust in a realistic way
      </li>
      <li>Write down the mapping from XML to RDF logic to allow a
      theorem to be proved from the raw XML (and similarly define
      the XML syntax in logic to allow a theorem to be proved from
      the byte stream), and using it;
      </li>
    </ul>
    <p>
      The sort of rule it is tempting to write is such as to allow
      the inference of an RDF triple from a message whose semantic
      content one can algebraically derive that triple.
    </p>
    <pre>
forall message,t, r, x, y (
  (signed(message,K)
    &amp; derivable(t, message)
    &amp; subject(t, x)
    &amp; predicate(t, r)
    &amp; object(t, y))
   -&gt; r(x,y)
)
</pre>
    <p>
      (where K is a specific constant public key, and t is a
      triple)
    </p>
    <p>
      This breaks the boundary between the premises which deal with
      the mechanics of the language and the conclusion which is
      about the subject-matter of the language. Do we really need
      to do this, or can we get by with several independent levels
      of machinery, letting one machine prepare a "believable"
      message stream and parse it into a graph, and then a second
      machine which shares no knowledge space with the first, do
      the reasoning on the result? To me this seems hopeless, as
      one will in practice want to direct the front end's search
      for new documents from the needs of the reasoning by the back
      end. But this is all hunch.
    </p>
    <p>
      Peregrin tries to categorize the needs for and problems with
      higher order logic (HOL) in [Peregrin]. His description of
      Henkinian Understanding of HOL in which predicates are are
      subclass of objects ("individuals") seems to describe my
      current understanding of the mapping of RDF into logic, with
      RDF predicates, binary relations, being subclass of RDF
      nodes. Certainly in RDF the "property" type can be deduced
      from the use of any URI as a predicate:
    </p>
    <p>
      forall p,x,y p(x,y) -&gt; type(p, property)
    </p>
    <p>
      and we assume that the "assert" predicate
      &lt;rdf:property&gt; is equivalent to the predicate itself.
    </p>
    <p>
      forall p,x,y assert(p,x,y) &lt;--&gt; p(x,y)
    </p>
    <p>
      so we are moving between second-order formulation and
      first-order formulation.
    </p>
    <p>
      (2000) The experience of the [<a href="#PCA">PCA</a>] work
      seems to demonstrate that higher order logic is a very
      realistic way of unifying these systems.
    </p>
    <p>
      (2001) The treatment of contexts in [<a href="#CLA">CLA</a>]
      seems consistent with the design we've implemented.
    </p>
    <h2>
      <a name="L736">Induction, primitive recursion, and
      generalizing to infinitely many cases</a>
    </h2>
    <p>
      It seems clear that FOL is insufficient in that some sort of
      induction seems necessary.
    </p>
    <blockquote>
      <p>
        I agree with Tait (Finitism, J. of Philosophy, 1981, 78,
        524-546) that PRA is THE NECESSARY AND SUFFICIENT logic for
        talking about logics and proofs
      </p>
      <p>
        <a href=
        "http://theory.stanford.edu/people/uribe/mail/qed.messages/22.html">
        Robert S. Boyer, 18 Apr 93</a>
      </p>
    </blockquote>
    <p>
      also: <a href="/2001/03swell/pra.n3">pra.n3</a>, an N3
      transcription of <a href=
      "http://www.earlham.edu/~peters/courses/logsys/recursiv.htm">Peter
      Suber, Recursive Function Theory</a>
    </p>
    <p>
      also: ACL2: <a href=
      "http://www.cs.utexas.edu/users/moore/publications/km97a.ps.gz">
      A Precise Description of the ACL2 Logic</a> Kaufmann and
      <a href="http://www.cs.utexas.edu/users/moore/">Moore</a> 22
      Apr 1998, <a href=
      "http://rdfig.xmlhack.com/2002/03/26/2002-03-26.html#1017177958.271019">
      rdf scratchpad entry 26Mar</a>
    </p>
    <p>
      (for another sort of induction, i.e. as opposed to deduction,
      see: <a href=
      "http://www-formal.stanford.edu/jmc/circumscription.html">Circumscription</a>
      by McCarthy, 1980.)
    </p>
    <hr />
    <h2>
      Yet to be addressed (1999)
    </h2>
    <p>
      I personally need to understand what the limitations are on
      higher order logics...and which are hard limitations and
      which are unresolved areas.
    </p>
    <p>
      Was Frege's formulation of second order logic (which should
      be of course a formulation of common sense) buggy in that
      Russel found it could derive a contradiction, or are we
      really finding it is not possible to represent a mathematical
      system to follow common sense reasoning? Yes, Frege seems to
      have used the classical logic in which any proposition must
      be true or false.
    </p>
    <p>
      When we say that there are valid sentences which cannot be
      derived - what exactly do we mean by "valid"? In predicate
      logic, validity can be defined by a truth table, ie
      evaluation for all truth evaluations of the variables
      [Burris]. In fact this method equates to a resolution by
      mapping algebraically into disjoint (say) normal form, and so
      is not an operation "outside" the language. In any logic with
      unbounded sets this is not practical. Typically we fall back
      on some logic such as common sense outside the language under
      discussion. In higher order logic, intuition suggests we
      should be able to to construct such a proof of validity in
      the logic itself. Goedel's incompleteness theorem
      specifically addresses the difference between validity and
      derivability, so perhaps a good exposition of that [DanC
      recommends The Unknownable@@]would contain the necessary
      distinctions.
    </p>
    <p>
      I wonder whether the answers can be found on the web...
    </p>
    <p>
      After a hallway chat with Peter SP-S, DMAL meeting 2001/2/14:
    </p>
    <p>
      The paradox problem lies not simply in being able to express
      a paradox, but the logic being such that merely considering a
      paradox leads one to be able to deduce a falsehood. One can
      take issue then, with either the ability to express the
      paradox, or with the "p or not p" axiom. Logics which try to
      get above first order logic can be divided into two groups in
      this way. There are sets of logics which use sets, but limit
      them in some way - for example, by requiring sets to be "well
      formed", meaning they are either empty or contain at least
      one element which is disjoint with the set itself. These
      tricks limit the logic so that you can't in fact have the
      Russel paradox set (of all sets which are not members of
      themselves) as you exclude such things as not well-formed. As
      Peter admitted, on the web it seems silly to use this route
      when people will be expressing paradoxes.
    </p>
    <p>
      The other route alsob has many followers. Some of them are,
      people say, complicated. But it seems that the path can only
      be in that direction. Meanwhile, back at the ranch, I notice
      that most of the semantic web applications I have been
      playing with involve rules and axioms which are not at all
      complete. Each application has its own sets of axioms and can
      be individually proved consistent (or not). So one way
      forward for standards would be to instantiate that, allowing
      each document and message on the semantic web to have a
      pointer to the vocabulary its uses, including its varieties
      of logical connectives and their assocaited axioms.
    </p>
    <p>
      @@@
    </p>
    <p>
      <a href="http://www.altavista.com/" add_date="910926204"
      last_visit="917062905" last_modified="910926194"></a>
    </p>
    <p>
      Mapping RDB and SM models - defining URIs; nulls; etc.
    </p>
    <h2>
      <a name="reasoning">On reasoning from contradictions and the
      excluded middle</a>
    </h2>
    <p>
      <em>@@prose</em>
    </p>
    <p>
      Pat Hayes mentioned a logic in which the law of the excluded
      middle does not exist - which si important as you can always
      considera paradox which is neither true nor false. See email
      2000/09/01
    </p>
    <blockquote>
      <p>
        I think that a more suitable 'basic' logic might be
        generalized horn clause logic. This is similar to the
        horn-clause subset of FOL which is used in logic
        programming, but instead of full negation (which is
        computationally expensive) or negation-by-failure (which is
        cheap but grubbily nonmonotonic), it uses an elegant
        intermediate idea called intuitionistic (or constructive)
        negation (which is like full negation but rules out proofs
        by contradiction: basically, it insists on a higher
        standard of provability than classical negation. In
        intuitionistic reasoning, Holmes' famous advice about
        eliminating the impossible does not always apply.) It is
        computationally tractable, completely monotonic, and has a
        neat, elegant semantic theory behind it. It escapes some of
        the limitations of Horn logic while retaining a lot of its
        advantages. Its been implemented and seems to have been
        applied to some neat applications. It might be just what
        you need. For details see <a href=
        "http://citeseer.nj.nec.com/hogan98putting.html">the
        papers</a> at
        http://citeseer.nj.nec.com/hogan98putting.htmlhttp://citeseer.nj.nec.com/hogan98putting.html,
        especially <a href=
        "http://citeseer.nj.nec.com/158146.html">the 1996
        "meta-programming...."</a> and the <a href=
        "http://citeseer.nj.nec.com/hogan98putting.html">main
        citation at the top</a>. For the raw logic follow the
        McCarty 88 references.
      </p>
    </blockquote>
    <p>
      quoted without permission. DanC recommends <a href=
      "http://www.earlham.edu/~peters/courses/logsys/pnc-pem.htm">Non-Contradiction
      and Excluded Middle</a> in Peter Suber's <a href=
      "http://www.earlham.edu/~peters/courses/logsys/lshome.htm">Logical
      Systems</a> course notes as an explanation of intuitionistic
      logic. TimBL noted the same bit that I did on first reading:
    </p>
    <blockquote>
      <p>
        In the world of metamathematics, the intuitionists are not
        at all exotic, despite the centrality of the PEDC (hence
        the PEM) to the ordinary sense of consistency. Their
        opponents do not scorn them as irrationalists but, if
        anything, pity them for the scruples that do not permit
        them to enjoy some "perfectly good" mathematics.
      </p>
    </blockquote>
    <p>
      <em>is socratic completeness, as in [</em><a href=
      "#Crawf90"><em>Crawf90</em></a><em>], directly relevant?
      hmmm@@</em>
    </p>
    <p>
      solutions to paradoxes around wtr in KIF: <a href=
      "http://meta2.stanford.edu/kif/Hypertext/node35.html#SECTION00093000000000000000">
      complex one in KIFv3</a>, <a href=
      "http://logic.stanford.edu/kif/dpans.html#10.3">simpler, more
      limited one in later KIF spec</a>. (<a href=
      "/2000/07/hs78/KIF.html">KIF/RDF stuff</a> from Jul 2000)
    </p>
    <hr />
    <h2>
      References
    </h2>
    <p>
      [<a name="PCA">PCA</a>] <cite>Proof-Carrying
      Authentication</cite>. Andrew W. Appel and Edward W. Felten,
      6th ACM Conference on Computer and Communications Security,
      November 1999. (<a href=
      "http://www.cs.princeton.edu/sip/projects/pca/">background</a>)
      <a href=
      "http://lists.w3.org/Archives/Public/sw99/2000JanMar/0005.html">
      Mar 2000 discovery</a>. based on [<a href="#LF">LF</a>]
    </p>
    <p>
      [<a name="Burris">Burris</a>] Stanley N. Burris, Logic for
      Mathematics and Computer Science, Prentice Hall 1998.
    </p>
    <p>
      [<a name="BurrisSup">BurrisSup</a>]<a href=
      "http://thoralf2.uwaterloo.ca/htdocs/stext.html">Supplementary
      text to the above</a>.
    </p>
    <p>
      [<a name="Cheng">Cheng</a>] The ERM model paper, available in
      [<a href="#Laplante">Laplante</a>]
    </p>
    <p>
      [<a name="Connolly">ConnollySoLREK</a>] <a href=
      "/Collaboration/knowledge">Dan Connolly's home page for this
      sort of stuff. "A Study of Linguistics: Representation and
      Exchange of Knowledge"</a>.
    </p>
    <p>
      [<a name="Crawf90">Crawf90</a>]
    </p>
    <blockquote>
      <a href=
      "ftp://ftp.cs.utexas.edu/pub/qsim/papers/Crawford+Kuipers-sowa-91.ps.Z">
      ALL: Formalizing Access Limited Reasoning</a><br />
      J. M. Crawford jc@cs.utexas.edu<br />
      Benjamin Kuipers kuipers@cs.utexas.edu<br />
      Department of Computer Sciences<br />
      The University of Texas At Austin<br />
      Austin, Texas 78712<br />
      25 May 1990
      <p>
        Abstract
      </p>
      <p>
        Access-Limited Logic (ALL) is a language for knowledge
        representation which formalizes the access limitations
        inherent in a network-structured knowledge base. Where a
        classical deductive method or logic programming language
        would retrieve all assertions that satisfy a given pattern,
        an access-limited logic retrieves all assertions reachable
        by following an available access path. The complexity of
        inference is thus independent of the size of the
        knowledge-base and depends only on its local connectivity.
        Access-Limited Logic, though incomplete, still has a well
        defined semantics and a weakened form of completeness,
        <em>Socratic Completeness</em>, which guarantees that for
        any query which is a logical consequence of the
        knowledge-base, there exists a series of queries after
        which the original query will succeed. This chapter
        presents an overview of ALL, and sketches the proofs of its
        Socratic Completeness and polynomial time complexity.
      </p>
    </blockquote>
    <p>
      <a href=
      "http://www.cs.utexas.edu/users/qr/algernon.html#references">algernon
      papers</a>
    </p>
    <p>
      [<a name="Crawf91">Crawf91</a>] J. M. Crawford and B.J.
      Kuipers, "Negation and proof by Contradiction in Access
      Limited Logic", in <em>Proceedings of the Ninth National
      Conference on Artificial Intelligence (AAA1-91)</em>,
      Cambridge MA, 1991
    </p>
    <p>
      [<a name="Date">Date</a>] An Introduction to Database
      Systems, 6th Edn, Adison-Wesley, 1995
    </p>
    <p>
      [Davis] Randall Davis and Howard Schrobe, "<a href=
      "http://www.ai.mit.edu/courses/6.871/">6.871: Knowledge Based
      Application Systems</a>" course supporting information, 1999
    </p>
    <p>
      [<a name="CLA">CLA</a>] <a href=
      "http://www-formal.stanford.edu/guha/">Contexts: A
      Formalization and Some Applications</a>, Ramanathan V. Guha,
      1991 Stanford PhD thesis. see also: <a href=
      "/XML/9711theory/ContextLogic.lsl">ContextLogic.lsl</a>, May
      2001 transcription into larch.
    </p>
    <p>
      [<a name="HOLintro">HOLintro</a>]M. J. C. Gordon and T. F.
      Melham, "Introduction to HOL A theorem proving environment
      for higher order logic", Cambridge University Press, 1993
      ISBN 0 521 44189 7
    </p>
    <p>
      [<a name="Laplante">Laplante</a>] Phillip Laplante (Ed.),
      "Great Papers in Computer Science", West, 1996,
      <small>ISBN:0-314-06365-X</small>
    </p>
    <p>
      [<a name="Marchiori9">Marchiori98</a>] M. Marchiori, Metalog
      paper at QL98
    </p>
    <p>
      [<a name="Peregrin">Peregrin</a>] Jaroslav Peregrin,
      "<a href="http://dec59.ruk.cuni.cz/~peregrin/HTMLTxt/hol.htm">What
      Does One Need, When She Needs "Higher-Order Logic?</a>"
      <em>Proceedings of LOGICA'96, FILOSOFIA, Praha,</em> 1997,
      75-92
    </p>
    <p>
      [<a name="VLFM">VLFM</a>] J. P. Bowen, <a href=
      "http://www.comlab.ox.ac.uk/archive/formal-methods.html">Formal
      Methods, in WWW Virtual Library</a>
    </p>
    <p>
      @@
    </p>
    <p>
      Much of this is following in the wake of Dan Connolly's
      investigations, and so many of the references were provided
      directly or indictly by Dan.Other pointers from Dan Connolly
    </p>
    <ul>
      <li>DC on the need for HOL - <a href=
      "http://lists.w3.org/Archives/Team/w3t-tech/1998OctDec/0314.html">
        private communication</a>
      </li>
    </ul>
    <p>
      [<a name="LF">LF</a>] Harper, Honsell, Plotkin "<a href=
      "http://www.dcs.ed.ac.uk/lfcsreps/91/ECS-LFCS-91-162/">A
      Framework for Defining Logics</a>", Journal of the ACM,
      January 1993 appears to be the seminal paper on ELF paper
      (<a href=
      "http://www.acm.org/pubs/articles/journals/jacm/1993-40-1/p143-harper/p143-harper.pdf">ACM
      pdf</a>)<br />
      <a href="/XML/9711theory/ELF">connolly's (attempted)
      transcription into larch</a>
    </p>
    <p>
      <a href=
      "http://www.coginst.uwf.edu/~phayes/research.html">Pat Hayes,
      Research Interests and Selected Papers</a> contains stuff on
      time modelsImpl
    </p>
    <p>
      <a name="Reiter">Reiter</a>, R., On Closed World Data Bases,
      in: H. Gallaire and J Minker (eds.), <cite>Logic and Data
      Bases</cite>, Plenum, 1978, pp. 55-76. Defines Cloased World
      Assumption? Ref from CIL
    </p>
    <p>
      <a name="Perlis85">Perlis</a>, D., Languages with
      Self-Reference I: Foundations. (or: We Can Have Everything in
      First-Other Logic!). <cite>Artificial Intelligence</cite>,
      25:301-322, 1985.
    </p>
    <h2>
      Footnotes
    </h2>
    <p>
      "Second-order logic is rather difficult to study, since it
      lacks a decent set of axioms and rules of inference, so we
      have little to say about it" -- Footnote to preface to
      [Burris], p.xii
    </p>
    <p>
      @@
    </p>
    <hr />
    <p>
      <a href="Overview.html">Up to Design Issues</a>
    </p>
    <p>
      <a href="../People/Berners-Lee">Tim BL</a>
    </p>
  </body>
</html>