Logic.html
31.8 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
<html xmlns="http://www.w3.org/1999/xhtml">
<head>
<meta name="generator" content=
"HTML Tidy for Mac OS X (vers 31 October 2006 - Apple Inc. build 13), see www.w3.org" />
<title>
Logic on the Web -- Web architecture
</title>
<link rel="Stylesheet" href="di.css" type="text/css" />
<meta http-equiv="Content-Type" content=
"text/html; charset=us-ascii" />
</head>
<body bgcolor="#DDFFDD" text="#000000" lang="en">
<address>
Tim Berners-Lee<br />
Date: 1998, last change: $Date: 2009/08/27 21:38:07 $
$Author: timbl $<br />
Status: personal view only. Editing status: first draft.
</address>
<p>
<a href="./">Up to Design Issues</a>
</p>
<hr />
<h1>
The Semantic Web as a language of logic
</h1>
<p>
This looks at the Semantic Web design in the light a little
reading on formal logic, of the Access Limited Logic system,
in particular, and in the light of logical languages in
general. A problem here is a that I am no logician, and so I
am am having to step like a fascinated reporter into this
world of which I do not possess intimate experience.
</p>
<h2>
Introduction
</h2>
<p>
The <a href="Toolbox.html">Semantic Web Toolbox</a> discusses
the step from the web as being a repository of flat data
without logic to a level at which it is possible to express
logic. This is something which knowledge representation
systems have been wary of.
</p>
<p>
The Semantic Web has a different set of goals from most
systems of logic. As Crawford and Kuipers put it in [<a href=
"#Crawf90">Crawf90</a>],
</p>
<blockquote>
[...]a knowledge representation system must have the
following properties:
<ol>
<li>It must have a reasonably compact syntax.
</li>
<li>It must have a well defined semantics so that one can
say precisely what is being represented.
</li>
<li>It must have sufficient expressive power to represent
human knowledge.
</li>
<li>It must have an efficient, powerful, and understandable
reasoning mechanism
</li>
<li>It must be usable to build large knowledge bases.
</li>
</ol>
<p>
It has proved difficult, however, to achieve the third and
fourth properties simultaneously.
</p>
</blockquote>
<p>
The semantic web goal is to be a unifying system which will
(like the web for human communication) be as un-restraining
as possible so that the complexity of reality can be
described. Therefore item 3 becomes essential. This can be
achieved by dropping 4 - or the parts of item 4 which
conflict with 3, notably a single, efficient reasoning
system. The idea is that, within the global semantic web,
there will be a subset of the system which will be
constrained in specific ways so as to achieve the
tractability and efficiency which is no necessary in real
applications. However, the semantic web itself will not
define a reasoning engine. It will define valid operations,
and will require consistency for them. On the semantic web in
general, a party must be able to follow a proof of a theorem
but is not expected to generate one.
</p>
<p>
(This fundamental change goals from KR systems to the
semantic web is loosely analogous with the goal change from
conventional hypertext systems to the original Web design
dropping link consistency in favor of expressive flexibility
and scalability.The latter did not prevent individual web
sites from having a strict hierarchical order or matrix
structure, but it did not require it of the web as a whole.)
</p>
<p>
If there is a <em>semantic web machine</em>, then it is a
proof validator, not a theorem prover. It can't find answers,
it can't even check that an answer is right, but it can
follow a simple explanation that an answer is right. The
Semantic Web as a source of data should be fodder for
automated reasoning systems of many kinds, but it as such not
a reasoning system.
</p>
<p>
Most knowledge representation systems distinguish between
inference "rules" and other believed information. In some
cases, this is because the rules (such as substitution in a
formula) cannot be written in the language - they are defined
outside the language. In fact the set of rules used by the
system is often not only formally quite redundant but
arbitrary. However, a universal design such as the Semantic
Web must be minimalist. We will ask all logical data on the
web to be expressed directly or indirectly in terms of the
semantic web - a strong demand - so we cannot constrain
applications any further. Different machines which use data
from the web will use different algorithms, different sets of
inference rules. In some cases these will be powerful AI
systems and in others they will be simple document conversion
systems. The essential this is that the results of either
must be provably correct against the same basic minimalist
rules. In fact for interchange of proof, the set of rules is
an engineering choice.
</p>
<p>
There are many related ways in which subsystems can be
created
</p>
<ul>
<li>The semantic web language can be subsetted, by the
removal of operations and axioms and rules;
</li>
<li>The set of statements may be limited to that from
particular documents or web sites;
</li>
<li>The form of formulas used may be constrained, for example
using document schemata;
</li>
<li>Application design decisions can be made so as to
specifically guarantee tractable results using common
reasoning engines.
</li>
<li>Proofs can be constructed by completely hand-built
application-specific programs
</li>
</ul>
<p>
For example, Access Limited Logic is restricted (as I
understand it) to relations r(a,b) available when r is
accessed, and uses inference rules which only chain forward
along such links. There is also a "partitioning" of the Web
by making partitioning the rules in order to limit
complexity.
</p>
<p>
For the semantic web as a whole, then, we do require
tractable
</p>
<ul>
<li>Consistency, that it must not be possible to deduce a
contradiction (without having been given one)
</li>
<li>Strength in that all applications must be subsets
</li>
</ul>
<h3>
Grounding in Reality
</h3>
<p>
Philosophically, the semantic web produces more than a set of
rules for manipulation of formulae. It defines documents on
the Web has having a socially significant meaning. Therefore
it is not simply sufficient to demonstrate that one can
constrain the semantic web so as to make it isomorphic to a
particular algebra of a given system, but one must ensure
that a particular mapping is defined so that the web
representation of that particular system conveys is semantics
in a way that it can meaningfully be combined with the rest
of the semantic web. Electronic commerce needs a solid
foundation in this way, and the development of the semantic
web is (in 1999) essential to provide a rigid framework in
which to define electronic commerce terms, before electronic
commerce expands as a mass of vaguely defined semantics and
ad hoc syntax which leaves no room for automatic treatment,
and in which the court of law rather than a logical
derivation settles arguments.
</p>
<p>
Practically, the meaning of semantic web data is grounded in
non-semantic-web applications which are interfaced to the
semantic web. For example, currency transfer or ecommerce
applications, which accept semantic web input, define for
practical purposes what the terms in the currency transfer
instrument mean.
</p>
<h2>
Axiomatic Basis
</h2>
<p>
<em>@@I [DanC] think this section is outdated by recent
thoughts [2002] on</em> <a href="#reasoning"><em>paradox and
the excluded middle</em></a>
</p>
<p>
To the level of first order logic, we don't really need to
pick one set of axioms in that there are equivalent choices
which lead to demonstrably the same results.
</p>
<p>
(A cute one at the propositional logic level seems [Burris,
p126] to be Nicod's set in which nand (in XML toolbox
<not>..</not> and below [xy]) is the Sheffer
(sole) connective and the only rules of inference are
substitution and the <em>modus ponens</em> equivelent that
from F and [F[G H]] one can deduce H, and the single axiom
[[P[QR]][[S[SS]][[UQ][[PU][PU]]]].)
</p>
<p>
Let us assume the properties of first order logic here.
</p>
<p>
If we add anything else we have to be careful that it should
either be definable in terms of the first order set or that
the resulting language is a subset of a well proven logical
system -- or else we have a lot of work to do in establishing
a new system!
</p>
<h2>
Intractability and Undecidability
</h2>
<p>
These are two goals to which we explicitly do not aspire in
the Semantic Web in order to get in return expressive power.
(We still require consistency!). The world is full of
undecidable statements, and intractable problems. The
semantic web has to give the power to express such things.
</p>
<p>
Crawford and Kuipers The same explain in the introduction
their Negation in ALL paper,
</p>
<blockquote>
"Experience with formally specified knowledge representation
systems has revealed a trade-off between the expressive power
of knowledge representation systems and their computational
complexity. If, for example, a knowledge representation
system is as expressive as first order predicate calculus,
then the problem of deciding what an agent could logically
deduce from its knowledge base is unsolvable"
</blockquote>
<p>
Do we need in practice to decide what an agent could deduce
from its logic base? No, not in general. The agent may have
various kinds of reasoning engine, and in practice also
various amounts of connectivity, storage space, access to
indexes, and processing power which will determine what it
will actually deduce. Knowing that a certain algorithm may be
nondeterministic polynomial in the size of the entire Web may
not be at all helpful, as even linear time would be quite
impractical. Practical computability may be assured by
topological properties of the web, or the existence of know
shortcuts such as precompiled indexes and definitive
exclusive lists.
</p>
<p>
Keeping a language less powerful than first order predicate
calculus is quite reasonable within an application, but not
for the Web.
</p>
<h2>
Decidability
</h2>
<p>
A dream of logicians in the last century to find languages in
which all sentences were either true or false, and provably
so. This involved trying to restrict the language so as to
avoid the possibility of (for example) self-contradictory
statements which can not be categorized as a true or not
true.
</p>
<p>
On the Semantic Web, this looks like a very academic problem,
when in fact one anyway operates with a mass of untrustworthy
data at any point, and restricts what one uses to a limited
subset of the web. Clearly one must not be able to derive a
self-contradictory statement, but there is no harm in the
language being powerful enough to express it. Indeed,
endorsement systems must give us the power to say "that
statement is false" and so loops which if believed prove
self-contradictory will arise by accident or design. A
typical response of a system which finds a self-contradictory
statement might be similar to the response to finding a
contradiction, for example, to cease to trust information
from the same source (or public key).
</p>
<h3>
Reflection: Quoting, Context, and/or Higher Order Logic
</h3>
<p>
<em>@@hmm... better section heading? maybe just quoting, or
contexts? one place where we really do seem to need more than
HOL is <a href="#L736">induction</a>.</em>
</p>
<p>
The fact that there is [Burris p___] "no good set of axioms
and rules for higher order logic" is frustrating not only in
that it stumps the desire to write common sense
mathematically, but also because operations which seem
natural for electronic commerce seem at first sight to demand
higher order logic. There is also a fundamental niceness to
having a system powerful enough to describe its own rules, of
course, just as one expects to be able to write a compiler
for a programming language in the same language <em>(@@need
to study</em> <a href=
"http://lists.w3.org/Archives/Public/www-archive/2002Apr/0057.html">
<em>references from Hayes</em></a><em>, esp "Tarski's results
on meta-descriptions (a consistent language can't be the same
expressive power as its own metatheory), Montague's paradox
(showing that even quite weak languages can't consistently
describe their own semantics)"</em>. When Frege tried
second-order logic, I understand, Russel showed that his
logic was inconsistent. But can we make a language in which
is consistent (you can't derive a contradiction from its
axioms) and yet allows enough to for example:-
</p>
<ul>
<li>Model human trust in a realistic way
</li>
<li>Write down the mapping from XML to RDF logic to allow a
theorem to be proved from the raw XML (and similarly define
the XML syntax in logic to allow a theorem to be proved from
the byte stream), and using it;
</li>
</ul>
<p>
The sort of rule it is tempting to write is such as to allow
the inference of an RDF triple from a message whose semantic
content one can algebraically derive that triple.
</p>
<pre>
forall message,t, r, x, y (
(signed(message,K)
& derivable(t, message)
& subject(t, x)
& predicate(t, r)
& object(t, y))
-> r(x,y)
)
</pre>
<p>
(where K is a specific constant public key, and t is a
triple)
</p>
<p>
This breaks the boundary between the premises which deal with
the mechanics of the language and the conclusion which is
about the subject-matter of the language. Do we really need
to do this, or can we get by with several independent levels
of machinery, letting one machine prepare a "believable"
message stream and parse it into a graph, and then a second
machine which shares no knowledge space with the first, do
the reasoning on the result? To me this seems hopeless, as
one will in practice want to direct the front end's search
for new documents from the needs of the reasoning by the back
end. But this is all hunch.
</p>
<p>
Peregrin tries to categorize the needs for and problems with
higher order logic (HOL) in [Peregrin]. His description of
Henkinian Understanding of HOL in which predicates are are
subclass of objects ("individuals") seems to describe my
current understanding of the mapping of RDF into logic, with
RDF predicates, binary relations, being subclass of RDF
nodes. Certainly in RDF the "property" type can be deduced
from the use of any URI as a predicate:
</p>
<p>
forall p,x,y p(x,y) -> type(p, property)
</p>
<p>
and we assume that the "assert" predicate
<rdf:property> is equivalent to the predicate itself.
</p>
<p>
forall p,x,y assert(p,x,y) <--> p(x,y)
</p>
<p>
so we are moving between second-order formulation and
first-order formulation.
</p>
<p>
(2000) The experience of the [<a href="#PCA">PCA</a>] work
seems to demonstrate that higher order logic is a very
realistic way of unifying these systems.
</p>
<p>
(2001) The treatment of contexts in [<a href="#CLA">CLA</a>]
seems consistent with the design we've implemented.
</p>
<h2>
<a name="L736">Induction, primitive recursion, and
generalizing to infinitely many cases</a>
</h2>
<p>
It seems clear that FOL is insufficient in that some sort of
induction seems necessary.
</p>
<blockquote>
<p>
I agree with Tait (Finitism, J. of Philosophy, 1981, 78,
524-546) that PRA is THE NECESSARY AND SUFFICIENT logic for
talking about logics and proofs
</p>
<p>
<a href=
"http://theory.stanford.edu/people/uribe/mail/qed.messages/22.html">
Robert S. Boyer, 18 Apr 93</a>
</p>
</blockquote>
<p>
also: <a href="/2001/03swell/pra.n3">pra.n3</a>, an N3
transcription of <a href=
"http://www.earlham.edu/~peters/courses/logsys/recursiv.htm">Peter
Suber, Recursive Function Theory</a>
</p>
<p>
also: ACL2: <a href=
"http://www.cs.utexas.edu/users/moore/publications/km97a.ps.gz">
A Precise Description of the ACL2 Logic</a> Kaufmann and
<a href="http://www.cs.utexas.edu/users/moore/">Moore</a> 22
Apr 1998, <a href=
"http://rdfig.xmlhack.com/2002/03/26/2002-03-26.html#1017177958.271019">
rdf scratchpad entry 26Mar</a>
</p>
<p>
(for another sort of induction, i.e. as opposed to deduction,
see: <a href=
"http://www-formal.stanford.edu/jmc/circumscription.html">Circumscription</a>
by McCarthy, 1980.)
</p>
<hr />
<h2>
Yet to be addressed (1999)
</h2>
<p>
I personally need to understand what the limitations are on
higher order logics...and which are hard limitations and
which are unresolved areas.
</p>
<p>
Was Frege's formulation of second order logic (which should
be of course a formulation of common sense) buggy in that
Russel found it could derive a contradiction, or are we
really finding it is not possible to represent a mathematical
system to follow common sense reasoning? Yes, Frege seems to
have used the classical logic in which any proposition must
be true or false.
</p>
<p>
When we say that there are valid sentences which cannot be
derived - what exactly do we mean by "valid"? In predicate
logic, validity can be defined by a truth table, ie
evaluation for all truth evaluations of the variables
[Burris]. In fact this method equates to a resolution by
mapping algebraically into disjoint (say) normal form, and so
is not an operation "outside" the language. In any logic with
unbounded sets this is not practical. Typically we fall back
on some logic such as common sense outside the language under
discussion. In higher order logic, intuition suggests we
should be able to to construct such a proof of validity in
the logic itself. Goedel's incompleteness theorem
specifically addresses the difference between validity and
derivability, so perhaps a good exposition of that [DanC
recommends The Unknownable@@]would contain the necessary
distinctions.
</p>
<p>
I wonder whether the answers can be found on the web...
</p>
<p>
After a hallway chat with Peter SP-S, DMAL meeting 2001/2/14:
</p>
<p>
The paradox problem lies not simply in being able to express
a paradox, but the logic being such that merely considering a
paradox leads one to be able to deduce a falsehood. One can
take issue then, with either the ability to express the
paradox, or with the "p or not p" axiom. Logics which try to
get above first order logic can be divided into two groups in
this way. There are sets of logics which use sets, but limit
them in some way - for example, by requiring sets to be "well
formed", meaning they are either empty or contain at least
one element which is disjoint with the set itself. These
tricks limit the logic so that you can't in fact have the
Russel paradox set (of all sets which are not members of
themselves) as you exclude such things as not well-formed. As
Peter admitted, on the web it seems silly to use this route
when people will be expressing paradoxes.
</p>
<p>
The other route alsob has many followers. Some of them are,
people say, complicated. But it seems that the path can only
be in that direction. Meanwhile, back at the ranch, I notice
that most of the semantic web applications I have been
playing with involve rules and axioms which are not at all
complete. Each application has its own sets of axioms and can
be individually proved consistent (or not). So one way
forward for standards would be to instantiate that, allowing
each document and message on the semantic web to have a
pointer to the vocabulary its uses, including its varieties
of logical connectives and their assocaited axioms.
</p>
<p>
@@@
</p>
<p>
<a href="http://www.altavista.com/" add_date="910926204"
last_visit="917062905" last_modified="910926194"></a>
</p>
<p>
Mapping RDB and SM models - defining URIs; nulls; etc.
</p>
<h2>
<a name="reasoning">On reasoning from contradictions and the
excluded middle</a>
</h2>
<p>
<em>@@prose</em>
</p>
<p>
Pat Hayes mentioned a logic in which the law of the excluded
middle does not exist - which si important as you can always
considera paradox which is neither true nor false. See email
2000/09/01
</p>
<blockquote>
<p>
I think that a more suitable 'basic' logic might be
generalized horn clause logic. This is similar to the
horn-clause subset of FOL which is used in logic
programming, but instead of full negation (which is
computationally expensive) or negation-by-failure (which is
cheap but grubbily nonmonotonic), it uses an elegant
intermediate idea called intuitionistic (or constructive)
negation (which is like full negation but rules out proofs
by contradiction: basically, it insists on a higher
standard of provability than classical negation. In
intuitionistic reasoning, Holmes' famous advice about
eliminating the impossible does not always apply.) It is
computationally tractable, completely monotonic, and has a
neat, elegant semantic theory behind it. It escapes some of
the limitations of Horn logic while retaining a lot of its
advantages. Its been implemented and seems to have been
applied to some neat applications. It might be just what
you need. For details see <a href=
"http://citeseer.nj.nec.com/hogan98putting.html">the
papers</a> at
http://citeseer.nj.nec.com/hogan98putting.htmlhttp://citeseer.nj.nec.com/hogan98putting.html,
especially <a href=
"http://citeseer.nj.nec.com/158146.html">the 1996
"meta-programming...."</a> and the <a href=
"http://citeseer.nj.nec.com/hogan98putting.html">main
citation at the top</a>. For the raw logic follow the
McCarty 88 references.
</p>
</blockquote>
<p>
quoted without permission. DanC recommends <a href=
"http://www.earlham.edu/~peters/courses/logsys/pnc-pem.htm">Non-Contradiction
and Excluded Middle</a> in Peter Suber's <a href=
"http://www.earlham.edu/~peters/courses/logsys/lshome.htm">Logical
Systems</a> course notes as an explanation of intuitionistic
logic. TimBL noted the same bit that I did on first reading:
</p>
<blockquote>
<p>
In the world of metamathematics, the intuitionists are not
at all exotic, despite the centrality of the PEDC (hence
the PEM) to the ordinary sense of consistency. Their
opponents do not scorn them as irrationalists but, if
anything, pity them for the scruples that do not permit
them to enjoy some "perfectly good" mathematics.
</p>
</blockquote>
<p>
<em>is socratic completeness, as in [</em><a href=
"#Crawf90"><em>Crawf90</em></a><em>], directly relevant?
hmmm@@</em>
</p>
<p>
solutions to paradoxes around wtr in KIF: <a href=
"http://meta2.stanford.edu/kif/Hypertext/node35.html#SECTION00093000000000000000">
complex one in KIFv3</a>, <a href=
"http://logic.stanford.edu/kif/dpans.html#10.3">simpler, more
limited one in later KIF spec</a>. (<a href=
"/2000/07/hs78/KIF.html">KIF/RDF stuff</a> from Jul 2000)
</p>
<hr />
<h2>
References
</h2>
<p>
[<a name="PCA">PCA</a>] <cite>Proof-Carrying
Authentication</cite>. Andrew W. Appel and Edward W. Felten,
6th ACM Conference on Computer and Communications Security,
November 1999. (<a href=
"http://www.cs.princeton.edu/sip/projects/pca/">background</a>)
<a href=
"http://lists.w3.org/Archives/Public/sw99/2000JanMar/0005.html">
Mar 2000 discovery</a>. based on [<a href="#LF">LF</a>]
</p>
<p>
[<a name="Burris">Burris</a>] Stanley N. Burris, Logic for
Mathematics and Computer Science, Prentice Hall 1998.
</p>
<p>
[<a name="BurrisSup">BurrisSup</a>]<a href=
"http://thoralf2.uwaterloo.ca/htdocs/stext.html">Supplementary
text to the above</a>.
</p>
<p>
[<a name="Cheng">Cheng</a>] The ERM model paper, available in
[<a href="#Laplante">Laplante</a>]
</p>
<p>
[<a name="Connolly">ConnollySoLREK</a>] <a href=
"/Collaboration/knowledge">Dan Connolly's home page for this
sort of stuff. "A Study of Linguistics: Representation and
Exchange of Knowledge"</a>.
</p>
<p>
[<a name="Crawf90">Crawf90</a>]
</p>
<blockquote>
<a href=
"ftp://ftp.cs.utexas.edu/pub/qsim/papers/Crawford+Kuipers-sowa-91.ps.Z">
ALL: Formalizing Access Limited Reasoning</a><br />
J. M. Crawford jc@cs.utexas.edu<br />
Benjamin Kuipers kuipers@cs.utexas.edu<br />
Department of Computer Sciences<br />
The University of Texas At Austin<br />
Austin, Texas 78712<br />
25 May 1990
<p>
Abstract
</p>
<p>
Access-Limited Logic (ALL) is a language for knowledge
representation which formalizes the access limitations
inherent in a network-structured knowledge base. Where a
classical deductive method or logic programming language
would retrieve all assertions that satisfy a given pattern,
an access-limited logic retrieves all assertions reachable
by following an available access path. The complexity of
inference is thus independent of the size of the
knowledge-base and depends only on its local connectivity.
Access-Limited Logic, though incomplete, still has a well
defined semantics and a weakened form of completeness,
<em>Socratic Completeness</em>, which guarantees that for
any query which is a logical consequence of the
knowledge-base, there exists a series of queries after
which the original query will succeed. This chapter
presents an overview of ALL, and sketches the proofs of its
Socratic Completeness and polynomial time complexity.
</p>
</blockquote>
<p>
<a href=
"http://www.cs.utexas.edu/users/qr/algernon.html#references">algernon
papers</a>
</p>
<p>
[<a name="Crawf91">Crawf91</a>] J. M. Crawford and B.J.
Kuipers, "Negation and proof by Contradiction in Access
Limited Logic", in <em>Proceedings of the Ninth National
Conference on Artificial Intelligence (AAA1-91)</em>,
Cambridge MA, 1991
</p>
<p>
[<a name="Date">Date</a>] An Introduction to Database
Systems, 6th Edn, Adison-Wesley, 1995
</p>
<p>
[Davis] Randall Davis and Howard Schrobe, "<a href=
"http://www.ai.mit.edu/courses/6.871/">6.871: Knowledge Based
Application Systems</a>" course supporting information, 1999
</p>
<p>
[<a name="CLA">CLA</a>] <a href=
"http://www-formal.stanford.edu/guha/">Contexts: A
Formalization and Some Applications</a>, Ramanathan V. Guha,
1991 Stanford PhD thesis. see also: <a href=
"/XML/9711theory/ContextLogic.lsl">ContextLogic.lsl</a>, May
2001 transcription into larch.
</p>
<p>
[<a name="HOLintro">HOLintro</a>]M. J. C. Gordon and T. F.
Melham, "Introduction to HOL A theorem proving environment
for higher order logic", Cambridge University Press, 1993
ISBN 0 521 44189 7
</p>
<p>
[<a name="Laplante">Laplante</a>] Phillip Laplante (Ed.),
"Great Papers in Computer Science", West, 1996,
<small>ISBN:0-314-06365-X</small>
</p>
<p>
[<a name="Marchiori9">Marchiori98</a>] M. Marchiori, Metalog
paper at QL98
</p>
<p>
[<a name="Peregrin">Peregrin</a>] Jaroslav Peregrin,
"<a href="http://dec59.ruk.cuni.cz/~peregrin/HTMLTxt/hol.htm">What
Does One Need, When She Needs "Higher-Order Logic?</a>"
<em>Proceedings of LOGICA'96, FILOSOFIA, Praha,</em> 1997,
75-92
</p>
<p>
[<a name="VLFM">VLFM</a>] J. P. Bowen, <a href=
"http://www.comlab.ox.ac.uk/archive/formal-methods.html">Formal
Methods, in WWW Virtual Library</a>
</p>
<p>
@@
</p>
<p>
Much of this is following in the wake of Dan Connolly's
investigations, and so many of the references were provided
directly or indictly by Dan.Other pointers from Dan Connolly
</p>
<ul>
<li>DC on the need for HOL - <a href=
"http://lists.w3.org/Archives/Team/w3t-tech/1998OctDec/0314.html">
private communication</a>
</li>
</ul>
<p>
[<a name="LF">LF</a>] Harper, Honsell, Plotkin "<a href=
"http://www.dcs.ed.ac.uk/lfcsreps/91/ECS-LFCS-91-162/">A
Framework for Defining Logics</a>", Journal of the ACM,
January 1993 appears to be the seminal paper on ELF paper
(<a href=
"http://www.acm.org/pubs/articles/journals/jacm/1993-40-1/p143-harper/p143-harper.pdf">ACM
pdf</a>)<br />
<a href="/XML/9711theory/ELF">connolly's (attempted)
transcription into larch</a>
</p>
<p>
<a href=
"http://www.coginst.uwf.edu/~phayes/research.html">Pat Hayes,
Research Interests and Selected Papers</a> contains stuff on
time modelsImpl
</p>
<p>
<a name="Reiter">Reiter</a>, R., On Closed World Data Bases,
in: H. Gallaire and J Minker (eds.), <cite>Logic and Data
Bases</cite>, Plenum, 1978, pp. 55-76. Defines Cloased World
Assumption? Ref from CIL
</p>
<p>
<a name="Perlis85">Perlis</a>, D., Languages with
Self-Reference I: Foundations. (or: We Can Have Everything in
First-Other Logic!). <cite>Artificial Intelligence</cite>,
25:301-322, 1985.
</p>
<h2>
Footnotes
</h2>
<p>
"Second-order logic is rather difficult to study, since it
lacks a decent set of axioms and rules of inference, so we
have little to say about it" -- Footnote to preface to
[Burris], p.xii
</p>
<p>
@@
</p>
<hr />
<p>
<a href="Overview.html">Up to Design Issues</a>
</p>
<p>
<a href="../People/Berners-Lee">Tim BL</a>
</p>
</body>
</html>