Lecture.html 27.5 KB
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359 360 361 362 363 364 365 366 367 368 369 370 371 372 373 374 375 376 377 378 379 380 381 382 383 384 385 386 387 388 389 390 391 392 393 394 395 396 397 398 399 400 401 402 403 404 405 406 407 408 409 410 411 412 413 414 415 416 417 418 419 420 421 422 423 424 425 426 427 428 429 430 431 432 433 434 435 436 437 438 439 440 441 442 443 444 445 446 447 448 449 450 451 452 453 454 455 456 457 458 459 460 461 462 463 464 465 466 467 468 469 470 471 472 473 474 475 476 477 478 479 480 481 482 483 484 485 486 487 488 489 490 491 492 493 494 495 496 497 498 499 500 501 502 503 504 505 506 507 508 509 510 511 512 513 514 515 516
<?xml version="1.0" encoding="iso-8859-1"?>
<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Transitional//EN"
    "http://www.w3.org/TR/xhtml1/DTD/xhtml1-transitional.dtd">
<html xmlns="http://www.w3.org/1999/xhtml">
<head>
  <title>Lecture - Japan prize 2002 Commemorative Lecture -
  Berners-Lee</title>
  <style type="text/css">

blockquote {background-color: #EFFFC1}</style>
  <link rel="stylesheet" href="/StyleSheets/base.css" />
  <!-- Changed by: tbl 19990524-->
</head>

<body xml:lang="en" lang="en">
TimBL 

<h2>Commemorative Lecture</h2>

<h1>The World Wide Web - Past Present and Future</h1>

<p>Exploring Universality</p>

<h3>Abstract</h3>

<blockquote>
  <p>The most important thing about the World Wide Web is that it is
  universal. By exploring this idea along its many axes we find a framework
  for considering its history, its role today, and guidance for future
  developments.</p>
</blockquote>

<p>Hardware independence, which once meant running on mainframes,
minicomputers and microcomputers, now extends to a multitude of devices from
watches and speech devices to big screen televisions. The separation of the
essential meaning of the information from the form in which it is conveyed
helps this independence, and also makes the Web accessible to people with
disabilities. Software independence, which is so important to prevent
fragmentation into many disconnected proprietary webs, is under as much
pressure as ever. That the Web must be independent of nation and location is
nowhere so clear as in an international gathering such as this, where
character set, language, and culture can be barriers which the technology
helps us to bridge. As we look forward, we are tempted to distinguish between
the multimedia world of information targeted for human perception, and the
well-defined world of data which machines handle. A Web which encompasses
both these extremes and all the rich land in between is the one which will
help us best fulfill our hopes for society, for understanding between
peoples, and finding a balance between the diversity and commonality in this
rich world.</p>

<h3>Introduction</h3>

<blockquote>
  <p>The concept of the Web integrated many disparate information systems, by
  forming an abstract imaginary space in which the differences between them
  did not exist. The Web had to include all information of any sort on any
  system. The only common idea needed to tie it all together was the
  <em>Universal Resource Identifier</em>(URI) identifying a document. From
  that cascaded a series of designs of protocols (such as HTTP) and data
  formats (such as HTML) which allowed computers to exchange information,
  mapping their own local formats into standards which provided global
  interoperability.</p>
</blockquote>

<p>Back in 1989, before the World Wide Web, many different information
systems existed. They ran on different sorts of computers, each running
different operating systems, connected by different networks, and using quite
different programs to give to the user very different ways of accessing
information. Thus, while the information on two systems might be very
relevant, the path between them was very long. And yet, in fact, each of the
computer systems was very likely to be connected to some sort of network. And
that network was very likely to be connected to another network, so that in
fact there was a path from bit of data on one computer through a series of
networks to the other computer. So there was, finally, no fundamental reason
why these barriers to communication should exist.</p>

<p>The first breakthrough was the Internet, and I can't emphasize too often
that I didn't invent the Internet! There were many networks, but they were of
different types, some small, some large, and they used different sorts of
connection. A computer could be on more than one network, and it was Vint
Cerf and his colleagues who realized that a computer connected to more than
one network could act as a kind of postal sorting office, and be used to
forward information between the networks. Even though the little networks
might use different numbering schemes for different computers, they imagined
that each computer was on some global "Inter-network" and gave each computer
a number. To describe things simply, the information is passed around in
little packets (rather, as Vint says, like postcards) and each has on it a
number that is the address of the computer to which it has to be delivered.
The forwarding computers just look at the address number on each packet to
figure out which network to send it over next. In this way, all you have to
do is send off a packet with the right address number on it, and sooner or
later it will arrive at the right place. The Internet was invented around the
1970s. I was fortunate in that in 1989, when I was looking at the problems of
networked information systems, it was deployed across the US and to a certain
extent in Europe.</p>

<p>The way the Web works is very simple. When you see a link in a Web page,
it might be underlined, or blue, but however the computer indicates there is
a link, that means that, in a special hidden code inside the document, there
is the URI of the document to which the link goes. What happens when you
click? You computer looks at that URI, and if (like most URIs at the moment)
it starts "http:", then it looks at the next bit, something like
"www.w3.org". That is the <em>domain name</em> of the publishing authority,
but what it needs is the number of a computer. Fortunately, a large number of
<em>domain name servers</em> exist, computers which collaborate to hold a
list of which domain name correspond to which computer address. Your Web
browser sends, to one of these domain name servers, a packet containing the
name of "www.w3.org", and received in return the address, "18.23.7.7". Your
browser then sends off a packet to that address asking for the URI. The
server responds by transmitting the new Web page back across the Internet to
the browser. The browser receives the document, decodes the HTML tags in it,
and displays a fresh Web page on your screen.</p>

<p>The Web required everyone to give a URI to their documents: a large
request. To attain its universality, the design of the Web could not impose
any extra constraint on how data was represented or organized. In fact, the
first Web-specific communications protocol (HTTP) and data format (HTML),
designed at the same time as the URI, were very successful and used for a
very large amount of the web. However, the Web still was designed to only
fundamentally rely on one specification: the Universal Resource
Identifier.</p>

<h3>Device independence</h3>

<blockquote>
  <p>That the same information should be accessible from many devices is a
  core rule of the Web. Once the choices were 80 character terminals or the
  new personal computers. Now, the number of pixels on a computer screen has
  steadily increased, but mobile devices have small screens or voice input
  and output. Our ability to represent information independently of the
  hardware we use is more important than ever.</p>
</blockquote>

<p>The direct impact of the Web was seen in its ability to cross hardware and
software boundaries. Before the web, at CERN, academic papers and
administrative data were kept on a mainframe computer, but much live
information and "help" information was available on minicomputers. Most
people had replaced the terminals in their offices with personal computers,
but still kept a window open logged onto the mainframe simply to access the
phone book. Unexciting though it was, access to the phone book from computers
of all sorts was an early incentive for browser adoption at CERN: for these
people it was the <em><span style="font-style: normal">critical
application</span></em><span style="font-style: normal"></span>which
convinced them.</p>

<p>A crucial factor in the design of the Web was the use of markup languages
which transmitted the intent of the markup instead of the actual form for
display. For example, tags for <em>heading level one</em> rather than
<em>centered bold big text</em> allowed the same information to be displayed
on color terminals with only one font, as well as black and white multifont
windows, or whatever was available. This concept, the <strong>separation of
form and content</strong>, is very important still for today's Web
designers.</p>

<p>Interestingly, the Web spread so fast that it was not apparent to many
designers how limiting it would be to make assumptions about which device a
user had. Many sites proclaimed that they were "best viewed using 800x600
pixel screens". A few years later, as typical screens increased to 1024x768
pixels but many users were still using old 640x480 screens, the mistake
became apparent.</p>

<p>More recently, the need for device independence has taken on new
dimensions as the long-promised dawn of practical speech recognition software
becomes a reality. Speech interaction breaks the user interface metaphor
assumption which graphic user interfaces introduced, the idea that the
computer and human share a view of a document. Speech interfaces bring us
back to the conversational style which in fact computers used in the old days
of the command line program. This change is more than just one of screen
size. When we try to generalize a user's interaction in a way that may
include mixtures all these modes, it causes significant rethinking, in which
the community is currently (2002) engaged.</p>

<h3>Software Independence</h3>

<blockquote>
  <p>Many different forms of software provide and consume Web information,
  and no one program was critical to the whole Web. This decentralization of
  software development was and always will be crucial to its unimpeded
  growth. It also prevents the Web itself from coming under the control of a
  given company or government through control of the software. Communication
  standards give people a choice of software, but we must all learn to be
  aware of when our experience is being controlled by software with a
  bias.</p>
</blockquote>

<p>The Web was deployed not as a program, but as a set of protocols.</p>

<p>The initial diagrams made it clear that those specifications -- URI, HTTP,
HTML, and others -- would form a sort of "bus" connecting the many different
sort of user programs ("clients") and many different sorts of information
provider programs ("servers").</p>

<p><img alt="Clients communicate with servers using a common connection bus."
src="../../Talks/04-sweb/Architecture.crop.gif" /></p>

<p>The initial client for the Web ran on a NeXT computer, at that time the
most sophisticated platform available. The second client was a simple
terminal-oriented command-line program for use on systems which didn't have a
graphic interface at all. Between the two they demonstrated the concept of
software independence.</p>

<p>The market situation around Web software has been though many phases, but
this issue has always been important, and still is today. Now that so much
money, and human attention - which is quickly turned into money, flows
through the web-human connection, control of any aspect of the interaction
with a human can be very lucrative.</p>

<p>Soon companies tried to find ways to influence and control the user's
choice of information. Computers came with free software, and software came
with built-in bias toward certain Web pages, and certain search engines.
Users thinking they are just "searching the Web" use a specific search engine
which points them to specific information, views, and products. Not only must
the technology support a choice of software, but a competitive market must
exist, and users must be informed and aware of what is going on.</p>

<h3>Internationalization</h3>

<blockquote>
  <p>From its beginning in a laboratory run by over a dozen collaborating
  countries, the Web had to be independent of any inherent bias toward one
  given country. XML, being firmly based on Unicode, now allows all kinds of
  characters. Internationalization must take into account much more: the
  direction in which text moves across the page, hyphenation conventions, and
  even cultural assumptions about the way people work and address each other,
  and the forms of organization they make.</p>
</blockquote>

<p></p>

<p>In 1994, in response to pressure for a body to coordinate development of
interoperable standards for the Web became intense. The World Wide Web
Consortium (W3C) was founded at the Laboratory for Computer Science (LCS) at
the Massachusetts Institute of Technology (MIT) to lead the technical
evolution of the Web and ensure its interoperability by developing common
protocols. A lot of effort is spent at the W3C to justify the first two W's
of "WWW'.</p>

<p>The first HTML documents were unfortunately (due to my ignorance of
<em>Unicode</em>) capable of representing only Western European languages.
Since then, the new version, XHTML is based on XML, which is based on the
<em>Unicode</em> standard.</p>

<p>Nowadays, the Consortium's Internationalization Working Group reviews new
technology to try to spot areas in which a national, linguistic, or cultural
bias may have crept into the design. We are very pleased to be hosted in
parallel by INRIA in France and Keio University in Japan, as well as the
Massachusetts Institute for Technology in the United States of America.</p>

<p>For all this work, the English language still tends to dominate the
Internet. From the technical point of view, the Internet had been installed
across the US when the Web started, but had not spread so much in other
countries. From the market point of view, the US provides a single-language
block which is a huge market for a new Web site, in contrast to Europe where
the need for translation into many languages is a hinderance to the explosive
uptake of a new site. I certainly hope that the Web will allow many cultures
and languages to flourish, and that we will not sink to that common subset of
expression which we can all understand.</p>

<h3>Multimedia</h3>

<blockquote>
  <p>Multimedia is not just a buzz-word, it stands for an important dimension
  of variety - the palette of technologies available to human creativity.
  Even the early demos of the web included sounds and music. What has changed
  since then is that the capacity of typical computers to handle graphics and
  sound has increased, and for some, the bandwidth even allows video to be
  sent. Because many things can still be done with plain text, the exotic and
  the mundane will always coexist on the Web.</p>
</blockquote>

<p>The first Web pages were displayed in a variety of fonts and formatting
options, but images, sounds and movies were separate documents linked from
the text page. Marc Andreessen's <em>Mosaic</em> browser led the way in
integrating images, and Pei Wei's <em>Viola</em> browser demonstrated the
power of scripting. Now, single Web pages can integrate text, photographs,
line drawings, and mathematical formulae. The image technology has advanced
with standards in <em>Scalable Vector Graphics</em> which allows a drawing to
be sent as an abstract collection of graphic objects, and rendered on arrival
into the appropriate style and resolution for each device, whether a large
computer or a small phone. This gives much better results than the use of
pixel graphics such as GIF and the later PNG. With the Synchronized
Multimedia Integration Language (SMIL) there is now a standard for how all
manner of multmedia things should be combined into a single experience.
Unfortunately, in streaming audio, standards are less clear.</p>

<p>It is still the case, as a decade ago, that bandwidth and processor power
limit what is practical, especially for video. But always, plain text, which
needs neither of these, is all one needs for poetry and for most electronic
commerce.</p>

<h3>Accessibility</h3>

<blockquote>
  <p>Just as people differ in the language, characters and cultures to which
  they are used, so they differ in terms of their capacities, for example, in
  vision, hearing, motor or cognition. The universality which we expect of
  the Web includes making sure that, as much as we can, we make the Web a
  place which people can use irrespective of disabilities. There are
  guidelines for Web site designers to help with this now, and a site which
  follows them will typically be easier for anyone to use, and easier to
  index and search.</p>
</blockquote>

<p>The separation of form and content, referred to above, is also a key to
making the Web accessible to those with disabilities. To communicate well, we
need not only to master each multimedia genre as effectively as we can, but
we must also allow people a choice of medium, as users get on more easily
with some than others. Soundtracks have subtitles, images have descriptions,
mouse movements have keyboard alternatives, and so on.</p>

<p>It turns out that this work overlaps very much with other areas.
Accessibility is enhanced when we have separated form and content, and the
forms of available include different media. It takes a bit of extra work, for
example in using the text of a to the video in the form of captions for the
visually impaired, and in making up textual explanations of the contents of
images but it is important and worth the extra effort.</p>

<h3>Rhyme and Reason</h3>

<blockquote>
  <p>There is another axis along which information varies. At one end of the
  axis is the poem, at the other the database table. The poem, or for that
  matter the 15 second TV commercial, is designed to connect to a human brain
  using all its complex series of associations in clever and powerful ways
  which we can never fully analyze. The database is designed to be queried
  and processed by a machine. It has well-defined values of information
  regularly arranged in columns which, hopefully, has well-defined meaning.
  Databases can be joined and split, combined and repurposed. Human beings
  use different sides of the brain for dealing with these types of
  information. Most information on the Web now contains both elements. The
  Web technology must allow information intended for a human to be
  effectively presented, and also allow machine processable data to be
  conveyed. Only then can we start to use computers as tools again.</p>
</blockquote>

<p>When I first was at CERN, the computing division was known as the DD, or
<em>Data and Documents</em> division. That name was later deemed outmoded,
and the usual phrases such as <em>Management Information Systems</em>,
<em>Computing and Networking</em>, and <em>Information Technology</em> were
used. However, the old name draws a useful distinction. One can think of
documents as information items, multimedia possibly, for people. Data, on the
other hand, is for machines; hard, well defined, the stuff of computation.</p>

<p>The elegance of the WWW browser as a computer application was that it
almost completely hid its workings from the user. The user never saw the HTML
and, in the first browser, never saw the URIs. The job of the machine is to
keep a low profile, to leave the user alone in an abstract space of
documents. And so it should be, as machines cannot really do much else in the
realm of documents. They cannot understand them, and therefore cannot work
with them.</p>

<p>In the realm of data, things are different. Numbers can be crunched. Rules
can be applied. Data can be sifted and correlated by machines very
effectively. This is what the late Michael Dertouzos, Director of the LCS at
MIT, described as the "heavy lifting" of information work. The analogy is
with building work, like when machines can shift the earth much better than
we can, though they add no creativity as to how to do it. The Web at the
moment lets us down in the area of data, because the data is not in a form
which machines can use. It isn't well identified in terms of the way it
should be combined. All a computer can do is to pretend to be a person
browsing the Web, and then guess what each Web page means!</p>

<p>The Semantic Web development adds to the Web formats for represneting data
and its <em>semantics</em> - the meaning for a machine in terms of what rules
can be applied and how it can be transformed into other data. This will lead
to much greater clarity in complex communications, when an invoice is sent
with some accoumpanying simple mathematics which describes its role in
commerce transaction. It will lead to much greater re-use of data, and much
easier analysis of what is going on.</p>

<p></p>

<h3>Quality</h3>

<blockquote>
  <p>Many documentation systems used to be designed for particular
  collections of information, and one could assume that the information in
  such a system had achieved a certain quality. However, the Web itself
  cannot enforce any single notion of quality. Such notions are very
  subjective, and change with time. To support this -- to allow users to
  actually use the web even though it contains junk as well as gems -- the
  technology must allow powerful filtering tools which, combining opinions
  and information about information from many sources, are completely under
  the control of the user.</p>
</blockquote>

<p>It is understood that a collection of works, such as a set of technical
reports or a library, only includes articles reaching a certain standard, and
some early dial-up information services similarly amassed information
according to some quality criterion. Some people miss that with the Web --
hence the need for portals which provde a filtered view. However useful
people find such portals, though, it is important that the Web itself doesn't
try to promote a single notion of quality.</p>

<p>The Web has to be able to carry, uncomplaining, beauty and ugliness,
honesty and lie. Users who find all of this of course complain, and sometimes
ask for it all to be organized and filtered. However, not only would one
central authority for quality be socially a disaster, but also, any one
single categorization of data would be only one person's view. Human
knowledge is not a tree, it is a web. How can we give the user the subjective
perception of higher quality, while maintaining an open Web for people whose
criteria are different?</p>

<p>The answer is through filtering. Unlike censorship, which is the forceful
prevention of one person's communication by another, filtering is the control
by the reader of what he or she reads. The trick is to allow the user to
chose another person, or another group's, criteria of selection. This is what
happens when a user selects from one of a choice of portals. More
sophisticated systems involve white lists of "desirable" sites, or black
lists of "undesirable" sites to be selected. This sort of information about
information is known as metadata. Metadata in general includes all the
information which catalogers and publishers and librarians keep about
information. The Semantic Web langauges (such as RDF) allow metadata to be
exchanged freely between different parties. As the richness of metadata
grows, so users will be able to combine criteria to hone their searches and
guide their browsing. And the Web will be left unconstrained by a central
authority deciding what information is appropriate for everyone.</p>

<p>There will, always be trash out there, and gems. Remember that you don't
have to read the junk. And also remember that the unimportant notes of today
maybe the foundation of revolutionary new ideas tomorrow.</p>

<h3>Independence of Scale</h3>

<blockquote>
  <p>The Web is described as a global phenomenon, and it is, but we must
  remember that personal information systems, and family and group
  information systems are part of it too. There should be no information
  boundary which would prevent a link from my personal diary to a public
  meeting. We know we need harmony on a global scale for peace, but that
  peace will only be stable so long as social groups of all sizes are
  respected. Starting at the individual, a group of one, one can think of
  institutions and ad-hoc groups of all sizes. The Web must support all of
  those, allowing privacy of personal information to be negotiated, and
  groups to feel safe in controlling access to their spaces. Only in such a
  balanced environment can we develop a sufficiently complex a many-layered
  fractal structure which will respect the rights of every human being, and
  allow all the billions of us to live in peace.</p>
</blockquote>

<p></p>

<p>When people express to me nervousness about the Web, there are two
concerns I hear repeatedly.</p>

<p>The first is that the Web will become one giant MacDonalds, and
international only as a mono-language, mono-cultural block. The French feared
that the transatlantic Internet cable would cause the culture of the Louvre
to be trampled over by the culture of Disney. People fear that only one
portal will end up surpassing all the others and become the only lens through
which all people see the world. It's a serious concern that if we have a
global network, we will homogenise our culture. It would be horrible if
language began to contain only those concepts that are sufficiently bland to
be understandable by absolutely everybody. We would lose a great deal of
richness. We need diverse pool of ideas for solving the unknown problems
ahead of us as teh human race.</p>

<p>The other fear is the opposite. One can chose what Websites one reads. One
can filter one's email, so that one sees no information except from a small
group, a clique, or possibly a group of deluded dangerous fanatics. A person
can operate in a virtual world without reality-checks from friends and
neighbors. The danger for people who operate without interaction with the
larger world is that the only common language they have with those different
from themselves may be violence. Our world at the moment desperately needs
enough common understanding to bring peace.</p>

<p>So, it is important as well, that while we have diversity, that there is a
balance between the small scale culture and the large scale culture - and all
scales in between. It seems that it is not only society which clearly needs a
balance: in its way, a lot of nature does as well. Nature is filled with
fractal patterns. This can be seen for example in ferns or coastlines. One
might approach a coastline, and as one gets closer above the coast, it has a
certain interesting structure. Then, closer, to a tenth of the altitude, it
still has an interesting structure. Closer and closer, until the point where
the seaweed curling around a few of the pebbles is visible, it still has an
interesting structure. It has structure in all levels. I have a deep feeling
that society needs to be like that. It can't be a simple structure which
operates just at one level. We need a complicated structure, which is fractal
in some way. That means that our society and the technology which we use to
support it has to work at each of these levels.</p>

<p>The development of the World Wide Web is a great example of human endeavor
in which many people participated, driven by individual excitement and a
common vision. There was no global management plan to make the World Wide
Web. It happened because a very diverse group of people, connected by the
Internet, wanted it to happen. The process was great fun, and still is. From
the fact that is it worked, I draw great hope for all our futures. May we now
use every ability we have to communicate to build a society in which mutual
respect, understanding and peace occur at all scales, between people and
between nations.</p>

<p></p>
<hr />

<p></p>

<p>Tim Berners-Lee</p>

<p>For Japan Prize Commemorative Lecture, 2002.</p>

<p></p>

<p>with thanks to Amy van der Heil for helping to put this together.</p>

<p></p>

<p></p>

<p><small>Last change <!-- keep -->
 $Id: Lecture.html,v 1.9 2003/01/22 18:59:17 amy Exp $ <!-- /keep-->
</small></p>
<address>
  <a href="/People/Berners-Lee">TimBL</a>
</address>
</body>
</html>