Saturday, 27 July 2013

The Woodstock of the Web: Geneva 25-27 May, 1994

Background: about this article (27 July 2013)

In 1994 I was based in Oxford University Computing Services, clinging post-doctorally to academic life through a variety of research posts at the intersection of computing and the humanities (creating the Collate program with Leverhulme support, starting off the Canterbury Tales Project with minimal formal support but lots of enthusiasm, advising Cambridge University Press about the new world of digital publication, etc.)  Because I had no regular post, people threw me the crumbs they could not eat themselves ("people" being mostly Susan Hockey, Lou Burnard and Marilyn Deegan).  At that time, Lou and Harold Short were writing a report outlining the need for a computing data service for the humanities: this report served as the template for what became the Arts and Humanities Data Service (now alas deceased, killed by British Rover -- that's another story). You might still be able to get a copy of this report online.  They had got some money for travel to research the report, more travel than Lou, for his part, could manage: so several times that year Lou (whose office was beside mine in 6 Banbury Road) said to me, "Peter, how would you like to go to..".  So I got to go to Dublin to see John Kidd commit public academic hari-kari at the James Joyce conference.  And, I got to go to Geneva for the first 'WWW' conference, ever. Some time in that period -- 1993, it must have been -- Charles Curran called me into this office at OUCS to show me the latest whizzy thing: a system for exchanging documents on the internet, with pictures! and links! This was the infant web, and OUCS had (so legend goes) server no. 13.

The total triumph of the web as invented at CERN in those years (that is, http + html, etc) makes it easy to forget that at that time, it was one among many competing systems, all trying to figure out how to use the internet for information exchange. Famously, a paper proposed by Tim Berners-Lee and others for a conference on hypertext in 1992 was rejected; Ted Ritchie, who ran OWL which looked just like the web before the Web, has a nice TED video telling how he was unimpressed by Tim Berners-Lee's idea.  So in early 1994, it was not a slam-dunk that the Web would Win. I had already met TBL: he came to Oxford, I think in late 1993 or early 1994, with the proposal that Oxford should be a host of what became the WWW3 consortium. I was present (with Lou, and others) at some of a series of meetings between TBL and various Oxford people.  The Oxford reception was not rapturous, and he ended by going to MIT (as this report notes).

The report is completely un-edited: I've highlighted headings, otherwise the text is exactly as I sent it to -- who? presumably Lou and Harold.  And, I still have, and occasionally wear, the conference t-shirt.

****My trip report****

Summary:

The first World Wide Web conference showed the Web at a moment of transition.
It has grown enormously, on a minimum of formal organization, through the
efforts of thousands of enthusiasts working with two cleverly devised interlocking
systems: the HTML markup language that allows resource to call resource across
networks, and the HTTP protocol that allows computer servers to pass the
resources about the world.  However, the continued growth and health of the Web
as a giant free information bazaar will depend on sounder formal organization, and
on its ability to provide efficient gateways to commercial services.  The challenge
for the Web is to harmonize this organization and increasing commercialism with
the idealistic energies which have driven it so far, so fast.  For the Arts and
Humanities Data Archive, the promise of the Web is that it could provide an ideal
means of network dissemination of its electronic holdings, blending free access with
paid access at many different levels.

From Wednesday 25th to Friday 27th I braved the clean streets and expensive
chocolates of Geneva, in search of an answer to a simple question: what is World
Wide Web? and what will it be, in a month, a year, a decade?

Once more, I donned my increasingly-unconvincing impersonation of Lou
Burnard.  This time, I was the Lou Burnard of the Arts and Humanities Data
Archive feasibility study, in quest of masses of Arts and Humanities data which
might need a secure, air-conditioned, (preferably clean and chocolate-coated)
environment for the next few millennia.  In the past months, WWW has become
possibly the world's greatest data-generator, with seemingly the whole of academe
and parts far beyond scrambling to put up home pages with links to everything.  So
my mission: what effect is the enormous success of WWW going to have on the
domains of data archives in particular, and on library practice and electronic
publishing in general?

The conference was held at CERN, the European Centre for Nuclear Research.
Underneath us, a 27km high-vacuum tunnel in which a positron can travel for three
months at the speed of light without ever meeting another atomic particle.
Ineluctably, the comparison presents itself: one can travel for weeks on WWW
without encountering anything worth knowing.  The quality of the Web was a
major subtext of the conference.  One thing the Web is not (yet) is any sort of
organization: it is a mark-up language (HTML), a network protocol by which
computers can talk to each other (HTTP) and it is a lot of people rushing in
different directions.  Anyone with a computer and an internet account can put
anything on the Web, and they do.  In no time at all, the Web has grown, and
grown...

Here follow questions I took with me, and some answers...

How big is the Web?

From the conference emerged a few remarkable figures on the growth of the Web:
in June 1993 there were 130 Web servers.  This had doubled by November 1993 to
272, then doubled again within a month to 623, then doubled again by March this
year to 1265 (source: David Eichmann conference paper 113, 116).  Other speakers
at the conference indicated that between March and June this year the number of
servers has doubled and redoubled again, and is probably now somewhere over
5000.  Another speaker estimated that there were around 100,000 'artifacts'
(articles, pictures etc etc...) on the Web (McBryan; Touvet put the figure at
'billions').  The volume of information is matched by the level of use.  The NCSA
home page currently receives around one million requests a week.  Around 500 gb
a week are shifted around the Web, and WWW now ranks sixth in total internet
traffic and will accelerate to first if it continues its current growth.

Who is the Web?

All this, from an idea in the head of Tim Berners-Lee, with no full-time staff or
organization to speak of.  I met (at lunch on Friday) Ari Luotonen, one of the two
staff at CERN who have direct responsibility for running the CERN server, the
first home of the Web, and still its European hub.  Simultaneously, they are
running the server, writing software etc. for it, and refining the file transfer
protocol HTTP which underpins the whole enterprise.  Before I went, I met Dave
Raggett of Hewlett-Packard, Bristol, who is busy simultaneously inventing the next
incarnation of HTML+ and writing a browser for it, and I spoke several more
times to him during the conference.  HTML, the Web's 'native' markup language, is
to the software of the Web what HTTP is to the hardware: without these two,
nothing.

To one used to the rigours of the Text Encoding Initiative, which has
ascetically divided design from practice in search of truly robust and elegant
solutions, the spectacle of the base standards of the Web being vigorously
redesigned by the same people who are writing its basic software is exhilerating
(there is a lot of: that's a nice idea! I'll have that) and also rather scary.  There is a
glorious mix of idealism and hard-headed practicality about the people at the heart
of the Web.  This is nowhere better epitomized than in Tim Berners-Lee, whose
address on the first morning had machines manipulating reality, with the Web being
'a representation of human knowledge'.  Along the way, Tim tossed out such
treasures as the Web allowing us to play dungeons and dragons in the Library of
Congress, or the Web finding the right second-hand car, or putting together a
skiing party including Amanda.  And, he said, you could buy a shirt on the Web.
If the Web is to continue to grow and prosper, and become more than a
computer hobbyist's super toy, it will need to be useful.  Or will it?  Can we only
have all human knowledge on the Web if we can buy shirts on it, too?  Useful, to
who?

Who uses the Web?

One paper (Pitkow and Recker) reported the results of a survey on who uses the
Web, based on 4500 responses to a posting on the Web in January 1994: over 90
per cent of its users are male, and 60 % are under 30, and 69% are in North
American universities.  One could say that this survey was rather skewed in that the
survey relied on the 'forms' feature in Mosaic, and as few people had access to this
on anything other than Unix 93% of respondents (the same 94 % who were male?)
were Unix/Mosaic users.  In any case, the conference was overwhelmingly male,
and strongly American in accent.  And the papers at the conference were
overwhelmingly technical: concerned with bandwidth, Z39.50 protocols, proxy
servers, caching, algorithms for 'walking the Web', and so on.  In part this
technical bias is necessary.  The very success of the Web threatens Internet
meltdown.  But extreme technical skill in one area can mean frightening naievety in
another, and vast quantities of naievety were on view over the three days.  For
many people, the Web is the world's greatest computer playground: hacker heaven
in 150 countries at once.  And so, many of the papers were exercises in ingenuity:
look what I have done! without reflection as to whether it was worth doing in the
first place.  But, ingenuity certainly there was.  Here are some examples...

How do we find out what is on the Web?

The sudden and huge size of the Web has created a monster: how do you find just
what you want amid so much, where there are no rules as to what goes where, no
catalogue or naming conventions.  The whole of Thursday morning, and many
other papers outside that time, were devoted to 'taming the Web': resource
discovery methods for locating what you want on the Web.  It is typical of the
extraordinary fertility of invention which the Web provokes (and this just might be
the Web's greatest gift to the world) that no two of these papers were even
remotely alike in the methods they proposed.  One system (SOLO, presented by
Touvet of INRIA) focussed on directory services: indexing all the file names
accessed through the web and constructed a 'white pages' for the Web which maps
all the file names and their servers to resource names, so that one need only point at
the SOLO directory to find an object, whereever it is.  Again, something like this is
badly needed as lack of 'name persistence' -- the distressing tendency for files to
disappear from where they are supposed to be -- is a major Web shortcoming.
Fielding (UC at Irvine) outlined a different solution to this particular problem: a
'MOMspider' which would walk the Web and find places where files have been
mislaid, and hence compute the Web's health and alert document owners of
problems.  Fielding argued that living infostructures must be actively maintained to
prevent structural collapse: so, who is to do the maintenance?

A quite different approach, though still centred on whole documents, is
ALIWEB, presented by Martijn Koster of NEXOR Ltd.  This uses index files
prepared according to standard formats, containing information on who made the
file, what it is, and some key-words, etc.  These files must be 'hand-prepared' by
someone at each server site; ALIWEB then collects all these files together and loads
them into a database sent back to all the servers.  This is modelled on the working
of Archie, where servers similarly compile and exchange information about ftp
resources, etc.  There is no doubting the value of something like ALIWEB.  But
how many of the people running those 5000-plus servers (especially, the 4000 or so
which have joined in the last three months) even know of ALIWEB?
Another system using 'hand-prepared' data based on whole-document
description was outlined by McBryan (University of Colorado).  This is GENVL,
or 'the mother of all bulletin boards'.  This uses a very clever system of nested
'virtual libraries' deposited in GENVL by outside users.  I can make up my own
'virtual library' of Web resources in my own area of interest using the format
supplied by GENVL.  This library can then be nested within the existing GENVL,
and point at yet other virtual libraries supplied by other people. People can find my
library and the documents and libraries it points at by travelling down the
hierarchy, or by WAIS searching on the whole collection.  But, who guarantees the
accuracy and timeliness of the data in GENVL?  No-one, of course: but it seems on
its way towards being the biggest yellow pages in the world.  McBryan's
demonstration of this was far the best demonstration seen at the conference: you can
get into it for yourself by
http://www.cs.colorado.edu/homes/mcbryan/public_html/bb/summary_html.  I must
say the competition for the best demonstration was less intense than might have
been supposed.  Indeed, there were more presentational disasters than I have ever
seen at any conference.  Many presenters seemed so struck by the wonder of what
they had done that they thought it was quite enough to present their Web session
live, projected onto a screen, in happy ignorance of the fact that the ergonomics of
computer graphics meant that hardly anyone in the auditorium could read what was
on the screen.  This enthusiastic amateurism is rather typical of a lot which goes on
in the Web.

Other papers in this strand looked into the documents themselves.   The
Colorado system described by McBryan supplements the hand-prepared 'virtual
libraries' by an automated 'Worm' (it appears the Web is breeding its own parasitic
menagerie; in Web-speak, these are all varieties of 'robots') which burrows its way
through every document on the Web, and builds a database containing index entries
for every resource on the Web (as at March 1994, 110,000 resources, 110,000
entries).  It does not try to index every word; rather it indexes the titles to every
resource, the caption to every hypertext link associated with that resource, and the
titles of every referenced hypertext node.   This is clever, and again was most
impressive in demonstration: you simply dial up the worm with a forms aware
client (Mosaic 2.0, or similar) and type in what you are looking for.  Each of
McBryan's systems (GENVL and the Worm) currently receives over 2000 requests
a day.  They were far and away the most impressive, and most complete, of the
search systems shown at the conference.

We met many more of these 'robots' during the conference: beasts bred by
computer science departments to graze their way through the web.  They are
creatures of controversy: they can add considerably to the network load and there
are some who would ban all robots from the web. Eichmann of the University of
Houston (Clear Lake) described his 'spider' which again walked the whole Web,
like McBryan's worm.  However, his system performs a full-text index of each
Web document, not just an index of titles and hypertext captions.  This is a
modified WAIS index which works by throwing away the most frequently
referenced items.  However, the Web is so large that the index of even a part of it
soon grew to over 100 megs.  The sheer size of the Web, and the many near-
identical documents in some parts of it, can have entertaining consequences.
Eichmann related the sad tale of the time his spider fell down a gravitational well.
It spent a day and a half browsing a huge collection of NASA documents (actually a
thesaurus of NASA terms) and got virtually nothing back you could use in a search:
all it found was NASA, ad infinitum, and so it threw the term away from the index
and .... eventually Eichmann had to rescue his spider.  As NASA were sponsoring
this research they were not amused to find that if you asked the spider to find all
documents containing 'NASA' you got precisely nothing.

Every conference on computing must contain a paper with the title 'lost in
hyperspace': this one was from Neuss (Fraunhofer Institute, Germany).  I was by
now rather fatigued by techies who thought that running GREP across the Web
actually was a decent search strategy.  Neuss did a little more than this (using 'fuzzy
matching') but what he had was essentially our old friend the inverted file index.
Which of course when faced with gigabytes finds it has rather too many files to
invert and so tends to invert itself, turtle-like, which is why both Neuss and
Eichmann found themselves concluding that their methods would really only be
useful on local domains of related resources, etc.  At last, after all this we had a
speaker who had noticed that there is a discipline called 'information retrieval',
which has spent many decades developing methods of relevance ranking, document
clustering and vectoring, relevance feedback, 'documents like...', etc.  He pointed
out the most glaring weaknesses in archie, veronica, WAIS and the rest (which had
all been rather well demonstrated by previous speakers).  However, it appeared he
had good ideas but no working software (the reverse of some of the earlier
speakers).

Finally, in this strand, the true joker of the pack: De Bra of Eindhoven
University with -- hold your hat -- real-time full-text searching of the web.  That
is: you type in a search request and his system acually fires it off simultaneously to
servers all over the Web, getting them to zoom through their documents looking
for matches.  It is not quite as crude as this, and uses some very clever 'fishing'
algorithms to control just how far the searches go, and also relies heavily on
'caching' to reduce network load.  But it is exactly this type of program which
rouses the ire of the 'no-robots' Web faction (notably, Martijn Koster).

So: masses of ways of searching the Web.  But is there anything there worth
finding?

How serious is the Web? -- Electronic publishing and the Web

There were times in these three days when I felt I had strayed into some vast
electronic sandpit.  I had come, on the part of the Arts and Humanities archives, in
search of masses of data which might need archiving.  What I found instead was a
preoccupation with how the Web worked, and very little interest in the quality of
material on it.  On the Wednesday morning, Steffen Meschkat (ART + COM,
Berlin) described his vision of active articles in interactive journals.  These would
present a continuous, active model of publication.  The boundaries between
reviewed and published articles would disappear, and the articles themselves would
merge with the sea of information.  To Meschkat, the idea of publication as a
defining act has no meaning; there is no place in his world for refereed publication,
so important to University tenue committees.  There is no doubt that the Web is a
superb place for putting up pre-publication drafts of material for general comment,
and many scholars are making excellent use of it for just that.  But to argue that this
is a completely satisfactory model of publication is absurd.  A journal whose only
defining strategy is the whims of its contributers (how do you 'edit' such a
journal?) is not a journal at all; it is a ragbag.  I am pleased to report that the
centrepiece of Meschkat's talk, a demonstration of the active journal, failed to
work.

In general, I was surprised by the level of ignorance shown about what the
Web would need if it were to become a tool for serious electronic publishing of
serious materials.  It is the first conference I have been to where the dread word
copyright hardly appeared.  Further, there was hardly any discussion of what the
Web would need to become a satisfactory electronic publishing medium: that it
would have to tackle problems of access control, to provide multilayered and fine-
grained means of payment, to give more control over document presentation.  A
SGML group met on Wednesday afternoon to discuss some of these issues, but did
not get very far.  There were representatives of major publishers at this group
(Springer, Elsevier) but none yet are doing more than dip a toe in the Web.
From the talk at the conference, from the demonstrations shown, and what I
have seen of the Web, it is not a publication system.  It is a pre-publication system,
or an information distribution system.  But it is as far from a publishing system as
the Thursday afternoon free paper is from OUP.

How serious is the Web? free information, local information

The free and uncontrolled nature of the Web sufficiently defines much of what is
on it.  Most of what is on the Web is there because people want to give their
information away.  Because it is free, it does not matter too much that the current
overloading of the internet means that people may not actually be able to get the
information from the remote server where it is posted.  Nor does it matter that the
rght files have disappeared, or that the information is sloppily presented, eccentric
or inaccurate.  These things would be fatal to a real electronic publishing system,
but do not matter in the give-away information world which is much of the Web.

There is an important category of Web use which does not fit this
description.  That is its use for distribution of local information systems: for
carrying the local 'help' or 'info' system for a university, or a company.   The most
impressive instances of Web use offered at the conference fitted this model.  Here,
we are dealing with information which is valuable but free: valuable, because it is
useful to a definable group of people.  Typically, the information is owned by the
local organization and so there are no copyright problems to prevent its free
availability.  Further, because the information is most valuable to the local user,
and rather less valuable to anyone beyond, it is quite valid to optimize the local
server and network for local access; those beyond will just have to take their
chances.  The Web server currently being set up in Oxford is an excellent example
of the Web's power to draw together services previously offered separately
(gopher, info, etc.) into a single attractive parcel.  Several conference papers told
of similar successful applications of the Web to local domains: a 'virtual classroom'
to teach a group of 14 students dispersed across the US, but linked to a single Web
server (Dimitroyannis, Nikhef/Fom, Amsterdam); an on-line company-wide
information system by Digital Equipment (Jones, DEC).  Most impressive was the
PHOENIX project at the University of Chicago (Lavenant and Kruper).  This aims
to develop, on the back of the Web, a full teaching environment, with both students
and teachers using the Web to communicate courses, set exercises, write papers etc.
Three pilot courses have already been run, with such success that a further 100
courses are set to start later this year.  Because it is being used for formal teaching,
PHOENIX must include sophisticated user-authentication techniques.  This is done
through the local server; a similar method (though rather less powerful) is in use at
the City University, London (Whitcroft and Wilkinson).

Such local applications resolve the problem of access to remote servers
which threatens to strangle the web (and with it, the whole Internet).  They also
solve the problems of accountability and access: someone is responsible for the
information being accurate and available.  As for access: if restriction is required,
this can be done through the local server, so avoiding all the problems of trying to
enforce control through servers on the other side of the world.  The Web is so well
suited to these local information services that these alone will see the Web continue
to burgeon.  Much of the Web will be a network of islands of such local systems,
upon which the rest of the world may eavesdrop as it will.  Among these islands,
we can expect to see many more distance learning providers, on the model of
PHOENIX but distributed across world-wide 'virtual campuses'.  Butts et al.
described the 'Globewide Network Academy' (GNA), which claims to be the
world's first 'virtual organization', and appears to intend to become a global
university on the network.  This seems rather inchoate as yet.  But already 2500
people a day access GNA's 'Web Tree'; where they are going, others will certainly
follow.  We could also expect to see many more museums, archives, libraries, etc.
using the Web as an electronic foyer to their collections.  There was only one such
group presenting at the conference, from the Washington Holocaust museum
(Levine): that there should be only this one tells us a great deal about the lack of
computer expertise in museums.  The Web is an ideal showplace for these.  Expect
to see many more museums,etc., on the Web as the technology descends to their
range.

How can I make sure I can get through to the Web?

For days on end, I have tried to get through to McBryan's server at Colorado, or
the NCSA home page, or indeed anywhere outside the UK.  Unless you are
prepared to time it for a moment when all the rest of the world is in bed (and you
should be too), you cannot get through.  This alone appears as if it could be doom
for the Web's aspirations.  Certainly, it could not be a commercial operation at this
basis: noone is going to pay to be told that the server they want is unavailable.  Of
course, this does not matter if all you want is local information, as described in the
last section.  But this is far from the vision of the world talking to itself promoted
so enthusiastically by Berners-Lee and others.

Help is at hand.  A surprising number of papers were devoted to problems
of making the Web work.  The answer is simple: if lack of network bandwidth
means we cannot all dial into America at once, set up clones of America all over the
world.  There are many ways of doing this: 'mirror' servers, etc, as long known in
the FTP world, for example.  There are obvious problems with mirror servers in
the fast-moving world of the Web, with some documents both changing and being
accessed very frequently.  The method that seems most enthusiastically pursued is
'caching', otherwise 'proxy' servers.  Instead of dialling America direct,  you dial
the proxy server.  The server checks if it has a copy of the document you want; if it
has, then it gives you the document itself.  If it has not, it gets the document from
America, passes it to you and keeps a copy itself so the next person who asks for it
can get it from the cache.  Papers by Luotonen, Glassman, Katz and Smith
described various proxy schemes, several of them based on Rainer Post's Lagoon
algorithms.  It is the sort of topic which fascinates computer scientists, with lots of
juicy problems to be solved.  For example: once someone has dialled into a proxy
and then starts firing off http calls from the documents passed over by the proxy,
how do we make sure all those http calls are fed to the proxy and not to the 'home'
servers whose address is hard-wired into the documents?  or, how do you operate a
proxy within a high-security 'firewall' system, without compromising the security?
how do you make sure the copy the proxy gives you is up to date?  and so on.

These systems work, and work well: since I have caught on to dialling the
Web via the HENSA proxy at Kent, life has suddenly become possible.  An
interesting statistic, drawn from Smith of Kent's talk: over an average week, it
seems that around two thirds of the calls to the Web by all those hundreds of
thousands of people are to some 24,000 documents (about a quarter of the total on
the Web, as of May), amounting to just 450 Meg of data.  The Web is both
enormous, and surprisingly small.  One point is clear: proxies and similar devices
can only work given sophisticated co-operation between servers.

Has the Web a future?

If the Web stays as it is, it will slowly strangle itself with too much data, too little
of which is of any value, and with even less of what you want available when you
want it.  I had more than a few moments of doubt at the conference: there were too
many of the hairy-chested 'I wrote a Web server for my Sinclair z99 in turbo-
Gaelic in three days flat' to make me overwhelmingly confident.  The Web will not
stay as it is. The Web has grown on the enthusiastic energies of thousands of
individuals, working in isolation.  But the next stages of its growth, as the Web
seeks to become more reliable and to grow commercial arms and legs, will require
more than a protocol, a mark-up language and enthusiasm.  Servers must talk to
one another, so that sophisticated proxy systems can work properly.  Servers will
have to make legal contracts with one another, so that money can pass around the
Web.  Richer mark-up languages, as HTML evolves, will require more powerful
browsers, and create problems of standardization, documentation and maintenance.
All this will require co-operation, and organization.  The organization is on the
way.  After much hesitation, CERN has decided to continue to support the Web,
and the EC is going to fund a stable European structure for the Web, probably
based at CERN and possibly with units in other European centres (including,
perhaps, Oxford).  Berners-Lee himself is going to MIT, where he is going to work
with the team which put together XWindows.

What will the Web look like, then, in a few years?  The momentum for
much of the Web, at least, to remain free and uncontrolled (even, anarchic) will
continue: it will become the world's free Thursday afternoon paper, a rag-bag of
advertizing, vanity publishing, plain nonsense and serious information.  There will
be many local islands of information -- help systems, archive galleries, educational
material.  Eccentrics, like the Norwegian family who have put their whole house on
the Web so you can dial up a video camera in their living room and watch what
they are doing, will continue to find a home on the Web (Ludvigsen).  But
increasingly the Web will provide a gateway to other, specialist and typically
commercial services: electronic publishing, databases, home shopping, etc.  The
presence of these satellite services will relieve the Web itself from the need to be all
things to all  people, and allow it to continue to be the lingua franca to the
electronic world -- demotic, even debased, but easy, available and above all, free.
The Web will not be the only gateway to services which can do what the Web
cannot, but it may well be the biggest and the most widely used.  For organizations
like Universities, libraries, archives of all sorts (and, especially, the AHDA), which
give some information away and sell other information, the Web and its satellites
will be an ideal environment.  This aspect of the Web has been barely touched, to
now.  But it will happen, and it is our business to make sure it happens the right
way.

What have I not described?

Even this rather long account omits much.  Briefly, a few things I have not
covered:
1.  HTML tools: there was discussion about HTML and its future (Raggett; various
workshops); several HTML editors were described (Williams and Wilkinson;
Kruper and Lavenant; Rubinsky of SoftQuad announced HotMeTaL); various
schemes for converting documents to HTML were outlined (from FrameMaker:
Stephenson; Rousseau; from Latex: Drakos).
2.  Connectivity tools for linking the Web to other resources: to CD-ROMs
(Mascha); to full-featured hypertext link servers (Hall and the Microcosm team); to
computer program teaching systems (Ibrahim); to databases (Eichmann)
and much, much more.

Thirty years after McLuhan, the global village is here.

No comments:

Post a Comment