Center "Leo Apostel", Free University of Brussels,
Pleinlaan 2, B-1050 Brussels
email: firstname.lastname@example.org, email@example.com
If society is viewed as a super-organism, communication networks play the role of its brain. This metaphor is developed into a model for the design of a more intelligent global network. The World-Wide Web, through its distributed hypermedia architecture, functions as an "associative memory", which may "learn" by the strengthening of frequently used links. Software agents, exploring the Web through spreading activation, function as problem-solving "thoughts". Users are integrated into this "super-brain" through direct man-machine interfaces and the reciprocal exchange of knowledge between individual and Web.
Yet, there is at least one domain where integration seems to be moving full speed ahead: the development of ever more powerful communication media. In the society as super-organism metaphor, the communication channels play the role of nerves, transmitting signals between the different organs and muscles [Turchin, 1977]. In more advanced organisms, the nerves develop a complex mesh of interconnections, the brain, where sets of incoming signals are integrated and processed. After the advent in the 19th century of one-to-one media, like telegraph and telephone, and in the first half of this century of one-to-many media, like radio and TV, the last decade in particular has been characterized by the explosive development of many-to-many communication networks. Whereas the traditional communication media link sender and receiver directly, networked media have multiple cross-connections between the different channels, allowing complex sets of data from different sources to be integrated before being delivered to the receivers. For example, a newsgroup discussion on the Internet will have many active contributors as well as many people just `listening in'. Moreover, the fact that the different `nodes' of the digital network are controlled by computers allows sophisticated processing of the collected data, reinforcing the similarity between the network and the brain. This has led to the metaphor of the world-wide computer network as a `global brain' [Mayer-Kress & Barczys, 1995; Russell, 1995].
In organisms, the evolution of the nervous system is characterized by a series of metasystem transitions producing subsequent levels of complexity or control [Turchin, 1977; Heylighen, 1995, 1991b]. The level where sensors are linked one-to-one to effectors by neural pathways or reflex arcs is called the level of simple reflexes. It is only on the next level of complex reflexes, where neural pathways are interconnected according to a fixed program, that we start recognizing a rudimentary brain. This paper will argue that the present global computer network is on the verge of undergoing similar transitions to the subsequent levels of learning, characterized by the automatic adaptation of connections, thinking, and possibly even metarationality. Such transitions would dramatically increase the network's power, intelligence and overall usefulness. They can be facilitated by taking the "network as brain" metaphor more seriously, turning it into a model of what a future global network might look like, and thus helping us to better design and control that future. In reference to the super-organism metaphor for society this model will be called the "super-brain".
The distributed hypermedia paradigm is a synthesis of three ideas [Heylighen, 1994]. 1) Hypertext refers to the fact that WWW documents are cross-referenced by `hotlinks': high-lighted sections or phrases in the text, which can be selected by the user, calling up an associated document with more information about the phrase's subject. Linked documents (`nodes') form a network of associations or `web', similar to the associative memory characterizing the brain. 2] Multimedia means that documents can present their information in any modality or format available: formatted text, drawings, sound, photos, movies, 3-D `virtual reality' scenes, or any combination of these. This makes it possible to choose that presentation best suited for conveying an intuitive grasp of the document's contents to the user, if desired, bypassing abstract, textual representations for more concrete, sensory equivalents. 3) Distribution means that linked documents can reside on different computers, maintained by different people, in different parts of the world. With good network connections, the time needed to transfer a document from another continent is not noticeably different from the time it takes to transfer a document from the neighbouring office. This makes it possible to transparently integrate information on a global scale.
Initially the Web was used for passive browsing through existing documents. The addition of `electronic forms', however, made it possible for users to actively enter information, allowing them to create documents and query specialized computer programs anywhere on the net. At present the World-Wide Web can be likened to a huge external memory, where stored information can be retrieved either by following associative links, or by explicitly entering looked-for terms in a search engine.
A first step to make the `Web as memory' more efficient is to let the Web itself discover the best possible organization. In the human brain knowledge and meaning develop through a process of associative learning: concepts that are frequently used together become more strongly connected (Hebb's rule for neural networks). It is possible to implement similar mechanisms on the Web, creating associations on the basis of the paths followed by the users through the maze of linked documents. The principle is simply that links followed by many users become `stronger', while links that are rarely used become `weaker'. Simple heuristics can then propose likely candidates for new links: if a user moves from A to B to C, it is probable that there exists not only an association between A and B but also between A and C (transitivity), and between B and A (symmetry). In this manner, potential new links are continuously generated, while only the ones that gather sufficient `strength' are retained and made visible to the user. This process was tested by us in an adaptive hypertext experiment, where a web of randomly connected words self-organized into a semantic network, by learning from the link selections made by its users. [See Bollen & Heylighen, 1996, for more details about learning algorithms and experimental results].
The strength of such associative learning mechanisms is that they work locally (they only need to store information about documents at most two steps away), but the self-organization they produce is global: given enough time, documents which are an arbitrary number of steps away from each other can become directly connected if a sufficient number of users follow the connecting path. We could imagine extending this method by more sophisticated techniques, which e.g. compute a degree of similarity between documents on the basis of the words they contain, and use this to suggest similar documents as candidate links from a given document. The expected result of such associative learning processes is that documents that are likely to be used together will also be situated near to each other in the topology of `cyberspace'.
If such learning algorithms could be generalized to the Web as a whole, the knowledge existing in the Web could become structured into a giant associative network which continuously learns from its users. Each time a new document is introduced, the links to and from it would immediately start to adapt to the pattern of its usage, and new links would appear which the author of the document never could have foreseen. Since this mechanism in a way assimilates the collective wisdom of all people consulting the Web, we can expect the result to be much more useful, extended and reliable than any indexing system generated by single individuals or groups.
A first such mechanism can be found in WAIS-style search engines [e.g. Lycos, http://lycos.cs.cmu.edu/]. Here the users enters a combination of keywords that best reflect his or her query. The engine scans its index of web documents for documents containing those keywords, and scores the `hits' for how well they match the search criteria. The best matches (e.g. containing the highest density of desired words) are proposed to the user. For example, the input of the words "pet" and "disease" might bring up documents concerning veterinary science. This method only works if the documents effectively contain the proposed keywords. However, many documents may discuss the same subject using different words (e.g. "animal" and "illness"), or use the same words to discuss different subjects (e.g. PET tomography).
Some of these problems may be overcome through a direct extension of the associative memory metaphor, the mechanism of spreading activation [Jones, 1986; Salton & Buckley, 1988]: activating one concept in memory activates its adjacent concepts which in turn activate their adjacent concepts. Documents about pets in an associative network are normally linked to documents about animals, and so a spread of the activation received by "pet" to "animal" may be sufficient to select all documents about the issue. This can be implemented as follows. Nodes get an initial activation value proportional to an estimate of their relevance for the query. This activation is transmitted to linked nodes. The total activation of a newly reached node is calculated as the sum of activations entering through different links, weighted by the links' strength. This process is repeated, with the activation diffusing along parallel paths, until a satisfactory solution is found (or the activation value becomes too low).
A simple way to conceptualize the function of an agent is through the concept of vicarious selector [Campbell . A vicarious selector is a delegate mechanism, which explores a variety of situations and selects the most adequate ones, in anticipation of the selection that would eventually be carried out by a more direct mechanism. For example, echo-location in bats and dolphins functions through the broadcast of an acoustic signal, which is emitted blindly in all directions, but which is selectively reflected by objects (e.g. prey or obstacles). The reflections allow the bat to locate these distant objects in the dark, without need for direct contact. Similarly, an agent may be `broadcast' over the Web, exploring different documents without a priori knowledge of where the information it is looking for will be located. The documents that fulfil the agent's selection criteria can then be `reflected' back to the user. In that way, the user, like the bat, does not need to personally explore all potentially important locations, while still being kept informed of where the interesting things are.
A web agent might contain a combination of possibly weighted keywords that represents its user's interest. It would evaluate the documents it encounters with respect to how well they satisfy the interest profile, and return the ones that score highest scoring to the user. Agents can moreover implement spreading activation: an agent encountering different potentially interesting directions (links) for further exploration, could replicate or divide itself into different copies, each with a fraction of the initial `activation', depending on the strengths of the links and the score of their starting document. When different copies arrive in the same document, their activations are added in order to calculate the activation of the document. In order to avoid epidemics of virus-like agents spreading all over the network, a cut-off mechanism should be built in, so that no further copies are made below a given threshold activation, and so that the initial activation supply of an agent is limited, perhaps in proportion to the amount of computer resources the user is willing to invest in the query.
An agent's selection criteria may be explicitly introduced by the user, but they can also be learnt by the agent itself [Maes, 1994]. An agent may monitor its user's actions and try to abstract general rules from observed instances. For example, if the agent notes that many of the consulted documents contain the word "pet", it may add that word to its search criteria and suggest to the user to go and collect more documents about that topic. Learning agents and the learning Web can reinforce each others effectiveness. An agent that has gathered documents related according to its built-in or learned selection criteria can signal that to the Web, allowing the Web to create or strengthen links between these documents. Reciprocally, by creating better associations, the learning Web will facilitate the agents' search, by guiding the spread of activation or by suggesting related keywords (e.g. "animal" in addition to "pet"). Through their interaction with a shared associative web, agents can thus indirectly learn from each other, though they may also directly exchange experiences [Maes, 1994].
We can safely assume that in the following years virtually the whole of human knowledge will be made available on the Web. If that knowledge is organized as an associative or semantic network, `spreading' agents should be capable to find the answer to practically any question for which an answer somewhere exists. The spreading activation mechanism allows questions that are vague, ambiguous or ill-structured: you may have a problem, but not be able to clearly formulate what it is you are looking for.
For example, imagine the following situation: your dog is regularly licking the mirror in your home. You worry whether that is just normal behavior, or perhaps a symptom of a disease. So, you try to find more information by entering the keywords "dog", "licking" and "mirror" into a web search agent. If there would be a `mirror-licking' syndrome described in the literature about dog diseases, such a search would immediately find the relevant documents. However, that phenomenon may just be an instance of the more general phenomenon that certain animals like to touch glass surfaces. A traditional search on the above keywords would never find a description of that phenomenon, but the spread of activation in a semantically structured web would reach "animal" from "dog", "glass" from "mirror" and "touching" from "licking", selecting documents that contain all three concepts. Moreover, a smart agent would assume that documents discussing possible diseases would be more important to you than documents that just describe observed behavior, and would retrieve the former with higher priority.
This example can be generalized to the most diverse problems. Whether it has to do with how to decorate your house, how to reach a certain place, or how to combat stress: whatever the problem you have, if some knowledge about the issue exists, spreading agents should be able to find it. For the more ill-structured problems, the answer may be reached only after a number of steps. Formulating part of the problem brings up certain associations that make you or the agent reformulate the problem (e.g. excluding documents about tomography), in order to better select relevant documents. The Web will not only provide straight answers but general feedback to guide you in your efforts to get closer to the solution.
Coming back to our brain metaphor, the agents searching the Web, exploring different regions, creating new associations by the paths they follow and the selections they make, and combining the found information into a synthesis or overview, which either solves the problem or provides a starting point for a further round of reflection, seem wholly analogous to thoughts spreading and recombining over the network of associations in the brain. This would bring the Web into the metasystem level of thinking, which is characterized by the capability to combine concepts without the need for an a priori association between these concepts to exist in the network [Turchin, 1977; Heylighen, 1991b, 1995].
Many different techniques are available to support such discovery of general principles, including different forms of statistical analysis, genetic algorithms, inductive learning and conceptual clustering, but these still lack integration. The controlled development of knowledge requires a unified metamodel: a model of how new models are created and evolve [Heylighen, 1991b]. A possible approach to develop such a metamodel might start with an analysis of the building blocks of knowledge, of the mechanisms that (re)combine building blocks to generate new knowledge systems, and of a list of selection criteria, which distinguish `good' or `fit' knowledge from `unfit' knowledge [Heylighen, 1992].
In order to most effectively use the cognitive power offered by an intelligent Web, there should be a minimal distance between the user's wishes and desires and the sending out of web-borne agents. At present, we are still using computers connected the network by phone cables, creating queries by typing in keywords in specifically selected search engines. This is quite slow and awkward when compared to the speed and flexibility with which our own brain processes thoughts. Several mechanisms can be conceived to accelerate that process.
The quick spread of wireless communication and portable devices promises the constant availability of network connections, whatever the user's location. We already mentioned multimedia interfaces, which attempt to harness the full bandwidth of 3-dimensional audio, visual and tactile perception in order to communicate information to the user's brain. The complementary technologies of speech or gesture recognition make the input of information by the user much easier. We also mentioned the learning agents, which try to anticipate the user's desires by analysing his or her actions. But even more direct communication between the human brain and the Web can be conceived.
There have already been experiments in which people managed to steer images on a computer screen simply by thinking: their brain waves associated with focused thoughts (such as "up", "down", "left" or "right") are registered by sensors, interpreted by neural network software, and translated into commands, which are executed by the computer. Such set-ups use a two-way learning process: the neural network learns the correct interpretation of the registered brain-wave patterns, while the user, through bio-feedback, learns to focus thoughts so that they become more understandable to the computer. An even more direct approach can be found in neural interface research, the design of electronic chips that can be implanted in the human body and connected to nerves, so as to register neural signals [Kovacs et al., 1994]. Once these technologies have become more sophisticated, we could imagine the following scenario: at any moment a thought might form in your brain, then be translated automatically via a neural interface to an agent or thought in the external brain, continue its development by spreading activation, and come back to your own brain in a much enriched form. With a good enough interface, there should not really be a border between `internal' and `external' thought processes: the one would flow naturally and immediately into the other. It would suffice that you think about your dog licking mirrors to see an explanation of that behavior pop up before your mind's eye.
In a sense, the brains of the users themselves would become nodes in the Web: stores of knowledge linked to the rest of the Web, which can be consulted by other users or by the Web itself. Eventually, the individual brains may become so strongly integrated with the Web that the Web would literally become a `brain of brains': a super-brain. A thought might run from one user to the Web, to another user, back to the Web, and so on. Thus, billions of thoughts would develop in parallel over the super-brain, creating ever more knowledge in the process.
The question remains whether individuals would agree to be so intimately linked into a system they only partially control. On the one hand, individuals might refuse answering requests from the super-brain. On the other hand, no one would want to miss the opportunity to use the unlimited knowledge and intelligence of the super-brain for solving one's own problems. However, the basis of social interaction is reciprocity. People will stop answering your requests if you never answer theirs. Similarly, one could imagine that the intelligent Web would be based on the simple condition that you can use it only if you provide some knowledge in return.
In practice, such conditions may come out of the economic constraints of the `knowledge market', which make that people must provide services in order to earn the resources they need to sustain their usage of other services. Presently, there is a rush of commercial organizations moving to the Web in order to attract customers. The best way to convince prospective clients to consult their documents, will be to make these documents as interesting and useful as possible. Similarly, the members of the academic community are motivated by the `publish or perish' rule: they try to make their ideas as widely known as possible, and are most likely to succeed if these results are highly evaluated by their peers on the Web. Thus, we might expect a process where the users are maximally motivated both to make use of the Web's existing resources and to add new resources to it. This will make the Web-user interaction wholly two-way, the one helping the other to become more competent.
However, there remains the problem of intellectual property (e.g. copyright or patents): though it might be in the interest of society to immediately make all new knowledge publicly available, it is generally in the interest of the developer of that knowledge to restrict access to it, because this makes it easier to get compensation for the effort that went into developing it. An advantage of the global network is that it may automatize compensation, minimize the costs of development and transaction of knowledge, and foster competition between knowledge providers, so that the price of using a piece of knowledge developed by someone else might become so low as to make it practically free. A very large number of users paying a very small sum may still provide the developer with a sufficient reward for the effort.
As to the fair distribution of material resources over the world population, it must be noted that their value (as contrasted with intellectual resources) is steadily decreasing as a fraction of the total value of products or services. Moreover, the super-brain may facilitate the emergence of a universal ethical and political system, by promoting the development of shared ideologies that transcend national and cultural boundaries [cf. Heylighen & Campbell, 1995], and by minimizing the distance between individuals and government. However, these questions are very subtle and complex, and huge obstacles remain to any practical implementation, so that it seems impossible to make predictions at this stage.
Yet, there are the many unfulfilled promises from the 40 year history of Artificial Intelligence to remind us that problems may be much more serious than they initially appeared. It is our impression that the main obstacles hindering AI have been overcome in the present model. First, AI was dogged by the fact that intelligent behavior requires the knowledge of an enormous mass of common-sense facts and rules. The fact that millions of users in parallel add knowledge to the super-brain eliminates this bottleneck. The traditional symbolic AI paradigm moreover made the unrealistic demand that knowledge be formulated as precise, formal rules. Our view of the super-brain rather emphasizes the context-dependent, adaptive and fuzzy character of associative networks, and is thus more reminiscent of the connectionist paradigm. Finally, traditional AI tended to see knowledge as a mapping or encoding of outside reality, a philosophy that runs into a host of practical and epistemological problems [Bickhard & Terveen, 1995]. The present model, on the other hand, is constructivist or selectionist: potential new knowledge is generated autonomously by the system, while the environment of users selects what is adequate.
It will only become clear in the next few years whether these changes in approach are sufficient to overcome the technical hurdles. At this stage, we can only conclude that extensive research will be needed in order to further develop, test and implement the ideas underlying the present model for a future network.
Bollen J. & Heylighen F. (1996): "Algorithms for the Self-Organization of Distributed Multi-user Networks", in: R. Trappl (ed.)Cybernetics and Systems '96 (this volume).
Campbell D.T. (1974): "Evolutionary Epistemology", in: The Philosophy of Karl Popper, Schilpp P.A. (ed.), (Open Court Publish., La Salle, Ill.), p. 413-463.
Fayyad U.M. & Uthurusamy R. (eds.) (1995): Proc. 1st Int. Conference on Knowledge Discovery and Data Mining (AAAI Press, Menlo Park, CA).
Heylighen F. (1991a): "Design of a Hypermedia Interface Translating between Associative and Formal Representations" Int J. Man-Machine Studies 35, p. 491.
Heylighen F. (1991b): "Cognitive Levels of Evolution", in: The Cybernetics of Complex Systems, F. Geyer (ed.), (Intersystems, Salinas, CA), p. 75-91.
Heylighen F. (1993): "Selection Criteria for the Evolution of Knowledge", Proc. 13th Int. Cong. on Cybernetics (Int. Ass. of Cybernetics, Namur), p. 524-528.
Heylighen F. (1994): "World-Wide Web: a distributed hypermedia paradigm for global networking",Proc SHARE Europe, Spring 1994, (Geneva), p. 355-368.
Heylighen F. (1995): "(Meta)systems as constraints on variation", World Futures 45, p. 59-85.
Heylighen F. & Campbell D.T. (1995): "Selection of Organization at the Social Level", World Futures: the Journal of General Evolution 45, p. 181-212.
Jones W. P. (1986): "On the Applied Use of Human Memory Models", International Journal of Man-Machine Studies 25, p. 191-228.
Kovacs G.T., Storment C.W., Halks-Miller M. (1994): "Silicon-Substrate Microelectrode Arrays for Parallel Recording of Neural Activity in Peripheral and Cranial Nerves", IEEE Trans. Biomed..Engin. 41, p. 567.
Krol E. (1993): The Whole Internet (O'Reilly, Sebastopol, CA).
Maes P. (1994): "Agents that Reduce Work and Information Overload", Comm. of the ACM 37 (3).
Mayer-Kress G. & C. Barczys (1995): "The Global Brain as an Emergent Structure from the Worldwide Computing Network", The Information Society 11 (1).
Russell, P. (1995): The Global Brain Awakens: Our Next Evolutionary Leap (Miles River Press).
Salton G. & Buckley C. (1988): "On the Use of Spreading Activation Methods in Automatic Information Retrieval", Proc. 11th Ann. Int. ACM SIGIR Conf. on R&D in Information Retrieval (ACM), p. 147-160.
Stock G. (1993): Metaman: the merging of humans and machines into a global superorganism, (Simon & Schuster, New York).
Turchin V. (1977): The Phenomenon of Science. A cybernetic approach to human evolution (Columbia University Press, New York ).