💾 Archived View for spam.works › mirrors › textfiles › virus › hormones.vir captured on 2023-06-16 at 21:03:01.

View Raw

More Information

-=-=-=-=-=-=-

Date: Thu, 16 Mar 89 20:56:18 +0100
From: David Stodolsky <stodol@diku.dk>

Net Hormones: Part 1 - 
Infection Control assuming Cooperation among Computers

Copyright (c) 1989 David S. Stodolsky, PhD. All rights reserved.

1.   Abstract

A new type of infection control mechanism based upon contact tracing is 
introduced. Detection of an infectious agent triggers an alerting 
response that propagates through an affected network. A result of the 
alert is containment of the infectious agent as all hosts at risk 
respond automatically to restrict further transmission of the agent. 
Individually specified diagnostic and treatment methods are then 
activated to identify and destroy the infective agent. The title "Net 
Hormones" was chosen to indicate the systemic nature of this programmed 
response to infection.

2.   Introduction

A new type of infection control mechanism that is based upon network-
wide communication and that depends upon cooperation among computer 
systems is presented. Neither diagnosis nor treatment is necessary for 
the operation of the mechanism. The mechanism can automatically trigger 
responses leading to effective containment of an infection. The 
identification and destruction of the infectious agent is determined by 
individual actions or programs. This permits a highly desirable 
heterogeneity in diagnostic and treatment methods.

Definition: "Hormone . . . 1: a product of living cells that circulate 
in body fluids or sap and produces a specific effect on the activity of 
cells remote from its point of origin; especially one exerting a 
stimulatory effect on a cellular activity. 2: a synthetic substance 
that acts like a hormone (Webster's new collegiate dictionary, 1976)." 
The analogy here is between each network node or computer system and 
the cell. In biological systems hormones attach to specialized 
receptors on the cell surface resulting in cell activation. In the 
system described here, a match between a code in a system archive and a 
code delivered as part of an alerting message results in activation. 
Alerting messages circulated electronically serve the role of hormones. 

Epidemiology has traditionally had three major approaches to the 
control of infectious agents:

:1 - Treatment of the sick (e. g., penicillin)

:2 - Contact tracing (e. g., social-work notification programs, laws 
forcing the reporting of certain diseases and of contacts of infected 
persons)

:3 - Prevention (e. g., vaccination, public information campaigns)

In computer system terms:

:1 - Treatment of infections (e. g., various programs and manually 
installed patches and fixes)

:2 - Contact tracing (e. g., software "recall", and other manual 
operations)

:3 - Prevention (e. g., various programs for blocking virus 
replication, alerting users, and for logging suspicious events)

Contact tracing has been neglected with computer systems, although it 
could be argued it is much easier with computer systems than with 
biological systems. Currently such tracing depends upon people reading 
reports and determining if their system is subject to infection, 
performing diagnostic tests, determining a treatment method, obtaining 
software, and so on. This is chancy and time consuming, requiring most 
often people with the highest level of expertise. As computers and 
networks speed up, an infectious agent could spread through a network 
in hours or minutes. "Once a virus has infected a large number of 
computers on a network, the number of infected removable media elements 
will begin to skyrocket. Eventually, if the virus continues to go 
undetected, a stage is reached in which the probability of identifying 
and recovering all of the infected media is virtually zero (McAfee, 
1989)." An automated contact tracing system thus seems essential in the 
future if infectious agents are to be controlled.

3.   Threats

"The modification of an existing virus to incorporate a long term delay
(such as 6 months or even a year) coupled with a totally destructive
manipulation task (such as a FAT, Boot sector scribble followed by a
complete format) is a fairly simple task. Such an action would convert
even a crude virus strain such as the Lehigh 1 virus into a
devistating (sic) strain. (Eg the comment by Ken that the modified 
version of the Lehigh virus is now far more dangerous due to 
modification of the delay in activation of its manipulation task) 
(Ferbrache, 1989)."

Scott (1989) requested comments on:
 
"A little future speculation here... currently we seem to be fighting a
losing battle against virus detection and as viruses improve it's
unlikely that that will change.  If we want the capability to download
shareware, etc, from bulletin boards, etc, then we must assume that we
cannot check the software for a virus with 100% success before running
it.  In general, you can't know the output of a program given the
input without running it, except in special cases.
 
We can check for *known* viruses; but how long before shape-changing
and mutating viruses hit the scene that defeat all practical
recognition techniques?"

An inapparent infection could spread rapidly, with damage noted only 
much later. Consider a worm that is constructed to carry a virus. The 
worm infects a system, installs the virus and then infects other nearby 
systems on the net. Finally, it terminates erasing evidence of its 
existence on the first system. The virus is also inapparent, it waits 
for the right moment writes some bits and then terminates destroying 
evidence of its existence. Later the worm retraces its path reads some 
bits, then writes some bits and exits. The point is that an inapparent 
infection could spread quite widely before it was noticed. It also 
might be so hard to determine whether a system was infected or not, 
that it would not be done until damage was either immanent or apparent. 
This analysis suggests response to network-wide problems would best be 
on a network level. 

4.   Theory of operation

Computers generate (in the simplest case) random numbers which are used 
to label transactions. A transaction is defined as an interaction 
capable of transmitting an infectious agent. After each transaction 
both systems therefore have a unique label or code for that 
transaction. In the event that a system is identified as infected, the 
transaction codes which could represent transactions during which the 
agent was transmitted are broadcast to all other computers. If a 
receiving computer has a matching code, then that system is alerted to 
the possibility of the agent's presence, and can broadcast transaction 
codes accumulated after the suspect contact. This iterates the process, 
thus identifying all carriers eventually. The effect is to model the 
epidemiological process, thereby identifying all carriers through 
forward and backward transaction tracking (Stodolsky, 1979a; 1979b; 
1979c; 1983; 1986). 

5.   The process of infection control

The process can be broken down into routine and alerting operations. 
During routine operations, each file transfer is labeled in a way that 
does not identify the systems involved. These labels are time stamped 
(or have time stamps encoded in them). They are written into archives 
on each system, ideally write-once/read-many times devices or some 
other type of storage that could not easily be altered. 

Alerting procedures are invoked when an infectious agent is noted or 
when a suspect transaction code is received that matches one in the 
system's archive. The earliest time the agent could have arrived at the 
system and latest time (usually the moment the agent is noted or a 
received suspect transaction code is matched) it could have been 
transmitted from the system are used to delimit suspect transaction 
codes. These codes are broadcast to alert other systems to the 
potential presence of the agent.

In the simplest and most common case, if a system gets an alert that 
indicates, "You could have been infected at time one," then the system 
automatically packages the transaction codes between time one and the 
present time to generate a new alert indicating the same thing to other 
systems with which it has had contact. 

Another automatic response could be to immediately cut off 
communications in progress, thus reducing the risk of infection. A 
further benefit of such a reaction would be the possibility of 
disrupting the transfer of an infectious agent. Such a disrupted agent 
would be harmless and easily identified and evaluated. Reestablishment 
of communication could occur immediately with new procedures in force 
that could warn new users that an alert was in progress as well as 
limiting the type of transfers that could take place.

5.1.   Practical considerations

Direct identification, as opposed to identification through forward 
tracing notification, does not delimit effectively the earliest time 
that an agent could have been present on a system. Thus an alert from 
an originating system could include all transaction codes written prior 
to the identification (or some default value). This could generate 
excessive reaction on the network. This reaction could be controlled if 
another system in a later alert indicated it had originated the 
infection on the system originating the alert. Thus, protection of 
identity which reduces any inhibition about reporting infection is 
important. The type of reaction discussed here might be called a panic 
reaction, because an excessive number of systems might be notified of 
potential infection in the first instance. 

A more restricted response could be generated if persons at the alert 
originating system analyzed the causative agent, thereby hopefully 
establishing the earliest time the agent could have been present on 
that system. In this case, the suspect transactions could be delimited 
effectively and all systems that could have been infected would be 
notified, as would the system that had transmitted the agent to the 
system originating the alert (assuming one exists). Ideally, each 
notified system would be able to determine if it had received or 
originated the infection and respond accordingly. 

5.2.   Forward tracing assumption

Assume, however, that rapid response is desired. Each notified system 
would then react as if it had been notified of an infection transmitted 
to it. It would package the transaction codes that had been written 
later than the suspect transaction code it had received and issue a 
secondary alert. This forward tracing assumption would lead to quite 
effective control because of the exponential growth in the number of 
infected hosts in epidemics (and exponential growth of alerts resulting 
>From forward tracing). That is, a system can infect many others as a 
result of a single infective agent transmitted to it. Forward tracing 
would alert all systems that the alerting system could have infected. 
These newly alerted systems would also issue forward trace alerts, and 
this would continue until containment was reached under the forward 
tracing assumption.

5.3.   Backward tracing of suspect contacts and diagnosis

As a result of this rapid forward tracing response, it is likely that 
more active infections would be identified. The resulting new 
information could be used to more effectively characterize the life 
cycle of the agent, thereby hopefully permitting effectively delimited 
backward tracing. Also as a result of accumulated information, positive 
tests for the agent would become available. Once this stage had been 
reached the focus of action could shift from control of suspect 
transactions to control of transactions known to facilitate the 
transmission of the agent.

6.   Feasibility and Efficiency

Both technical and social factors play a key role in the operation of 
the control mechanism. Contact tracing is probably most effective for 
sparsely interacting hosts. The rate of transfer of the infectious 
agent as compared to the rate of transfer of the suspect transaction 
codes is also a critical factor. Recording of transactions can be 
comprehensive on computer networks, however, unregistered transactions 
will be a factor in most cases. Once the infectious agent has been 
identified, the type of transactions capable of transmitting the agent 
can be delimited. This could increase efficiency. 

6.1.   Social organization of alerts

Another major efficiency factor is errors in origination of alerts. 
Since protected messages would trigger network-wide alerts, it is 
important that false alarms are controlled effectively. On the other 
hand, failure to report an infection could permit an infectious agent 
to spread in an uncontrolled manner and could increase the number of 
systems unnecessarily alerted. Successful operation of the mechanism 
described above assumes voluntary cooperation among affected systems. 
This assumption could be relaxed by application of an enforcement 
mechanism. It would require substantially greater complexity and 
greater centralization of coordination. In other words, if cooperation 
was not forthcoming "voluntarily", users would likely be treated to a 
complicated, restrictive, and resource intensive mechanism that would 
be developed to enforce it. "Estimates of the damages inflicted by 
November's Internet infection alone ranged upward of $100 million . . . 
(McAfee, 1989)." Costs of this magnitude make it very likely that even 
expensive enforcement mechanisms will be developed if they are made 
necessary.

The simplest organizational strategy would assume that protection of 
identity was not needed, but this would also be likely to inhibit 
alerting. True anonymity, however, permits irresponsible behavior to go 
unchecked. A reputation preserving anonymity (pseudonymity) would be 
desirable to ensure both protection and accountability and thereby 
promote cooperation. Pseudonyms would best be the property of persons 
(in association with a computer system). 

Even sincere cooperation, however, would not eliminate inefficiencies 
resulting from false alarms or failure to alert. Both inadequate 
training and poor judgement are likely sources of these errors. If 
users realize that there are reputational costs associated with these 
failures, then they are likely to be motivated to minimize them. False 
alarms are already a major problem because of user inexperience and the 
high level of defects in widely used software. A reputational mechanism 
would motivate increased user education and more careful software 
selection, with a corresponding pressure on software publishers to 
produce well behaved and carefully documented products. 

6.2.   Enforcing cooperation

Crypto-protocols could be used to ensure that a non-cooperator could 
not communicate freely with others using the infection control 
mechanism. This type of communication limiting could be used routinely 
to ensure that a system requesting connection was not infected. In 
effect, systems would exchange health certificates before file 
exchanges, to ensure that they would not be infected. A system that 
could not show a health certificate could be rejected as a conversation 
partner due to risk of infection. This would no doubt enforce 
cooperation. The mechanism (Stodolsky, 1986) is beyond the scope of 
this note.

6.3.   Non-network transfers

While the discussion above has focused on transfers through networks, 
the same principles could be applied to disk or tape transfers. The 
originating system would write a transaction code on the medium with 
each file. Protection of identity would possibly be reduced under this 
type of transfer. Since there is no question about the directionality 
of transmission of an infectious agent in off-line transfers, non-
network transmission is likely to be easier to control. Several other 
factors, such as the rate of spread of the agent, are likely to make 
such infections less troublesome. 

7.   Summary and Benefits 

The idea behind Net Hormones is to make immanent danger apparent. More 
precisely Net Hormones permit the visualization of infection risk. 

7.1.   Control of unidentified infectious agents.

Net Hormones work by permitting isolation of infectious hosts from 
those at risk. Identification of the infectious agent is not required 
for action. Therefore, new and as yet unidentified agents can be 
effectively controlled.

7.2.   Rapid response

Hosts could automatically respond to alerts by determining if they had 
been involved in suspect contacts, and generate new alerts that would 
propagate along the potential route of infection. 
 
7.3.   Protection of identity

The mechanism could function without releasing the identity of an 
infected host. This could be crucial in the case an institution that 
did not wish it to be publicly know that its security system had been 
compromised, or in the case of use of unregistered software. More 
precisely, software obtain by untraceable and anonymous file transfers 
could be protected by this mechanism without release of users' 
identity.

7.4.   Distributed operation

Operation is not dependent upon a centralized register or enforcement 
mechanism. Some standardization would be helpful, however, and a way to 
broadcast alerts to all potential hosts would be valuable.

8.   References

Ferbrache, David J. (1989, February 10). Wide area network worms. 
VIRUS-L Digest, V. 2 : Issue 44. [<davidf@CS.HW.AC.UK> <Fri, 10 Feb 89 
11:45:37 GMT>]

McAfee, J. D. (1989, February 13). In depth: Managing the virus threat. 
Computerworld, 89-91; 94-96.

Scott, Peter. (1989, February 10). Virus detection. VIRUS-L Digest, V. 
2 : Issue 44. [PJS%naif.JPL.NASA.GOV@Hamlet.Bitnet 
<pjs@grouch.jpl.nasa.gov>. <Fri, 10 Feb 89 10:46:21 PST>]

Stodolsky, D. (1979a, April 9). Personal computers for supporting 
health behaviors. Stanford, CA: Department of Psychology, Stanford 
University.  (Preliminary proposal)

Stodolsky, D. (1979b, May 21). Social facilitation supporting health 
behaviors. Stanford, CA: Department of Psychology, Stanford University. 
(Preliminary proposal)

Stodolsky, D. (1979c, October). Systems approach to the epidemiology 
and control of sexually transmitted diseases. Louisville, KY: System 
Science Institute, University of Louisville. (Preliminary project 
proposal)

Stodolsky, D. (1983, June 15). Health promotion with an advanced 
information system. Presented at the Lake Tahoe Life Extension 
Conference. (Summary)

Stodolsky, D. (1986, June). Data security and the control of infectious 
agents. (Abstracts of the cross disciplinary symposium at the 
University of Linkoeping, Sweden: Department of Communication Studies). 

Webster's new collegiate dictionary. (1976). Springfield, MA: G. & C. 
Merriam

-------------------------------------------------------------

David Stodolsky                          diku.dk!stodol@uunet.UU.NET
Department of Psychology                       Voice + 45 1 58 48 86
Copenhagen Univ., Njalsg. 88                    Fax. + 45 1 54 32 11
DK-2300 Copenhagen S, Denmark                         stodol@DIKU.DK