💾 Archived View for aphrack.org › issues › phrack25 › 4.gmi captured on 2022-03-01 at 15:36:18. Gemini links have been rewritten to link to archived content
⬅️ Previous capture (2021-12-03)
-=-=-=-=-=-=-
==Phrack Inc.== Volume Three, Issue 25, File 4 of 11 =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= =-= =-= =-= S P A N =-= =-= =-= =-= Space Physics Analysis Network =-= =-= =-= =-= Brought To You by Knight Lightning =-= =-= =-= =-= March 15, 1989 =-= =-= =-= =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= Preface ~~~~~~~ In the spirit of the Future Transcendent Saga, I continue to bring forth information about the wide area networks. The information presented in this file is based primarily on research. I do not have direct access to SPAN other than through TCP/IP links, but this file should provide you with general information with which to properly use the Space Physics Analysis Network. Introduction ~~~~~~~~~~~~ The Space Physics Analysis Network (SPAN) has rapidly evolved into a broadly based network for cooperative, interdisciplinary and correlative space and Earth science data analysis that is spaceflight mission independent. The disciplines supported by SPAN originally were Solar-Terrestrial and Interplanetary Physics. This support has been expanded to include Planetary, Astrophysics, Atmospherics, Oceans, Climate, and Earth Science. SPAN utilizes up-to-date hardware and software for computer-to-computer communications allowing binary file transfer, mail, and remote log-on capability to over 1200 space and Earth science computer systems in the United States, Europe, and Canada. SPAN has been reconfigured to take maximum advantage of NASA's Program Support Communication Network (PSCN) high speed backbone highway that has been established between its field centers. In addition to the computer-to-computer communications which utilizes DECnet, SPAN provides gateways to the NASA Packet Switched System (NPSS), GTE/Telenet, JANET, ARPANET, BITNET and CSNET. A major extension for SPAN using the TCP/IP suite of protocols has also been developed. This file provides basic information on SPAN, it's history, architecture, and present guidelines for it's use. It is anticipated that SPAN will continue to grow very rapidly over the next few years. Several existing wide-area DECnet networks have joined with SPAN to provide a uniform internetwork structure and more will follow. History Of The SPAN and the Data Systems Users Working Group (DSUWG) ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ A considerable evolution has occurred in the past two decades in the way scientific research in all disciplines is done. This is particularly true of NASA where early research was centered around exploratory missions in which measurements from individual scientific instruments could be meaningfully employed to advance the state of knowledge. As these scientific disciplines have progressed, a much more profound and interrelated set of questions is being posed by researchers. The result is that present-day investigations are generally much more complex. For example, within the space science community large volumes of data are acquired from multiple sensors on individual spacecraft or ground-based systems and, quite often, data are needed from many institutions scattered across the country in order to address particular physical problems. It is clear that scientific research during the late 1980s and beyond will be devoted to intense multi-disciplinary studies aimed at exploring very complex physical questions. In general, the need for researchers to exchange data and technical information in a timely and interactive way has been increasing. The problems of data exchange are exacerbated by the lack of standards for scientific data bases. The net result is that, at present, most researchers recognize the value of multi-disciplinary studies, but the cost in time and effort is devastating to their research efforts. This trend is antithetical to the needs of the NASA research community. SPAN is only one of many research networks that are just beginning to fill a need for access to remote capabilities that are not obtainable locally. In May of 1980 the Space Plasma Physics Branch of the Office of Space Science of NASA Headquarters funded a project at Marshall Space Flight Center (MSFC) to investigate ways of performing correlative space plasma research nationwide on a daily basis. As a first step, a user group was formed called the Data Systems Users Working Group (DSUWG) to provide the space science community interaction and direction in the project. After the first meeting of the DSUWG in September 1980, it was decided that the approach would be to design, build, and operate a spacecraft mission independent science network as a test case. In addition, the construction of the system would be designed to use existing data analysis computer systems at space physics institutions and to take full advantage of "off-the-shelf" software and hardware. The Space Physics Analysis Network (SPAN) first became operational in December 1981 with three major nodes: o University of Texas at Dallas o Utah State University o MSFC Since that time it has grown rapidly. Once operational, SPAN immediately started to facilitate space-data analysis by providing electronic mail, document browsing, access to distributed data bases, facilities for numeric and graphic data transfer, access to Class VI machines, and entry to gateways for other networks. The DSUWG continues to provide guidance for SPAN growth and seeks to identify, promote, and implement appropriate standards for the efficient management and exchange of data, related information, and graphics. All SPAN member organizations are expected to participate in the DSUWG. The basic composition of the DSUWG is a representative scientist and computer systems manager (who has the networking responsibility) at each of the member institutions. DSUWG meetings are held regularly at approximately nine month intervals. The DSUWG is structured along lines conducive to addressing major outstanding problems of scientific data exchange and correlation. There is a chairman for each subgroup to coordinate and focus the group's activities and a project scientist to oversee the implementation of the DSUWG recommendations and policies. The working group itself is divided into several subgroups which address issues of policy, networking and hardware, software and graphics standards, and data base standards. The DSUWG is a dynamic, evolving organization. We expect members to move in (or out) as appropriate to their active involvement in data related issues. We also realize that at present SPAN and the DSUWG are dealing with only a limited portion of the whole spectrum of problems facing the NASA research community. As present problems are solved, as the network evolves, and as new issues arise, we look to the DSUWG to reflect these changes in it's makeup, structure, and focus. The SPAN is currently managed by the National Space Science Data Center (NSSDC) located at Goddard Space Flight Center (GSFC). All SPAN physical circuits are funded by the Communication and Data Systems Division at NASA Headquarters. Personnel at the NSSDC facility, at the NASA SPAN centers, and the remote institutions work in unison to manage and maintain the network. Network Configuration and Evolution ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ The initial topology for SPAN was a modified star where all communication with the remote institutions came to a major central switching or message routing node at MSFC. This topology served the network well until many new nodes were added and more scientists became accustomed to using the network. As data rate demands on the network increased, it was apparent that a new topology using lines with higher data rates was needed. Toward this end, a new communication architecture for SPAN was constructed and implemented. The current structure of SPAN in the United States is composed of an interconnected four-star, mesh topology. Each star has, as its nucleus, a SPAN routing center. The routing centers are located at GSFC, MSFC, Jet Propulsion Lab (JPL), and Johnson Space Center (JSC). The routing centers are linked together by a set of redundant 56 kbps backbone circuits. Tail circuits, at speeds of 9.6 kbps (minimum line speed), are connected to each routing center and and into the SPAN backbone. Most remote institutions have local area networks that allow a number of different machines to be connected to SPAN. Regardless of a machine's position in the network, all computers on SPAN are treated logically equal. The main goal of the new SPAN architecture is for a node that is located across the country through two routing centers to be as transparently accessible as a SPAN node sharing the same machine room with the originating system. This ease of use and network transparency is one of SPAN's greatest assets. The new configuration allows for rapid expansion of the network via the addition of new tail circuits, upgrade to existing tail circuits, and dynamic dialing of higher data-rate backbone circuits Implementation of this new configuration began in July 1986, and the new topology was completed in November 1986, although there are new circuits being added on a continuing basis. It is expected that a fifth routing center located at Ames Research Center. Nearly all of the machines on SPAN are linked together using the commercially available software package DECnet. DECnet allows suitably configured computers (IBM-PCs and mainframes, SUN/UNIX workstations, DEC/PROs, PDPs, VAXs, and DECSYSTEMs) to communicate across a variety of media (fiber optics, coax, leased telephone lines, etc.) utilizing a variety of low level protocols (DDCMP, Ethernet, X.25). There are also several institutions that are connected through Janus hosts which run more then one protocol. SPAN links computers together and touches several other networks in the United States, Europe, and Canada that are used for data analysis on NASA spaceflight missions and other NASA related projects. At this time, there are well over 1200+ computers that are accessible through SPAN. DECnet networks has been accomplished by the unprecedented, successful cooperation of the network management of the previously separate networks. For example, the International High Energy Physics Network (HEPNET), the Canadian Data Analysis Network (DAN) and the Texas University Network (TEXNET) now have nonconflicting network addresses. Every node on each of these networks is as accessible to SPAN users as any other SPAN node. The mutual cooperation of these WANs has given enhanced capabilities for all. There are several capabilities and features that SPAN is developing, making it unique within the NASA science community. The SPAN system provides remote users with access to science data bases and brings scientists throughout the country together in a common working environment. Unlike past NASA mission networks, where the remote sites have only remote terminals (supporting one person at the remote site at a time), SPAN supports many users simultaneously at each remote node through computer-to-remote computer communications software. Users at their institutions can participate in a number of network functions involving other remote computer facilities. Scientific papers, data and graphics files can easily be transferred between network nodes. This significantly reduces the time it takes to perform correlative work when authors are located across the country or ocean. As an introduction to SPAN's network wide capabilities. More advanced users are referred to the DEC DECnet User's Manual. SPAN will continue to be used as a test case between NASA science investigators with the intent of exploring and employing modern computer and communication technology as a tool for doing NASA science research. This can be accomplished because SPAN is not a project dependent system that requires a static hardware and software configuration for the duration of a mission. SPAN has provided a quick reaction capability for several NASA and ESA missions. Each of these missions needed to rapidly move near real-time ground and spacecraft observations to a variety of destinations for analysis and mission planning. Because of SPAN's great success, new NASA spaceflight missions are seriously looking into creating networks with similar capabilities that are internetworked with SPAN. Within the next few years, new developments in software and hardware will be implemented on SPAN that will continue to aid NASA science research. It is anticipated that SPAN will greatly improve its access to gateways into Europe and other locations throughout the world. As a natural evolution, SPAN will migrate toward the International Standards Organization's (ISO) Open Systems Interconnect (OSI) protocol as the software becomes available. It is expected that the ISO/OSI protocol will greatly enhance SPAN and increase the number of heterogeneous computer systems accessible. Security And Conduct On The Network ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ Misconduct is defined as: 1. Any unauthorized access or use of computers on the network, 2. Attempts to defeat computer security systems (e.g. violating a captive account), 3. Repeated login failures to computers or privileged accounts to which the user is not authorized to use, 4. Massive file transfers from a given site without prior consent and coordination with the appropriate SPAN routing centers. The network is monitored very closely, and it is relatively simple to spot an attempted break-in and then track down the source. When a violation is found, the matter will be reported to the DSUWG steering committee and the SPAN line will be in immediate danger of being disconnected. If the situation cannot be resolved to the satisfaction of both the DSUWG steering committee and network management, the SPAN line to the offending site will be reviewed for the possibility of permanent disconnection. In short, NASA pays for the communications lines and will not tolerate misconduct on the network. SPAN Network Information Center (SPAN-NIC) ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ The SPAN-NIC is located at the National Space Science Data Center in Greenbelt, Maryland. The purpose of the SPAN-NIC is to provide general user services and technical support to SPAN users via telephone, electronic mail, and postal mail. As SPAN has grown exponentially over recent years, it was realized that a central organization had to be developed to provide users with technical assistance to better utilize the resources that the network provides. This is accomplished by maintaining and distributing relevant technical documents, providing user assistance on DECnet related questions, monitoring traffic on the network, and maintaining an online data base of SPAN node information. More specific information on becoming a SPAN site, beyond that provided in this document, can also be obtained through SPAN-NIC. The SPAN-NIC uses a VAX 8650 running VMS as its host computer. Users wishing to use the online information services can use the account with the username SPAN_NIC. Remote logins are capable via SET HOST from SPAN, TELENET from ARPANET and by other procedures detailed later. SPAN-NIC DECnet host address: NSSDCA or 6.133 SPAN-NIC ARPANET host address: NSSDC.ARPA or 128.183.10.4 SPAN-NIC GTE/TELENET DTE number: 311032107035 An alternative to remote login is to access online text files that are available. These text files reside in a directory that is pointed to by the logical name "SPAN_NIC:". Example commands for listing this directory follow: From SPAN: $ DIRECTORY NSSDCA::SPAN__NIC: From ARPA: FTP> ls SPAN__NIC: The available files and a synopsis of their contents can be found in the file "SPAN_NIC:SPAN_INDEX.TXT". Once a file is identified, it can be transferred to the remote host using the VMS COPY command, or the FTP GET command. It is important to note that this capability will be growing significantly not only to catch up to the current SPAN configuration but also keep current with its growth. DECnet Primer ~~~~~~~~~~~~~ The purpose of the SPAN is to support communications between users on network nodes. This includes data access and exchange, electronic mail communication, and sharing of resources among members of the space science community. Communication between nodes on the SPAN is accomplished by means of DECnet software. DECnet software creates and maintains logical links between network nodes with different or similar operating systems. The operating systems currently in use on SPAN are VAX/VMS, RSX, and IAS. DECnet provides network control, automatic routing of messages, and a user interface to the network. The DECnet user interface provides commonly needed functions for both terminal users and programs. The purpose of this section of the file is to provide a guide on the specific implementation of DECnet on SPAN and is not intended to supercede the extensive manuals on DECnet already produced by DEC. DECnet supports the following functions for network users: 1. TASK-TO-TASK COMMUNICATIONS: User tasks can exchange data over a network logical link. The communicating tasks can be on the same or different nodes. Task-to- task communication can be used to initiate and control tasks on remote nodes. 2. REMOTE FILE ACCESS: Users can access files on remote nodes at a terminal or within a program. At a terminal, users can transfer files between nodes, display files and directories from remote nodes, and submit files containing commands for execution at a remote node. Inside a program, users can read and write files residing at a remote node. 3. TERMINAL COMMUNICATIONS: RSX and IAS users can send messages to terminals on remote RSX or IAS nodes. This capability is available on VMS nodes by using the PHONE utility. 4. MAIL FACILITY: VMS users can send mail messages to accounts on remote VMS nodes. This capability is currently available for RSX and IAS nodes but is not supported by DEC. There are slight variations for RSX and IAS network mail compared to VMS mail. 5. REMOTE HOST: VMS, RSX, and IAS users can log-on to a remote host as if their terminals were local. Network Implementations For DECnet ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ The SPAN includes implementations for RSX, IAS and VAX/VMS operating systems. DECnet software exists at all the SPAN nodes and it allows for the communication of data and messages between any of the nodes. Each of the network nodes has a version of DECnet that is compatible with the operating system of that node. These versions of DECnet have been presently developed to different extents causing some nodes to have more or less capabilities than other nodes. The version or "phase" of the DECnet, as it is called, indicates the capability of of that node to perform certain levels of communication. Since RSX and IAS implementations are almost identical, they are described together. Users need not have any special privileges (VAX/VMS users will need the NETMBX privilege on their account) to run network tasks or create programs which access the network. However users must supply valid access control information to be able to use resources. The term "access control" refers to the user name and password of an account (local or on a remote node). Online system documentation is a particularly important and valuable component of DEC systems. At the present, SPAN is comprised almost completely of DEC systems. An extensive set of system help files and libraries exists on all the SPAN DEC nodes. The HELP command invokes the HELP Utility to display information about a particular topic. The HELP utility retrieves help available in the system help files or in any help library that you specify. You can also specify a set of default help libraries for HELP to search in addition to these libraries. Format: HELP [keyword [...]] On many systems, new users can display a tutorial explanation of HELP by typing TUTORIAL in response to the "HELP Subtopic?" prompt and pressing the RETURN key. Utilities for DECnet-VAX ~~~~~~~~~~~~~~~~~~~~~~~~ VAX terminal users have several utility programs for network communications available from the VMS operating system. Documentation for most of these utilities can be found in the Utility Reference Manual of the VAX/VMS manual set, and each utility has extensive online help available. The following descriptions offer a brief introduction to these utilities: MAIL: The VAX/VMS mail utility allows you to send a message to any account or to a series of accounts on the network. To send a message, you must know the account name of the person you wish to contact and his node name or node number. (This will be covered more extensively later in this file). FINGER: The DECUS VAX/VMS Finger utility has been installed on a number of SPAN VAX/VMS systems. Finger allows a user to see who is doing what, both on his machine and on other machines on the network that support Finger. Finger also allows a user to find information about the location and accounts used by other users, both locally and on the network. The following is an example session using the FINGER utility. $ FINGER NSSDCA VAX 8600, VMS V4.3. Sunday, 28-Sep-1986 19:55,4 Users,0 Batch. Up since Sunday, 28-Sep-1986 14:28 Process Personal name Program Login Idle Location HILLS H.Kent Hills Tm 19:02 NSSDC.DECnet _RTA4: Dr. Ken Klenk Tm 17:55 NSSDC.DECnet _NVA1: Michael L. Gough Mail 15:13 SPAN Man Joe Hacker Finger 17:33 bldg26/111 $ FINGER SWAFFORD@NSSDCA [NSSDCA.DECnet] NSSDCA VAX/VMS, Sunday, 28-Sep-1986 19:55 Process Personal name Program Login Idle Location SPAN Man Finger 17:33 Logged in since: Sunday, 28-Sep-1986 17:33 Mail: (no new mail) Plan: Joe Hacker, SPAN Hackers Guild Telephone: (800)555-6000 If your VAX supports VMS Finger, further information can be found by typing HELP FINGER. If your system does not currently have the FINGER utility, a copy of it is available in the form of a BACKUP save set in the file: NSSDCA::SPAN_NIC:FINGER.BCK PHONE: The VAX/VMS PHONE utility allows you to have an interactive conversation with any current user on the network. This utility can only be used on video terminals which support direct cursor positioning. The local system manager should know if your terminal can support this utility. To initiate a phone call, enter the DCL command PHONE. This should clear the screen and set up the phone screen format. The following commands can be executed: DIAL nodename::username Places a call to another user. You must wait for a response from that user to continue. DIAL is the default command if just nodename::username is entered. ANSWER Answers the phone when you receive a call. HANGUP Ends the conversation (you could also enter a CTRL/Z). REJECT Rejects the phone call that has been received. DIR nodename:: Displays a list of all current users on the specified node. This command is extremely useful to list current users on other nodes of the network. FACSIMILE filename Will send the specified file to your listener as part of your conversation. To execute any of these commands during a conversation, the switch hook character must be entered first. By default, that character is the percent key. REMOTE FILE ACCESS: DCL commands that access files will act transparently over the network. For example, to copy a file from a remote node: $copy From: node"username password"::disk:[directory]file.lis To: newfile.lis This will copy "file.lis" in "directory" on "node" to the account the command was issued in and name it "newfile.lis". The access information (user name and password of the remote account) is enclosed in quotes. Note that you can also copy that same file to any other node and account you desire. For another example, to obtain a directory listing from a remote node, use the following command: $dir node::[directory] (if on the default disk) Utilities for DECnet-11M/DECnet-IAS ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ There are certain DECnet functions that can only be done on nodes that have the same type of operating systems, such as the MPB, TRW, SPRL, LASR, and UTD nodes all with an RSX-11M operating system. The capabilities offered to the RSX DECnet user can be broken down into two major categories: those functions for terminal users and those functions for FORTRAN programmers. DECnet-11M terminal users have several utility programs available to them which allows logging onto other machines in the network, file transfers, message communication, and network status information. REMOTE-LOGON: The REMOTE-LOGON procedure allows a user at a node to log-on to another node in the network. This capability is also called virtual terminal. The "SET /HOST=nodename" command allows the user to log-on to adjacent nodes in the network from a DECnet-11M node. This command is initiated by simply typing "SET /HOST=nodename". The "SET HOST" command on the SPAN-VAX also allows you to log-on to adjacent nodes. NETWORK FILE TRANSFER: NFT is the Network File Transfer program and is part of the DECnet software. It is invoked by typing NFT <CR> to file = from file or by typing NFT to file = from file. Embedded in the file names must be the node name, access information, and directory if it is different than the default conventions. Also note that file names can only be 9 (nine) characters long on RSX systems. Therefore, VAX/VMS files with more than 9 characters will not copy with default-file naming. In such a case you must explicitly name the file being copied to an RSX system. The following structure for the file names must be used when talking to the SPAN nodes with NFT. NODE/username/password::Dev:[dir.sub-dir]file.type The following NFT switches are very useful: /LI Directory listing switch. /AP Appends/adds files to end of existing file. /DE Deletes one or more files. /EX Executes command file stored on remote/local node. /SB Submits command file for execution (remote/local). /SP Spools files to the line printer (works only with "like" nodes). A particular use for NFT is for the display of graphics files on the network. It is important to note, however, that some device-dependent graphics files are not all displayable, such as those generated by IGL software. The graphic files generated by graphic packages that are displayable when residing at other nodes may be displayed by using the following input: NFT> TI:=SPAN/NET/NET::[NETNET.RIMS]D1364.COL Graphics files generated by IGL can be displayed by running either REPLAY or NETREP programs (see the net-library documentation). TERMINAL COMMUNICATIONS: TLK is the Terminal Communications Utility which allows users to exchange messages through their terminals. TLK somewhat resembles the RSX broadcast command but with more capabilities. TLK currently works only between RSX-11 nodes and within a RSX-11 node. There are two basic modes of operation for TLK: The single message mode and the dialogue mode. The single message mode conveys short messages to any terminal in the same node or remote node. The syntax for this operation is: >TLK TARGETNODE::TTn:--Message-- To initiate the the dialogue mode type: >TLK TARGETNODE::TTn<cr> When you receive the TLK> prompt, you can enter a new message line. Graphics Display Utilities ~~~~~~~~~~~~~~~~~~~~~~~~~~ One of the main objectives of the SPAN system project is to accommodate coordinated data analysis without leaving one's institution. Therefore, there is a strong need to develop the ability to have graphic images of data from any node to be displayed by any other node. The current inability to display data on an arbitrary graphics device at any node has been quickly recognized. As general network utilities are developed to support the display of device dependent and independent graphic images, the handbook SPAN Graphics Display Utilities Handbook will serve to document their use and limitations. The graphics handbook is a practical guide to those common network facilities which will be used to support network correlative studies from the one-to-one to the workshop levels. For each graphics software utility the handbook contains information necessary to obtain, use, and implement the utility. Network Control Program ~~~~~~~~~~~~~~~~~~~~~~~ NCP is the Network Control Program and is designed primarily to help the network manager. However, there are some NCP commands which are useful for the general user. With these commands, the user can quickly determine node names and whether nodes are reachable or not. Help can be obtained by entering NCP>HELP and continuing from there. For a complete listing of all the NCP commands that are available to nonpriviledged users, refer to the NCP Utility manual on VAXs, and the NCP appendix of the DECnet-11M manual for PDPs. The following two commands are probably the most beneficial to users: $ RUN SYS$SYSTEM:NCP !on VAXs -or- > RUN $NCP !on PDPs NCP> SHOW KNOWN NODES !show a list of all nodes ! defined in the volatile data base NCP> SHOW ACTIVE NODES !show a list of only currently reachable Please note that the second command cannot be used on "end nodes", that is, nodes that do not perform at least DECnet Level I routing. In addition, only nodes in the user's area will be displayed on either Level I or Level II routers. In the case of end nodes, users should find out the name of the nearest Level I or II routing node and issue the following command: NCP> TELL GEORGE SHOW ACTIVE NODES Mail ~~~~ As briefly discussed earlier all SPAN DEC nodes have a network mail utility. Before sending a mail message, the node name and user name must be known. To send a message to the project manager, you would enter the following commands: $ MAIL MAIL> SEND To: NSSDCA::THOMAS Subj: MAIL UTILITY TEST Enter your message below. Press ctrl/z when complete ctrl/c to quit: VALERIE, OUR NETWORK CONNECTION IS NOW AVAILABLE AT ALL TIMES. WE ARE LOOKING FORWARD TO WORKING FULL TIME ON SPAN. THANKS FOR ALL YOUR HELP. FRED <CTRL/Z> MAIL>EXIT In order to send mail to more than one user, list the desired network users on the same line as the TO: command, separating each with a comma. Another way to accomplish this is to use a file of names. For example, in the file SEPAC.DIS, all SEPAC investigators on SPAN are listed: SSL::ROBERTS SSL::REASONER SSL::CHAPPELL SWRI::JIM TRW::TAYLOR STAR::WILLIAMSON The network mail utility will send duplicate messages to all those named in the above file by putting the file name on the TO: command line (TO: @SEPAC). A second option for the SEND command is to include a file name that contains the text to be sent. You will still be prompted for the To: and Subject: information. The following statements give a brief description of other functions of the MAIL utility: READ n Will list, on the terminal, the mail message corresponding to number n. If n is not entered, new mail messages will be listed. EXTRACT Saves a copy of the current message to a designated file. FORWARD Sends a copy of the current message to other users. REPLY Allows you to send a message to the sender of the current message. DIR Lists all messages in the current folder that you have selected. The sequence numbers can then be used with the READ command. DEL Delete the message just read. The message is actually moved to the WASTEBASKET folder until you exit the utility, when it is actually deleted. Therefore, you can retrieve a message that you have "deleted", up until you enter "exit" or ^Z to the MAIL> prompt. HELP Always useful if you're lost. Remote Node Information Files ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ All nodes on the SPAN are required to maintain two node specific information files in their DECnet default directories. The first file is a network user list file that contains specific information on each network user who has an account on the machine. At a minimum, the user list file should contain the name of the user, his electronic mail address, his account/project identifier, and his default directory. All of this information is easily obtained on VAX/VMS systems from the SYS$SYSTEM:SYSUAF.DAT file. (Note that the SYSUAF.DAT file is (and should be) only readable by the system manager.) The file is called USERLIST.LIS and resides in the node's DECnet default directory. A command procedure for creating this file is available in NSSDCA::SPAN_NIC:USERLIST.COM. This procedure should be executed from the SYSTEM account on the remote node for which it is to be compiled. Following is an example of displaying the USERLIST.LIS file on NSSDCA from a VAX/VMS system. $ TYPE NSSDCA::USERLIST Userlist file created at : 28-SEP-1986 22:06:01.71 Owner Mail Address Project Default Directory ---------------- ----------------- --------- ----------------- ROBERT HOLZER NSSDCA::HOLZER CD8UCLGU CDAW_C8USER:[HOLZER] RICHARD HOROWITZ NSSDCA::HOROWITZ ACQ633GU ACQ_USER:[HOROWITZ] CHERYL HUANG NSSDCA::HUANG CD8IOWGU CDAW_C8USER:[HUANG] DOMINIK P. IASCO NSSDCA::IASCONE PCDCDWPG CDAW_DEV:[IASCONE] ISADARE BRADSKY NSSDCA::IZZY DVDSARPG DAVID_DEV:[IZZY] WENDELL JOHNSON NSSDCA::JOHNSON DCSSARPG CODD_DEV:[JOHNSON] DAVID JOSLIN NSSDCA::JOSLIN SYSNYMOP OPERS_OPER:[JOSLIN] JENNIFER HYESONG NSSDCA::JPARK CAS130GU CAS_USER:[JPARK] HSIAOFANG HU NSSDCA::JUDY DVDSARPG DAVID_DEV:[JUDY] YOUNG-WOON KANG NSSDCA::KANG ADCSARGU ADC_USER:[KANG] SUSAN E. KAYSER NSSDCA::KAYSER ACQSARGU ACQ_USER:[KAYSER] DR. JOSEPH KING NSSDCA::KING ADM633MG ADM_USER:[KING] BERNDT KLECKER NSSDCA::KLECKER CD8MAXGU CDAW_C8USER:[KLECKER] KENNETH KLENK NSSDCA::KLENK PCDSARPG ADM_USER:[KLENK] Much like the user list, a node information listing is available for all nodes in their DECnet default account. This file is named NODEINFO.LIS. The following example is for the SSL node and should be taken as a template for the generic NODEINFO.LIS file that should be on each node in SPAN. $ TYPE SSL::NODEINFO Telenet Access To SPAN ~~~~~~~~~~~~~~~~~~~~~~ As SPAN grows, the number of users wishing to make use of its capabilities increases dramatically. Now it is possible for any user with a terminal and a 0.3 or 1.2 kbps modem to access SPAN from anywhere in the U.S. simply by making a local telephone call. There exists an interconnection between SPAN and the NASA Packet Switched Service (NPSS). The NPSS in turn has a gateway to the public GTE Telenet network which provides the local call access facilities. The user dials into one of Telenet's local access facilities and dials the NASA DAF (Data Access Facility) security computer. The user is then able to access SPAN transparently through the NSSDC or SSL machines. To find the phone number of a PAD local to the area you are calling from, you can call the Telenet customer service office, toll free, at 1-800-TELENET. They will be able to provide you with the number of the nearest Telenet PAD. The following outlines the steps that one must go through to gain access to SPAN through Telenet. 1. First dial into the local Telenet PAD. 2. When the PAD answers, hit carriage return several times until the '@' prompt appears. <CR><CR><CR> @ 3. Next enter the host identification address of the NASA DAF (security computer). This identification was not yet available at publication time, but will be made available to all users requesting this type of access. @ID ;32100104/NASA 4. You will then be prompted for a password (which will be made available with the identification above). PASSWORD = 021075 (Note: Tthe password will not be echoed) 5. Then type <CR>. You will be connected to the NASA DAF computer. The DAF will tell you which facility and port you succeeded in reaching, along with a "ready" and then an asterisk prompt: NASA PACKET NETWORK - PSCN TROUBLE 205/544(FTS 824)-1771 PAD 311032115056 *1 ready * All entries to the DAF must be in capital letters, and the USERID and PASSWORD will undoubtedly be echoed on the screen. *LOGON ENTER USERID> LPORTER ENTER PASSWORD> XXXXXXX ENTER SERVICE> SPANSSL NETWORK CONNECTION IN PROGRESS connected Alternatively, you may enter NSSDC for the "Service>" request. 6. You should now get the VMS "Username" prompt: Username: SPAN 7. You will then be prompted for the name of the SPAN host destination. For instance, if you are a Pilot Land Data System user on the NSSDC VAX 11/780, you would enter NSSDC and hit the carriage return in response to the prompt for host name. SPAN host name? NSSDC 8. Finally, continue with normal logon procedure for the destination host. The SPAN X.25 gateways have also been used extensively for internetwork communications to developing networks in Europe and Canada. The traffic from the United States to Europe was so extensive that a dedicated link between the GSFC and ESOC routing centers. This link became operational in January 1987. Configuration Of SPAN/TELENET Gateway ---------- | dial-up| | user | ---------- | ------------------------- | TELENET | ------------------------- | gateway ------------------------- | NPSS | ------------------------- | | ----------- ----------- | SSL | | NSSDC | | VAX 780 | | VAX 8650| ----------- ----------- | | ------------------------- | SPAN | ------------------------- | | | | ------ ------ ------ ------ |SPAN| |SPAN| |SPAN| |SPAN| |node| |node| |node| |node| ------ ------ ------ ------ SPAN/ARPANET/BITNET/Public Packet Mail Gateways ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ SPAN supports several gateways both to and from several major networks. The following gives the current syntax for forming an address to another user on another network. There are several similar gateways at other SPAN nodes that are not included in this list. Stanford is used here only as a typical example. If it is necessary for you to use the Stanford mail gateway on an occasional basis, you should obtain permission from the system manager on the STAR node (or any other non-NASA gateway node). Currently, there is no restriction on the NSSDCA gateway usage. SPAN-to-ARPANET: NSSDC Gateway . . To: NSSDCA::ARPA%"arpauser@arpahost" JPL Gateway . . . To: JPLLSI::"arpauser@arpahost" Stanford Gateway. To: STAR::"arpauser@arpahost" ARPANET-to-SPAN: NSSDC Gateway . . To: spanuser%spanhost.SPAN@128.183.10.4 JPL Gateway . . . To: spanuser%spanhost.SPAN@JPL-VLSI.ARPA Stanford Gateway. To: spanuser%spanhost.SPAN@STAR.STANFORD.EDU [Note: 128.183.10.4 is MILNET/ARPANET address for the NSSDC] SPAN-to-BITNET: NSSDC Gateway. . .To: NSSDCA::ARPA%"bituser%bithost.BITNET@CUNY.CUNYVM.EDU" JPL Gateway. . . .To: JPLLSI::"bituser%bithost.BITNET@CUNY.CUNYVM.EDU" Stanford Gateway .To: STAR::"bituser%bithost.BITNET@CUNY.CUNYVM.EDU" BITNET-to-SPAN: Stanford Gateway. . . . To: spanuser%spanhost.SPAN@SU-STAR.ARPA The following gateways allow users on a VAX that supports a connection to a public packet switch system (virtually anywhere in the world) to reach SPAN nodes and vice-versa. Note that this will transmit mail only to and from VAXs that support DEC PSI and PSI incoming and outgoing mail. SPAN-to-Public Packet VAX NSSDC Gateway. To: NSSDCA::PSI%dte_number::username SSL Gateway. . To: SSL::PSI%dte_number::username Public Packet VAX-to-SPAN node NSSDC Gateway. To: PSI%311032107035::span_node_name::username SSL Gateway. . To: PSI%311032100160::span_node_name::username It is possible for remote terminal access and mail between users on England's Joint Academic Network (JANET) and SPAN. JANET is a private X.25 network used by the UK academic community and is accessible through the two SPAN public packet switched gateways at MSFC and at the NSSDC. List Of Acronyms ~~~~~~~~~~~~~~~~ ARC - Ames Research Center ARPANET - Advanced Research Projects Agency network BITNET - Because It's Time Network CDAW - Coordinated Data Analysis Workshop CSNET - Computer Science Network DDCMP - DEC "level II" network protocol DEC - Digital Equipment Corporation DECnet - DEC networking products generic family name DSUWG - Data System Users Working Group ESOC - European Space Operations Center ESTEC - European Space Research and Technology Center GSFC - Goddard Space Flight Center GTE - General Telephone and Electic HEPNET - High Energy Physics Network INFNET - Instituto Nazional Fisica Nucleare Network ISAS - Institute of Space and Astronautical Science ISO/OSI - International Standards Organization/Open Systems Interconnection (network protocol) ISTP - International Solar Terrestrial Physics JANET - Joint Academic Network (in United Kingdom) JPL - Jet Propulsion Laboratory JSC - Johnson Space Center kbps - Kilobit per second LAN - Local area network LANL - Los Alamos National Laboratory MFENET - Magnetic Fussion Energy Network MILNET - Defence data network (originally part of ARPANET) MSFC - Marshall Space Flight Center NCAR - National Center for Atmospheric Research NFT - Network File Transfer (program on RSX/IAS systems) NIC - Network Information Center NPSS - NASA Packet Switched System (using X.25 protocol) NSSDC - National Space Science Data Center (at GSFC) PDS - Planetary Data System PSCN - Program Support Communications Network SESNET - Space and Earth Science Network (at GSFC) SPAN - Space Physics Analysis Network SSL - Space Science Laboratory (at MSFC) RVT - Remote virtual terminal program for RSX or IAS systems TCP/IP - Transmission Control Protocol/Internet Protocol Telenet - A public packed switched network owned by GTE TEXNET - Texas Network (Academic network) WAN - Wide area network X.25 - A "level II" communication protocol for packet switched networks _______________________________________________________________________________