💾 Archived View for spam.works › mirrors › textfiles › computers › disks3 captured on 2023-06-14 at 16:01:38.
-=-=-=-=-=-=-
???????????????????????????????????????????????????????????????? ? ? ? The Logical Structure, Organization, ? ? and Management of Hard Disk Drives ? ? ? ? by ? ? Steve Gibson ? ? GIBSON RESEARCH CORPORATION ? ? ? ? Portions of this text originally appeared in Steve's ? ? InfoWorld Magazine TechTalk Column. ? ? ? ???????????????????????????????????????????????????????????????? As our operating systems and application software have continued to grow in size, their memory requirements have increased steadily. A vital memory in our system is hard disk storage. Bound within the hard disk's structure lie the answers to questions like: What is a low level format? What does FDISK do? What is a hard disk partition and why does DOS limit us to 32 megabytes in a partition? What does it mean to have "lost cluster chains" or "cross-linked files?" What does it mean to have our disks "defragmented?" Let's explore MS-DOS and PC-DOS hard disk organization to answer these questions and others. The first stage in preparing any hard disk for operation is known as low level formatting. Low level formatting takes any hard disk from its virgin "fresh from the factory" state and prepares it for operation with a particular hard disk controller and computer system. Low level formatting divides each circular track into equal size SECTORS by placing SECTOR ID HEADERS at uniform positions around each track. The start of a sector ID is marked with a special magnetic pattern which cannot be generated by normal recorded data. This ADDRESS MARK allows the beginning of each sector to be uniquely discriminated from all recorded data. The sector ID information, which immediately follows the address mark contains each sector's Cylinder, Head, and Sector number which is completely unique for each sector on the disk. When the hard disk controller is late reading or writing to these disk sectors, it compares the sector's pre-recorded cylinder number to make sure that the heads haven't "mis-stepped" and that they're flying over the proper cylinder. It then compares the head number to verify that unreliable cabling is not causing an improper head to be selected and waits for the proper sector to start by comparing the pre-recorded sector number as it passes by with the sector number for which it is searching. Since many hard disk surfaces are not flawless, low level formatting programs include a means for entering the hard disk drive's defect list. The defect list specifies tracks (by cylinder and head number) that the manufacturer's sensitive drive certification equipment found to stray from the normal which indicates some form of physical flaw that might prevent data from being reliably written and read. The list of such defects is typically printed and attached to the outside of the drive. When these tracks are entered into the low level formatter, the defective tracks receive a special code in their sector ID headers which indicates that the track has been flagged as bad and cannot be used for any data storage. Later, as we shall see, high level formatting moves this defective track information into the system's File Allocation Table (FAT) to prevent the operating system from allocating files within these defective regions. When the low level format has been established, we have a completely empty drive, devoid of stored information, which can accept and retrieve data with the specification of any valid cylinder, head, and sector number. There's an important issue about the low level formatting of a hard disk which is frequently overlooked, but which can be quite important to appreciate. Since the hard disk controller works in intimate concert with its hard disk drive to transfer the data within its numbered sectors to and from the computer's memory, the exact details of the address mark, sector ID header, and rotational sector timing can be completely arbitrary for any controller and drive. Since these details are initially established when the drive receives its low level formatting, they are forever hence agreed upon by both the hard disk drive and the controller. But more importantly, there's absolutely no reason to assume that the relatively arbitrary low level formatting specifics used by any particular hard disk controller would be compatible with any other model of hard disk controller. In practice this means that differing makes or models of hard disk controllers are completely unable to read, write, or interpret the formatted information created by any other make or model of controller. Consequently, whenever it is necessary or desirable to exchange hard disk controllers, a complete backup of the hard disk's data, while attached to the initial controller, MUST BE followed by creating a new low level format with the new controller on the drive before any of the backed-up information can be restored to the drive with the new controller. So we've given our drives a low level format, since we see that it is this process which first establishes "communication" between a hard disk and its controller by creating 512-byte "sectors" where none existed before. Now lets take up the next phase of hard disk structuring: The hard disk PARTITION. The notion of hard disk (or "fixed disk" as IBM calls them) partitions was created to allow a hard disk based computer system to contain and "boot up" several completely different operating systems. Partitioning divides a single physical hard disk into multiple LOGICAL partitions. A birthday cake is divided into multiple pieces by slicing it radially whereas a hard disk's divisions are circular. For example, a drive's first partition might extend from cylinder zero through 299 with the second partition beginning on cylinder 300 and extending through 599. This circular partitioning is far more efficient since it minimizes the disk head travel when moving within a single partition. The partitions on a drive, even if there's only one, are managed by a special sector called the PARTITION TABLE which is located at the very beginning of every hard disk. It defines the starting and ending locations for each of the disk's partitions and specifies which of the partitions is to gain control of the system during system boot up. When the hard disk drive is booted a tiny program at the beginning of the partition table locates the partition which is flagged as being the "bootable partition" in the table and executes the program located in the first sector, the "boot sector," of that partition. This boot sector loads the balance of the partition's operating system then transfers control to it. Each partition on a hard disk is blind to the existence of any other. By universal agreement, the operation of software inside a partition is completely contained within the bounds of the partition. Adherence to this agreement prevents multiple operating systems from colliding and allows strange environments to cohabitate on a single hard disk. The sectors within a partition are numbered sequentially starting at zero and extending to the end of the partition. In kind with DOS's original belief that 640K of RAM would be more than we'd EVER need, there was a time in the not-so-distant past when a ten megabyte hard disk was an unheard of luxury and was considered huge. How could any single person ever fill up 10 megabytes? No way. Consequently DOS was designed to access sectors within its hard disk partition with a single sixteen-bit quantity. One "word" was set aside for the specification of partition sectors. As many of you know, a single sixteen-bit binary word can represent values from 0 through 65,535. So this limited a partition's total sector count to 65,536. Since hard disk sectors are 512 bytes long, a partition could contain 33,554,432 bytes. When you remember that binary megabytes are really 1,048,576 bytes each, that's exactly 32 megabytes. This is the origin of DOS's infamous 32 megabyte barrier. Today of course we have affordable drives with capacities well exceeding DOS's 32 megabyte limit. The industry has invented three solutions to this partition size dilemma. The first solution invented to the partition size problem utilizes DOS's inherent extendibility with external device drivers. Programs such as OnTrack's DISK MANAGER, Storage Dimensions' SPEEDSTOR, and Golden Bow's VFEATURE DELUXE utilize a clever trick to circumvent the 32 megabyte DOS limit: They trick DOS into believing that sectors are larger than 512 bytes! By interposing themselves between DOS and the hard disk, these partitioning device drivers lead DOS to believe that individual sectors are much larger than they really are. Then when DOS asks for one "logical" 4k-byte sector they hand DOS eight 512-byte physical sectors. This transforms the 65,536 sector count limit into a single partition containing more than 268 megabytes! The second solution was introduced by IBM's PC-DOS 3.3 operating system with its ability to allow DOS to have simultaneous access to multiple logical partitions on a single drive. With DOS 3.3, the standard FDISK command can establish any number of 32- megabyte or smaller partitions on a drive. While this doesn't create a single unified huge partition, it also doesn't require any external resident device drivers. The final solution has recently been introduced by Compaq Computer with their introduction of DOS 3.31. Being big enough to get away with sacrificing some software compatiblity, Compaq has redefined the way DOS numbers its partition sectors thereby removing the limitation at its source. So now our hard disks have a low level format, with "addressability" to the disk's individual physical sectors established. We have also defined and established partitions on our drive, which gives DOS a sub-range of the hard disk within which to build its filing system. Now let's examine the structure of MS-/PC-DOS filing systems. The following discussion also applies to DOS diskettes which aren't partitioned but otherwise have an identical structure. Let's begin by looking at the problem that DOS's filing system solves: Its task is to allow us, through the vehicle of DOS application programs, to create named collections of bytes of data, called files, and to help with their management by providing directories of these named files. The directory entry for any DOS file contains the file's name and extension, the date and time when the file was last written and closed, an assortment of Yes/No "attributes" which indicate whether the file has been modified since last backup, whether it can be written to, whether it's even visible in the directory, etc. The directory entry for the file also contains the address of the start of the file. We already know that hard disks are divided into numbered sectors 512 bytes in length. Since most of the files DOS manages are much larger than a single sector, disk space is allocated in "clumps" of sectors called clusters. Various versions of DOS utilize clusters of 4, 8 or 16 sectors each, or 2048, 4096, or 8192 bytes in length. When a hard disk is completely empty, its clusters of sectors are all available for storing file data. As files are created and deleted on the hard disk, a bookkeeping system is needed which keeps track of which clusters are in use by which existing files, and which clusters are still available for allocation to new or growing files. This is the vital role played by the File Allocation Table. The "FAT," as it's frequently called, is the table DOS uses to manage the allocation of space on the hard disk. As we know, the hard disk is arranged as a long stream of sectors. After being clumped together into clusters, it can be viewed as a long stream of clusters. Now picture a table consisting of a long stream of entries, with one entry in the table for each cluster on the disk. The first FAT table entry corresponds to the first hard disk cluster, and the last FAT entry corresponds to the last hard disk cluster. Now imagine that DOS needs to create a new text or spreadsheet file for us. It must first find a free cluster on the hard disk, so it searches through the File Allocation Table looking for an empty FAT table entry, which corresponds to an empty hard disk cluster. When DOS finds the empty table entry it memorizes its number, then places a special "end of chain" marker in the FAT entry to show that this cluster has been allocated and is no longer free for use. DOS then goes out to the sectors which comprise this cluster and writes the file's new data there. This is all great until the file grows longer than a single cluster of sectors. DOS now needs to allocate a second cluster for this file. So it once again searches through the File Allocation Table for a free cluster. When found, it again places the special "end of chain" marker in this cluster and memorizes its number. Now things begin to get interesting... and just a little bit tricky. Since files might be really long, consisting of thousands of individually allocated clusters, there's no way for DOS to memorize all of the clusters used by each file. So DOS uses each File Allocation Table entry to store the number of the file's next cluster! Following along with our example, after finding and allocating the second cluster for the growing file, DOS goes back to the first cluster's FAT entry where it had placed that first "end of chain" marker and replaces it with the number of the file's second cluster. If a third cluster were then needed, its FAT entry would be marked "not available" by placing the special "end of chain" marker in it, then this third cluster number would be placed into the second cluster's FAT entry. Get it? This creates a "chain" of clusters with each cluster entry pointing to the next one, and the last one containing a special "end of chain" entry which signals that the end of the file's allocation chain has been reached. Finally, when the file is "closed," an entry is created in a DOS directory which names the file and contains the number of the file's first cluster. Then, using that first cluster's FAT entry, the entire allocation "chain" can be "traversed" to find the clusters which contain the file's data. So now let's do a bit of review.... The allocation of file space within a DOS partition is recorded and maintained within DOS's File Allocation Tables (FATs). The FATs make up a map of the utilization of space on any floppy or hard disk with one entry in the FAT for each allocatable cluster of sectors. Each entry in the FAT can indicate one of four possible conditions for the clusters of sectors it represents: It can be unused and available for allocation, unused and marked as bad to prevent its use, in use and pointing to the next cluster of the file, or in use as the last cluster of a file. If each entry in the FAT points to the next, who points to the first entry? This is the role of the file's directory entry. It contains the name of the file, the file's exact length, the time and date of the file's last modification, file attribute flags, and the identity of file's first cluster. In a sense, a file's directory entry forms the head of the file's allocation chain with each link thereafter pointing to the next link in the chain. This system, while quite workable and efficient, does have its dangers. These dangers center around the fact that the FAT contains the ONLY record of disk space utilization and a stubborn failure to correctly read a single sector of the FAT could render hundreds of files unrecoverable. This danger explains the popularity of several utility programs which create a back-up copy of the File Allocation Table and Root Directory with each system boot-up. They provide some hope of recovery from the cataclysmic loss of the FAT's data. The original designers of DOS were aware of the importance of the FAT and do provide a duplicate copy immediately following the first, but its physical proximity to the original renders it little better than none, and DOS has long been notorious for failing to intelligently utilize this extra copy of FAT information even in the event of a primary FAT failure. (DOS 3.3 seems to be much smarter in this regard.) Important as FAT reliability is, it's not generally the prime source of DOS file corruption, since even with perfect data retrieval, it's still possible to scramble DOS's files like crazy. The primary cause of DOS file system troubles are user error, program bugs, and "glitches." The advent of TSR "rule breaking" resident multitasking-style software has further complicated the scene. When a new file is created or "opened," information about it is maintained inside DOS. The file's name, status, and first cluster are all held in internal tables. Then, as the file grows, free clusters are "checked out" of the File Allocation Table and allocated to the file's chain of clusters. Now here's the crucial fact which causes so much trouble: No matter how big the newly created file becomes, a directory entry for the file is ONLY created when the file is finally and properly CLOSED. Until then the file exists only as a chain of allocated clusters filled with the file's data. If anything occurs to prevent the error-free closing of this file we have a real problem because the file's data is occupying a chain of "checked out" disk clusters, but there is no anchoring directory entry to point to the first cluster in the chain! A chain of clusters without an anchoring directory entry is called a "lost chain." It exists, it contains data, but there's no record of the file's name, exact size, or purpose. Lost cluster chains are frequently created when programs abort abnormally, when TSR's crash the system suddenly, when the computer user forgets to write a TSR's files out to disk before shutting the system down, or when a task in a multi-tasking system is not terminated. (It's easy to forget that a file was left open in a suspended background task.) Additionally, any damage to DOS's root directory or subdirectories can "liberate" chains of lost clusters. DOS provides the CHKDSK (pronounced Check Disk) command to help its users keep an eye on just these sorts of problems. CHKDSK provides a comprehensive verification of DOS's filing system integrity and provides a means for straightening things out. When the CHKDSK command is given, the parentage of all cluster chains is checked, allocation chains are "followed" to be sure they don't cross over other chains (creating cross-linked files), and several other system integrity checks are performed. In the case of lost chains, CHKDSK will offer to convert these into files by anchoring them to the root directory. Then any suitable text editor can be used to open these new files for the sake of identifying them and moving them back to where they belong. Unfortunately the structure of DOS filing systems lacks the fundamental redundancy required to provide simple and error-free recovery from many forms of damage. Even the tools and techniques available from third party suppliers can't surmount these problems. The best bet is to understand DOS's weak spots, make certain that all opened files are closed successfully, perform a weekly CHKDSK command to collect accumulating file fragment "debris" and back up your hard disks regularly. "Disk Optimizers" which promise to increase the throughput and performance of old and well used hard disk drives number among the most popular of the general use hard disk utilities. We've seen how DOS's file allocation system operates. Files are composed of clusters which in turn are composed of sectors. And while the group of sectors which comprise a cluster are by definition contiguous, the cluster linking scheme which DOS employs allows a file's clusters to be scattered across the disk's surface. Since the file's directory entry specifies the file's first cluster, and each succeeding cluster entry in the file allocation table specifies the next one, the file's contents could be literally anywhere on the disk. The term "file fragmentation" refers to the condition where a file's clusters are not consecutively numbered. Let's first examine how a disk's files might become fragmented. When a file is deleted from a disk, its directory entry is flagged as unused and each cluster which the file occupied is flagged in the system's FAT as being free for use. If the surrounding clusters are still in use by other files, this creates a "hole" of free space in the disk. Now suppose that a new file is copied from a floppy disk onto the hard disk. As DOS reads the new file's data from the floppy, it must allocate space for this file on the hard disk. So each time another cluster of sectors is needed, DOS searches through the file allocation table to find the next available cluster. In our example, DOS would discover the clusters which had been freed by the first file we deleted and allocate them for use by the new file. Then, when all of the clusters in the free space hole had been used, DOS would be forced to continue its search deeper into the drive. When space was found further in, the file's contents would be partially stored near the beginning of the disk and partially nearer to the end. The file would then consist of at least two fragments. During the normal course of daily computer usage, many files are being constantly created, copied, extended, deleted, and replaced. When a wordprocessor creates an automatic backup file, the original file is typically renamed to identify it as a backup file and a new file is created. Every new file creation is an opportunity for fragmentation. The files which are being modified most often are most subject to extensive fragmentation since any search by DOS for a free file cluster is almost guaranteed to produce a new discontinuity. With continued use, it's typical for much of the disk's file data to become haphazardly scattered across the surface of the disk drive. But since DOS's cluster allocation scheme was specifically designed to manage such scattering, what's the problem? Any time the drive's head moves, two things occur: Time is consumed, and the drive experiences some mechanical wear and tear. If a file's data is scattered across the surface of the disk, the drive's head is forced to move a large distance many times to read a single file. If the file is a database whose records are being accessed at random, this excessive head motion can degrade the overall system performance tremendously and induce many other wear-related disk drive problems. The extra time wasted in cluster fragment chasing is directly proportional to the drive's average head access time. The prior generation of 65 to 80 millisecond stepping motor drives lose far more performance to fragmentation than the latest sub-28 millisecond drives. Disk optimizers like SoftLogic Solutions' DISK OPTIMIZER, Norton's SPEEDDISK, Central Point's COMPRESS, and Golden Bow's VOPT operate by physically rearranging the allocation of files on the disk. They relocate file cluster fragments while simultaneously updating the system's File Allocation Tables to reflect the new cluster locations. When finished, every file on the disk consists of a single contiguous run of consecutively numbered clusters. Once the disk drive's head has been positioned to the beginning of the file, the entire file can be read or randomly accessed with an absolute minimum of head motion. Besides improving the system's overall performance, file defragmentation minimizes the mechanical wear and tear placed upon the drive's hardware. If some disaster should befall your system's Root Directory or File Allocation Table, contiguous files are also much easier to find and recover than files with severe fragmentation. Since file fragmentation is a continually occurring fact of living with DOS, periodic defragmentation, like hard disk backup, should become part of every serious DOS user's regimen. - The End - Copyright (c) 1989 by Steven M. Gibson Laguna Hills, CA 92653 **ALL RIGHTS RESERVED **