💾 Archived View for spam.works › mirrors › textfiles › law › mar90_5.law captured on 2023-06-16 at 19:00:10.

View Raw

More Information

-=-=-=-=-=-=-

March 1990                                                        
                                                                  
           MANAGING POLICE BASIC TRAINING CURRICULUM                         
                                                                  
                              By

                       Rene A. Browett             
                      Curriculum Manager                                 
          Northern Virginia Criminal Justice Academy                        
                     Arlington, Virginia                                  
                                                                 
     The scenario might go something like the following.  It's 
test day at the police academy. The recruits have just completed 
their first major examination and are anxious to find out how 
they performed.  The training staff is sequestered in a large 
room, seated at long tables covered with stacks of exams in 
front of them.  Calculators are in clear evidence.  The grueling 
task of hand grading the exams begins.  Question by question, 
they trudge through the exam.  To validate test questions (to 
find out if more than half the class has missed a certain 
question), the leader calls out the questions by number to see 
how many of the graders have papers in which a student has missed 
a specific question.  Hands go up, a count is taken, numbers 
recorded, calculators figure averages--some right, some wrong.  
The process goes on for hours, even days.  Meanwhile, the 
students wait.                         
                           
     Does this often-repeated scene have to be?  No--not if the 
recruit curriculum testing-and-evaluation vehicle involves some 
computer assistance.  Such a system exists at the Northern 
Virginia Criminal Justice Academy (NVCJA).  The NVCJA basic 
training staff has developed a systematic process whereby days of 
effort by a training staff are reduced to less than 2 hours using 
one or two people.                                   
             
     This article will discuss how a basic police training 
curriculum can be quickly and efficiently managed with an 
effective, programmatic approach.  Using the NVCJA as a case 
study, the article will first provide some background about the 
academy and then will discuss the hardware, software and the 
process involved in managing the basic recruit curriculum.    

ACADEMY BACKGROUND                                                

     Established in 1965, the Northern Virginia Criminal Justice 
Academy provides training for over 25 criminal justice 
jurisdictions in the Northern Virginia area.  Staffed with 32 
full-time employees, the academy has a leadership cadre of six 
executives, including the director, all of whom are former police 
officers.  Along with a permanent support staff, the academy is 
augmented with officers from each of the participating 
jurisdictions which provide them as instructors on assignment for 
up to 3 years.  As one of nine regional academies in the State of 
Virginia, the academy is governed by a Board of Directors 
comprised of the chiefs of police, sheriffs and city/county 
managers from the larger participating jurisdictions.             

     The academy provides both recruit and inservice training for 
each of the participating police departments and sheriff's 
offices in the region.  Consequently, recruit training consists 
of subject matter leading to three certifications mandated by the 
Commonwealth of Virginia:  Basic Law Enforcement, Basic Civil 
Process-Court Security, and Basic Jailors.  Each year the academy 
graduates approximately 300 students after completion of a 14- to 
18-week course of instruction.                                    

     In order to graduate, each student must successfully 
complete all State- and academy-mandated tests and related 
requirements.  Developed by the Department of Criminal Justice 
Services (DCJS) headquartered in Richmond, VA, State-mandated 
requirements are commonly known as performance objectives (POs). 
These State mandates (POs) are the end result of a formal job 
task analysis, commissioned by the department (DCJS), where the 
various functions of police officers and sheriffs were 
identified.  Developed from this study were over 400 performance 
objectives which form the basis for State-mandated police 
training.  Each training academy must teach and test every 
performance objective and also retest any objective missed by 
students.                                                         

     The academy (NVCJA) also provides State-mandated minimum 
inservice requirements (MIR) training.  The inservice staff 
coordinates with DCJS to ensure that every 2 years, all officers 
receive at least 40 hours of State-mandated training, to include 
instruction in the law.  Taught almost exclusively by outside 
instructors and coordinated by a professional staff, the academy 
offers overs 100 inservice training classes.  While inservice 
training is not the focus of this article, information is 
provided to give a more complete profile of the academy. 

DEVELOPING THE ``STAR'' SYSTEM                                    

     Because the State requires that every student successfully 
pass all of the performance objectives, several problems 
immediately became evident when the academy staff first 
approached the problem of more efficient curriculum management. 
First, how could the academy successfully track each objective 
through the training process to ensure accountability?  Second, 
what type of test construction would be needed to assure the 
administration that mandated objectives would be adequately 
tested and validated?  Third, how could performance-based tests 
be graded within a few hours, not a few days?                     

     With these basic questions in mind, the recruit staff 
concluded that a computer application might offer a workable 
solution.  After preliminary analysis and over 5 years of 
refinement, the Student Testing and Records (STAR) System was 
developed and is presently used at the academy.  With this 
program as the basic software package, the testing system 
incorporates several components:                                  

     *  An optical mark reader (SCANTRON), which automatically 
        scores each exam and feeds the raw data directly into a
        computer  

     *  An additional software program which provides both
        database and spreadsheet capabilities

     *  An IBM compatible personal computer with 640K RAM memory
        and a hard drive, and finally

     *  A laser printer for letter quality reports                   

     This system costs less than $5,000.  By using this 
relatively simple but highly effective system, the curriculum 
manager is now able to better manage the basic training 
curriculum from tracking to testing to validation of each 
State-mandated performance objective.                             

MANAGING THE PROCESS                                              

     At the core of the academy's curriculum and testing system 
are the POs.  Simply put, the tests must ensure that each student 
masters each State-mandated PO.  Thus, test construction and 
administration are vital to the integrity of the academy's 
curriculum management process.  Performance-objective 
accountability and the testing process are the primary 
responsibilities of the academy's curriculum manager.  From 
lesson plan review to test construction and administration, the 
curriculum manager is the academy's point man with regard to 
accuracy and accountability.                           

Constructing Tests                                                

     All basic training examinations are constructed by the 
curriculum manager.  It is also his responsibility to analyze and 
validate all test results for each recruit.  The testing process 
begins with the basic lesson plan, which is then reviewed and 
approved by academy management.  Written by staff instructors, 
each lesson plan must be revised and updated at the end of each 
training session.  They must also contain those specific POs 
mandated by the State and appropriate for that block of 
instruction.  Test questions flow from and can be directly 
tracked to POs found in each lesson plan, thus assuring test 
accountability.  Staff instructors, accountable to both the 
students and the curriculum manager, are responsible for ensuring 
each PO is adequately taught.  Student performance, at test 
time, usually will reflect whether this situation has, in fact, 
occurred.                                                         

     At the end of each specified testing time period, the 
curriculum manager begins to prepare a test that spans several 
disciplines and many instructors.  How the test construction 
takes place mechanically is simple and is coordinated by the 
curriculum manager.  First, he speaks with all instructors to 
verify that their test questions, which are based on the mandated 
POs, have been taught and are part of a pre-existing database.    

     Second, the curriculum manager constructs a rough-draft test 
based on pertinent subject area questions stored in the database.  
The draft exam is then reviewed by all the respective instructors 
for their final updates and edits.  This phase of the process is 
accomplished with a high degree of attention to exam security.  
At this point, if an instructor did not teach a specific 
objective, the instructor advises the curriculum manager who 
deletes that PO from the current exam.  It will, however, be 
tested on a later exam, so as to comply with State mandates.      

     Third, the draft exam is then edited by the curriculum 
manager based on specific verbal and written feedback from each 
instructor.  The final exam is then constructed, with the rough 
draft copy kept on file for documentation and accountability.     

Administering Tests                                               

     To ensure uniformity and test security, all exams are passed 
out simultaneously to proctoring staff members who immediately 
take them to the test sites.  Proctors physically remain at each 
test site for the exam's duration to ensure test integrity.  Once 
the tests are passed out, a staff member reads a test cover sheet 
containing complete test instructions.  Once the instructional 
sheet has been read, the students begin their exams.              

     The tests are primarily multiple choice with very few 
true-and- false questions.  Students fill out their answers on an 
answer sheet with a #2 pencil so that it can easily be read by 
the optical mark reader.                                          

     When each recruit section is finished, all exams are 
returned and accounted for by the curriculum manager.  A single 
missing exam is treated as a compromise to the test's integrity 
and the results are then deemed invalid.  This has yet to happen 
at the academy while using this system.                    

Scoring Tests                                                     

     To score each examination, the assistant director for basic 
training and the curriculum manager work as a team to complete 
the effort.  First, the curriculum manager determines each answer 
sheet is properly completed.  If an answer sheet is incomplete, 
the recruit officer is called in and asked to make the required 
corrections.  This pertains only to basic identification 
information on the form and not to incomplete test answers. 
Should a defective answer sheet get into the system, the computer 
will automatically reject it when it is scanned.  Multiple 
answers, unclear erasures, or answer spaces left blank can cause 
an answer sheet to be defective.  In each case, the computer 
indicates the nature of the problem, the location of the problem, 
and will inquire what the user wishes it to do regarding the 
defect.                                                           

     To prepare the system for the grading mode, a master answer 
sheet (previously prepared by the curriculum manager from the 
master exam) is scanned.  This sheet will provide the test basis 
from which all student answer sheets will be graded.  Each 
student's answer sheet is then quickly fed into the scanner, with 
the data automatically stored on the recruit section's 
information disk.  The scanning process takes approximately 3-5 
minutes for a section of 25 to 30 recruits.  After the scanning 
step is completed, the computer produces a raw scores average 
sheet which is a rank-order listing of each class member based on 
the computer-generated averages of correct responses.  This 
immediately shows the curriculum manager what the statistical 
range of the class is and if he has any individual failures to 
review.                                                           

     The next step in the grading process is item analysis and 
test validation.  To determine the relative fairness of each exam 
question, the computer automatically produces a distractor 
analysis document.  The computer automatically views each 
question, and where 50% or more of the class get a question 
wrong, the question is reviewed and the instructor consulted.  If 
he or she feels the material was adequately covered, the question 
will remain.  If the question is tough, but fair, it stays in the 
exam.  However, where there exists reasonable doubt that the 
students were genuinely confused by the question, it is 
eliminated from the overall test score.  The benefit of a doubt 
is always given to the student.  After all the eliminated 
questions are determined, they are subtracted from the original 
total to formulate a net basis for computing the final test 
scores.  Students must score 70% or better to pass.      

Generating Reports                                                

     As an integral part of the process, the computer generates 
several other reports.  First, it produces a subject mastery 
report (a report card) to each student which tells them how they 
did in each subject area.  Second, it produces the same report 
but in a cumulative format, which is used by staff members for 
counseling and remedial training.  Third, a test answer sheet is 
generated which tells the recruits not only the correct answer 
but also what answer they put on their answer sheet.  This report 
also indicates that questions are State-mandated and require 
retesting if missed.  This capability provides a vehicle that    
allows identification of missed POs and retest on a timely basis 
by the curriculum manager, a procedure required by the State.     

     In short, using the software in the STAR program, the 
computer can generate any aspects of the testing process into a 
hard copy for the student staff member within an average of 2 
hours--from start to finish.                                      

Maintaining Security                                             

      The testing, grading, question database, lesson plans and 
section data records are maintained on both floppy disk and hard 
drive.  As an added security feature, the curriculum manager's 
office is locked during non-office hours and backup disks and 
access codes are secured.  Hard copies of exams, rough drafts, 
and actual lesson plans are likewise kept secure.                 

CONCLUSION                                                        

     A well-managed curriculum begins with a good job task 
analysis and performance objectives that arise from such 
analysis.  In turn, lesson plans and student activities should be 
based on those performance objectives and must be tested 
accurately.  Unfortunately, all too often, testing and tracking 
of such objectives get so cumbersome that administrators of an 
academy or educational institution do what they can, not what 
they should.  However, if applied meaningfully to the task, the 
computer offers welcome relief to training administrators.  The 
NVCJA has, over the years, tried to develop and refine a process 
that adequately manages a complex curriculum process without 
hamstringing the staff--a compromise that is working very well.