Skip navigation
Skip Navigation
small header image
Click for menu... About NAEP... Click for menu... Subject Areas... Help Site Map Contact Us Glossary NewsFlash
Sample Questions Analyze Data State Profiles Publications Search the Site
NAEP Technical Documentation
The Nation's Report Card (home page)

Table of Contents  |  Search Technical Documentation  |  References

NAEP Scoring → Scoring NAEP Science Assessments

Scoring NAEP Science Assessments

      

Number of 2000 Constructed-Response Items by Score-Point Level

Number of 1996 Constructed-Response Items Rescored in 2000 by Score-Point Level

The NAEP science items that are not scored by machine are constructed-response items—those for which the student must write in a response rather than selecting from a printed list of multiple choices. Each constructed-response item has a unique scoring guide that identifies the range of possible scores for the item. To measure longitudinal trends in science, NAEP requires trend scoring—replication of scoring from prior assessment years—to demonstrate statistically that scoring is comparable across years.

Students' constructed responses are scored on computer workstations using an image-based scoring system. This allows for item-by-item scoring and online, real-time monitoring of science interrater reliabilities and the performance of each individual rater. In the 2000 assessment, some of these items—those that appeared in large-print booklets—required scoring by hand. The 2000 science assessment included 295 discrete constructed-response items. The total number of constructed responses scored was 4,398,021. The number of raters working on the science assessment and the location of the scoring are listed here:

Location of scoring activities, science assessment: 2000
Scoring location Start date End date Number of raters Number of scoring supervisors
Iowa City, Iowa 3/13/2000 6/04/2000 115 16
Tucson, Arizona 4/13/2000 4/29/2000 40 4
SOURCE: U.S. Department of Education, Institute of Education Sciences, National Center for Education Statistics, National Assessment of Educational Progress (NAEP), 2000 Science Assessment.

One unique aspect of the science assessment is the use of "hands-on" tasks that are given to students as a part of the assessment. Each student who performs a hands-on task is given a kit with all of the materials needed to conduct the experiment. For the 2000 assessment, a total of 9 hands-on tasks (3 per grade) originally designed for the 1996 assessment were chosen for use, although the actual kits used by the students were new. During scoring of the hands-on task items, raters actually performed the experiment as part of their training. Each student's experiment was scored as a unit because of the inter-connectivity of the questions the student had to answer.

Each item's scoring guide identifies the range of possible scores for the item and defines the criteria to be used in evaluating student responses. During the course of the project, each team scores the items using a 2-, 3-, 4-, or 5-point scale as outlined below:

Dichotomous Items
3 = complete
1 = unsatisfactory/incorrect

Short Three-Point Items
3 = complete
2 = partial
1 = unsatisfactory/incorrect

Extended Four-Point Items
4 = complete
3 = essential
2 = partial
1 = unsatisfactory/incorrect

Extended Five-Point Items
5 = complete
4 = essential
3 = adequate
2 = partial
1 = unsatisfactory/incorrect

In some cases, student responses do not fit into any of the categories listed in the scoring guide. Special coding categories for the unscorable responses are assigned to these types of responses. These categories are only assigned if no aspect of the student's response could be scored. Scoring supervisors and/or trainers are consulted prior to the assignment of any of the special coding categories. The unscorable categories used for science are outlined below.

Categories for unscorable responses, science assessment: 2000
Label Description
B Blank responses, random marks on paper, word underlined in prompt but response area completely blank, mark on item number but response area completely blank
X Completely crossed out, completely erased
IL Completely illegible response
OT Off task, off topic, comments to the test makers, refusal to answer, "Who cares," language other than English (unless otherwise noted)
? "I don't know," "I can't do this," "No clue," "I don't understand," "I forget"
NOTE: Because the NAEP scoring contractor's database recognizes only alphanumeric characters and sets a single-character field for the value for each score, the label "IL" appears in the database file as "I," the label "OT" appears as "T," and the label "?" appears as "D."
SOURCE: U.S. Department of Education, Institute of Education Sciences, National Center for Education Statistics, National Assessment of Educational Progress (NAEP), 2000 Science Assessment.
Last updated 15 April 2008 (TS)

Printer-friendly Version

1990 K Street, NW
Washington, DC 20006, USA
Phone: (202) 502-7300 (map)