skip navigational linksDOL Seal - Link to DOL Home Page
Photos representing the workforce - Digital Imagery© copyright 2001 PhotoDisc, Inc.
www.dol.gov

Previous Section

Content Last Revised:
---DISCLAIMER---

Next Section

CFR  

Code of Federal Regulations Pertaining to ESA

Title 41  

Public Contracts and Property Management

 

Chapter 60  

Office of Federal Contract Compliance Programs, Equal Employment Opportunity, Department of Labor

 

 

Part 60-3  

Uniform Guidelines on Employee Selection Procedures (1978)


41 CFR 60-3.14 - Technical standards for validity studies.

  • Section Number: 60-3.14
  • Section Name: Technical standards for validity studies.

    The following minimum standards, as applicable, should be met in 
conducting a validity study. Nothing in these guidelines is intended to 
preclude the development and use of other professionally acceptable 
techniques with respect to validation of selection procedures. Where it 
is not technically feasible for a user to conduct a validity study, the 
user has the obligation otherwise to comply with these guidelines. See 
sections 6 and 7 of this part.
    A. Validity studies should be based on review of information about 
the job. Any validity study should be based upon a review of information 
about the job for which the selection procedure is to be used. The 
review should include a job analysis except as provided in section 
14B(3) of this section with respect to criterion-related validity. Any 
method of job analysis may be used if it provides the information 
required for the specific validation strategy used.
    B. Technical standards for criterion-related validity studies--(1) 
Technical feasibility. Users choosing to validate a selection procedure 
by a criterion-related validity strategy should determine whether it is 
technically feasible (as defined in section 16) to conduct such a study 
in the particular employment context. The determination of the number of 
persons necessary to permit the conduct of a meaningful criterion-
related study should be made by the user on the basis of all relevant 
information concerning the selection procedure, the potential sample and 
the employment situation. Where appropriate, jobs with substantially the 
same major work behaviors may be grouped together for validity studies, 
in order to obtain an adequate sample. These guidelines do not require a 
user to hire or promote persons for the purpose of making it possible to 
conduct a criterion-related study.
    (2) Analysis of the job. There should be a review of job information 
to determine measures of work behavior(s) or performance that are 
relevant to the job or group of jobs in question. These measures or 
criteria are relevant to the extent that they represent critical or 
important job duties, work behaviors or work outcomes as developed from 
the review of job information. The possibility of bias should be 
considered both in selection of the criterion measures and their 
application. In view of the possibility of bias in subjective 
evaluations, supervisory rating techniques and instructions to raters 
should be carefully developed. All criterion measures and the methods 
for gathering data need to be examined for freedom from factors which 
would unfairly alter scores of members of any group. The relevance of 
criteria and their freedom from bias are of particular concern when 
there are significant differences in measures of job performance for 
different groups.
    (3) Criterion measures. Proper safeguards should be taken to insure 
that
scores on selection procedures do not enter into any judgments of 
employee adequacy that are to be used as criterion measures. Whatever 
criteria are used should represent important or critical work 
behavior(s) or work outcomes. Certain criteria may be used without a 
full job analysis if the user can show the importance of the criteria to 
the particular employment context. These criteria include but are not 
limited to production rate, error rate, tardiness, absenteeism, and 
length of service. A standardized rating of overall work performance may 
be used where a study of the job shows that it is an appropriate 
criterion. Where performance in training is used as a criterion, success 
in training should be properly measured and the relevance of the 
training should be shown either through a comparsion of the content of 
the training program with the critical or important work behavior(s) of 
the job(s), or through a demonstration of the relationship between 
measures of performance in training and measures of job performance. 
Measures of relative success in training include but are not limited to 
instructor evaluations, performance samples, or tests. Criterion 
measures consisting of paper and pencil tests will be closely reviewed 
for job relevance.
    (4) Representativeness of the sample. Whether the study is 
predictive or concurrent, the sample subjects should insofar as feasible 
be representative of the candidates normally available in the relevant 
labor market for the job or group of jobs in question, and should 
insofar as feasible include the races, sexes, and ethnic groups normally 
available in the relevant job market. In determining the 
representativeness of the sample in a concurrent validity study, the 
user should take into account the extent to which the specific 
knowledges or skills which are the primary focus of the test are those 
which employees learn on the job.

Where samples are combined or compared, attention should be given to see 
that such samples are comparable in terms of the actual job they 
perform, the length of time on the job where time on the job is likely 
to affect performance, and other relevant factors likely to affect 
validity differences; or that these factors are included in the design 
of the study and their effects identified.
    (5) Statistical relationships. The degree of relationship between 
selection procedure scores and criterion measures should be examined and 
computed, using professionally acceptable statistical procedures. 
Generally, a selection procedure is considered related to the criterion, 
for the purposes of these guidelines, when the relationship between 
performance on the procedure and performance on the criterion measure is 
statistically significant at the 0.05 level of significance, which means 
that it is sufficiently high as to have a probability of no more than 
one (1) in twenty (20) to have occurred by chance. Absence of a 
statistically significant relationship between a selection procedure and 
job performance should not necessarily discourage other investigations 
of the validity of that selection procedure.
    (6) Operational use of selection procedures. Users should evaluate 
each selection procedure to assure that it is appropriate for 
operational use, including establishment of cutoff scores or rank 
ordering. Generally, if other factors reman the same, the greater the 
magnitude of the relationship (e.g., correlation coefficent) between 
performance on a selection procedure and one or more criteria of 
performance on the job, and the greater the importance and number of 
aspects of job performance covered by the criteria, the more likely it 
is that the procedure will be appropriate for use. Reliance upon a 
selection procedure which is significantly related to a criterion 
measure, but which is based upon a study involving a large number of 
subjects and has a low correlation coefficient will be subject to close 
review if it has a large adverse impact. Sole reliance upon a single 
selection instrument which is related to only one of many job duties or 
aspects of job performance will also be subject to close review. The 
appropriateness of a selection procedure is best evaluated in each 
particular situation and there are no minimum correlation coefficients 
applicable to all employment situations. In determining whether a 
selection procedure is appropriate for operational use the following
considerations should also be taken into account: The degree of adverse 
impact of the procedure, the availability of other selection procedures 
of greater or substantially equal validity.
    (7) Overstatement of validity findings. Users should avoid reliance 
upon techniques which tend to overestimate validity findings as a result 
of capitalization on chance unless an appropriate safeguard is taken. 
Reliance upon a few selection procedures or criteria of successful job 
performance when many selection procedures or criteria of performance 
have been studied, or the use of optimal statistical weights for 
selection procedures computed in one sample, are techniques which tend 
to inflate validity estimates as a result of chance. Use of a large 
sample is one safeguard: cross-validation is another.
    (8) Fairness. This section generally calls for studies of unfairness 
where technically feasible. The concept of fairness or unfairness of 
selection procedures is a developing concept. In addition, fairness 
studies generally require substantial numbers of employees in the job or 
group of jobs being studied. For these reasons, the Federal enforcement 
agencies recognize that the obligation to conduct studies of fairness 
imposed by the guidelines generally will be upon users or groups of 
users with a large number of persons in a job class, or test developers; 
and that small users utilizing their own selection procedures will 
generally not be obligated to conduct such studies because it will be 
technically infeasible for them to do so.
    (a) Unfairness defined. When members of one race, sex, or ethnic 
group characteristically obtain lower scores on a selection procedure 
than members of another group, and the differences in scores are not 
reflected in differences in a measure of job performance, use of the 
selection procedure may unfairly deny opportunities to members of the 
group that obtains the lower scores.
    (b) Investigation of fairness. Where a selection procedure results 
in an adverse impact on a race, sex, or ethnic group identified in 
accordance with the classifications set forth in section 4 of this part 
and that group is a significant factor in the relevant labor market, the 
user generally should investigate the possible existence of unfairness 
for that group if it is technically feasible to do so. The greater the 
severity of the adverse impact on a group, the greater the need to 
investigate the possible existence of unfairness. Where the weight of 
evidence from other studies shows that the selection procedure predicts 
fairly for the group in question and for the same or similar jobs, such 
evidence may be relied on in connection with the selection procedure at 
issue.
    (c) General considerations in fairness investigations. Users 
conducting a study of fairness should review the A.P.A. Standards 
regarding investigation of possible bias in testing. An investigation of 
fairness of a selection procedure depends on both evidence of validity 
and the manner in which the selection procedure is to be used in a 
particular employment context. Fairness of a selection procedure cannot 
necessarily be specified in advance without investigating these factors. 
Investigation of fairness of a selection procedure in samples where the 
range of scores on selection procedures or criterion measures is 
severely restricted for any subgroup sample (as compared to other 
subgroup samples) may produce misleading evidence of unfairness. That 
factor should accordingly be taken into account in conducting such 
studies and before reliance is placed on the results.
    (d) When unfairness is shown. If unfairness is demonstrated through 
a showing that members of a particular group perform better or poorer on 
the job than their scores on the selection procedure would indicate 
through comparison with how members of other groups perform, the user 
may either revise or replace the selection instrument in accordance with 
these guidelines, or may continue to use the selection instrument 
operationally with appropriate revisions in its use to assure 
compatibility between the probability of successful job performance and 
the probability of being selected.
    (e) Technical feasibility of fairness studies. In addition to the 
general conditions needed for technical feasibility for the conduct of a 
criterion-related study (see section 16, below) an investigation of 
fairness requires the following:
    (1) An adequate sample of persons in each group available for the 
study to achieve findings of statistical significance. Guidelines do not 
require a user to hire or promote persons on the basis of group 
classifications for the purpose of making it possible to conduct a study 
of fairness; but the user has the obligation otherwise to comply with 
these guidelines.
    (2) The samples for each group should be comparable in terms of the 
actual job they perform, length of time on the job where time on the job 
is likely to affect performance, and other relevant factors likely to 
affect validity differences; or such factors should be included in the 
design of the study and their effects identified.
    (f) Continued use of selection procedures when fairness studies not 
feasible. If a study of fairness should otherwise be performed, but is 
not technically feasible, a selection procedure may be used which has 
otherwise met the validity standards of these guidelines, unless the 
technical infeasibility resulted from discriminatory employment 
practices which are demonstrated by facts other than past failure to 
conform with requirements for validation of selection procedures. 
However, when it becomes technically feasible for the user to perform a 
study of fairness and such a study is otherwise called for, the user 
should conduct the study of fairness.
    C. Technical standards for content validity studies--(1) 
Appropriateness of content validity studies. Users choosing to validate 
a selection procedure by a content validity strategy should determine 
whether it is appropriate to conduct such a study in the particular 
employment context. A selection procedure can be supported by a content 
validity strategy to the extent that it is a representative sample of 
the content of the job. Selection procedures which purport to measure 
knowledges, skills, or abilities may in certain circumstances be 
justified by content validity, although they may not be representative 
samples, if the knowledge, skill, or ability measured by the selection 
procedure can be operationally defined as provided in paragraph 14C(4) 
of this section, and if that knowledge, skill, or ability is a necessary 
prerequisite to successful job performance.

A selection procedure based upon inferences about mental processes 
cannot be supported solely or primarily on the basis of content 
validity. Thus, a content strategy is not appropriate for demonstrating 
the validity of selection procedures which purport to measure traits or 
constructs, such as intelligence, aptitude, personality, commonsense, 
judgment, leadership, and spatial ability. Content validity is also not 
an appropriate strategy when the selection procedure involves 
knowledges, skills, or abilities which an employee will be expected to 
learn on the job.
    (2) Job analysis for content validity. There should be a job 
analysis which includes an analysis of the important work behavior(s) 
required for successful performance and their relative importance and, 
if the behavior results in work product(s), an analysis of the work 
product(s). Any job analysis should focus on the work behavior(s) and 
the tasks associated with them. If work behavior(s) are not observable, 
the job analysis should identify and analyze those aspects of the 
behavior(s) that can be observed and the observed work products. The 
work behavior(s) selected for measurement should be critical work 
behavior(s) and/or important work behavior(s) constituting most of the 
job.
    (3) Development of selection procedures. A selection procedure 
designed to measure the work behavior may be developed specifically from 
the job and job analysis in question, or may have been previously 
developed by the user, or by other users or by a test publisher.
    (4) Standards for demonstrating content validity. To demonstrate the 
content validity of a selection procedure, a user should show that the 
behavior(s) demonstrated in the selection procedure are a representative 
sample of the behavior(s) of the job in question or that the selection 
procedure provides a representative sample of the work product of the 
job. In the case of a selection procedure measuring a knowledge, skill, 
or ability, the knowledge, skill, or ability being measured should be 
operationally defined. In the case of a selection procedure measuring a
knowledge, the knowledge being measured should be operationally defined 
as that body of learned information which is used in and is a necessary 
prerequisite for observable aspects of work behavior of the job. In the 
case of skills or abilities, the skill or ability being measured should 
be operationally defined in terms of observable aspects of work behavior 
of the job. For any selection procedure measuring a knowledge, skill, or 
ability the user should show that (a) the selection procedure measures 
and is a representative sample of that knowledge, skill, or ability; and 
(b) that knowledge, skill, or ability is used in and is a necessary 
prerequisite to performance of critical or important work behavior(s). 
In addition, to be content valid, a selection procedure measuring a 
skill or ability should either closely approximate an observable work 
behavior, or its product should closely approximate an observable work 
product. If a test purports to sample a work behavior or to provide a 
sample of a work product, the manner and setting of the selection 
procedure and its level and complexity should closely approximate the 
work situation. The closer the content and the context of the selection 
procedure are to work samples or work behaviors, the stronger is the 
basis for showing content validity. As the content of the selection 
procedure less resembles a work behavior, or the setting and manner of 
the administration of the selection procedure less resemble the work 
situation, or the result less resembles a work product, the less likely 
the selection procedure is to be content valid, and the greater the need 
for other evidence of validity.
    (5) Reliability. The reliability of selection procedures justified 
on the basis of content validity should be a matter of concern to the 
user. Whenever it is feasible, appropriate statistical estimates should 
be made of the reliability of the selection procedure.
    (6) Prior training or experience. A requirement for or evaluation of 
specific prior training or experience based on content validity, 
including a specification of level or amount of training or experience, 
should be justified on the basis of the relationship between the content 
of the training or experience and the content of the job for which the 
training or experience is to be required or evaluated. The critical 
consideration is the resemblance between the specific behaviors, 
products, knowledges, skills, or abilities in the experience or training 
and the specific behaviors, products, knowledges, skills, or abilities 
required on the job, whether or not there is close resemblance between 
the experience or training as a whole and the job as a whole.
    (7) Content validity of training success. Where a measure of success 
in a training program is used as a selection procedure and the content 
of a training program is justified on the basis of content validity, the 
use should be justified on the relationship between the content of the 
training program and the content of the job.
    (8) Operational use. A selection procedure which is supported on the 
basis of content validity may be used for a job if it represents a 
critical work behavior (i.e., a behavior which is necessary for 
performance of the job) or work behaviors which constitute most of the 
important parts of the job.
    (9) Ranking based on content validity studies. If a user can show, 
by a job analysis or otherwise, that a higher score on a content valid 
selection procedure is likely to result in better job performance, the 
results may be used to rank persons who score above minimum levels. 
Where a selection procedure supported solely or primarily by content 
validity is used to rank job candidates, the selection procedure should 
measure those aspects of performance which differentiate among levels of 
job performance.
    D. Technical standards for construct validity studies--(1) 
Appropriateness of construct validity studies. Construct validity is a 
more complex strategy than either criterion-related or content validity. 
Construct validation is a relatively new and developing procedure in the 
employment field, and there is at present a lack of substantial 
literature extending the concept to employment practices. The user 
should be aware that the effort to obtain sufficient empirical support 
for construct validity is both an extensive and arduous effort involving 
a series of research studies, which include criterion related
validity studies and which may include content validity studies. Users 
choosing to justify use of a selection procedure by this strategy should 
therefore take particular care to assure that the validity study meets 
the standards set forth below.
    (2) Job analysis for construct validity studies. There should be a 
job analysis. This job analysis should show the work behavior(s) 
required for successful performance of the job, or the groups of jobs 
being studied, the critical or important work behavior(s) in the job or 
group of jobs being studied, and an identification of the construct(s) 
believed to underlie successful performance of these critical or 
important work behaviors in the job or jobs in question. Each construct 
should be named and defined, so as to distinguish it from other 
constructs. If a group of jobs is being studied the jobs should have in 
common one or more critical or important work behav- iors at a 
comparable level of complexity.
    (3) Relationship to the job. A selection procedure should then be 
identified or developed which measures the construct identified in 
accord with subparagraph (2) of this section. The user should show by 
empirical evidence that the selection procedure is validly related to 
the construct and that the construct is validly related to the 
performance of critical or important work behavior(s). The relationship 
between the construct as measured by the selection procedure and the 
related work behavior(s) should be supported by empirical evidence from 
one or more criterion-related studies involving the job or jobs in 
question which satisfy the provisions of paragraph 14B of this section.
    (4) Use of construct validity study without new criterion-related 
evidence--(a) Standards for use. Until such time as professional 
literature provides more guidance on the use of construct validity in 
employment situations, the Federal agencies will accept a claim of 
construct validity without a criterion-related study which satisfies 
paragraph 14B of this section only when the selection procedure has been 
used elsewhere in a situation in which a criterion-related study has 
been conducted and the use of a criterion-related validity study in this 
context meets the standards for transportability of criterion-related 
validity studies as set forth above in section 7. However, if a study 
pertains to a number of jobs having common critical or important work 
behaviors at a comparable level of complexity, and the evidence 
satisfies subparagraphs 14B (2) and (3) of this section for those jobs 
with criterion-related validity evidence for those jobs, the selection 
procedure may be used for all the jobs to which the study pertains. If 
construct validity is to be generalized to other jobs or groups of jobs 
not in the group studied, the Federal enforcement agencies will expect 
at a minimum additional empirical research evidence meeting the 
standards of subparagraphs section 14B (2) and (3) of this section for 
the additional jobs or groups of jobs.
    (b) Determination of common work behaviors. In determining whether 
two or more jobs have one or more work behavior(s) in common, the user 
should compare the observed work behavior(s) in each of the jobs and 
should compare the observed work product(s) in each of the jobs. If 
neither the observed work behavior(s) in each of the jobs nor the 
observed work product(s) in each of the jobs are the same, the Federal 
enforcement agencies will presume that the work behavior(s) in each job 
are different. If the work behaviors are not observable, then evidence 
of similarity of work products and any other relevant research evidence 
will be considered in determining whether the work behavior(s) in the 
two jobs are the same.
Previous Section

Next Section



Phone Numbers