Psychological Testing: Chapter 8: Test Development
This flashcard set outlines the key stages involved in developing a psychological test, from initial conceptualization to revision. It includes constructing test items, conducting tryouts, and using statistical item analysis to ensure reliability, validity, and overall test quality.
Stages in the Process of Developing a Test
Test Conceptualization
Test Construction
Test Tryout
Item Analysis
Test RevisionKey Terms
Stages in the Process of Developing a Test
Test Conceptualization
Test Construction
Test Tryout
Item Analysis
Test RevisionTest Construction
Drafting of items for the test
Test Tryout
First draft of the test is then tried out on a group of sample testtakers
Item Analysis
When statistical procedures are employed to assist in making judgments about which items are good as they are, which items need to be revised, and ...
Analysis of the Test’s Items Include
Analyses of item reliability
Item Validity
Item Discrimination
Test Conceptualization
There ought to be a test designed to measure (____) in a (____) way; stimulus could be anything; review of related literature on existing tests
Related Flashcard Decks
Study Tips
- Press F to enter focus mode for distraction-free studying
- Review cards regularly to improve retention
- Try to recall the answer before flipping the card
- Share this deck with friends to study together
| Term | Definition |
|---|---|
Stages in the Process of Developing a Test | |
Test Construction | Drafting of items for the test |
Test Tryout | First draft of the test is then tried out on a group of sample testtakers |
Item Analysis | When statistical procedures are employed to assist in making judgments about which items are good as they are, which items need to be revised, and which items should be discarded |
Analysis of the Test’s Items Include | Analyses of item reliability |
Test Conceptualization | There ought to be a test designed to measure (____) in a (____) way; stimulus could be anything; review of related literature on existing tests |
Preliminary Questions to Ask During Test Conceptualization | What is the test designed to Measure? |
What is the test designed to Measure? | Closely linked to how the test developer defines the construct being measured and how that definition is the same as or different from other tests purporting to measure the same construct |
What is the objective of the test? | In service of what goal will the test be employed? In what way or ways is the objective of this test the same as or different from other tests with similar goals? What real-world behaviors would be anticipated to correlate with testtaker responses? |
Is there a need for this test? | Are there any other tests purporting to measure the same thing? In what ways will the new test be better than or different from existing ones? Will there be more compelling evidence for its reliability or validity? Will it be more comprehensive? Will it take less time to administer? In what ways would this test not be better than existing tests? |
Preliminary Questions to be Addressed | Who will use this test? |
Who will use this test? | Clinicians? Educators? Others? For what purpose or purposes would this test be used? |
Who will take this test? | Who is this test for? Who needs to take it? Who would find it desirable to take it? For what age range of testtakers is the test designed? What reading level is required of a testtaker? What cultural factors might affect the testtaker response? |
What content will the test cover? | Why should it cover this content? Is this coverage different from the content coverage of existing tests with the same or similar objectives? How and why is the content area different? To what extent is this content culture-specific? |
How will the test be administered? | Individually or in groups? Is it amenable to both group and individual administration? What differences will exist between individual and group administrations of this test? Will the test be designed for or amenable to computer administration? How might differences between versions of the test be reflected in test scores? |
Should more than one form of the test be developed? | On the basis of a cost-benefit analysis, should alternate or parallel forms of this test be created? |
What special training will be required of test users for administering or interpreting the test? | What background and qualifications will a prospective user of data derived from an administration of this test need to have? What restrictions, if any, should be placed on distributors of the test and on the test’s usage |
What types of responses will be required of testtakers? | What kind of disability might preclude someone from being able to take this test? What adaptations or accommodations are recommended for persons with disabilities? |
Who benefits from an administration of this test? | What would the testtaker learn, or how might the testtaker benefit, from an administration of this test? What would the test user learn, or how might the test user benefit? What social benefit, if any, derives from an administration of this test? |
Is there any potential for harm as the result of an administration of this test? | What safeguards are built into the recommended testing procedure to prevent any sort of harm to any of the parties involved in the use of this test? |
How will meaning be attributed to scores on this test? | Will a testtaker’s score be compared to others taking the test at the same time? To others in a criterion group? Will the test evaluate masters of a particular content area? |
Good item on a Norm-referenced Test | An item for which high scorers on the test respond correctly; low scorers on the test tend to respond to that same item incorrectly |
Good item on a Criterion-Oriented Test | High scorers on the test get a particular item right whereas low scorers on the test get that same item wrong; each item should address the issue of whether the testaker has met certain criteria |
Pilot Work/Pilot Study/Pilot Research | Refers to the preliminary reserach surrounding the creation of a prototype of the test; test items may be piloted to evaluate whether they should be included in the final form of the instrument; May involve open-ended interviews with research subjects believed for some reason (perhaps on the basis of an existing test); developer attempts to determine how best to measure a targeted construct |
Pilot Work Process | |
Scaling | Assignment of numbers according to rules; defined as the process of setting rules for assigning numbers in measurement; process by which a measuring device is designed and calibrated and by which numbers (or other indices)-scale values- are designed to different amounts of the trait, attribute, or characteristic being measured |
Age-Based Scale | If the Testtaker's test performance as a function of age is of critical interest |
Grade-Based Scale | If the testtaker's test performance as a function of grade is of critical interest |
Stanine Scale | If all raw scores on the test are to be transformed into scores that can range from 1 to 9 |
Categorization of a Test Scale | Unidimentional vs. Multidimensional | Comparative vs. Categorical |
Rating Scale | Defined as a grouping of words, statements, or symbols on which judgments of the strength of a particular trait, attitude, or emotion are indicated by a testtaker; can be used to record judgments of oneself, other, experiences, or objects, and that they can take several forms |
Summative Scale | When final test score is obtained by summing the ratings across all the items |
Likert Scale | used extensively in psychology, usually to scale attitudes; relatively easy to construct |
Method of Paired Comparisons | Testtakers are presented with pairs of stimuli which they are asked to compare; select one of the stimuli according to some rule; the rule that they agree more with one statement than the other, or the rule that they find one stimulus more appealing than the other; |
Comparative Scaling | Entails judgments of a stimulus in comparison with every other stimulus on the scale |
Categorical Scaling | Scaling system that relies on sorting; stimuli are place into one of two or more alternative categories that differ quantitatively with respect to some continuum |
Guttman Scale | Another scaling method that yields ordinal-level measures; items range sequentially from weaker to stronger expressions of the attitude, belief, or feeling being measured; all respondents who agree with the stronger statements of the attitude will also agree with milder statements |
Scalogram Analysis | Item-analysis procedure and approach to test development that involves a graphic mapping of a testtaker's responses; Objective for the developer of a measure of attitudes is to obtain an arrangement of items wherein endorsement of one item automatically connotes endorsement of less extreme positions |
How to Create a Scale using Thurstone's equal-appearing interval method | Collect a reasonably large number of statements reflecting positive and negative attitudes towards a topic are collected |
Scaling Method Employed Depends on | Variables being measured |
Questions to Ask for the Test Blueprint | What range of content should the items cover? |
Item Pool | Reservior or well from which test items will or will not be drawn for the final version of the test |
Item Format | Include variables such as the form, plan, structure, arrangement, and layout of individual test items |
Types of Response Formats | Selected-Response Format | Constructed-Response Format |
Selected-Response Format | Require testtakers to select a response from a set of alternative responses |
Constructed-Response Format | Require the testtakers to supply or to create the correct answer, not merely to select it |
Types of Selected-Response Item Formats | Multiple Choice |
Elements of Multiple-Choice Format | Stem |
Characteristics of a good multiple-choice item in an achievement test | Has one correct alternative |
Matching Item | Testtaker is presented with two columns: Premises on the left and responses to the right; |
Binary Choice Item | Multiple-choice item that contains only two possible responses |
True-False Item | The most familiar binary-choice item; type of selected-response item which takes the form of a sentence that requires the testtaker to indicate whether the statement is or is not a fact |
Good Binary Choice Item | Contains a single idea, is not excessively long, and is not subject to debate; correct response must undoubtedly be one of the two coices |
Completion Item | Requires the examinee to provide a word or phrase that completes a sentence; also known as Short-Answer Item |
Good Completion Item | Should be worded so that the correct answer is specific; Should be written clearly enough that the testtaker can respond succinctly (with a short answer) |
Essay Item | Useful when the test developer wants the examinee to demonstrate a depth of knowledge about a single topic; permits restating of learned material and allows for the creative integration and expression of the material in the testtaker's own words; subjective and inter-scorer differences |
Item Bank | Relatively large and accessible collection of test questions; advantage is accessibility to a large number of test items conveniently classified by subject area, item statistics, or other variables |
Item Branching | Technique with the ability to individualize testing; ability of the computer to tailor the content and order of presentation of test items on the basis of responses to previous items |
Computerized Adaptive Testing (CAT) | Refers to an interactive, computer-administered testtaking procedure wherein items presented to the testtaker are absed in part on the testtaker's performance on previous items; tends to reduce floor effects and ceiling effects |
Floor effect | Refers to the diminished utility of an assessment tool for distinguishing testtakers at the low end of the ability, trait, or other attribute being measured |
Ceiling Effect | Refers to the diminished utility of an assessment tool for distinguishing testtakers at the high end of the ability, trait, or other attribute being measured |
Class or Category Scoring | Employs testtaker responses which earn credit toward placement in a particular class or category with other testtakers whose pattern of responses is presumably similar in some way; used by dome diagnostic systems wherein individuals must exhibit a certain number of symptoms to qualify for a specific diagnosis |
Ipsative Scoring | Comparing a testtaker's score on one scale within a test to another scale within that same test |
Edwards Personal Preference Schedule | EPPS designed to measure the relative strength of different psychological needs |
Formal Item-Analysis | Cross Validation |
Tests Due For Revision When The Following Conditions Exist | Stimulus materials look dated and current testtakers cannot relate to them. |
Cross-Validation | Refers to the revalidation of a test on a sample of testtakers other than those on whom test performance was originally found to be a valid predictor of some criterion |
Validity Shrinkage | The decrease in item validities that inevitably occurs after cross-validation of findings; expected and viewed as integral to the test development process; infinitely preferable to a scenariou wherein high item validities are published in a test manual as a result of inappropriately using the identical sample of testtakers for test standardization and cross-validation of findings |
Test Manual | Should outline the test development procedures used | Reliability information, including test-retest validity and Internal consistency estimates |
Co-Validation | Defined as a test validation process conducted on two or more tests using the sample of testtakers |
Co-Norming | Process that occurs when co-validation is used in conjunction with the creation of norms or the revision of existing norms |
Anchor Protocol | Test protocol scored by a highly authoritative scorer that is designed as a model for schoring and a mechanism for resolving scoring discrepancies |
Scoring Drift | A discrepancy between scoring in an anchor protocol and the scoring of another protocol |
Roles of IRT in Test Construction | Evaluating existing tests for the purpose of mapping test revisions |
IRT Information Curves | Help test developers evaluate how well an individual item (or entire test) is working to measure different levels of the underlying construct |
Differential Item Functioning (DIF) | Phenomenon wherein an item functions differently in one group of testtakers as compaered to another group of testtakers known to have the same (or similar) level of the underlying Trait |
DIF Analysis | A process by which test developers scrutinize group-by-group item response curves, looking for DIF Items; used to evaluate the effect of different test administration procedures and item ordering effects |
DIF Items | Items that respondents from different groups at the same level of the underlying traid have different probabilities of endorsing as a function of their group membership |