首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 156 毫秒
1.
Contrary to the adage warning against changing test answers, mean gain from changing has been an invariant research finding. Consistency of this gain was tested for students instructed about the research results, and composition of the gain was analyzed by examining the students' reasons for changing. Students in six graduate measurement classes instructed about the answer-changing literature responded to three exams and a questionnaire. Mean gain remained positive and consistent with gain for previously studied uninstructed groups; amount of change was also stable. "Rethinking the item and conceptualizing a better answer" was the most frequent reason given for changing. "Rereading the item and better understanding the question" was the second most cited reason, followed by "rereading/rethinking" combined, and "making a clerical error." For each frequently used reason, wrong-to-right (WR) changes were in the majority. Implications for research and practice are discussed.  相似文献   

2.
Answer Changing on Multiple-Choice Test Items Among Eighth-Grade Readers   总被引:1,自引:1,他引:0  
This study was done to examine the effect of answer changing on multiple-choice test performance among good and poor readers in the eighth grade. Although the gains of poor readers were higher than those of good readers, all subjects profited significantly from changing their answers on items. For all subjects, when a single response was changed, there was a two-to-one chance that the new response would raise rather than lower the final score. Gains from answer changing on test items were slightly higher for poor readers as a group than were those for good readers. However, the result was determined not to be significant. More important, this hypothesis is strengthened by the fact that all subjects profited from answer changing. Therefore, the results were interpreted as lending support to the notion that answer-changing response among young examinees should be encouraged if there is a reasonable doubt about their “first impression.”  相似文献   

3.
Are some students advantaged when changing multiple-choice answers? The authors of this investigation assessed the importance of an examinee's cognitive style in the answer-changing process. Two separate studies were conducted using undergraduates (n = 125 and n = 84). One set of variables consisted of a measure of field dependence/field independence, a measure of impulsivity/reflectivity, and an introductory psychology unit examination made up of multiple-choice items. A second set of variables was formed by gathering two answer-changing scores for each subject using the scannable forms from the unit examination—one reflecting the effect of answer changes and the other representing the number of changes. Canonical correlation analysis was used to describe the relationship between the two sets of variables. Only the first canonical correlation coefficient was statistically significant in each study. The structure coefficients indicated that the cognitive-style variables had little impact on the canonical solution and that a combination of the effect of answer changes, the number of changes, and unit examination scores were the most influential components of the first canonical variates.  相似文献   

4.
Changing a small number of answers to multiple-choice questions reliably improves test-scores, although it remains unclear how examinees select which initial answers to change and whether answer-changing behaviour is susceptible to instruction. We tested the effect of an instructional intervention on the number of changes made by examinees on a mock-exam in a controlled experimental design. We also examined how examinees' confidence with their initial answers, and their judgement of how difficult each exam question was, predicted their answer-changing behaviour. We found that the number of changes made increased, to a small extent, through instruction, without increasing the rate of errors. The likelihood of changing an initial answer decreased with examinees’ feeling of confidence, and increased with their feeling of difficulty. This is consistent with the theory that examinees use metacognitive experiences to select which initial answers to change on exams.  相似文献   

5.
Six undergraduate and three graduate classes were given multiple-choice tests with subsequent evaluation of answer changes. The 300 students were tested twice, once before and once after instruction on answer changing. After each test, students were asked to complete two forms. The forms evaluated attitude toward answer changing, reasons for changing, and confidence in final answers. Students showed a significant increase in favorability toward answer changing after instruction. No significant change was found in number of answers changed. Psychology students were found to change significantly more items than were business students. Mean gain score did not change significantly after instruction. It was concluded that although instruction does lead to a change in attitude in answer changing, the number of changes and overall gain due to answer changing do not change. It was also determined that students continue to make significant gains even when their confidence in the final answer is less than 50 on a 100-point scale.  相似文献   

6.
Several researchers have concluded that changing answers to multiple-choice questions is typically beneficial for students. This research attempts to validate these earlier results and to determine if students accurately perceive the outcome of their own answer-changing behavior. The results support the conclusions of the earlier studies that when students change answers on multiple-choice questions they tend to gain roughly 3 points for every point lost. This study also found that students predominantly underestimate the benefit of their own answer-changing behavior.  相似文献   

7.
Recent studies have shown that restricting review and answer change opportunities on computerized adaptive tests (CATs) to items within successive blocks reduces time spent in review, satisfies most examinees' desires for review, and controls against distortion in proficiency estimates resulting from intentional incorrect answering of items prior to review. However, restricting review opportunities on CATs may not prevent examinees from artificially raising proficiency estimates by using judgments of item difficulty to signal when to change previous answers. We evaluated six strategies for using item difficulty judgments to change answers on CATs and compared the results to those from examinees reviewing and changing answers in the usual manner. The strategy conditions varied in terms of when examinees were prompted to consider changing answers and in the information provided about the consistency of the item selection algorithm. We found that examinees fared best on average when they reviewed and changed answers in the usual manner. The best gaming strategy was one in which the examinees knew something about the consistency of the item selection algorithm and were prompted to change responses only when they were unsure about answer correctness and sure about their item difficulty judgments. However, even this strategy did not produce a mean gain in proficiency estimates.  相似文献   

8.
According to a popular belief, test takers should trust their initial instinct and retain their initial responses when they have the opportunity to review test items. More than 80 years of empirical research on item review, however, has contradicted this belief and shown minor but consistently positive score gains for test takers who changed answers they found to be incorrect during review. This study reanalyzed the problem of the benefits of answer changes using item response theory modeling of the probability of an answer change as a function of the test taker’s ability level and the properties of items. Our empirical results support the popular belief and reveal substantial losses due to changing initial responses for all ability levels. Both the contradiction of the earlier research and support of the popular belief are explained as a manifestation of Simpson’s paradox in statistics.  相似文献   

9.
In an article in the Winter 2011 issue of the Journal of Educational Measurement, van der Linden, Jeon, and Ferrara suggested that “test takers should trust their initial instincts and retain their initial responses when they have the opportunity to review test items.” They presented a complex IRT model that appeared to show that students would be worse off by changing answers. As noted in a subsequent erratum, this conclusion was based on flawed data, and that the correct data could not be analyzed by their method because the model failed to converge. This left their basic question on the value of answer changing unanswered. A much more direct approach is to simply count the number of examinees whose scores after an opportunity to change answers are higher, lower, or the same as their initial scores. Using the same data set as the original article, an overwhelming majority of the students received higher scores after the opportunity to change answers.  相似文献   

10.
The hypothesis that it is unwise to change answers to multiple choice questions was tested using the technique of multiple regression analysis. The net number of correct answers as a result of changing responses was regressed against final grade in the course, numeric score on the exam, percent of total answers changed for all questions and for analytical questions, sex of the student, and scope of the exam.
The results show that there are gains to be made by changing responses. The variables which proved to be significant indicated that students who did well on the test changed a large percentage of answers, and that those who were taking a final exam tended to gain more. Final grades, sex of the student, and analytical questions had no significant impact on gains from changing responses. On the basis of the results gathered, the authors reject the hypothesis that changing responses is unwise.  相似文献   

11.
Abstract

In an attempt to identify some of the causes of answer changing behavior, the effects of four tests and item specific variables were evaluated. Three samples of New Zealand school children of different ages were administered tests of study skills. The number of answer changes per item was compared with the position of each item in a group of items, the position of each item in the test, the discrimination index and the difficulty index of each item. It is shown that answer changes were more likely to be made on items occurring early in a group of items and toward the end of a test. There was also a tendency for difficult items and items with poor discriminations to be changed more frequently. Some implications of answer changing in the design of tests are discussed.  相似文献   

12.
It has been reasonably well established that test takers can, to varying degrees, answer some reading comprehension questions without reading the passages on which the questions are based, even for carefully constructed measures like the Scholastic Aptitude Test (SAT). The aim of this study was to determine what test-taking strategies examinees use, and which are related to test performance, when reading passages are not available. The research focused on reading comprehension questions similar to those that will be used in the revised SAT, to be introduced in 1994. The most often cited strategies involved choosing answers on the basis of consistency with other questions and reconstructing the main theme of a missing passage from all of the questions and answers in a set. These strategies were more likely to result in successful performance on individual test items than were any of many other possible (and less constructrelevant) strategies.  相似文献   

13.
Abstract

An attempt was made to extend and clarify prior research which had demonstrated consistently that changed answers to objective test items tend to be correct. Results extended the basic effect of profiting from changed answers to Air Force personnel responding to multiple-choice questions regarding technical skills; the profit from changes was very similar to that observed in a university group responding to relatively "academic" items. Secondly, most individuals in both groups profited from changes. Third, individuals with the highest test scores tended to profit more from changes than those with the lowest test scores. Fourth, neither Airman Qualifying Exam scores (for the military personnel) nor Scholastic Aptitude Test scores (for the university students) were related to profit. Finally, a systematic case against the popular belief that one should not change answers on objective tests was made, based on an integration of the research to date.  相似文献   

14.
Abstract

The purpose of this study was to examine the prevalence and effect of answer-changing on objective test performance. 5s were 77 college students of both sexes. All types of response changes were recorded and analysed by sex and by caliber of student. Results showed that when a response was changed there was a three-to-one chance that the new response would Improve rather than lower the final score. Females were more Inclined to make changes than were males, and their overall test performance was superior to that of males. The brightest 5s of both sexes made the greatest mean number of answer changes. It was concluded that the fear frequently expressed by students that answer changes will reduce test performance was unfounded In this instance.  相似文献   

15.
Open–ended counterparts to a set of items from the quantitative section of the Graduate Record Examination (GRE–Q) were developed. Examinees responded to these items by gridding a numerical answer on a machine-readable answer sheet or by typing on a computer. The test section with the special answer sheets was administered at the end of a regular GRE administration. Test forms were spiraled so that random groups received either the grid-in questions or the same questions in a multiple-choice format. In a separate data collection effort, 364 paid volunteers who had recently taken the GRE used a computer keyboard to enter answers to the same set of questions. Despite substantial format differences noted for individual items, total scores for the multiple-choice and open-ended tests demonstrated remarkably similar correlational patterns. There were no significant interactions of test format with either gender or ethnicity.  相似文献   

16.
Background: Number sense is a key topic in mathematics education, and the identification of children’s misconceptions about number is, therefore, important. Information about students’ serious misconceptions can be quite significant for teachers, allowing them to change their teaching plans to help children overcome these misconceptions. In science education, interest in children’s alternative conceptions has led to the development of three- and four-tier tests that not only assess children’s understandings and misconceptions, but also examine children’s confidence in their responses. However, there are few such tests related to mathematical content, especially in studies of number sense.

Purpose: The purpose of this study was to investigate children’s performance and misconceptions with respect to number sense via a four-tier diagnostic test (Answer Tier → Confidence rating for Answer Tier → Reason Tier → Confidence rating for Reason Tier).

Design and method: A total of 195 fifth graders (10–11 years old) from Taiwan participated in this study. The four-tier test was web-based and contained 40 items across five components of number sense.

Findings: The results show that (1) students’ mean confidence rating for the answer tier was significantly higher than for the reason tier; (2) an average of 68% of students tended to have equal confidence ratings in both answer and reason tiers; (3) students who chose correct answers or reasons had higher mean confidence ratings in most items (36 out of 40) than those who did not; and (4) 16 misconceptions were identified and most of them were at a strong level.

Conclusion: The four-tier test was able to identify several misconceptions in both the answer and reason tier and provide information about the confidence levels. By using such information, teachers may be better positioned to understand the nature of learners’ misconceptions about number sense and therefore support their pupils’ progress in mathematics.  相似文献   

17.
We investigated the effect of two visual aids in representational illustrations on pupils’ realistic word problem solving. In part 1 of our study, 288 elementary school pupils received an individual paper-and-pencil task with seven problematic items (P-items) in which realistic considerations need to be made to come to an appropriate reaction. These items were presented together either with representational illustrations, representational illustrations in which an element was added to make the realistic modelling complexity more apparent, or representational illustrations in which this element was cued. In part 2, the pupils received the same P-items together with a realistic and a non-realistic answer option, with the request to choose the best answer. The findings show that there was no positive effect of the visual aids on the number of realistic reactions in part 1 and that when reviewing possible answers to P-items in part 2, there again was no positive effect.  相似文献   

18.
A statistical test for the detection of answer copying on multiple-choice tests is presented. The test is based on the idea that the answers of examinees to test items may be the result of three possible processes: (1) knowing, (2) guessing, and (3) copying, but that examinees who do not have access to the answers of other examinees can arrive at their answers only through the first two processes. This assumption leads to a distribution for the number of matched incorrect alternatives between the examinee suspected of copying and the examinee believed to be the source that belongs to a family of "shifted binomials." Power functions for the tests for several sets of parameter values are analyzed. An extension of the test to include matched numbers of correct alternatives would lead to improper statistical hypotheses.  相似文献   

19.
At the Medical College of Wisconsin, a procedure was developed to allow computerized grading and grade reporting of laboratory practical examinations in the Clinical Human Anatomy course. At the start of the course, first year medical students were given four Lists of Structures. On these lists, numbered items were arranged alphabetically; the items were anatomical structures that could be tagged on a given lab practical examination. Each lab exam featured an anatomy laboratory component and a computer laboratory component. For the anatomy lab component, students moved from one question station to another at timed intervals and identified tagged anatomical structures. As students identified a tagged structure, they referred to a copy of the list (provided with their answer sheet) and wrote the number corresponding to the structure on their answer sheet. Immediately after the anatomy lab component, students were escorted to a computer instruction laboratory where they typed their answer numbers into a secured testing component of a learning management system that recorded their answers for automatic grading. After a brief review of examination scores and item analysis by faculty, exam scores were reported to students electronically. Adding this brief computer component to each lab exam greatly reduced faculty grading time, reduced grading errors and provided faster performance feedback for students without changing overall student performance. Anat Sci Ed 1:220–223, 2008. © 2008 American Association of Anatomists.  相似文献   

20.
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号