首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 305 毫秒
1.
王晓龙 《天津教育》2014,(11):40-41
汉语作文智能测评系统是指使用计算机、智能手机等对电子稿作文进行评分与评级以及自动给出评语和建议的软件系统。这在我国可能是一种前所未有的汉语作文评价系统,是一项语文评价工具现代化的重要的学术和技术探索。  相似文献   

2.
计算机自动评分(CAS)用于自学考试外语类课程的翻译测验评分,能够有效提高评分效率及客观性。本研究对72名自考学习者翻译测验作答数据的计算机自动评分结果与人工评分结果进行相关分析及配对样本t检验,并将两种评分方式的诊断结果进行比较。研究发现,计算机自动评分与人工评分结果高度相关,两种评分方式的翻译测验总分无显著差异,总体而言本次翻译测验自动评分结果是可靠的;但计算机自动评分与人工评分对自考学习者的翻译能力结构诊断结果有一定差异。  相似文献   

3.
本研究主要针对非英语专业学生,通过问卷调查和访谈,了解他们对英语自动作文评分系统的使用情况及满意度。研究结果表明,学生普遍对该评分系统比较满意。自动评分系统可以有效增加学生修改作文的频次,但对作文的思想内容、结构方面的反馈有待完善。教师应加强引导,建立多元反馈机制,以充分发挥人工评改和智能评改的优势,促进学生写作兴趣和能力的提高。  相似文献   

4.
《现代教育技术》2018,(2):75-80
随着国内自动评分技术的飞速发展,写作自动评分方法不断涌现,i Write英语写作教学与评阅系统2.0应运而生。该系统拥有包含写作教学系统、智能批改引擎和自动评分引擎在内的多项自主知识产权技术,能够实现对英语作文的语言、内容、篇章结构和技术规范四个维度的机器智能评阅。文章从真实语言测试中获取了大量英语作文语料,充分利用一致性和一致率算法,将作文的人工评分与i Write 2.0的机器评分做了多角度对比,并对i Write 2.0在线英语作文评分的信度进行了实证研究。  相似文献   

5.
本研究运用语料库语言学的方法, 以大学英语英汉翻译测试的译文语料及其正式人工评分为研究对象, 通过语料标注和统计分析, 揭示翻译测试人工评分中评分员关注的主要因素, 并运用多元线性回归方法自动构建评估模型加以验证, 其评价结果与人工评分的皮尔森相关系数达到0.78. 本文以实证的方法初步揭示大学英语翻译测试人工评分中评分员关注的反映译文质量的文本特征项, 为英汉翻译测试译文的自 动量化评价研究提供一些有益的参考 。  相似文献   

6.
本实验检验了《新视野大学英语》作文自动评分系统的评分效度。实验数据表明:系统自动评分与人工评分之间的相关性较显著;系统评分中高分档的评分准确率较低,其他档的评分准确率较高;内容板块较语言板块、篇章结构板块对作文总体评分的影响显著。系统自动评分与适量的人工参与相结合,系统整体评分与细化的文本特征项相结合有助于提高系统的评分效度。  相似文献   

7.
吴晓萍 《文教资料》2014,(34):38-39
基于计算机网络的作文自动评分系统能够在一定程度上解决教师作文评阅负担重和人工评分存在主观影响的问题。本文对国内外出现的作文自动评分系统及相关研究现状做了评述,认为依靠计算机系统评阅作文的信度和效度还有待进一步证实。  相似文献   

8.
本研究通过实验探索人工智能评测技术在人工网上评卷质量监控中的应用及其他相关应用。实验数据采集自2017年安徽省高考语文作文和英语作文共计841 610份试卷,对智能阅卷产生的机器评分、普通高考人工网上评卷产生的人工1评和人工2评以及报道分进行平均分、标准差、相关度、评分一致率等多个维度的数据分析,将智能阅卷产生的异常作答样本和大分差数据样本反馈给学科专家组进行质检评分。结果表明:智能阅卷基本上达到了与评卷教师相当的水平;智能阅卷始终采用统一的评分标准,更具客观公正性,能为人工网上评卷提供有效的质量监控。  相似文献   

9.
本研究通过应用AntConc3.2.1w和SPSS11.5计算机软件,以中国学习者英语语料库(CLEC)和兰开斯特汉语普通话语料库(LCMC)为基础,并且以本族语者语料库LOB&BROWN为参照,针对CONSE-QUENCE的搭配特点进行量化分析,通过对它使用搭配多样性、类联接等多方面对比分析,进而揭示出中国英语学习者使用CONSEQUENCE语义韵的特点。研究表明,与本族语者相比,中国英语学习者表现出正确使用CONSEQUENCE的消极语义韵的特点。通过对汉语和英语CONSEQUENCE语义韵对比分析,本研究发现:汉语的语言迁移是造成中国英语学习者正确使用CONSEQUENCE语义韵的主要原因。  相似文献   

10.
1988年广东省语文高考标准化试验中作文命题及评分的质量如何?我们通过抽取2234份试卷,分别摘抄考生在每条作文上的“内容”、“语言”和“结构”的分数,然后用电子计算机进行了数据处理。结果如下: 从上表中我们可以得到如下信息: 1、在作文的效度方面。考生在作文1的得分与全卷总分的相关为0.64,在作文2的得分与全卷总分的相关为0.71(见最后一行)。经统计检验表明,它们均达到极为显著的水平(P<0.01)。这就是说,考生在这二道作文题上的得分与总分的关系极为密切;或者说考  相似文献   

11.
人机评阅作文的关系探讨促进了自动作文评价系统研究开始从自动作文评分转向计算机辅助写作。大量计算机辅助写作工具软件的引入和应用使得自动作文评价系统在写作教学中越来越充满活力和希望。  相似文献   

12.
Research on Automated Essay Scoring has become increasing important because it serves as a method for evaluating students’ written responses at scale. Scalable methods for scoring written responses are needed as students migrate to online learning environments resulting in the need to evaluate large numbers of written-response assessments. The purpose of this study is to describe and evaluate three active learning methods that can be used to minimize the number of essays that must be scored by human raters while still providing the data needed to train a modern Automated Essay Scoring system. The three active learning methods are the uncertainty-based, the topological-based, and the hybrid method. These three methods were used to select essays included in the Automated Student Assessment Prize competition that were then classified using a scoring model that was trained with the bidirectional encoder representations from a transformer language model. All three active learning methods produced strong results, with the topological-based method producing the most efficient classification. Growth rate accuracy was also evaluated. The active learning methods produced different levels of efficiency under different sample size allocations but, overall, all three methods were highly efficient and produced classifications that were similar to one another.  相似文献   

13.
ABSTRACT

The literature on Automated Essay Scoring (AES) systems has provided useful validation frameworks for any assessment that includes AES scoring. Furthermore, evidence for the scoring fidelity of AES systems is accumulating. Yet questions remain when appraising the scoring performance of AES systems. These questions include: (a) which essays are used to calibrate and test AES systems; (b) which human raters provided the scores on these essays; and (c) given that multiple human raters are generally used for this purpose, which human scores should ultimately be used when there are score disagreements? This article provides commentary on the first two questions and an empirical investigation into the third question. The authors suggest that addressing these three questions strengthens the scoring component of the validity argument for any assessment that includes AES scoring.  相似文献   

14.
自动作文评分系统的技术优势为英语写作教学模式的创新改革提供一个良好的平台。本研究对基于自动作文评分系统的英语写作教学模式进行了设计与教学实践,包括写前阶段、初稿和同伴互评阶段、修改和自动评阋阶段、课堂讲评和定稿阶段的设计。为期一年的写作教学实验表明:新的写作教学模式督促学生写,保持写作的频率,激发学生的写作兴趣,培养学生自主写作能力,提高学生英语写作水平。  相似文献   

15.
赵慧  唐建敏 《教育技术导刊》2019,18(11):168-171
英语写作能力培养一直是大学英语教学的重点和难点,目前自动作文评分AES(Automated Essay Scoring)技术已得到广泛应用,但如何将其与大学英语写作教学有效结合仍有待深入研究。鉴于此,根据我国大学英语写作教学现状,结合L2(Second Language)语言学习特点,在分析AES技术相关原理基础上,对大学英语写作教学模式进行分析研究。结果表明,当前中国大学英语写作教学需结合AES技术和L2语言学习特点,构建基于AES的大学英语教学模式,以激发学生学习兴趣,提升学生英语写作能力。  相似文献   

16.
Automated summary evaluation is proposed as an alternative to rubrics and multiple-choice tests in knowledge assessment. Inbuilt rubric is a recent Latent Semantic Analysis (LSA) method that implements rubrics in an artificially-generated semantic space. It was compared with classical LSA’s cosine-based methods assessing knowledge in a within-subjects design regarding two validation sources: a comparison with the results of rubric scores and multiple-choice tests, and the sensitivity of predicting the academic level of the test-taker. Results showed a higher reliability for inbuilt rubric (from Pearson correlation coefficient .81 to .49) over the classical LSA method (from .61 to .34), and a higher sensitivity using binary logistic regressions and effect sizes to predict academic level. It is concluded that inbuilt rubric has a qualitatively higher reliability and validity than classical LSA methods in a way that is complementary to models based on semantic networks. Thus, it is concluded that new automated summary evaluation approaches such as the inbuilt rubric method can be practical in terms of reliability and efficiency, and, thus, they can offer an affordable and valuable form of knowledge assessment in different educational levels.  相似文献   

17.
A Latent Semantic Index (LSI) was constructed from arguments made by navy officers concerning events in an anti-air warfare scenario. A model based on LSI factor values predicted level of domain expertise with 89% accuracy. The LSI factor space was reduced using MDS to five dimensions: aircraft route, aircraft response, kinematics, localization, and an unclassifiable element. Arguments in the localization category were reliably more common among officers with the greatest expertise. Automated classification of arguments into these elements achieved 84% accuracy. LSI may be a useful tool for automating aspects of modeling expertise and diagnosing knowledge deficiencies.  相似文献   

18.
几个英语作文自动评分系统的原理与评述   总被引:8,自引:0,他引:8  
本文介绍目前美国在大规模考试和英语教学中最为流行的几个作文自动评分系统的基本原理并对这些系统进行简单的评述。所涉及的系统包括Project Essay Grader(PEG),Intelligent Essay Assessor (IEA),E-rater和Criterion,IntelliMetric和MY Access!,Bayesian Essay Test Scoring System(BETSY)。  相似文献   

19.
五四文学革命,各种文学体裁都在掊击传统与借鉴外国中寻求自身的话语定位。散文也是如此,一面批判桐城古文,要求摆脱古文学束缚,一面向外借鉴英法Essay建立现代散文的话语模式,形成了现代散文的两种风格:重叙事抒情的随笔和重哲理议论的杂感。这也正呼应了英国近代散文的两种风格,就其在当时的发展来看,叙事抒情的“兰姆风”影响更大一些。而重哲理议论的“培根风”在五四之初也得到很大发展。从Es-say对中国现代散文话语建立的影响以及现代散文对这一影响有选择的接受这两个方面来讨论现代散文的话语定位。  相似文献   

20.
Automated Essay Scoring (AES) has garnered a great deal of attention from the rhetoric and composition/writing studies community since the Educational Testing Service began using e-rater® and the Criterion® Online Writing Evaluation Service as products in scoring writing tests, and most of the responses have been negative. While the criticisms leveled at AES are reasonable, the more important, underlying issues relate to the aspects of the writing construct of the tests AES can rate. Because these tests underrepresent the construct as it is understood by the writing community, such tests should not be used in writing assessment, whether for admissions, placement, formative, or achievement testing. Instead of continuing the traditional, large-scale, commercial testing enterprise associated with AES, we should look to well-established, institutionally contextualized forms of assessment as models that yield fuller, richer information about the student's control of the writing construct. Such tests would be more valid, as reliable, and far fairer to the test-takers, whose stakes are often quite high.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号