[ReadWrite 2479] Re: Writing assessments
Archived Content Disclaimer
This page contains archived content from a LINCS email discussion list that closed in 2012. This content is not updated as part of LINCS’ ongoing website maintenance, and hyperlinks may be broken.
Fri Apr 13 12:10:18 EDT 2012
- Previous message: [ReadWrite 2478] Writing assessments
- Next message: [ReadWrite 2480] Annoucement: Upcoming Guest Discussion on the Assessment Discussion List
- Messages sorted by: [ date ] [ thread ] [ subject ] [ author ]
Hello Daphne and everyone.
I'm sorry I'm not answering your question but I wanted to highlight a new report featured in Education Week
The study shows that automated computer scoring of essays is as reliable and accurate as scoring by a human. The study was funded by the Hewlett Foundation
I wonder to what extent the computer is able to take into account the special characteristics that tend to appear in the writing of non-native writers of English and weigh those differently than they might score the mistakes in grammar and structure made by native speakers of English
Heide Spruck Wrigley
From: readwrite-bounces at lincs.ed.gov [mailto:readwrite-bounces at lincs.ed.gov] On Behalf Of Daphne Greenberg
Sent: Thursday, April 12, 2012 8:46 AM
To: The Reading and Writing Skills Discussion List
Subject: [ReadWrite 2478] Writing assessments
As many of you know, from time to time, I highlight a resource found in the LINCS Basic Skills Resource Collections
The resource I would like to share with you today is called: Study Guide for "Some Consequences of Writing Assessment"
The following are questions that are mentioned in this resource, that would be helpful to think about, and I wonder how people on this list would respond to them:
How are your students' writing abilities assessed?
When/how often do these assessments take place?
How are the results used? Who sees the results?
-------------- next part --------------
An HTML attachment was scrubbed...