Difference between revisions of "Assessment"

Difference between revisions of "Assessment"

From Learning and training wiki

Share/Save/Bookmark
Jump to: navigation, search
m (added link to Peer-to-Peer Assessment entry)
 
(24 intermediate revisions by 5 users not shown)
Line 1: Line 1:
{{Term|ASSESSMENT|Continuous process with the objective of understanding and increasing learning by defining, selecting, designing, collecting, analyzing and interpreting the information obtained through observation and measurement of the learners. Assessments provide faculty with a better understanding of what learners are learning and engage them more deeply in the process of learning content. Learning and assessment are correlated, since the learning tasks stimulate the learners to develop some competencies and the assessment makes them demonstrate these competencies. To make an assessment more effective, the learners should be given the opportunity to practice with that form of evaluation before it becomes effective. By using assessment strategies that draw learners into the assessment process, they learn more of the content while getting the benefits of learning skills that will be useful to them in future. It also provides information to be used as feedback to modify the teaching and learning activities in which learners are engaged.<ref> [http://serc.carleton.edu serc.carleton.edu ](15 April 2008), [http://www.ltscotland.org.uk www.ltscotland.org.uk](15 April 2008)</ref>
+
{{Term|ASSESSMENT|A generic term used to describe any systematic method, e.g., quizzes, tests, surveys and exams, whose purpose is to evaluate and measure learners’ knowledge, skills, abilities and attitudes by collecting, selecting, analyzing and interpreting the information obtained through performance observation.
  
Fadel, Honey and Pasnik (2007) believe that we are "shifting toward an innovation-based economy" where successful intelligence will be composed of "a three-point foundation of analytical, practical and creative skills".  They suggest that future assessment should be created to be "assessments ''for'' learning, rather than assessments ''of'' learning", so that the assessment process benefits students more effectively.  They also recommend that assessment "be performance-based" (so as to be practical), that it "makes students' thinking visible" (so that it can be observed and measured), "that it generate data that can be acted upon" (by teachers and administrators in improving the learning process), "that it build capacity in both teachers and students, and be part of a comprehensive and well-aligned curriculum".<ref>Fadel, C., Honey, M., & Pasnik, S. (2007, May 23). Assessment in the Age of Innovation. Education Week, 34, 40.</ref>
+
Learning and assessments are correlated since the learning tasks stimulate the development of some competencies/skills/abilities and the assessment makes learners demonstrate their achievement.<ref> [http://serc.carleton.edu serc.carleton.edu ](15 April 2008), [http://www.ltscotland.org.uk www.ltscotland.org.uk](15 April 2008)</ref>
  
When computers are involved in education and assessment, special attention must be paid to the way this is done in order to keep the goal of learning in mind.  Lin and Dwyer (2006) discuss the two orders of the fingertip effect (perpetual access to computers and the Internet), stating that the first-order effect changes how people accomplish tasks, but not what they are attempting to do, whereas the second-order effect changes the way people think and what they are attempting to accomplish.  Considering these effects, it is important to determine which effects the computer-based assessment has achieved, and to strive to use computers not only for assessment administration (creation and correction), but also for changing the type of assessment given.<ref>Lin, H., & Francis, D. (2006). The Fingertip Effects of Computer-based Assessment in Education. TechTrends, 50 (6), 27-31.</ref>
+
There are five different styles of assessments which have different purposes as described below:
  
It is also important to not overdo assessment and to keep in mind its ultimate goal: to enhance learning.  Pelligrino (1999) points out that the majority of the most common assessments serve the needs of audiences other than the learner (SATs, for example).  In an example of diagnostic assessment, meant to help learning, Pelligrino cites a multiple choice question where the learner is asked not only what the correct answer is, but the reason why their chosen answer is correct. In this way, the instructor can see the different ways in which students may be misunderstanding a concept.<ref>Pelligrino, J. W. (1999). The Evolution of Educational Assessment : Considering the Past and Imagining the Future.  William H. Anghoff Memorial Lecture Series (pp. 1-19). New Jersey: Educational Testing Service, Princeton University.</ref>
+
'''Diagnostic Assessments''': identify the needs and prior knowledge of learners with the purpose of directing them to the most appropriate learning experience.
 +
Pre-learning assessments (pre-tests, placement tests, personality assessments)
  
See also: [[Assessment Standardization]]; [[Peer-to-Peer Assessment]]; [[Performance Assessment]]; [[Evaluation]]; [[Formative Evaluation]]; [[Instructional Design (ID)]]; [[Learning Objectives]]; [[Summative Evaluation]]
+
'''Formative Assessments''': strengthen memory recall providing practice for search and retrieval from memory, correct misconceptions and promote confidence in learners’ knowledge.
 +
Practice tests and exams (quizzes during learning, self assessment of knowledge, skills, attitudes)
 +
 
 +
'''Needs Assessments''': determine the needs or "gaps" between learners’ current and desired conditions to improve present performance or to correct a deficiency.
 +
Needs Analysis Surveys
 +
 
 +
'''Reaction Assessments''': determine the satisfaction level with a learning or assessment experience.
 +
Course evaluation Surveys, Employee Opinion Surveys, Customer/Partner satisfaction Surveys
 +
 
 +
'''Summative Assessments''': measure or certify knowledge, skills and aptitudes.
 +
Post Course Tests, Exams during study, Internal Exams, Open Exams, Licensing Exams, Pre-Employment Tests<ref> Shepherd, E., Godwin J., “Assessments through the learning process”, QuestionMark White Paper, 2010, pp. 3-9.</ref>
 +
 
 +
See also: [[A.D.D.I.E Model]], [[Assessment Standardization]]; [[Peer-to-Peer Assessment]]; [[Performance Assessment]]; [[Formative Evaluation]]; [[Instructional Design (ID)]]; [[Learning Objectives]]; [[Summative Evaluation]]
  
 
}}
 
}}
Line 13: Line 26:
  
  
{{Addlink|Below you have the link to a slideshow that summarizes some of the concepts related to learning objectives:}}
+
{{Tool|General Guidelines for developing assessments|
 +
Good assessments correspond to well-written [[Learning_Objectives|learning objectives]].
 +
The following list  shows how early in the instructional process they should be designed:
 +
 
 +
*Identify learning objectives;
 +
*Design and build assessments;
 +
*Design and build content and activities;
 +
*Conduct [[Formative_Evaluation|formative evaluation]];
 +
*Revise assessments, contents, and activities;
 +
*Complete development;
 +
*Conduct [[Summative_Evaluation|summative evaluation]];
 +
*Maintain the course.<ref> Patti Shank, Develop valid assessments, Infoline December 2009, ASTD Press, p. 2</ref>
 +
 
 +
=='''Design and build assessments'''==
 +
*'''Assessment methods''': after having determined key learning objectives, it is necessary to identify which type of assessment is appropriate to determine the level of knowledge/ performance achieved as a result of the learning activity.
 +
 
 +
If the objective is a knowledge objective which calls for recalling or selecting, test items can be used. Below is a list<ref>[http://www.businessballs.com/bloomstaxonomyoflearningdomains.htm www.businessballs.com] (10 November 2011) </ref>  based on [[Bloom's_Taxonomy|Bloom’s Taxonomy]] matching cognitive objectives with appropriate test assessments:
 +
 
 +
 
 +
[[Image:‎Bloom's_Taxonomy_Table.jpg|center]]
 +
 
 +
 
 +
If the objective calls for performance, learners should be asked to actively demonstrate their knowledge.  The goal of [[Performance_Assessment|Performance Assessment]] is to test learners in real or realistic situations. In those circumstances, learners need to perform, not merely recall or select information.
 +
 
 +
*'''Assessment Plan''': use a table that shows the format in which each learning objective will be assessed and the number of necessary assessments to test the range of conditions presented by learning objectives.
 +
 
 +
*'''Passing Grades''': determine cut-off scores for assessments, for example:
 +
 
 +
- Common sense cut-off (considering the lowest level of acceptable performance);
 +
 
 +
- Percentage of total (identifying a passing grade for the entire assessment and a minimum grade for each  learning objective in the task. Typically learners must pass both in order to pass the assessment. This choice is selected when a learning objective is more important than others).
 +
 
 +
*'''Design test or performance assessments''': the five most common test question types are true/false, short answers, fill-in-the blank, matching, and multiple choice questions. Bryan Hopkins, a training consultant with over 20 years of experience in developing effective training programmes, shared a comprehensive article on writing questions for training programmes: [[Media:Writing_Questions_for_Training_Programmes_Bryan_Hopkings.pdf‎|Writing Questions for Training Programmes]]. He distinguishes between formative and summative questions. The former ones help someone test their understanding, while the latter ones check the learner’s overall mastery of a subject.
 +
 
 +
Some examples of performance assessment are: [[Simulation|simulation]], games, group projects, individual projects, internships, laboratory problems, probationary work assignments. }}
 +
 
 +
=='''Job Aid'''==
 +
 
 +
[[Image:pdf.png]] [[Media:Toolkit_Assessment.pdf‎‎‎|General Guidelines for Developing Assessments.pdf‎]]
 +
 
 +
 
 +
{{Addmaterial|}}
 
{|border=1; width= 100%
 
{|border=1; width= 100%
!Link
+
!width= 200pt|Document
!Content
+
!width= 575pt|Content
 
|-
 
|-
|[http://www.slideshare.net/kernlearningsolutions/assessments-in-e-learning Assessments in e-Learning ]
+
|[[Media:Writing_Questions_for_Training_Programmes_Bryan_Hopkings.pdf‎|Writing Questions for Training Programmes written by Bryan Hopkins]]
|This slideshow presents some of the basic concepts related to the development of good learning objectives and to the importance of measuring learning. It also presents different types of assessment.
+
|General principles for writing good and effective questions for training programmes.
 
|}
 
|}
  
  
  
{{Tool|Implementing Effective Assessment|
 
== Some points to keep in mind when designing assessment: ==
 
*Determine clearly defined learning objectives.
 
*Create assessment that is aligned with those learning objectives.
 
*Use assessment types that allow proper measurement of level of attainment of objectives.  For instance, if your objective is to describe events which lead to a particular war, have the learner describe the events in an essay question, not pick them from a list in a multiple-choice question.
 
  
}}
+
{{Addlink|Below you have a list of selected websites where you can find additional information:}}
 +
{|border=1; width= 100%
 +
!Link
 +
!Content
 +
|-
 +
|[http://www.slideshare.net/kernlearningsolutions/assessments-in-e-learning Assessments in e-Learning ]
 +
|This slideshow presents some of the basic concepts related to the development of good learning objectives and to the importance of measuring learning. It also presents different types of assessment.
 +
|-
 +
|[http://www.youtube.com/user/eLearnerEngaged#p/u/19/plo61m_4-S4 Developing powerful Training Assessments]
 +
|Video about what to consider when performing assessments in training and teaching. It also describes what a powerful assessment is and why it is important in instructional design.
 +
|-
 +
|[http://www.edutopia.org/comprehensive-assessment-introduction How Should We Measure Student Learning? The Many Forms of Assessment]
 +
|Article stating the different existing forms of measuring students' abilities.
 +
|-
 +
|[http://www.edutopia.org/comprehensive-assessment-overview-video Performance assessments]
 +
|Video explaining performance assessments approach in evaluating what students know and can do.
 +
|-
 +
|[http://capacitydevelopmentindex.pbworks.com/f/WBI_Learning+Design+Model.pdf WBI Learning Design Process]
 +
|This document contains the learning design process illustrated by the Word Bank Institute. Five main steps for learning design are identified, going from the definition of outcomes and objectives, to the identification of methods, the selection of tools, the assembling of the programme and finally the evaluation of the tool. For each of these fundamentals steps, additional information is provided. A specific section is dedicated to identifying methods for assessing learners' knowledge, skills, abilities and attitudes.
 +
|-
 +
|[http://anethicalisland.files.wordpress.com/2013/05/event82.png Assessment Time, Infographic]
 +
|How do you assess your students' performance? Was the assessment of the performance reliable and valid? Follow these 27 tips to reliably and validly assess the performance.
 +
|}
  
 
== References ==
 
== References ==
 
<references/>
 
<references/>

Latest revision as of 11:52, 5 November 2013

Term2.png ASSESSMENT
A generic term used to describe any systematic method, e.g., quizzes, tests, surveys and exams, whose purpose is to evaluate and measure learners’ knowledge, skills, abilities and attitudes by collecting, selecting, analyzing and interpreting the information obtained through performance observation.

Learning and assessments are correlated since the learning tasks stimulate the development of some competencies/skills/abilities and the assessment makes learners demonstrate their achievement.[1]

There are five different styles of assessments which have different purposes as described below:

Diagnostic Assessments: identify the needs and prior knowledge of learners with the purpose of directing them to the most appropriate learning experience. Pre-learning assessments (pre-tests, placement tests, personality assessments)

Formative Assessments: strengthen memory recall providing practice for search and retrieval from memory, correct misconceptions and promote confidence in learners’ knowledge. Practice tests and exams (quizzes during learning, self assessment of knowledge, skills, attitudes)

Needs Assessments: determine the needs or "gaps" between learners’ current and desired conditions to improve present performance or to correct a deficiency. Needs Analysis Surveys

Reaction Assessments: determine the satisfaction level with a learning or assessment experience. Course evaluation Surveys, Employee Opinion Surveys, Customer/Partner satisfaction Surveys

Summative Assessments: measure or certify knowledge, skills and aptitudes. Post Course Tests, Exams during study, Internal Exams, Open Exams, Licensing Exams, Pre-Employment Tests[2]

See also: A.D.D.I.E Model, Assessment Standardization; Peer-to-Peer Assessment; Performance Assessment; Formative Evaluation; Instructional Design (ID); Learning Objectives; Summative Evaluation


Toolkit.png General Guidelines for developing assessments

Good assessments correspond to well-written learning objectives. The following list shows how early in the instructional process they should be designed:

  • Identify learning objectives;
  • Design and build assessments;
  • Design and build content and activities;
  • Conduct formative evaluation;
  • Revise assessments, contents, and activities;
  • Complete development;
  • Conduct summative evaluation;
  • Maintain the course.[3]

Design and build assessments

  • Assessment methods: after having determined key learning objectives, it is necessary to identify which type of assessment is appropriate to determine the level of knowledge/ performance achieved as a result of the learning activity.

If the objective is a knowledge objective which calls for recalling or selecting, test items can be used. Below is a list[4] based on Bloom’s Taxonomy matching cognitive objectives with appropriate test assessments:


Bloom's Taxonomy Table.jpg


If the objective calls for performance, learners should be asked to actively demonstrate their knowledge. The goal of Performance Assessment is to test learners in real or realistic situations. In those circumstances, learners need to perform, not merely recall or select information.

  • Assessment Plan: use a table that shows the format in which each learning objective will be assessed and the number of necessary assessments to test the range of conditions presented by learning objectives.
  • Passing Grades: determine cut-off scores for assessments, for example:

- Common sense cut-off (considering the lowest level of acceptable performance);

- Percentage of total (identifying a passing grade for the entire assessment and a minimum grade for each learning objective in the task. Typically learners must pass both in order to pass the assessment. This choice is selected when a learning objective is more important than others).

  • Design test or performance assessments: the five most common test question types are true/false, short answers, fill-in-the blank, matching, and multiple choice questions. Bryan Hopkins, a training consultant with over 20 years of experience in developing effective training programmes, shared a comprehensive article on writing questions for training programmes: Writing Questions for Training Programmes. He distinguishes between formative and summative questions. The former ones help someone test their understanding, while the latter ones check the learner’s overall mastery of a subject.
Some examples of performance assessment are: simulation, games, group projects, individual projects, internships, laboratory problems, probationary work assignments.

Job Aid

Pdf.png General Guidelines for Developing Assessments.pdf‎


MATERIAL.png Additional Materials
Document Content
Writing Questions for Training Programmes written by Bryan Hopkins General principles for writing good and effective questions for training programmes.



Link icon.png Web Resources
Below you have a list of selected websites where you can find additional information:
Link Content
Assessments in e-Learning This slideshow presents some of the basic concepts related to the development of good learning objectives and to the importance of measuring learning. It also presents different types of assessment.
Developing powerful Training Assessments Video about what to consider when performing assessments in training and teaching. It also describes what a powerful assessment is and why it is important in instructional design.
How Should We Measure Student Learning? The Many Forms of Assessment Article stating the different existing forms of measuring students' abilities.
Performance assessments Video explaining performance assessments approach in evaluating what students know and can do.
WBI Learning Design Process This document contains the learning design process illustrated by the Word Bank Institute. Five main steps for learning design are identified, going from the definition of outcomes and objectives, to the identification of methods, the selection of tools, the assembling of the programme and finally the evaluation of the tool. For each of these fundamentals steps, additional information is provided. A specific section is dedicated to identifying methods for assessing learners' knowledge, skills, abilities and attitudes.
Assessment Time, Infographic How do you assess your students' performance? Was the assessment of the performance reliable and valid? Follow these 27 tips to reliably and validly assess the performance.

References

  1. serc.carleton.edu (15 April 2008), www.ltscotland.org.uk(15 April 2008)
  2. Shepherd, E., Godwin J., “Assessments through the learning process”, QuestionMark White Paper, 2010, pp. 3-9.
  3. Patti Shank, Develop valid assessments, Infoline December 2009, ASTD Press, p. 2
  4. www.businessballs.com (10 November 2011)