Course Catalog | Chapter Directory
Module D > Data sources
Consider a cluster of questions when deciding how to measure an indicator:
Data sources are tools, documents, and locations for information that can be designed to show what happened in your target audience.
Anecdotes:
Surveys or feedback forms:
Behavior observation or assessment:
Participant projects:
Using participant projects, student and participant skills can be assessed through the use of rubrics that evaluate the level of skills. Six steps to creating a rubric are: 1. Write the student/participant learning outcome; 2. Define the indicators comprising the learning outcome; 3. Identify the levels of quality for each indicator; 4. Assign a point value to each level of quality; 5. Describe the criteria for each level of quality; and 6. Create the rubric table.
Examples of rubrics are: Learning Outcome Rubrics for Cultivation Techniques for Crops and Vegetables for (1) Students demonstrating information literacy skills, and (2) Teachers organizing information into usable database/textbook; a Research Process Rubric for Middle School; and an Information Literacy Value Rubric.
Status observation or assessment:
You’ll learn a lot by doing your logical Planning Model. Remember that you not only have the instructional modules but also the examples (available in the Cases section) to give you ideas and show you how to apply them. When you submit your logical Planning Model to the Wiki discussion page, you will get feedback and suggestions. Remember too your evaluation should be logical and honest, but will not be held to the standards of publishable research. For in-depth guidance on different forms of evaluation, see the references for Module D by clicking the Resources tab at the top of the screen.
For each outcome type, can you rate the usefulness of each type of data source? (for each type of data source, provide good/fair/weak for single-choice selection)
Skills/Knowledge/Behavior (short-term to medium-term):
Behavior Observation or Assessment
Project work products
Surveys or feedback forms
Anecdotes
Observation or Assessment results of status
Attitudes (short-term):
Surveys or feedback forms
Anecdotes
Observation or Assessment results of status
Behavior Observation or Assessment
Project work products
Status/Conditions (long-term)
Observation or Assessment results of status
Surveys or feedback forms
Anecdotes
Behavior Observation or Assessment
Project work products
Your common sense and knowledge of the situation will suggest what data to use. The kinds of data sources explained give the most common types. Remember that Outcomes Based Planning and Evaluation programs do not require formal research; although some programs may wish to get a consultant involved in evaluation. (References to further information can be found by clicking the Resources tab at the top of the screen.)
No kind of data source is better than another. The data source chosen should depend on what is being evaluated.
To indicate that children are developing a habit of reading: “# and % of Springfield students in the summer library reading program who spend at least an hour per day for independent reading for fun.
Anecdotes
Survey
Participant project
Organizational records
To indicate that West Dakota residents use public library databases as a preferred source of information: # and % of WD residents who say that they are likely or very likely to use the public library databases as one of their first 3 sources of health information.
Anecdotes
Survey
Participant project
Organizational records
To indicate that Student Bird Watchers learn bird-identification skills: “# and % of Student Bird Watchers who correctly identify five birds common to the area on a field trip.
Anecdotes
Survey
Participant project
Organizational records
To indicate 4th-8th grade teacher-participants who demonstrate ability to teach biodiversity with inquiry-based methods: “:# and % of teachers who implement completed science curriculum unit in the classroom.â€
Anecdotes
Survey
Participant project
Organizational records
Matrix of common needs analysis and evaluation data sources
Purpose Requirements Advantages Disadvantages Questionnaires and Surveys To collect standardized data from a large number of participants. Construction of survey. Paper, scannable forms, CBT embedded surveys, e-mail, intranet, or Internet can be used. If not proctored, participants cannot ask for clarification or instructions. Knowledge Assessments To assess participants' knowledge acquired through training or in the workplace or other environment. Construction of assessment test. Paper, scannable forms, CBT embedded tests, e-mail, intranet, or Internet can be used. If a pretest is used, participants may score higher on the posttest due to familiarity. Performance Assessments To assess participants' application of skills acquired through training or in the workplace or other environment. Construction of assessment (checklist, rating form). Paper, scannable forms, and interactive multimedia embedded tests can be used. Potential rater bias. Structured Observation To watch an activity and record what is seen. Construction of checklist or rating form. Objective of interest (for example, learner, designer, instructor, work sample) can be observed by a senior trainer, subject matter expert, designer/developer, evaluation specialist, supervisor, or manager. Potential observer bias. Focus Groups To explore a topic in-depth with a small number of participants. Development of session questions. Depth of inquiry possible. Potential group bias. Telephone Interviews To collect standardized reporting data over the telephone. Creation of interview transcript and recording form. Probing of incomplete answers is possible. Potential interview bias. Based on National Leadership Grant tutorial by the Institute of Museum and Library Services found at http://www.imls.gov/Project_Planning/index.asp
Tool
Explicit verbal and/or written instructions.
If not proctored, follow-up contact to increase return rate.
Variety of response alternatives can be used (Likert-type scales, multiple choice, open-ended).
Questions can be direct or indirect, general or specific
Easy and efficient to administer.
Participants may choose more than one response or give invalid responses.
Open-ended responses may be grammatically incorrect or ambiguous.
Participants may skip items.
Response rate may be low, if not proctored.
Assessment must have content validity and reliability.
Must use exact test or an equivalent test for pre- and post-assessment.
Easy to administer pre- and post-assessment.
Some participants may experience test anxiety and may not perform optimally.
Tasks or skills to be performed must be identified.
Criteria or standards of performance must be specified and realistic.
Assessment must have content validity and reliability.
Performance can be rated by instructor, participant, peer, team or supervisors.
Provides direct evidence or application of learning.
Can be obtrusive or unobtrusive.
Equipment and tools may not be available in the classroom.
If obtrusive, the observation and rating may affect performance due to nervousness or anxiety.
May not have time to measure all critical skills.
Pretest may not be practical given time constraints.
Explicit directions for the observer
Observer must be objective.
Observer should not give directions, answers, or nonverbal cues when observing performance.
Consistent use of the checklist.
Percentage agreement between observers should be calculated.
Observation of performance can be obtrusive or unobtrusive (if unobtrusive, observation can be naturalistic).
If obtrusive, observation of behavior may affect performance due to nervousness or anxiety.
Trained moderator.
Limited number of participants (6-12)
Different sessions for different groups of participants.
Opportunity for clarification.
Synergistic and snowball effects .
Potential moderator bias.
A few participants may dominate.
Results are not representative given the limited number of participants.
Trained interviewers.
Multiple contact attempts.
Clarification of misunderstandings is possible.
Interviewer has greater control over data collection.
Participants may give socially desirable answers.
Questions and response alternatives must be simple because participants are unable to see the questionnaire.
Participants may be difficult to contact.
Source: Falletta, Salvatore and Combs, Wendy. Info-line: Evaluating Technical Training: A Functional Approach. (September, 1997). Page 12-15. Alexandria, VA. ASTD (www.ASTD.org) (Used with permission.)
If your indicators are specific, usually they will suggest the data source to measure it as well as the people in the target audience it should be applied to.
What people or program data will the indicator be applied to? Consider the following issues.
Examine the visuals below to explore issues related to each program. For background on programs, click Cases.
Some or all of the target audience?
Remember that your target audience is usually stated in general characteristics and you may only involve part of that audience in your program as participants.
West Dakota Library Rx
All WD residents are part of the target audience, but they cannot all be surveyed. A random sample of residents is assumed to be representatives of the whole group.
Some or all of the participants?
Springfield Library Summer Reading Program
All the participants are students in an urban library district. Should reading tests be monitored for all participants? At-risk students only? Those who complete a specific number of activities?
People other than the participants?
Springfield Library Summer Reading Program
Are parents likely to be better sources of information about their children (such as number of hours spent reading and attitude toward reading) than the children themselves?
Remember: Collecting data costs time and money. Collect only enough information to figure out if your program is successful, so be specific and concrete. Consider the difference in costs between collecting information about “children†and “children who need after-school tutoring.†Or do you mean “children who participate in at least five tutoring sessions? If you expect 100 children to meet this criterion, should you say instead “a random sample of children who participate in at least five tutoring sessions�
For the following outcomes and indicators for the Memoirs program, try choosing data sources and how they should be applied.
Outcome 1: Participants show improvement in their writing.
Indicator : # and % of participants who revise five pieces, commenting on what they tried to improve in each revision.
Applied To : All workshop participants
Data Source : Writers’ portfolios of participants work
Indicator : # and % of participants whose revised pieces (two before-and-after versions) are judged better than the originals in a blind (no writer or dates given) overall grading by a creative writing specialist
Applied To : All workshop participants
Data Source : Expert evaluation of participants’ work
Indicator : # and % of participants whose revised pieces are judged better than the originals (for two sets) by a creative writing specialist when judged by the writer’s goals in the revision.
Applied To : All workshop participants
Data Source : Expert evaluation of participants’ work
Outcome 2: Participants demonstrate they feel themselves to be part of a community of writers.
Indicator : # and % of participants who can name three ways they feel a part of the community of writers.
Applied To : Program participants
Data Source : exit survey
Indicator : # and % of participants who act as part of a community of writers after the program (produce writing, continue library group or join another, attend readings, read memoirs, read regularly about authors’ concerns)
Applied To :Program participants
Data Source : Phone interview with checklist of behaviors
创建自: JacmanChin .
最后修改: 星期日 22 of 5月, 2011 00:47:21 MDT 作者 JacmanChin .
原始文件位于 https://evergreeneducation.org/wikipub/tiki-index.php?page=Module%20D%20Data%20sources