Fullscreen
Loading...

菜单 [toggle]

Tikiwiki助手

感谢您安装Tikiwiki!

登录To begin configuring Tiki, please 登录 as the Admin.

The Tikiwiki CommunityTo learn more, visit: http://www.tikiwiki.org.

Tikiwiki DocumentationFor help, visit http://doc.tikiwiki.org.

打印

Course Catalog  |  Chapter Directory

NEXT

Module D > Data sources  
 

 

Data Sources 

 

How can you measure that?

Consider a cluster of questions when deciding how to measure an indicator:

  • What data sources can give evidence about an indicator?
  • What part of the target audience(s) is an indicator applied to?  
  • What’s the data interval (that is, timing) required for gathering data? (before and after? Every year of the program?)

 

Data sources

Data sources are tools, documents, and locations for information that can be designed to show what happened in your target audience.

  • Anecdotes
  • Surveys or feedback forms
  • Observation or assessment reports
  • Participant projects
  • Other organization’s records or test information

Anecdotes:

  • Okay for attitudes, although you may just hear from happy people. 
  • Worst for skills, as people will want to talk about what they’re good at.
  • Good for information on outcomes you hadn’t anticipated.
  • Good for quotable quotes.

Surveys or feedback forms:

  • Surveys depend on “self-reporting”—fine for confidence in a skill, or interest in a subject—but not objective measures of skill levels or of behaviors.
  • Best for attitudes because survey questions can include quantifiable rating scales.
  • Weak for behaviors (depending on time between the experience and the survey), and weakest for skills.
  • Surveys are good for important outcomes that you were able to anticipate.
  • Look for standardized surveys (professionally and expertly designed) that cover the outcomes you’re interested in.  This may save time, but some may need to be adapted.  

Behavior observation or assessment:

  • With observation, an expert or 3rd party professional organization with expertise watches participant behaviors—mostly during a particular program—and can document objectively whether skills or behaviors have actually changed. Professional assessments include data sources such as teacher checklist of student behavior, portfolio assessment.
  • Best for skills and behaviors because they provide objective assessment by third-party observers.
  • Bad for attitudes.

Participant projects:

  • Best for skills and some behaviors.
  • Weakest for attitudes because it is impersonal.
  • Example: Tip sheets created by participants in a bird identification workshop show that they’ve internalized knowledge about bird characteristics enough to articulate them clearly.                      

     

    Coach

    Using participant projects, student and participant skills can be assessed through the use of rubrics that evaluate the level of skills.  Six steps to creating a rubric are: 1. Write the student/participant learning outcome; 2. Define the indicators comprising the learning outcome; 3. Identify the levels of quality for each indicator; 4. Assign a point value to each level of quality; 5. Describe the criteria for each level of quality; and 6. Create the rubric table.    

    Examples of rubrics are: Learning Outcome Rubrics for Cultivation Techniques for Crops and Vegetables for (1) Students demonstrating information literacy skills, and (2) Teachers organizing information into usable database/textbook; a Research Process Rubric for Middle School; and an Information Literacy Value Rubric.                                         

     

Status observation or assessment:

  • Useful for status or condition. A 3rd party professional organization with expertise can provide observation or assessment results, which accurately reflect the status or condition of interest, e.g. reading scores, grades, graduation rates, attendance, employment.
  • Good for knowledge
  • Weakest for behavior and attitude
  • Many programs start as a way to improve some indicator that already exists but is lagging: reading scores, grades, graduation rates, attendance, employment. Using these scores is inexpensive because they already exist, but both the organization that creates them and the individual participants need to cooperate to make them available to the program. Note: To justify that a change in status or condition is induced by your program is difficult. Your program, if successful, may play only a small part in the change in scores.

 

Coach

You’ll learn a lot by doing your logical Planning Model. Remember that you not only have the instructional modules but also the examples (available in the Cases section) to give you ideas and show you how to apply them. When you submit your logical Planning Model to the Wiki discussion page, you will get feedback and suggestions. Remember too your evaluation should be logical and honest, but will not be held to the standards of publishable research. For in-depth guidance on different forms of evaluation, see the references for Module D by clicking the Resources tab at the top of the screen.

 

For each outcome type, can you rate the usefulness of each type of data source? (for each type of data source, provide good/fair/weak for single-choice selection)

Skills/Knowledge/Behavior (short-term to medium-term):
Behavior Observation or Assessment
Project work products
Surveys or feedback forms
Anecdotes
Observation or Assessment results of status

Attitudes (short-term):
Surveys or feedback forms
Anecdotes
Observation or Assessment results of status
Behavior Observation or Assessment
Project work products

Status/Conditions (long-term)
Observation or Assessment results of status
Surveys or feedback forms
Anecdotes
Behavior Observation or Assessment
Project work products

 

Answer

 

Data: What makes sense to use?

Your common sense and knowledge of the situation will suggest what data to use. The kinds of data sources explained give the most common types. Remember that Outcomes Based Planning and Evaluation programs do not require formal research; although some programs may wish to get a consultant involved in evaluation. (References to further information can be found by clicking the Resources tab at the top of the screen.)

No kind of data source is better than another. The data source chosen should depend on what is being evaluated.

 

To indicate that children are developing a habit of reading: “# and % of Springfield students in the summer library reading program who spend at least an hour per day for independent reading for fun.

Anecdotes
Survey
Participant project
Organizational records

To indicate that West Dakota residents use public library databases as a preferred source of information: # and % of WD residents who say that they are likely or very likely to use the public library databases as one of their first 3 sources of health information.

Anecdotes
Survey
Participant project
Organizational records

To indicate that Student Bird Watchers learn bird-identification skills: “# and % of Student Bird Watchers who correctly identify five birds common to the area on a field trip.

Anecdotes
Survey
Participant project
Organizational records

To indicate 4th-8th grade teacher-participants who demonstrate ability to teach biodiversity with inquiry-based methods: “:# and % of teachers who implement completed science curriculum unit in the classroom.”  

Anecdotes
Survey
Participant project
Organizational records 

 

Answer

 

   

Dig Deeper

Matrix of common needs analysis and evaluation data sources

Tool

Purpose

Requirements

Advantages

Disadvantages

Questionnaires and Surveys

To collect standardized data from a large number of participants.

Construction of survey. 
Explicit verbal and/or written instructions. 
If not proctored, follow-up contact to increase return rate.

Paper, scannable forms, CBT embedded surveys, e-mail, intranet, or Internet can be used. 
Variety of response alternatives can be used (Likert-type scales, multiple choice, open-ended). 
Questions can be direct or indirect, general or specific 
Easy and efficient to administer.

If not proctored, participants cannot ask for clarification or instructions. 
Participants may choose more than one response or give invalid responses. 
Open-ended responses may be grammatically incorrect or ambiguous. 
Participants may skip items.
Response rate may be low, if not proctored.

Knowledge Assessments

To assess participants' knowledge acquired through training or in the workplace or other environment.

Construction of assessment test. 
Assessment must have content validity and reliability. 
Must use exact test or an equivalent test for pre- and post-assessment.

Paper, scannable forms, CBT embedded tests, e-mail, intranet, or Internet can be used. 
Easy to administer pre- and post-assessment.

If a pretest is used, participants may score higher on the posttest due to familiarity. 
Some participants may experience test anxiety and may not perform optimally.

Performance Assessments

To assess participants' application of skills acquired through training or in the workplace or other environment.

Construction of assessment (checklist, rating form). 
Tasks or skills to be performed must be identified. 
Criteria or standards of performance must be specified and realistic. 
Assessment must have content validity and reliability.

Paper, scannable forms, and interactive multimedia embedded tests can be used. 
Performance can be rated by instructor, participant, peer, team or supervisors. 
Provides direct evidence or application of learning. 
Can be obtrusive or unobtrusive.

Potential rater bias. 
Equipment and tools may not be available in the classroom. 
If obtrusive, the observation and rating may affect performance due to nervousness or anxiety. 
May not have time to measure all critical skills. 
Pretest may not be practical given time constraints.

Structured Observation

To watch an activity and record what is seen.

Construction of checklist or rating form. 
Explicit directions for the observer 
Observer must be objective. 
Observer should not give directions, answers, or nonverbal cues when observing performance.
Consistent use of the checklist. 
Percentage agreement between observers should be calculated.

Objective of interest (for example, learner, designer, instructor, work sample) can be observed by a senior trainer, subject matter expert, designer/developer, evaluation specialist, supervisor, or manager. 
Observation of performance can be obtrusive or unobtrusive (if unobtrusive, observation can be naturalistic).

Potential observer bias. 
If obtrusive, observation of behavior may affect performance due to nervousness or anxiety.

Focus Groups

To explore a topic in-depth with a small number of participants.

Development of session questions. 
Trained moderator. 
Limited number of participants (6-12) 
Different sessions for different groups of participants.

Depth of inquiry possible. 
Opportunity for clarification. 
Synergistic and snowball effects .

Potential group bias. 
Potential moderator bias. 
A few participants may dominate. 
Results are not representative given the limited number of participants.

Telephone Interviews

To collect standardized reporting data over the telephone.

Creation of interview transcript and recording form. 
Trained interviewers. 
Multiple contact attempts.

Probing of incomplete answers is possible. 
Clarification of misunderstandings is possible. 
Interviewer has greater control over data collection.

Potential interview bias. 
Participants may give socially desirable answers. 
Questions and response alternatives must be simple because participants are unable to see the questionnaire. 
Participants may be difficult to contact.

Based on National Leadership Grant tutorial by the Institute of Museum and Library Services found at http://www.imls.gov/Project_Planning/index.asp
Source: Falletta, Salvatore and Combs, Wendy. Info-line: Evaluating Technical Training: A Functional Approach. (September, 1997). Page 12-15. Alexandria, VA. ASTD (www.ASTD.org) (Used with permission.)



Choosing data sources and the people they’re applied to

If your indicators are specific, usually they will suggest the data source to measure it as well as the people in the target audience it should be applied to.

What people or program data will the indicator be applied to? Consider the following issues.

Examine the visuals below to explore issues related to each program. For background on programs, click Cases.

Some or all of the target audience?
Remember that your target audience is usually stated in general characteristics and you may only involve part of that audience in your program as participants. 

West Dakota Library Rx
All WD residents are part of the target audience, but they cannot all be surveyed. A random sample of residents is assumed to be representatives of the whole group.

Some or all of the participants?

Springfield Library Summer Reading Program 
All the participants are students in an urban library district. Should reading tests be monitored for all participants? At-risk students only? Those who complete a specific number of activities?

People other than the participants?

Springfield Library Summer Reading Program
Are parents likely to be better sources of information about their children (such as number of hours spent reading and attitude toward reading) than the children themselves?

Remember: Collecting data costs time and money. Collect only enough information to figure out if your program is successful, so be specific and concrete. Consider the difference in costs between collecting information about “children” and “children who need after-school tutoring.” Or do you mean “children who participate in at least five tutoring sessions? If you expect 100 children to meet this criterion, should you say instead “a random sample of children who participate in at least five tutoring sessions”? 

 

Library example: Riverton Memoirs program indicators

For the following outcomes and indicators for the Memoirs program, try choosing data sources and how they should be applied. 

Outcome 1: Participants show improvement in their writing.

Indicator # and % of participants who revise five pieces, commenting on what they tried to improve in each revision

Applied To : All workshop participants

Data Source : Writers’ portfolios of participants work

Indicator : # and % of participants whose revised pieces (two before-and-after versions) are judged better than the originals in a blind (no writer or dates given) overall grading by a creative writing specialist 

Applied To : All workshop participants

Data SourceExpert evaluation of participants’ work

Indicator : # and % of participants whose revised pieces are judged better than the originals (for two sets) by a creative writing specialist when judged by the writer’s goals in the revision.

Applied To : All workshop participants

Data SourceExpert evaluation of participants’ work

Outcome 2: Participants demonstrate they feel themselves to be part of a community of writers.

Indicator # and % of participants who can name three ways they feel a part of the community of writers.

Applied To : Program participants

Data Sourceexit survey

Indicator # and % of participants who act as part of a community of writers after the program (produce writing, continue library group or join another, attend readings, read memoirs, read regularly about authors’ concerns)

Applied To :Program participants

Data SourcePhone interview with checklist of behaviors

 NEXT



创建自: JacmanChin403 points . 最后修改: 星期日 22 of 5月, 2011 00:47:21 MDT 作者 JacmanChin403 points .