Return to search form  

Session Title: Intermediate Consulting Skills: A Self-Help Fair
Think Tank Session 102 to be held in Liberty Ballroom Section B on Wednesday, November 7, 4:30 PM to 6:00 PM
Sponsored by the Independent Consulting TIG
Presenter(s):
Robert Hoke,  Independent Consultant,  robert@roberthoke.com
Discussant(s):
Mariam Azin,  Planning, Research and Evaluation Services Associates Inc,  mazin@presassociates.com
Victoria Essenmacher,  Social Program Evaluators and Consultants Inc,  vessenmacher@specassociates.org
Maura Harrington,  Lodestar Management/Research Inc,  mharrington@lmresearch.com
James Luther,  Luther Consulting LLC,  jluther@lutherconsulting.com
Emmalou Norland,  Institute for Learning Innovation,  norland@ilinet.org
Geri Peak,  Two Gems Consulting Services,  geri@twogemsconsulting.com
Kathryn Race,  Race and Associates Ltd,  race_associates@msn.com
Dawn Hanson Smart,  Clegg & Associates,  dsmart@cleggassociates.com
Abstract: This skill-building workshop will allow experienced independent evaluation consultants to interact with colleagues with less experience in order to demonstrate and share some of their hard-earned lessons. A series of eight Topic Tables will be set up, each with an experienced Table Leader who is prepared to share information about one consulting topic they enjoy and do well. Every 15 minutes, participants will circulate to a different Topic Table with a different Table Leader. Topics include: -Developing Other Lines of Business- -How Close is TOO Close? Client/Evaluator Relationships- -Managing Sub-Contractors- -Budget Development, Monitoring Cash Flow, and Tracking Expenditures.- -Putting Quality First, Putting Yourself Last- "Web-based Surveys: How Design Influences Response and Costs" -Time and Budget Management- "Tips for Long-Distance Evaluation." Table Leaders have more than three years' consulting experience.

Session Title: Evaluation and Learning in a Changing Landscape: How Changes to First 5 LA's Evaluation Framework are Integrated by Evaluation Contractors
Panel Session 103 to be held in Mencken Room on Wednesday, November 7, 4:30 PM to 6:00 PM
Sponsored by the Government Evaluation TIG
Chair(s):
Marc Davidson,  First 5 Los Angeles,  mdavidson@first5.org
Abstract: A panel will present on the new and improved First 5 LA Evaluation Framework and the manner in which the revised evaluation principles contained therein have tackled by two First 5 LA-funded evaluation contractors. A representative from First 5 LA will present on the new framework, and the two evaluation contractors will discuss their learning and the impact on evaluation as they have weathered the process of the ever-evolving First 5 LA evaluation framework over the years. First 5 LA was established through Proposition 10, passed by California voters passed in 1998, establishing a 50 cent-per-pack tax on tobacco products which generates approximately $700 million a year to be invested in the healthy development of children from prenatal to age 5. First 5 LA is the organization that was founded in Los Angeles County to disseminate these funds. As the largest of the 58 California Counties, Los Angeles receives the majority of Proposition 10 funds from the state. First 5 LA has devoted a significant proportion of their resources to promoting and supporting evaluation of funded programs. The panel will be comprised of a representative from First 5 LA and funded evaluation firms charged with conducting comprehensive evaluation of three First 5 LA initiatives. The issues to be explored are a) the process of developing a county-wide vision for evaluation; and b) realizing the core principles in a concrete manner for the purposes of real-life evaluation work.
The Evolution of First 5 LA's Evaluation Framework
Marc Davidson,  First 5 Los Angeles,  mdavidson@first5.org
First 5 LA made a strategic decision to hold itself accountable for using its resources to produce real, measurable results in improving the well being of all children prenatal through five and their families in Los Angeles County. The process of embracing accountability, with evaluation at the forefront, and the Evaluation Framework which was evolved for First 5 LA will be described. First 5 LA articulated a mission designed to conduct evaluation at four levels: 1) County, 2) Goal, 3) Grantee, and 4) Los Angeles County Children and Families Commission (the board charged with oversight). Dr. Davidson will describe First 5 LA's vision of a comprehensive, multi-level and introspective approach to accountability, and how First 5 LA works with its partners throughout Los Angeles County to track progress in achieving goals and objectives as well as the County's Five Desired Outcomes for Children.
First 5 LA (Los Angeles) and First 5 Kern County Evaluation Frameworks: Impact on Evaluation
Grant Power,  Semics LLC,  grantdpower@gmail.com
Community-Developed Initiatives (CDI) is one of two major funding approaches adopted by First 5 LA in its 2001-2004 Strategic Plan. CDI was designed as an open funding process supporting innovative community-driven strategies that address needs related to children 0-5 and their families. In September, 2003, Semics, LLC, (Semics) was contracted to conduct an 'initiative-wide' evaluation of CDI, to be completed over a three-year period. The objective of this evaluation is to determine how the experience of CDI grantees can contribute to a deeper understanding of the effects of project activities funded by First 5 LA. Using this approach, the CDI Evaluation is intended to elicit and facilitate various types of learning that can be derived from this open-ended solicitation approach to funding (contrasted with a Commission-driven initiative with a singular goal or focus). In addition, Dr. Power served as the evaluation lead for an external review of First 5 Kern County's county-wide evaluation, and will compare and contrast the evaluation needs related to First 5 LA versus First 5 Kern County. Differing evaluation frameworks between the two counties and the manner in which each framework translates into evaluation practice will be explored.
Evaluation of a County-Wide Hotline for Families With Young Children and Their Service Providers
Elizabeth Harris,  EMT Associates Inc,  eharris@emt.org
The First 5 LA Parent Help Line is a warm line designed to provide information, referral and support to families with children aged five and younger, their caregivers and their service providers in Los Angeles County. E.M.T. Associates, Inc. was awarded the contract to conduct the external evaluation. Coming from a utilization-focused approach to evaluation, EMT has provided formative feedback on an ongoing basis in an effort to assist the program in refining their efforts, make mid-course corrections, and generally improve programming. Feedback to the program grantee has been framed in the context of First 5 LA's goals. In addition to measuring accountability at the grantee and goal level, EMT has also focused on the impact county-wide, given the scope of the warm line. EMT has been a contractor since 2002, and has revised the scope of the evaluation at least three times to respond to changes in the evaluation framework at First 5 LA. EMT's learning from the changing evaluation framework and how learning has impacted the evaluation of the warm line will be explored in terms of the overall impact and each of the four required levels of evaluation (County, Goal, Grantee, and the Los Angeles County Children and Families Commission). The nature of evaluation in a changing landscape and the necessity of being able to learn and grow quickly will also be explored.

Session Title: Community Focus
Multipaper Session 104 to be held in Edgar Allen Poe Room  on Wednesday, November 7, 4:30 PM to 6:00 PM
Sponsored by the Cluster, Multi-site and Multi-level Evaluation TIG
Chair(s):
Martha Ann Carey,  Azusa Pacific University,  mcarey@apu.edu
Multi-site Action Research
Presenter(s):
Thomas Fuller-Rowell,  Cornell University,  tf42@cornell.edu
Abstract: Classical action research within single organizations has become a well established and differentiated approach since its inception more than six decades ago. Although new larger scale varieties of action research are beginning to develop there is still a clear need to expand the scope of action research practice (Greenwood, 2002). This paper presents multi-site action research (MAR) as a new variant of action research implementation with promising potential to fill the gap between the classical and coalition type varieties of action research. MAR is defined as involving the concurrent implementation of multiple distinct single-site action research processes, with a similar focus, and some level of coordination across sites. From a review of three relevant multi-site initiatives, a conceptual framework for the potential benefits of MAR is derived. The framework defines, conceptually, how MAR can improve action research implementation on the site-level and expand the overall effectiveness of an initiative. The paper also describes how MAR can build toward the implementation of more effective coalitions, and diagrams the sequence of the MAR implementation process.
Multi-level Evaluation Alignment: A Systematic Approach to the Evaluation of the Child Care Quality Improvement System
Presenter(s):
Xuejin Lu,  Children's Services Council of Palm Beach County,  kim.lu@cscpbc.org
Karen Brandi,  Children's Services Council of Palm Beach County,  karen.brandi@cscpbc.org
Lance Till,  Children's Services Council of Palm Beach County,  lance.till@cscpbc.org
Abstract: This presentation will demonstrate how to adopt a systematic evaluation approach (Multilevel Evaluation Alignment Model) to an evaluation for a child care Quality Improvement System (QIS) in Palm Beach County. We will describe the goals and key strategies of the QIS and highlight the need for this kind of alignment model in the QIS evaluation. We will also outline our four-step approach, using the Multilevel Evaluation Alignment Model and how it assisted in the development of a system-wide logic model and key evaluation questions. Discussion will include the benefits of applying this model to the QIS evaluation, as well as the learning thus far.
Finding the Threads Across Different Organizations: Different Community Service Activities: Identifying Common Measures
Presenter(s):
Kenneth Terao,  JBS International Inc,  kterao@jbsinternational.com
Anna Marie Schmidt,  JBS International Inc,  aschmidt@jbsinternational.com
Abstract: National and state funded volunteer programs have grown throughout the country during the last ten years. Community organizations, local governmental departments, faith-based organizations, and educational institutions use government and private funds to provide a broad spectrum of services from educational support to environmental conservation to housing and health assistance. Traditionally, each local project evaluates its success using an internal evaluation that focuses on the unique indicators and outcomes that demonstrate their accomplishments and results. State and federal funders want to understand the aggregate effect of their investments across programs. Initially, conducting a meta analysis of individual program impacts was the accepted practice, however, as funding becomes tighter and accountability for results increases, grant makers are looking for an approach which increases the ability to extract broad statewide evaluation conclusions. This presentation will describe how state departments can approach state-wide evaluation, measuring performance across a variety of community-based service activities.

Session Title: Awareness and Education: Did You Change Your Behavior This Week?
Think Tank Session 105 to be held in Carroll Room on Wednesday, November 7, 4:30 PM to 6:00 PM
Sponsored by the Health Evaluation TIG
Presenter(s):
Anna Kathryn Webb,  Catholic Relief Services,  awebb@crs.org
Discussant(s):
Jaime Dominguez,  Catholic Relief Services,  jdomingu@crs.org
Christy Lynch,  Partners in Evaluation and Planning,  colynch@verizon.net
Abstract: The purpose of the session is to discuss the effective measurement of behavior change in awareness and educational programs. In particular, Catholic Relief Services (CRS) seeks to determine the most effective and efficient means (quantitative and qualitative) to measure behavior change in its awareness and educational programs (e.g., Food Fast, a hunger awareness program for youth) in order to assess progress toward its goal of a world "transformed" by global solidarity, specifically, a world in which U.S. Catholics are in relationship with the poor, especially vulnerable groups in the 99 countries where CRS works. The key questions are what, how, when, and why to measure behavior change. The session will include: (a) a 15-minute overview of CRS behavior change indicators; (b) a 45-minute small group exercise on behavior change measurement, including a report out; and (c) a 30-minute plenary discussion on what has been learned.

Session Title: International Development Evaluation: Opportunities and Challenges for the Use of the Development Assistance Committee (DAC) Criteria
Think Tank Session 106 to be held in Pratt Room, Section A on Wednesday, November 7, 4:30 PM to 6:00 PM
Sponsored by the International and Cross-cultural Evaluation TIG
Presenter(s):
Paul Lamphear,  Western Michigan University,  paul.a.lamphear@pfizer.com
Tererai Trent,  Heifer International,  tererai.trent@heifer.org
Sheri Hudachek,  Western Michigan University,  sherihudachek@yahoo.com
Todd Harcek,  Western Michigan University,  todd.d.harcek@wmich.edu
Ryoh Sasaki,  Western Michigan University,  ryoh.sasaki@wmich.edu
Discussant(s):
Thomaz Chianca,  Western Michigan University,  thomaz.chianca@wmich.edu
Ronald Scott Visscher,  Western Michigan University,  ronald.s.visscher@wmich.edu
Krystin Martens,  Western Michigan University,  krystinm@etr.org
Michael Scriven,  Western Michigan University,  scriven@aol.com
Paul Clements,  Western Michigan University,  paul.clements@wmich.edu
Abstract: The Development Assistance Committee (DAC) Evaluation Criteria are probably the most commonly adopted criteria to evaluate development projects. Donor agencies and international non-governmental organizations have adopted these criteria in many parts of the world. The Interdisciplinary PhD in Evaluation at Western Michigan University has created a taskforce to review the DAC criteria and propose improvements to them. Some orienting questions include: Are the DAC criteria the right ones? Are common interpretations and uses of the criteria justified? What criteria, if any, are missing? Should each of the criteria be allocated the same weight and significance? This think-tank session aims to present the findings resulting from the taskforce's effort, get the reaction of two world-class experts in development evaluation, and engage other AEA members in a dialogue to identify ways to improve the quality of evaluation criteria in international development.

Session Title: Building Capacity to Strengthen the Evaluation of Safe Start Promising Approaches: An Evidence-based Approach
Panel Session 107 to be held in Pratt Room, Section B on Wednesday, November 7, 4:30 PM to 6:00 PM
Sponsored by the Alcohol, Drug Abuse, and Mental Health TIG
Chair(s):
Yvette Lamb,  Association for the Study and Development of Community,  ylamb@capablecommunity.com
Discussant(s):
Kristen Kracke,  United States Department of Justice,  kristen.kracke@usdoj.gov
Abstract: While 45 states are implementing evidence-based practices for mental health and substance abuse disorders, little is known about how to effectively implement and evaluate these practices in community settings. The proposed panel will discuss the unique role of the Safe Start Promising Approaches evaluation in building the capacity of sites implementing evidence-based practices to improve the implementation and evaluation of these practices. Safe Start Promising Approaches is the second of a multi-phase initiative funded by the Office of Juvenile Justice to identify the best program strategies to reduce the impact of children's exposure to violence. The processes and strategies used to launch the national evaluation, collect the needed data, and ensure successful implementation of interventions will be discussed. The learnings from this panel are likely to inform other cross-site evaluation efforts and thus improve our ability to monitor and evaluate new approaches to helping at-risk children and their families.
Launching a National Evaluation of the Safe Start Promising Approaches Initiative: Getting to Green Light
Lisa Jaycox,  RAND Corporation,  jaycox@rand.org
Kristen Kracke,  United States Department of Justice,  kristen.kracke@usdoj.gov
Yvette Lamb,  Association for the Study and Development of Community,  ylamb@capablecommunity.com
Dana Schultz,  RAND Corporation,  dana_schultz@rand.org
Laura Hickman,  RAND Corporation,  hickman@rand.org
Lauren Honess-Morreale,  RAND Corporation,  laurenhm@rand.org
Dionne Barnes,  RAND Corporation,  dbarnes@rand.org
Joie Acosta,  Association for the Study and Development of Community,  jacosta@capablecommunity.com
Safe Start Promising Approaches is the second of a multi-phase Safe Start Initiative funded by the Office of Juvenile Justice and Delinquency Prevention focusing on preventing and reducing the negative impacts of children's exposure to violence. The presentation will describe the year-long process of launching the Safe Start Promising Approaches phase of the Initiative including the selection of measures, development of data collection processes, and development and use of 'green light' criteria for evaluation and implementation. The 'Green Light' criteria included demonstration of specific capabilities and plans and were developed to insure that the sites had all the components in place for successful implementation of their program and for data collection. The learnings from this process are likely to inform other cross-site evaluation efforts and thus improve our ability to monitor and evaluate new approaches to helping at-risk children and their families.
Training in Data Collection for a National Evaluation
Lauren Honess-Morreale,  RAND Corporation,  laurenhm@rand.org
Lisa Jaycox,  RAND Corporation,  jaycox@rand.org
Suzanne Perry,  RAND Corporation,  suzanne_perry@rand.org
Dana Schultz,  RAND Corporation,  dana_schultz@rand.org
Laura Hickman,  RAND Corporation,  hickman@rand.org
Dionne Barnes,  RAND Corporation,  dbarnes@rand.org
In preparation for the Safe Start Promising Approaches national evaluation, we worked with 15 sites to develop mandatory and optional procedures that matched their program and evaluation designs, providing sample protocols for sites to modify for their own purposes. An intensive 2-day training for data collection supervisors and staff covered tracking of potential participants, enrolling participants, randomization (if applicable) and discussion of how to operationalize site-specific inclusion/exclusion criteria, consent forms and assent protocols, and field operations. In addition, data collectors were trained on the assessment instruments in English and Spanish. Follow-up to data collection training included telephone training of new staff, dissemination of frequently asked questions, review and feedback on assessments as they were sent to the national evaluator, and site-specific technical assistance on an as-needed basis. This presentation will demonstrate the flexibility that allows site-specific tailoring of methods, while still adhering to the requirements of the national evaluation.
Improving Quality Assurance When Implementing and Evaluating an Evidence-Based Intervention in a Community Setting
Joie Acosta,  Association for the Study and Development of Community,  jacosta@capablecommunity.com
Yvette Lamb,  Association for the Study and Development of Community,  ylamb@capablecommunity.com
Dana Schultz,  RAND Corporation,  dana_schultz@rand.org
A continuum of quality assurance was developed to provide a range of strategies to improve the implementation of 15 intervention programs based on the level of their resources and degree of rigor. Examples of how they improved their quality of implementation at each stage of the continuum are organized around four components of quality assurance: (1) Standardized intervention training and materials; (2) Ongoing supervision and feedback; (3) Quality assurance monitoring; and (4) Agency/organizational support. The presentation will also apply the continuum of quality assurance to a multiple-component (therapeutic intervention, case management, and hospital support) evidence-based intervention in a community setting. The lessons learned are likely to inform the implementation and evaluation of other evidence-based interventions in community settings and thus improve our ability to implement, monitor, and evaluate interventions to reduce the harmful effects of exposure to violence on young children.
Planning and Communication Strategies to Engage Children, Families, and Practitioners
Elena Cohen,  Safe Start Center,  ecohen@jbsinternational.com
Yvette Lamb,  Association for the Study and Development of Community,  ylamb@capablecommunity.com
Joie Acosta,  Association for the Study and Development of Community,  jacosta@capablecommunity.com
The purpose of Safe Start Promising Approaches is to implement and evaluate the effectiveness of multi-component evidence-based interventions in community settings. Although evidence-based practices are being implemented in at least 45 states, there is a lack of strategies for implementation. Safe Start Promising Approaches sites have the unique challenge of implementing evidence-based practices with a population of children and families exposed to violence, as well as engaging this population and their mental health practitioners in a rigorous evaluation that requires the use of a control or comparison group. To facilitate the implementation and evaluation of these evidence-based interventions a multi-faceted process was developed. This process, which was focused on helping grantees develop messages and strategies to engage children, families, and practitioners, included peer-support, technical assistance, and tools. This presentation will discuss the development and application of the tools, as well as strategies and lessons learned from this process.

In a 90 minute Roundtable session, the first rotation uses the first 45 minutes and the second rotation uses the last 45 minutes.
Roundtable Rotation I: Why do Evaluators use the Technology They do and Why are They not a Stronger Factor for Innovating new Technology for Use in Evaluation
Roundtable Presentation 108 to be held in Douglas Boardroom on Wednesday, November 7, 4:30 PM to 6:00 PM
Presenter(s):
Brian Chantry,  Brigham Young University,  brian_chantry@byu.edu
David Williams,  Brigham Young University,  david_williams@byu.edu
Abstract: In the field of evaluation, technology is implemented in many different ways and to varying degrees. From pencil and paper measures to the use of online data collection software, evaluation is benefiting from technology. While we evaluators are implementing what is being made available to us, it is not apparent that we are a driving force to innovate new technology that will advance the field. Are we content with what we have? Do we feel we cannot be innovators because we are not sure where to begin or what is being done? Or is there a real reason for evaluators to keep their distance from the cutting edge that might impact our abilities to effectively carry out evaluations? This roundtable session will provide an opportunity for participants to engage in discussion on how evaluators are using technology to enhance evaluations and where we might go in the future.
Roundtable Rotation II: The Power of Technology: Using Wikis, Blogs, and Online Tools for Evaluation
Roundtable Presentation 108 to be held in Douglas Boardroom on Wednesday, November 7, 4:30 PM to 6:00 PM
Presenter(s):
Cary Johnson,  Brigham Young University,  cary_johnson@byu.edu
Stephen Hulme,  Brigham Young University,  byusnowboarder@yahoo.com
David Williams,  Brigham Young University,  david_williams@byu.edu
Abstract: Online collaboration tools such as blogs and wikis are not only good for collaboration, but also evaluation. Educators typically use these tools with learners for project and paper collaboration, but they are sometimes overlooked as effective evaluation tools. This roundtable will feature a discussion of how these new technologies can be used as tools in evaluation. In addition, participants will discuss how the technologies provide opportunities for students to engage in critical thinking and develop their own evaluation skills.

Session Title: Using Evaluative Processes in Foundations: Challenges and Solutions
Think Tank Session 109 to be held in Hopkins Room on Wednesday, November 7, 4:30 PM to 6:00 PM
Sponsored by the Organizational Learning and Evaluation Capacity Building TIG and the Non-profit and Foundations Evaluation TIG
Presenter(s):
Astrid Hendricks,  The California Endowment,  ahendricks@calendow.org
Discussant(s):
Bill Bickel,  University of Pittsburgh,  bickel@pitt.edu
Catherine Nelson,  Independent Consultant,  catawsumb@yahoo.com
Jennifer Iriti,  University of Pittsburgh,  iriti+@pitt.edu
Abstract: Foundations are uniquely positioned to test new ideas and to distill knowledge from their efforts to inform their own work, and that of the field. Arguably, foundations have the potential to be "centers of learning" if they are systematic, reflective, and transparent in their grant making and strategic operations. Evaluation has an important role to play in supporting learning and knowledge production. Yet, the record is far from satisfactory (Kramer & Bickel, 2004). Numerous barriers exist to the use of evaluation in foundations (Bickel, Millett, Nelson, 2002). Research on barriers (Leviton & Bickel, 2004) will be used to frame discussion of this session's core questions: What barriers to effective use of evaluation to support learning are most prevalent in the experiences of the participants? How are these being overcome?

Session Title: Evaluation of Multi-Country Teacher Training Programs and Curriculum Policies
Multipaper Session 110 to be held in Peale Room on Wednesday, November 7, 4:30 PM to 6:00 PM
Sponsored by the International and Cross-cultural Evaluation TIG
Chair(s):
Peter Fisch,  European Commission,  peter.fisch@ec.europa.eu
Methodological Lessons and Planning Insights From a Formative Evaluation of a Pilot of Afghanistan's Teacher Education Program (TEP) Initiative
Presenter(s):
Mohammad Javad Ahmadi,  Creative Associates International Inc,  mohammadj@af.caii.com
Abstract: The Teacher Education Program (TEP) is Afghanistan's new national in-service teacher training initiative. It was implemented as a pilot in several provinces across Afghanistan. A formative evaluation was conducted to provide recommendations for expanding TEP nationwide. Several methods were used in this research in order to cross check the findings: survey research, in-depth interviews, and focus groups. In this paper, the challenges of using each research method will be discussed. Some of problems we encountered included language issues, lack of mutual trust (the result of years of civil war), and lack of analytic thinking ability by the participants. What we found from this research is that using questionnaires one can only rarely elicit people's true thoughts and opinions; using in-depth interviewing in a qualitative research context -despite some challenges- seems to be the most appropriate approach.
Designing Effective Multi-country Evaluations: Lessons Learned From a Large Scale Teacher Training Program
Presenter(s):
Roshni Menon,  Education Development Center Inc,  rmenon@edc.org
Daniel Light,  Education Development Center Inc,  dlight@edc.org
Abstract: This paper presentation focuses on designing effective international multi-site evaluations, a particularly relevant topic for today's increasingly global and interconnected world. The objectives of this presentation are to delineate the inherent challenges in developing an evaluation design that can be used in multiple countries and to offer the audience some proven strategies to think about when planning such an evaluation. Data from the evaluation of the Intel Teach Essentials Course, a professional development program for teachers, will be used to illustrate how an effective evaluation design can ensure maintenance of the program's core goals and quality in its localized delivery while generating relevant information for local stakeholders in each country. The Intel Teach Essentials Course is a public/private partnership between the Intel Foundation and the national ministries of education, and seeks to increase teachers' use of technology and change their pedagogical beliefs and practices to reflect more inquiry-driven, project-based approaches.
Tracking the Consistency: Evaluating the Curriculum Reform Policy in China
Presenter(s):
Wenhui Yuan,  Western Michigan University,  whyuan99@gmail.com
Abstract: China started the nationwide curriculum reform in basic education in 2001. Now the stage of pilot study has ended and the new curricula system is being fully implemented throughout the country in a top-down way. In this article, the local curriculum reform policies in four provinces with different socioeconomic status will be evaluated. The main method will be content analysis and the focus will be their consistency with the policy from the central government. The Program Evaluation Standards will be used as the frame of reference. Recommendations for implementing curriculum reform policy in China, as well as the use of The Program Evaluation Standards in Chinese cultural context, will be made with caution.

Session Title: Assessment in Higher Education TIG Business Meeting and Presentation: Evaluating Alaska Native-Serving/Native Hawaiian-Serving Institutions and Hispanic Serving Institutions of Higher Education
Business Meeting Session 111 to be held in Adams Room on Wednesday, November 7, 4:30 PM to 6:00 PM
Sponsored by the Assessment in Higher Education TIG
TIG Leader(s):
William Rickards,  Alverno College,  william.rickards@alverno.edu
Presenter(s):
Henry Doan,  United States Department of Agriculture,  hdoan@csrees.usda.gov
Saleia Afele-Faamuli,  United States Department of Agriculture,  sfaamuliwcsrees.usda.gov
Irma Lawrence,  United States Department of Agriculture,  ilawrencewcsrees.usda.gov
Discussant(s):
Deborah H Kwon,  The Ohio State University,  kwon.59@osu.edu
Abstract: The Cooperative State Research, Education, and Extension Service (CSREES) provides national leadership in identifying, developing, and managing programs that support university-based research, extension, and teaching activities in order to solve nation-wide agricultural issues. CSREES provides funding to promote and strengthen the ability of Alaska Native-Serving/Native Hawaiian-Serving Institutions to carry out education, applied research, and related community development programs. The Agency also provides funding to promote and strengthen the ability of Hispanic Serving Institutions to carry out educational programs that attract outstanding students and produce graduates capable of enhancing the Nation's food and agricultural scientific and professional work force. This paper presents a plan to monitor and evaluate the two programs. The plan is based on the logic models that delineate linkages of various program components that lead to activities, outputs and expected program outcomes.

In a 90 minute Roundtable session, the first rotation uses the first 45 minutes and the second rotation uses the last 45 minutes.
Roundtable Rotation I: Evolution of a First Year Seminar: Evaluation for Organizational Learning
Roundtable Presentation 112 to be held in Jefferson Room on Wednesday, November 7, 4:30 PM to 6:00 PM
Presenter(s):
Karen M Reid,  University of Nevada, Las Vegas,  reidk2@unlv.nevada.edu
Peggy Perkins,  University of Nevada, Las Vegas,  peggy.perkins@unlv.edu
Amy Morris,  University of Nevada, Las Vegas,  amy.morris@unlv.nevada.edu
Abstract: First-year seminars (FYS) have become a main stay of higher-education with over 90% of American colleges and universities offering them in some form. Current evidence indicates FYS participation can influence students' successful transition to college, academic performance, and a sizable array of college experiences known to relate to bachelor's degree completion. Purpose of this paper is to investigate issues associated with assessing the impact and outcomes of integrating current national FYS research at a large, urban university. How does adopting a research-based two-hour FYS contribute to students' perception of improvement in the skills and strategies necessary to succeed in college? How does conversion to a three-hour FYS contribute to students' perception of improvement in the skills and strategies necessary to succeed in college? The objective was to assess the effectiveness of these changes in order to deduce appropriate program theory changes from the evidence derived.
Roundtable Rotation II: Assessing Student Learning Outcomes: An Examination of a Process That Focuses Upon the Improvement of Teaching and Learning
Roundtable Presentation 112 to be held in Jefferson Room on Wednesday, November 7, 4:30 PM to 6:00 PM
Presenter(s):
Tanis Stewart,  University of Nevada, Las Vegas,  tanis.stewart@unlv.edu
Abstract: Processes implemented at a large university in the southwestern United States to assess student learning outcomes were evaluated in order to examine university assessment practice and determine how well academic units responded to established criteria. Data collected from academic units, including student learning outcomes, curriculum alignment, and assessment methodology were analyzed to evaluate process outcomes. Implications for theory-based process evaluation and learning outcome assessment models are discussed.

Session Title: New Evaluation Initiatives on Diabetes Prevention and Childhood Obesity: From the National to the School Level
Multipaper Session 113 to be held in Washington Room on Wednesday, November 7, 4:30 PM to 6:00 PM
Sponsored by the Health Evaluation TIG
Chair(s):
Eunice Rodriguez,  Stanford University,  er23@stanford.edu
The National Diabetes Education Program Evaluation Framework: A Case Study in Evaluation of a National Public Health Program
Presenter(s):
Joanne Gallivan,  National Institutes of Health,  joanne_gallivan@nih.gov
Clarice Brown,  Social & Scientific Systems Inc,  cbrown@s-3.com
Rachel Greenberg,  Marketing and Communications Consulting,  rachelgreenberg@starpower.net
Abstract: The National Diabetes Education Program (NDEP) was launched to reduce the mortality and morbidity from diabetes and its complications by improving the treatment and outcomes for people with diabetes, promoting early diagnosis, and preventing the onset of diabetes. The NDEP is co-sponsored by the National Institute of Diabetes and Digestive and Kidney Diseases of the National Institutes of Health (NIH) and the Centers for Disease Control and Prevention (CDC). Evaluation is an integral component of program planning and implementation of NDEP. NDEP's evaluation and is based on CDC's "Framework for Program Evaluation in Health". This framework was developed to help ensure that public health programs “remain accountable and committed to achieving measurable outcomes”. This presentation describes how NDEP has applied the CDC evaluation framework, providing a case study of how multi-faceted health communications programs can design program evaluations to help answer key questions on program processes and effects.
Use of Coordinated School Health Programs as a Strategy for Diabetes Prevention
Presenter(s):
Debra Pinkney,  University of Arizona,  dpinkney@email.arizona.edu
Lynne Borden,  University of Arizona,  bordenl@ag.arizona.edu
Stephen Russell,  University of Arizona,  strussell@arizona.edu
Ralph Renger,  University of Arizona,  renger@u.arizona.edu
Abstract: This paper outlines the evaluation of a primary prevention effort involving the implementation of the School Health Index (SHI), a Coordinated School Health Program (CSHP) designed by the Centers for Disease Control and Prevention (CDC) to assist schools in the assessment of health policies and practices. This school-based prevention program was launched in response to the increasing prevalence rates of diabetes in the United States and specifically within Hispanic communities in Tucson. Multiple entities including school personnel, parents, and community key informants were invited to assist in the completion of the program components focusing on health promotion policies and practices, nutrition, physical activity, family and community involvement. This evaluation indicated that the components were completed however the success of the project was influenced by numerous barriers including scheduling conflicts, parent availability and lack of involvement, competing priorities, and school perceptions of parents.
Program Fidelity: The Forgotten Variable in Program Evaluation
Presenter(s):
Janet Clinton,  University of Auckland,  j.clinton@auckland.ac.nz
Sarah Appleton,  University of Auckland,  sk.appleton@auckland.ac.nz
Abstract: This paper describes the measurement of program fidelity in the Let's Beat Diabetes (LBD) program. Program fidelity is described as the fit between program plans, and program implementation. Is often measured using a range of variables. LBD is a five year, life course community based program in New Zealand at long-term changes to prevent and/or delay the onset of Type II Diabetes. A number of constructs are measured. First, the degree of implementation, which asks how much of the program participants received. Second, program adherence and or adaptation, which is defined as the extent to which implementation of a particular activity and method is consistent with program plans. Third, organizational management, this includes areas such as organizational structures, collaboration, sustainability, and evaluation readiness. These variables are correlated to illustrate their inter-connectedness and their relationship with overall program progress. The paper demonstrates the importance of assessing program fidelity.
Evaluating an After-school Health Education Curriculum Using a Randomized Groups Experimental Design
Presenter(s):
Catherine A Oleksiw,  Academy for Educational Development,  coleksiw@aed.org
Paul L Johnson,  National Institutes of Health,  pjohnson@mail.nih.gov
Abstract: The National Institute for Child Health and Human Development's (NICHD) Media-Smart Youth (MSY) program represents a continuing effort by the NICHD to strengthen its prevention efforts in fighting obesity and increasing physical activity. An after-school program composed of ten lessons, MSY focuses on increasing knowledge, skills and behavioral intent in nutrition, physical activity and media literacy. Although program evaluation in field settings does not always lend itself to rigorous design, this evaluation employs a randomized groups experimental design, specifically an untreated control group design with treatment and control matched groups being measured at pre-test and post-test. The evaluation provides the first opportunity to assess the curriculum being executed by the after-school program providers with trained facilitators. Using a Time x Treatment repeated measures ANOVA, a significant difference was found in the treatment and control groups from the pretest to the posttest on the knowledge and skills assessed.
Real-time Evaluation of Burnley Food and Fitness Program Aimed at Tackling Childhood Obesity
Presenter(s):
Mansoor Kazi,  University at Buffalo,  mkazi@buffalo.edu
Jenny Slaughter,  East Lancashire Primary Care Trust,  jenny.slaughter@eastlancspct.nhs.uk
Nicole Tomasello,  University at Buffalo,  nicki.marie@gmail.com
Abstract: In 2004, the United Kingdom government set a target to halt the rise in childhood obesity by 2010. Although obesity prevention involves healthy eating and active living, achieving a shift in patterns of behavior is very complex. The Burnley project was a school based project aiming to change behavior through both curriculum and after-school activity. This paper reports on the findings from the project's first year, targeting 149 children from four primary schools. Using real-time evaluation strategies, a series of pretest posttest designs were used and it was found that the majority had improved on the Body Mass Index, levels of fitness, reported food intake and reported physical activity between January 05 and January 06. A partnership was developed between internal and external evaluation to enable the project workers to learn from the evaluation findings on demand in order to inform future practice in real-time.

Session Title: When Leadership Moves From I to We: Evaluating Collective Leadership Development Efforts
Multipaper Session 114 to be held in D'Alesandro Room on Wednesday, November 7, 4:30 PM to 6:00 PM
Sponsored by the Non-profit and Foundations Evaluation TIG
Chair(s):
Claire Reinelt,  Leadership Learning Community,  claire@leadershiplearning.org
Discussant(s):
Claire Reinelt,  Leadership Learning Community,  claire@leadershiplearning.org
Abstract: This session will focus on what the presenters have learned through evaluation about how collective leadership is fostered, and how collective leadership can build capacity to shift the culture and dynamics of teams, organizations, communities, and systems. This session will feature longitudinal case studies from three unique collective leadership initiatives including a large non-profit healthcare system, a municipal government, and a grant-funded regional economic development effort. Each presentation will focus on the evaluation approach, methods, lessons learned, challenges faced, and impact of each initiative. Then the case studies will be summarized to include what is common across these programs and findings, what is unique, any critical contextual factors, and what can be generalized to other collective leadership programs. It is our intention to finish with recommendations about what critical factors should be considered when designing and implementing an evaluation of collective leadership development.
Leadership in the City: How Individuals and Teams Impact a Community
Jessica Baltes,  Center for Creative Leadership,  baltesj@leaders.ccl.org
Twenty-three people representing a cross section of employees of various levels from multiple departments within a municipal government received individual leadership training by attending the Center for Creative Leadership's Developing the Strategic Leader open enrollment programs. In the spring of 2005, the municipal government and CCL began an initiative to engage the employees who had taken part in this individual training in a more in-depth group experience of strategic leadership concepts, frameworks, and skills, individualized to the organization's context and goals. The team's challenge was twofold, (1) For the individuals involved to continue their own personal growth as strategic leaders while supporting each other in that growth, and (2) To collectively leverage their learnings from the DSL program in a project of strategic significance for the City. This paper documents the efforts, outcomes, and impact of this team over a two year period.
Improving the Health of the System: A Case Study of Collective Leadership Within Catholic Healthcare Partners
Tracy Enright Patterson,  Center for Creative Leadership,  pattersont@leaders.ccl.org
Jennifer Martineau,  Center for Creative Leadership,  martineauj@leaders.ccl.org
Since 2000, the Center for Creative Leadership has collaborated with Catholic Healthcare Partners (CHP), a large US-based, mission-oriented health system to implement a long-term, multi-cohort leadership development initiative for high-potential CHP executives. Unlike traditional programs that focus on individual leadership development, this initiative aims to create an impact on the organization by increasing understanding and focus on the overall system and the mission, fostering work across organizational boundaries, and enhancing the organization's capacity to tackle strategic, complex, and critical issues. Action learning projects play a key role in the process. This paper describes the design and implementation of the evaluation of this initiative including strategies and challenges in measuring organizational impact over time. The evaluation examines changes in participants' work scope and responsibility (e.g., promotions, transfers), level of commitment, and specific competencies as well as system level connections and changes resulting from this effort.
Using Social Network Analysis to Evaluate Collective Leadership and Collaboration
Emily Hoole,  Center for Creative Leadership,  hoolee@leaders.ccl.org
Kimberly Fredericks,  Indiana State University,  kfredericks@indstate.edu
The Piedmont Region's WIRED initiative is a grant from the Federal Department of Labor to create collaborative efforts around regional economic building efforts. One of the critical components of the initiative is a regional leadership development effort. The key goals of the leadership initiative are to develop the system's capacity for open dialogue, development of diverse horizontal collaborative networks, a focus on collective learning, and shared values and culture. All of this is meant to foster a collective ability to set direction, create alignment and build commitment to compete as a region in the global economy. Social network analysis (SNA) is one of the evaluative tools utilized to assess changes in a system around collaborative networks over time. This presentation will focus on the use of SNA as an evaluative tool to assess such collaborative efforts and will address data collection, measurement issues and the longitudinal analysis of such data.

Session Title: Learning From Alternative Models of Evaluation
Multipaper Session 115 to be held in Calhoun Room on Wednesday, November 7, 4:30 PM to 6:00 PM
Sponsored by the Quantitative Methods: Theory and Design TIG
Chair(s):
Sanjeev Sridharan,  University of Edinburgh,  sanjeev.sridharan@ed.ac.uk
Applying Item Response Theory in the Evaluation of a Clinical Program
Presenter(s):
Mukaria Itang'ata,  Western Michigan University,  mukaria.itangata@wmich.edu
Abstract: This presentation will demonstrate the application of the item response theory (IRT) in a clinical program evaluation study to evaluate student abilities as they trained. The use of IRT was necessary in the evaluation of the program for the purposes of establishing entry-level baselines and to develop items to test students' proficiencies (abilities) over time.
A Mixed Method Approach to Evaluating Civic Learning Outcomes
Presenter(s):
Lisa O'Leary,  Tufts University,  lisa.o_leary@tufts.edu
Abstract: Recently, higher education has witnessed a renewed commitment to the mission of preparing students for lives of active citizenship (Boyer 1996; Checkoway 2001; Harkavy 2006). To this end, universities have started to infuse programs focused on active citizenship at their campuses for students, faculty, and staff. This paper highlights a multi-faceted evaluation that has been developed to gauge one university's success at cultivating “active citizens”, one of its core missions. The process associated with operationally defining civic engagement, developing and validating multiple measurement instruments through exploratory factor analysis, confirmatory factor analysis, and item-response theory, and implementing a mixed-method outcomes evaluation is highlighted. Specifically, the paper describes the methodology, instrumentation, and metrics of a multi-cohort, time-series evaluation that is investigating undergraduate's participation in and attitudes towards civic engagement which simultaneously evaluates a specific co-curricular program, while capturing formative data about students' development with regard to civic engagement.
Maximizing Evaluation Impact by Maximizing Methods: Social Network Analysis Combined With Traditional Methods for Measuring Collaboration
Presenter(s):
Carl Hanssen,  Hanssen Consulting LLC,  carlh@hanssenconsulting.com
Maryann Durland,  Durland Consulting,  mdurland@durlandconsulting.com
Abstract: This paper will review the Year 3 and preliminary findings from the Year 4 Evaluations of the Milwaukee Mathematics Partnership. The Milwaukee Mathematics Partnership is a comprehensive collaborative project designed to increase student achievement in mathematics. The logic from program design to outcomes- courses, relationships, math content, teaching strategies to student increased achievement in mathematics - is complex. However, one clear factor critical for project success is collaboration. The evaluation designs incorporated Social Network Analysis as an evaluation model for exploring collaboration. This review illustrates SNA measures and how they were linked to other metrics to provide a more comprehensive evaluation of the effectiveness of the implementation strategies. In addition, the review will describe the development of an MMP leadership map based on the SNA results and how the leadership criteria guided and supported existing practices and changes for project implementation strategies and indicators of success.
A Realist Synthesis Approach to Evaluating Complex Health Interventions
Presenter(s):
Sanjeev Sridharan,  University of Edinburgh,  sanjeev.sridharan@ed.ac.uk
Abstract: Using an ongoing example of an evaluation of a National Demonstration project, this class will discuss an iterative process by which findings from a single evaluation are set in the context of multiple evaluations. The process is informed by a realist synthesis framework but is focused on an evaluation of a single complex health intervention. Key features of the presentation include discussion on: (a) the role of evidence, innovations, and “muddling through” in learning from evaluations; (b) An iterative series of stakeholder dialogue to develop programme theory; (c) The use of data to test and explicate hypothesized links in a programme theory; (d) multiple types of learning that are possible from an evaluation including organizational learning, process learning, understanding risk landscapes of individuals, and individual-level impacts.

Session Title: Providing Meaningful Evaluations for Prevention Projects in Indigenous Communities
Panel Session 116 to be held in McKeldon Room on Wednesday, November 7, 4:30 PM to 6:00 PM
Sponsored by the Indigenous Peoples in Evaluation TIG
Chair(s):
Richard Nichols,  Colyer Nichols Inc Consulting,  colyrnickl@cybermesa.com
Discussant(s):
Richard Nichols,  Colyer Nichols Inc Consulting,  colyrnickl@cybermesa.com
Abstract: Effective evaluations provide new learning both to programs and to their evaluators. This panel, by 3 experienced evaluators in 3 very different indigenous communities, discusses their learnings and challenges in providing meaningful evaluation for violence prevention and substance abuse prevention programs, especially those funded by non-tribal agencies which may have predetermined evaluation requirements. Indigenous evaluators discuss their work with Maori in New Zealand and with American Indian Nations in the northwest and central United States.
Evaluation of Amokura: An Indigenous Family Violence Prevention Strategy
Fiona Cram,  Katoa Ltd,  fionac@katoa.net.nz
Amokura (the Amokura Family Violence Prevention Strategy) is an integrated community-based, 3-year initiative to address family violence in Northland - Tai Tokerau, New Zealand. The initiative is led by the Consortium (the Family Violence Prevention Consortium) which is made up of the Chief Executives of seven iwi (tribal) authorities. The initiative itself consists of four project areas that provide a whole of population approach to addressing family violence prevention and early intervention: research, education and promotion, professional development and training, and advocacy. At the end year 1 (2005) a formative evaluation was conducted and this led into the development of a research plan for years 2 & 3, and a summative evaluation plan for year 3. The evaluation and research work are both driven by theory and community aspirations, and have been developed collaboratively (involving the Consortium and their management team, the community, and the evaluator). This presentation will look at what has been learned from the intersections of theory and practice, as communities strive for the goal of zero tolerance of violence. A special focus will be on the evaluation of complex community initiatives.
An Evaluation's Identity Crisis: Implementing a Federally Funded Prevention Project Evaluation within a Sovereign Tribal Government Context
Nicole Bowman,  Bowman Performance Consulting LLC,  nbowman@nbowmanconsulting.com
Learn through two case studies how an evaluation can have an 'identity crisis' because it was funded by outside (non-Indian) government agency but was being implemented within a sovereign Indian nation. Developing the evaluation's identity for a project and using that process as a responsive and respectful teaching tool for evaluation stakeholders is the heart of collecting valid, reliable, and useful evaluation data. Participants will understand how a prevention project's evaluation identity crisis became a transformative process in which the Indian and non-Indian stakeholders created a unique community of practice and strengthened their own evaluation capacity along the way. During this presentation you will learn how this evaluation's identity crisis was the catalyst for a multi-ethnic stakeholder group to create new evaluation methods, responsive policies, and new practices. These innovations reflected a more self-determined model and Indigenized strategies for prevention programming and for the evaluation design itself. Theoretical sources, conceptual frameworks, and the empirical data, instruments, and evaluation processes will be shared. Using Indian and non-Indian methods, the presenter will compare the approach to the prevention evaluations in a side-by-side manner to highlight the differences in evaluation designs, share the collaborative transformation process, and allow the audience to leave with concrete ideas and empowering resources.
Indigenous Evaluation of the Implementation of the Center for Substance Abuse Prevention's Strategic Prevention Framework in a Community on the Yakama Nation
Jane Grover,  RMC Research Corporation,  jgrover@rmccorp.com
The Strategic Prevention Framework uses a community participatory process in examining data to determine local issues that lead to problem levels of underage drinking. In American Indian Communities this process necessarily involves the whole community and seeks especially the knowledge of elders and youth. Learn from the experience of a small isolated rural community on the Yakama Indian Reservation that is providing insight into the SPF process and its evaluation in Native American and Native Alaskan communities. Questions addressed include: what constitutes evidence of effective prevention programs in these settings? what does the indigenous community consider to be successful outcomes?

Session Title: Evaluation of Organizations as Enterprises: Approaches, Appropriate Outcome Expectations, and Potential Indicators
Panel Session 117 to be held in Preston Room on Wednesday, November 7, 4:30 PM to 6:00 PM
Sponsored by the Organizational Learning and Evaluation Capacity Building TIG
Chair(s):
Thomas Chapel,  Centers for Disease Control and Prevention,  tchapel@cdc.gov
Abstract: The ultimate goal of CDC, healthy people in healthy communities through prevention, is realized through joint efforts of CDC, component programs, grantees, and networks of state and local partners that turn national vision into local reality. Increasingly, CDC organizational components - centers, divisions, and programs - are asked to demonstrate how their efforts contribute to the whole. This presents two challenges. First, CDC efforts are often infrastructural, hence removed from frontline outcomes they intend to empower. And second, CDC might interact with autonomous participants to achieve distal outcomes, making its role more about rallying these participants in consistent, common efforts than about achievement of final outcomes. Three CDC programs will discuss how they approach evaluation of their efforts. In certain cases, this entails identifying and measuring more proximal outcomes; in other cases, the key has been development of a clear logic model that can provide consistency among approaches of autonomous programs and grantees.
These are the Voyages of Evaluating an Enterprise
Michael Schooley,  Centers for Disease Control and Prevention,  mschooley@cdc.gov
Rosanne Farris,  Centers for Disease Control and Prevention,  rfarris@cdc.gov
Jan Jernigan,  Centers for Disease Control and Prevention,  jjernigan1@cdc.gov
Barri Burrus,  RTI International,  barri@rti.org
James Hersey,  RTI International,  hersey@rti.org
Erika Fulmer,  RTI International,  fulmer@rti.org
Jeanette Renaud,  RTI International,  jrenaud@rti.org
Alton Dunlap,  RTI International,  adunlap@rti.org
Are we doing the right things? Are we doing them well? Are we having the intended impact? As 'we' grow in complexity and breadth, it becomes increasingly challenging to develop evaluation plans that answer these questions. Evaluation planning is accepted practice for specific programs and interventions, but not necessarily for organizational entities. At an organizational level, it is challenging to evaluate a portfolio of activities and projects, including various grant programs, data collection activities, research, partnership, and administrative activities. Recognizing the need to answer these questions, the CDC's Division for Heart Disease and Stroke Prevention, with assistance from RTI International, developed a division-wide evaluation plan. Plan development was informed by an array of stakeholders including CDC leadership and external partners. The presentation provides a description of the evaluation planning methodology and results, including a comprehensive logic model, evaluation priorities, major challenges and lessons learned.
Evaluating Strategy Execution: Lessons Learned in Implementing Human Capital Management Plans
Joan Cioffi,  Centers for Disease Control and Prevention,  jcioffi@cdc.gov
The most appreciable asset in an organization is its employees. One of CDC's human capital goals is fostering a learning culture that provides opportunities for continuous development and encourages employees to participate. Human capital strategies should be aligned with this goal, related to organizational objectives, and integrated into a strategic plan, performance elements, and budgets. This presentation focuses on the intended short- and midterm outcomes of selected human capital strategies. It will focus on individual learning accounts (ILAs), which provide annual funding for employees to use for individual career development, and on individual development plan (IDPs), which are required for employees to access ILA funds. The presentation will also offer the metrics employed, CDC's performance regarding those metrics, and actions taken to improve performance, highlighting the importance of evaluating plan execution as early as possible in any organizationwide strategy, particularly those related to complex human capital initiatives.
Planning for Partnerships and Strategic Alliances at the United States Centers for Disease Control and Prevention
Adam Skelton,  Centers for Disease Control and Prevention,  askelton@cdc.gov
Strategic planning in the public sector is critical to assuring public value. Many planning processes focus on building effective and efficient organizations. This discussion will focus on the implementation of a planning process around partnerships and strategic alliances that was used at the CDC. Critical capacities that must be in place to accomplish organizational objectives were identified. This seemed to fit the unique role that collaboration plays at the CDC as a 'means to an end' rather than the end itself. Planning, evaluation, and organizational performance were understood through; 1) Key planning activities and their relative value 2) A logic model to support the planning process 3) Identification and assurance of the key organizational capabilities leading to organizational success 4) Translating capabilities into operational programs with measurable attributes 5) Planning in large, complex organizations such as the CDC 6) Assuring the necessary development of an evaluation infrastructure and methodology

Session Title: Putting Context in Context With Examples in Strategic Planning and Measuring Fidelity
Multipaper Session 118 to be held in Schaefer Room on Wednesday, November 7, 4:30 PM to 6:00 PM
Sponsored by the AEA Conference Committee
Chair(s):
Cheri Levenson,  Cherna Consulting,  c.levenson@cox.net
Putting Context in Context
Presenter(s):
A Rae Clementz,  University of Illinois at Urbana-Champaign,  clementz@uiuc.edu
Abstract: In its broadest sense, context can include any detail that helps to make comprehensible the meanings of words, actions, events, and even objects we experience. In evaluation, these often include demographic or economic details, social and political dynamics, cultural norms, even structural and geographical features. Sensitivity to nuances in context often differentiates truly insightful evaluations from those that are not. Despite its importance in our work, context itself is a fluid, relativistic, and confounded construct. Context is influenced by what is being looked at, who is looking, and the purpose for looking. This paper takes up questions like: • What matters about context? • In what ways does context matter? • How much context is too much or not enough? • What are valid ways of representing context? in an attempt to put the notion of context, at least as it is used in evaluation, into context.
Measuring Fidelity of Implementation of a Coach-based Professional Development Model
Presenter(s):
Tara Pearsall,  University of South Carolina,  tmcpearsall@yahoo.com
Ching Ching Yap,  University of South Carolina,  ccyap@gwm.sc.edu
Ashlee Lewis,  University of South Carolina,  ashwee301@hotmail.com
Abstract: Assessing fidelity of implementation through systematic monitoring helps program developers and evaluators gain an understanding of how the quality of program implementation can affect the outcomes of professional development (PD) programs. Examination of program implementation quality allows program developers to determine how changes in implementation may affect a programs success or failure. To assess fidelity of implementation of a coach-based, 12-week PD program, evaluators systematically documented training sessions and developed measures to quantify the fidelity of those training sessions. The measures created enabled evaluators to gain detailed information regarding weekly fidelity through examination of the activity ratio as well as overall fidelity through examination of a fidelity index. With the information gained from the fidelity measures, program developers, in the field of education and beyond, can evaluate the success of their programs by determining whether outcomes reflect the PD model, the fidelity of implementation, or both.

Session Title: Evaluating the Teaching of Program Evaluation: Student and Teacher Assessments
Panel Session 119 to be held in Calvert Ballroom Salon B on Wednesday, November 7, 4:30 PM to 6:00 PM
Sponsored by the Teaching of Evaluation TIG
Chair(s):
Katherine McDonald,  Portland State University,  kmcdona@pdx.edu
Abstract: Program evaluation has grown significantly since the 1960's (Rossi, Lipsey & Freeman 2004). As such, educating new program evaluators is an important aspect of promoting healthy growth of the field. However, we have conducted relatively few evaluations of the effectiveness of opportunities to learn about program evaluation (Trevisan, 2004). In this presentation, we will share findings from an evaluation of graduate students' learning in a program evaluation seminar and generate a vision for critical next steps. We assessed learning through student and teacher judgments of knowledge and skill demonstration. In some instances, we will draw comparisons between students and their peers not enrolled in the seminar and other evaluation novices in professional training opportunities (Stufflebeam & Wingate, 2005). Audience ideas and insights will be solicited throughout the presentation and implications for the field will be discussed interactively.
Ideas for Evaluating the Teaching and Learning of Program Evaluation
Katherine McDonald,  Portland State University,  kmcdona@pdx.edu
We will begin the presentation by asking participants to brainstorm ideas about how we can evaluate the teaching and learning of program evaluation. For example, we will ask participants to identify what questions we can ask and ideas for how these questions might be addressed. These ideas will be tracked on easel paper with post-it notes and organized thematically. After ideas have been generated, we will also ask participants to rank order each type of evaluative question in terms of its importance. These individual rankings will be examined to produce a group ranking of most to least important questions to consider in evaluating the teaching of program evaluation. This list will be used again in the final presentation to generate next steps.
Student Assessment of Gains in Knowledge of Program Evaluation
Lauren Denneson,  Portland State University,  laured@pdx.edu
Students' learning of program evaluation can be studied by examining their assessments of their knowledge. We gathered data on students' self-assessment of program evaluation knowledge through three means. First, students completed the Self-Assessment of Program Evaluation Expertise (Stufflebeam, 2001) at the beginning and end of the seminar; a group of students not in the seminar also completed this assessment. Second, students completed weekly reflections on course readings. Third, students rated the quality of their performance on assignments and reflected on factors that contributed to their learning at the end of the term. Here, we will present findings on students' perceived knowledge of the field. As appropriate, we draw comparisons between self-assessed knowledge of students in the seminar, students not in the seminar and individuals in professional development program evaluation workshops (Stufflebeam & Wingate, 2005). We discuss the benefits and limitations of examining students' assessment of their learning.
Teacher Assessment of Gains in Knowledge of Program Evaluation
Tina Taylor-Ritzler,  University of Illinois, Chicago,  tritzler@uic.edu
Students' learning of program evaluation is also examined through teachers' judgments of students' knowledge and skill. Here, we examine evidence for students' knowledge and skill through four means: weekly reflections on course readings, pre-evaluation proposals, and written and oral evaluation proposals. Specifically, we developed coding schemas (in collaboration with established evaluators) to assess students' evaluation-specific knowledge and skills and applied the schema to each data source. We will present thematic findings of teacher judgment of knowledge and skill application. Lastly, we will discuss the benefits and limitations of examining teacher assessment of students' learning.
Relationships Between Student and Teacher Assessment of Gains in Knowledge of Program Evaluation
Margaret Braun,  Portland State University,  pdx01350@pdx.edu
Student and teacher assessments of students' learning complement one another to more fully illuminate what, and how, students learn about program evaluation in a graduate seminar (Trevisan, 2004). That is, each source of data provide unique information about what, and how, students make knowledge and skill gains. Here, we examine relationships between these two forms of learning assessment. Specifically, we will examine each source's unique contribution to addressing what, and how, graduate students learn about program evaluation in a seminar on the topic along places of convergence and divergence of findings.
Generating a Vision for the Evaluation of Teaching Program Evaluation
Shannon Myrick,  Portland State University,  shannonm@pdx.edu
Here, we will integrate two sources of information that inform the evaluation of teaching program evaluation: (1) Previous evaluation on the topic (e.g., Febey & Coyne, 2007; Trevisan, 2004) and (2) Findings from Presentations 2, 3, and 4. From this synthesis, we will examine the current state of knowledge about the evaluation of teaching program evaluation and solicit audience input on the interpretation of this knowledge base. Then, we will compare that knowledge base to the audience-rated priorities for questions to answer generated earlier in the presentation. Through a participatory process, we will identify and reflect on knowledge gaps and begin visioning next steps for answering important questions in the evaluation of teaching program evaluation.

Session Title: Facilitating Fast-paced Learning: Developmental Evaluation for Complex Emergent Innovations
Demonstration Session 120 to be held in Calvert Ballroom Salon C on Wednesday, November 7, 4:30 PM to 6:00 PM
Sponsored by the Presidential Strand
Presenter(s):
Michael Quinn Patton,  Utilization-Focused Evaluation,  mqpatton@prodigy.net
Abstract: Innovations being implemented in complex environments under emergent and uncertain conditions present special challenges for evaluation. To be effective in responding to rapidly changing conditions, innovators need to be able to learn quickly. That means evaluators have to be able to gather relevant data rapidly and provide real time feedback if the findings are to be useful. At the same time, the evaluator is inculcating evaluative thinking into the innovative process, which is its own challenge, because creative innovators are often more intuition-driven than data-driven. Using understandings from systems thinking and complexity science, this session will describe and give examples of an approach to evaluation " Developmental Evaluation (DE) " that makes rapid feedback for learning and adaptation the centerpiece of the evaluative process. Learning in DE includes both substantive learning (findings use) and learning to think evaluatively (process use.)

Session Title: Building Evaluation Capacity Within Organizations
Multipaper Session 121 to be held in Calvert Ballroom Salon E on Wednesday, November 7, 4:30 PM to 6:00 PM
Sponsored by the Extension Education Evaluation TIG
Chair(s):
Mary Arnold,  Oregon State University,  mary.arnold@oregonstate.edu
Reporting Extension Program Impacts: Slaying the Dragon of Resistance
Presenter(s):
Nancy Franz,  Virginia Cooperative Extension,  nfranz@vt.edu
Abstract: Virginia Cooperative Extension has responded to today's environment of enhanced accountability by improving the organization's program impact reporting. Successful strategies to enhance the quality and quantity of program impact reports include new hires, training faculty and administration, individual and small group technical assistance, development of reporting tools, and tying impact reporting to performance and recognition. This holistic approach resulted in enhanced reporting and use of program impacts as well as improved program design and evaluation. In this session, learn how Virginia Cooperative Extension put this approach into place, what was learned, and faculty perceptions of the approach.
Do Workshops Work for Building Evaluation Capacity Among Cooperative Extension Service Faculty?
Presenter(s):
Kathleen Kelsey,  Oklahoma State University,  kathleen.kelsey@okstate.edu
Abstract: The need for institutional accountability is essential for land-grant institutions. Program evaluation has long been in the land-grant university's tool box for ensuring accountability for programs delivered through the Cooperative Extension Service (CES). However, many CES faculty lack evaluation skills to conduct and report evaluation research. To fill the gap for building evaluation capacity, many CES faculty engage in self-study and workshops. It has been reported that self-study and workshops work best as supplemental learning and cannot substitute for in-depth study in program evaluation. Self-efficacy has been used as a variable to determine an individual's intention toward action. Thus, building evaluation capacity within the land-grant university depends on robust training, continuing education, and high self-efficacy toward applying lessons learned. Using survey design methods, this paper will report on a study that explored the impact of an evaluation workshop to build evaluation skills and self-efficacy among CES faculty.
A Framework for Evaluating 4-H National Initiatives
Presenter(s):
Benjamin Silliman,  North Carolina State University,  ben_silliman@ncsu.edu
Abstract: A framework is presented for evaluation of 4-H National Initiatives in Science, Engineering, and Technology (SET), Youth in Governance (YIG), and Healthy Living (HL). Four components are addressed: 1) description of the organizational context, or assets for program planning and reporting; 2) discussion of research on critical youth outcomes and program quality standards in each initiative area; 3) recommendations for targeted indicators of youth outcomes and program quality; and 4) implications of evaluating Initiatives for Extension organizational change, program planning patterns, partnerships, data collection, and marketing. Finally, the framework offers recommendations on use of management-oriented, objectives-oriented, and participant-oriented evaluation to guide National Initiative teams. Following the presentation there will be opportunities for discussion.

Session Title: When Does Evaluation Not Feel Like Evaluation? Embedding Evaluation Activities Into Programs
Panel Session 122 to be held in Fairmont Suite on Wednesday, November 7, 4:30 PM to 6:00 PM
Sponsored by the Pre-K - 12 Educational Evaluation TIG
Chair(s):
Leslie Goodyear,  Education Development Center Inc,  lgoodyear@edc.org
Discussant(s):
Sylvia James,  National Science Foundation,  sjames@nsf.gov
Abstract: Embedding evaluation within program activities is a way to encourage programs to engage in ongoing, continuous evaluation. These four presenters, evaluators of projects funded by the National Science Foundation's ITEST (Information Technology Experiences for Students and Teachers) program, will present the ways in which they have worked with projects to embed evaluation within project activities and the learnings, both programmatic and evaluative, that come from their experiences. The chair and discussant for the session will tie together these presentations with information about the ITEST program and the evaluation research work in which these presenters are involved.
Using Embedded Evaluation to Assist Teachers in Using Inquiry-based Modules That Integrate Math, Science and Information Technology (IT)
Roxann Humbert,  Fairmont State University,  roxann.humbert@fairmontstate.edu
The Comprehensive Information Technology Education in Rural Appalachia CITERA program is designed to help teachers and students learn about, experience, and diffuse Information Technology (IT) concepts within the context of existing Science, Technology, and Mathematics courses. A key objective of CITERA is to assist teachers in the design of inquiry-based modules that integrate Math and Science content, IT concepts and skills, and national and state standards. This presentation will share the embedded evaluation methods used by the CITERA program, such as peer, student, and expert reviews of the teaching units and a daily blog that is monitored by the evaluation team.
Embedding Evaluation Activities to Promote Learning
Ann Howe,  SUCCEED Apprentiship Program,  achowe@earthlink.net
The SUCCEED Apprenticeship program gives 8th -10th grade students, over a two-year period, the opportunity to have authentic and appropriate hands-on experiences in the use of technologies, techniques and tools of IT within the context of science, mathematics and engineering. Ongoing, continuous evaluation is part of the regular routine, and aimed at ensuring that each apprentice reaches a high level of competence in each topic in the curriculum so that he or she will be able to apply this knowledge and skill to complete the next required IT project. This presentation will focus on the specific techniques used by the evaluator to integrate evaluation into all aspects of the program in order to promote learning.
Using Engineering Notebooks as Embedded Evaluation
Neal Grandgenett,  University of Nebraska, Omaha,  ngrandgenett@mail.unomaha.edu
Neal Grandgenett is the evaluator of the SPIRIT educational robotics project in Nebraska. SPIRIT stands for the Silicon Prairie Initiative for Robotics in Information Technology. The SPIRIT project is striving to help middle school teachers learn how to teach science, technology, engineering and mathematics topics in the context of a flexible and 'scrounged-parts' robotics platform called the TekBot. As part of SPIRIT's ongoing evaluation process, an embedded evaluation instrument is now being used which is essentially the individual student's 'engineering notebook'. The engineering notebook format was developed by teachers in the project, and has been well embraced in the overall evaluation process. Selected examples of this embedded evaluation process will be given during the panel presentations.
Games as Embedded Assessments
Karen Peterman,  Goodman Research Group Inc,  peterman@grginc.com
Deborah Muscella,  Girls Get Connected Collaborative,  dbm@muscella.com
Since 2005, Goodman Research Group, Inc. (GRG) and the Girls Get Connected Collaborative (GGCC) have worked together to create innovative authentic assessments to evaluate students' technology skills and science knowledge related to a program called Technology at the Crossroads. As a summer and after-school program, GGCC wanted to create assessment tools that engaged students and were learning activities themselves. The assessment model used for this evaluation involves creating games and activities that require students to use the skills and knowledge they have gained from the program to compete in a series of field day competitions. These assessments have been used to capture both output and outcomes data. GRG and GGCC are now creating new assessments that can be linked with one another longitudinally. We will share examples of the games created for this evaluation as well as the lessons they have learned about using this methodology.

Session Title: Evaluating the Reading First Program: Best Practices and Lessons Learned
Panel Session 123 to be held in Federal Hill Suite on Wednesday, November 7, 4:30 PM to 6:00 PM
Sponsored by the Pre-K - 12 Educational Evaluation TIG
Chair(s):
Michael Long,  Macro International Inc,  michael.c.long@orcmacro.com
Discussant(s):
Michael Long,  Macro International Inc,  michael.c.long@orcmacro.com
Abstract: The Reading First program is a federal literacy grant that has had a tremendous impact on how reading is taught in elementary schools across the country. This panel will consist of four evaluators of state Reading First programs in Delaware, Indiana, Maryland, and Ohio. Each evaluator will discuss their work in their state, including the types of evaluation questions they are answering, the variables that they are measuring, and the instruments and methods they have found most effective. They will also address the methodological, practical, and political obstacles that they have encountered, and how they have addressed these challenges.
Evaluation of Policy and Practice in Delaware's Reading First Initiative
Linda Grusenmeyer,  University of Delaware,  lgrusen@udel.edu
No Child Left Behind has the Federal Reading First program as its model for early literacy instruction. Within this national context, Delaware's Reading First (DERF) program proposes to effect local change through teacher training in Scientifically Based Reading Research practices and effective use of data to inform instruction. As external providers of DERF's program evaluation, the Delaware Education Research and Development Center examines the implementation and impact of the five-year project at three levels: student level (attitudes and achievement), classroom/teacher level (teacher knowledge, beliefs, opinions and practices), and system level (alignment and efficacy of curriculum, professional development, state and district support, and teacher preparation programs.) With its rich and varied evaluation design, Reading First presents a range of challenges to Delaware's evaluation team. Among those issues that will be discussed will be maintaining balance between formative and summative measures and providing guidance in collecting and using data.
Evaluation of the Maryland Reading First Program: Successes and Challenges
Michael Long,  Macro International Inc,  michael.c.long@orcmacro.com
In this presentation Mr. Long will discuss Macro International's evaluation of the Maryland Reading First program. He will begin my describing the original design of the evaluation, as well as some of the factors and considerations that led to that design. He will then describe how this design has changed over time, and what has prompted these changes. Some of the issues that he will focus on are working with the state to develop clear research questions, instrument selection and development, and the challenges of managing a statewide evaluation in a state that has strong local control of schools. He will conclude by providing some strategies that have worked well, as well as recommendations for other evaluators based on his experience with this project.
Evaluation of the Indiana Reading First Program
Amy Kemp,  Indiana University,  amkemp@indiana.edu
Dr. Kemp will discuss the Center for Evaluation and Education Policy's evaluation of the Indiana Reading First (RF) program, which provides summative and formative feedback as well as capacity building and technical assistance. Guiding questions of this evaluation include: What is the role of the coach in the classroom? To what extent can the successful implementation of RF be attributed to Leadership Teams? To what extent are classroom instructional strategies guided by assessment results? Data are collected through 18 annual visits to schools; web-based surveys of teachers, principals, coaches, and interventionists; and phone interviews with district representatives. In addition, ISTEP+, Terra Nova, and DIBELS data are analyzed to determine growth over time; analyses compare the growth of at-risk students to overall student growth. In addition to describing her work, Dr. Kemp will provide advice and suggestions for evaluators involved in similar projects.
It Can't Be Either/Or: Using Internal Evaluation Methods and Procedures to Complement and Strengthen External Evaluation
James Salzman,  Cleveland State University,  j.salzman@csuohio.edu
In his presentation, Dr. Salzman will present a brief overview of the purposes, limitations and usefulness of internal evaluation from his perspective as Co-Director of the Reading First Ohio Center and supervisor of internal evaluation. He will describe why both internal and external evaluation cycles are necessary, valuable, and informative pieces of project implementation. He will then argue for a strong internal evaluation plan, including the use of traditionally summative tools and methods used for formative purposes, as central to making appropriate programmatic adjustments-especially in the early stages of implementation. Finally, he will discuss the need to build in ongoing conversations among both external evaluators and program deliverers for the purpose of clarifying the implementation and being able to use the results of the different types of evaluation to strengthen the communication to internal and external stakeholders.

Session Title: Closing the Loop: Mapping Value to Inform Research Management
Multipaper Session 124 to be held in Royale Board Room on Wednesday, November 7, 4:30 PM to 6:00 PM
Sponsored by the Research, Technology, and Development Evaluation TIG
Chair(s):
Neville Reeve,  European Commission,  neville.reeve@ec.europa.eu
A Framework for Evaluating Large Scale AIDS Clinical Research Networks
Presenter(s):
Jonathan Kagan,  National Institutes of Health,  jkagan@niaid.nih.gov
Mary Kane,  Concept Systems Inc,  mkane@conceptsystems.com
Kathleen M Quinlan,  Concept Systems Inc,  kquinlan@conceptsystems.com
William Trochim,  Cornell University,  wmt1@cornell.edu
Daniel Montoya,  Hill and Knowlton,  daniel.montoya@hillandknowlton.com
Melissa Burns,  Concept Systems Inc,  mburns@conceptsystems.com
Brendan Cole,  National Institutes of Health,  bcole@niaid.nih.gov
Abstract: The National Institute for Allergy and Infectious Diseases' Division of AIDS has completed a restructure of its multimillion dollar clinical trials networks. An evaluation system is being designed to support the goals of the restructure. To begin the system design, the primary research question focused on defining success of the newly coordinated clinical research networks. With the broadest participation of network communities, a structured conceptualization methodology was adapted to this context to create a collaboratively authored framework. The framework contains 100 detailed ideas organized into higher level themes. Major concepts include collaboration and communication, community input, setting research priorities, coordinated strategic planning of the research agenda, and progress on biomedical objectives. The map framework was transformed into a program logic model, from which specific measures are being developed, and evaluation data collection systems are being identified.
Analysis of Follow-up Evaluation Results of Research and Development (R&D) Projects Applying Logic Model to Elucidate the Process of Innovation
Presenter(s):
Kazuki Ogasahara,  Ministry of Economy, Trade and Industry,  ogasahara-kazuki@meti.go.jp
Osamu Nakamura,  National Institute of Advanced Industrial Science and Technology,  osamu.nakamura@aist.go.jp
Kazuyuki Inahashi,  Ministry of Economy, Trade and Industry,  inahashi-kazuyuki@meti.go.jp
Chikahiro Miyokawa,  Ministry of Economy, Trade and Industry,  miyokawa-chikahiro@meti.go.jp
Yoshitaka Kimura,  Ministry of Economy, Trade and Industry,  kimura-yoshitaka@meti.go.jp
Abstract: In the Ministry of Economy, Trade and Industry (METI), the follow-up evaluations of eleven big R&D projects have ever been implemented since 2000 to investigate and analyze the effectiveness of R&D projects to produce the socio-economic outcomes toward innovation. This time, we have applied the logic model for each project based on actual follow-up evaluation results of R&D projects to learn about the logical chains from project outputs to socio-economic outcomes through many steps of intermediate outcomes. With this study, useful lessons to implement R&D projects will be acquired to produce outcomes toward innovation. Our points of issue are to apply the logic model to analyze the results of follow-up evaluation of R&D projects to review the innovation processes, and thus to implement the strategic planning and evaluation of the promising R&D projects.
Contribution of Evaluation to Management of Research and Development (R&D) in the Process of Technology Transfer: A Knowledge Value Mapping Approach
Presenter(s):
Juan Rogers,  Georgia Institute of Technology,  juan.rogers@pubpolicy.gatech.edu
Abstract: This paper will review the current theoretical frameworks that are used in R&D evaluation and propose a knowledge value mapping approach to the realization of value from knowledge creation activities. Oftentimes, the desired impact of research is conceived of in terms of its application as a result of a process of technology transfer from its source in research to a new context of application either commercial or mission oriented. We will use several cases of scientific and technological research to propose more general categories of the realization of values from the creation and utilization of knowledge and apply them in a set of procedures that comprise a knowledge value mapping approach to R&D evaluation. The management benefits will be presented based on experiences of application in the context of rehabilitation research in the US and the design of an evaluation framework for the national policy of a South American country.

Session Title: Practicing Culturally-Based Evaluation: Learnings From the Field
Panel Session 125 to be held in Royale Conference Foyer on Wednesday, November 7, 4:30 PM to 6:00 PM
Sponsored by the Multiethnic Issues in Evaluation TIG
Chair(s):
Cheryl Blanchette,  Harder & Company Community Research,  cblanchette@harderco.com
Abstract: While evaluators agree on the importance of culturally relevant evaluation, there is still a great need to share knowledge and to identify promising practices within the field. In this session, four practicing evaluators will describe specific strategies and lessons learned in this regard based on their experiences conducting evaluations in diverse communities throughout California. Specifically, the panel will: (1) describe their engagement in an intentional and collaborative learning process around culturally-based evaluation; (2) present a checklist approach for systematically thinking through opportunities to increase the cultural relevance of an evaluative undertaking; (3) discuss challenges in using Likert scales with diverse populations and alternative measurement approaches; and (4) present recommendations for improving the quality of translated data collection instruments. The aim of this presentation is to share promising practices and invite discussion with other practitioners in the field.
Check It Out: The Development and Use of a Culturally Based Consulting Checklist
Kym Dorman,  Harder & Company Community Research,  kdorman@harderco.com
Evaluation checklists have been found to be an effective tool for thinking through various steps or factors to be considered in the conduct of an evaluation. The first panelist will present the application of this approach to evaluation consulting in diverse communities. Specifically, she will present a sequential checklist that identifies opportunities to consider and integrate culturally-based evaluation approaches and practices throughout all stages of an evaluation, including: 1) planning, 2) implementation, 3) analysis, and 4) dissemination of findings. While the use of a checklist approach does not guarantee the cultural relevance of an evaluation, it does provide an opportunity for evaluators to systematically think through elements of their practice and how they intersect with issues of culture and power. The bilingual panelist has worked in evaluation for over 7 years and has taken a lead role in the development of the checklist and other culturally-based consulting activities.
On a Scale of 1 to 5, How Well do you Understand the Question?
Maricela Piña,  Harder & Company Community Research,  mpina@harderco.com
The increasing importance of evaluation in determining the merit of various social programs coupled with the growing diversity of the United States results in a pressing need to assess the reliability and validity of various measurement techniques with culturally diverse populations. One topic seldom addressed in the literature is the use of instruments that primarily rely on Likert scales to solicit information from culturally diverse populations. This presentation will address methodological challenges associated with using Likert scales. It will include discussion of the following items: (1) cultural differences in response to Likert scales; (2) illiteracy and innumeracy issues; (3) social desirability; (4) construct validity; and (5) general difficulties encountered by respondents. Recommendations from direct field experience and the literature will also be presented. The bilingual/bicultural panelist has extensive experience conducting evaluations with diverse populations throughout California.
No Entiendo: Improving the Quality of Translation in Evaluation Practice
Nayeli Cerpas,  Harder & Company Community Research,  ncerpas@harderco.com
Through the use of culturally-based research methods, evaluators are increasingly responsive to the diverse cultures and communities in which they operate. This presentation will discuss language translation in evaluation practice as a critical component of culturally-based consulting, primarily when translating written materials and when collecting data. Items discussed will include approaches for ensuring that translated tools are culturally relevant and accessible, key decisions in planning for translation, use of different translation resources, and recommendations identified through the academic literature and in practice. The panelist has extensive experience in the development, implementation, and evaluation of culturally appropriate tools, and takes a special interest in issues concerning communities disproportionately affected by health and social inequities.

Session Title: Collaborative, Participatory and Empowerment Evaluation TIG Business Meeting
Business Meeting Session 126 to be held in Hanover Suite B on Wednesday, November 7, 4:30 PM to 6:00 PM
Sponsored by the Collaborative, Participatory & Empowerment Evaluation TIG
TIG Leader(s):
David Fetterman,  Stanford University,  profdavidf@yahoo.com
Liliana Rodriguez-Campos,  University of South Florida,  lrodriguez@coedu.usf.edu

Session Title: Assessing Randomized Control Trials and Alternatives
Multipaper Session 127 to be held in Baltimore Theater on Wednesday, November 7, 4:30 PM to 6:00 PM
Sponsored by the Quantitative Methods: Theory and Design TIG
Chair(s):
James Derzon,  Pacific Institute for Research and Evaluation,  jderzon@verizon.net
Using the Pre-test/Post-test Only Design for Evaluation of Training
Presenter(s):
Jack McKillip,  Southern Illinois University, Carbondale,  mckillip@siu.edu
Joan Rycraft,  University of Texas, Arlington,  rycraft@uta.edu
Steven Wernet,  Saint Louis University,  spwernet@netzero.net
Michael Patchner,  Indiana University Purdue University Indianapolis,  patchner@iupui.edu
Edmund Mech,  Indiana University Purdue University Indianapolis,  mechresearch.qwest.net
Abstract: The Pretest-Posttest only design was used for evaluation of 501 training sessions of 6579 health and pregnancy counseling professionals on Adoption Awareness by the Nation Center For Adoption using a standardized curriculum. Secondary analyses indicated that training effects were very large on measures of knowledge, confidence, and self-rated skills (ds between 1.35 and 2.68). Training effects were larger for 3-day than 1-day training. Other between-session variance in training effects was very small (<5% using HLM). Extended posttests indicated effects were lasting. Use of 2 pretests or no pretest showed no practice effects nor pretest sensitization, although potential self-presentation effects were seen on the day of training. Strengths and flexibility of this design are discussed.
Comparison of Variations in Retrospective Pre-test (RPT) and Pre-test/Post-test Surveys Measuring the Outcomes of an Anti-violence Education Program
Presenter(s):
James Riedel,  Girl Scout Research Institute,  jriedel@girlscouts.org
Abstract: This study's purposes are to test the reliability of the retrospective pretest (RPT) survey design both between and within subjects and to measure outcomes of a violence prevention program. Project Anti-Violence Education teaches girls skills and strategies which reduce their chances of becoming a perpetrator and/or victim of violence through bullying prevention/intervention, gang prevention, crime prevention, and internet safety curricula. Participants are randomly assigned to six measurement conditions –variations on the ordering of the question types in the RPT and pre-post surveys. (i.e., Before; After; Compared to now, before the program I . . .; and Compared to before the program, now I . . .) The instruments were designed to measure constructs including conflict resolution, personal safety, healthy relationships, and decision-making. Data analyses also examine the reliability of RPT compared to pre/posttest, while controlling for the confound of pre-testing. Additionally, the effect of item ordering and the validity of comparative post-testing are assessed.
Using Randomized Control Trials to Learn What Works in Prevention
Presenter(s):
James Derzon,  Pacific Institute for Research and Evaluation,  jderzon@verizon.net
Abstract: In an ideal world, or the idealized world of the analog study, the Randomized Control Trial (RCT) is an elegant and irrefutable design for drawing causal inferences. However, when applied in real world research, the approach has proven a slow, conservative, and limited path to learning what works to prevent many problem behaviors. RCTs are expensive and limit knowledge generation to a handful of scientists. They focuses attention to (a) units that can be randomized, (b) subject (instead of intervention) characteristics, (c) and a single outcome of (legitimate) concern – effectiveness. The value of RCTs is evaluated based on the internal consistency of the design and estimates of statistical significance at the expense of generalizability and potential population impact. These and other implications of faith in the RCT will be discussed using evidence from meta-analysis and systematic reviews and an alternative, real world approach to learning what works will be presented.
They May Glitter, but are They Gold? Randomized Control Trials in Evaluation
Presenter(s):
Sheila Arens,  Mid-continent Research for Education and Learning,  sarens@mcrel.org
Andrea Beesley,  Mid-continent Research for Education and Learning,  abeesley@mcrel.org
Abstract: Regardless of one's stance on how research ought to be conducted, evaluators working in organizations sometimes do not have the luxury of selecting what they might consider to be the most appropriate evaluation design. In this session, presenters will share experiences conducting randomized control trials (RCTs) and share challenges confronted when developing proposals for clients committed to utilizing this methodology. Methodological decisions affect every aspect of the evaluation. In this session, presenters will discuss ways to navigate methodological conversations with clients – including whether RCTs are reasonable given the program's age, fiduciary constraints, recruiting, and analysis and reporting. In addition, presenters will address what stands to be lost if evaluators fail to engage in such conversations.

Session Title: Crime, Violence and IRT/Rasch Measurement
Panel Session 128 to be held in International Room on Wednesday, November 7, 4:30 PM to 6:00 PM
Sponsored by the Quantitative Methods: Theory and Design TIG
Chair(s):
Susan Hutchinson,  University of Northern Colorado,  susan.hutchinson@unco.edu
Discussant(s):
Michael Dennis,  Chestnut Health Systems,  mdennis@chestnut.org
Abstract: This panel presents a variety of applications of the Rasch measurement model in the assessment of crime and violence (Global Appraisal of Individual Need, Dennis, 2003). Maps of crime and violence items and persons are presented that illustrate the symptom hierarchies on a linear, interval yardstick. Atypical cases are profiled, e.g., those who endorse very little crime and violence except some of the more severe instances such as prostitution, forgery and rape. Differential item functioning for adults vs. adolescents and men vs. women are presented. Each of these examples deals with methods to improve the reliability, validity and fairness of the measures we use to evaluate programs.
Interpreting Maps and Statistics of the Crime and Violence Scale
Karen Conrad,  University of Illinois, Chicago,  kmconrad@uic.edu
Barth Riley,  University of Illinois, Chicago,  bbriley@chestnut.org
Ya-Fen Chan,  Chestnut Health Systems,  ychan@chestnut.org
Kendon Conrad,  University of Illinois, Chicago,  kjconrad@uic.edu
Michael Dennis,  Chestnut Health Systems,  mdennis@chestnut.org
In the 1930's Thurstone used his new scaling method to estimate the seriousness of crimes. In this presentation, Rasch analysis will be used to estimate the seriousness of the 31 crime and violence items of the Crime and Violence Scale (CVS) of the Global Appraisal of Individual Need (Dennis et al., 2003). Including a brief introduction to Rasch analysis, the presentation will include discussions of person and item reliability of the CVS. Then it will focus on the person/item map that depicts distributions of persons and items on the Rasch ruler of crime and violence. The discussion will include a comparison of the Rasch results to Thurstone's findings as well as to other theoretical expectations regarding the seriousness of various crimes.
Crime and Violence: Differential Item Functioning by Age and Gender
Ya-Fen Chan,  Chestnut Health Systems,  ychan@chestnut.org
Karen Conrad,  University of Illinois, Chicago,  kmconrad@uic.edu
Kendon Conrad,  University of Illinois, Chicago,  kjconrad@uic.edu
Barth Riley,  University of Illinois, Chicago,  bbriley@chestnut.org
Michael Dennis,  Chestnut Health Systems,  mdennis@chestnut.org
Criminal and violent behaviors are prevalent among people presenting to substance abuse treatment. The specific types of crime and violence committed, however, might vary by gender and age. Using a Rasch differential item functioning test (DIF) this study aims to examine whether the severity of criminal and violent behaviors differed by gender and age among adolescents and adults who were treated for substance abuse. The DIF test showed that males and females differed significantly in their endorsement of items related to prostitution, illegal gambling, physical violence (slapped someone), and armed theft. Furthermore, adults were found to be more likely to endorse prostitution, rape, forgery, arson, property crime and DUI than youths. The presentation will demonstrate how measurement models can improve our understanding of crime and violence among substance users in treatment as well as how it can impact program planning.
Detecting Misfitting Persons with the Rasch Model
Barth Riley,  University of Illinois, Chicago,  bbriley@chestnut.org
Ya-Fen Chan,  Chestnut Health Systems,  ychan@chestnut.org
Karen Conrad,  University of Illinois, Chicago,  kmconrad@uic.edu
Kendon Conrad,  University of Illinois, Chicago,  kjconrad@uic.edu
Michael Dennis,  Chestnut Health Systems,  mdennis@chestnut.org
The GAIN Crime and Violence Scale (CVS) consists of four subscales: General Conflict Tactics, Property Crimes, Interpersonal Crimes, and Drug Crimes. Rasch principal components analysis of person residuals and person fit statistics were examined in order to determine the extent to which atypical patterns of responding to the CVS was present in a sample of 7,435 substance abuse treatment clients. Principal components analysis identified five components, with the first component explained 10.9% of residual variance. Among persons who loaded negatively on the first component (n=2,334, 31.4% of total sample), half (50.3%) had response patterns characterized by high outfit (MNSQ Outfit > 1.33). The average measure on the CVS for this subgroup (n=1,175) = -1.66 logits. Persons with misfitting response patterns tended to endorse previous involvement in prostitution (item difficulty=2.34 logits) or indicated that they had committed forgery (item difficulty=1.91 logits) without endorsement of less severe forms of crime and violence.

Session Title: Evaluation Methodology in Educational Technology Contexts
Multipaper Session 129 to be held in Chesapeake Room on Wednesday, November 7, 4:30 PM to 6:00 PM
Sponsored by the Distance Ed. & Other Educational Technologies TIG
Chair(s):
Michael Coe,  Northwest Regional Educational Laboratory,  coem@nwrel.org
Investigating the Validity and Reliability of the School Technology Needs Assessment (STNA)
Presenter(s):
Jeni Corn,  University of North Carolina, Greensboro,  jocorn@serve.org
Abstract: The Technology in Learning (TiL) program at the SERVE Center at UNCG developed School Technology Needs Assessment (STNA) for assessing school-level needs related to the use of technology for teaching and learning. This research study investigated the validity and reliability of the School Technology Needs Assessment (STNA), using existing data from over 2000 respondents from schools across the United States. This study includes a thorough literature review, estimates of internal consistency reliability, and results from a principal factor analysis. Data analyses showed each of STNA constructs and subconstructs to have high internal consistency reliability (alpha ranged from .807 to .967). The exploratory factor analysis of STNA response data helped to verify the facets included in STNA and assess whether different items on the survey belong together in one scale. The results of this study will be used to make specific recommendations to inform future revisions to the instrument.
Building the Theoretical Contribution of the Worldly Science: The Case for Longitudinal Engagement in the Evaluation of Programmes
Presenter(s):
Charles Potter,  University of the Witwatersrand,  pottercs@gmail.com
Gordon Naidoo,  Open Learning Systems Education Trust,  van@mail.ngo.za
Abstract: This paper suggests the value of longitudinal engagement in the evaluation of programmes. It provides a case study of the evaluation of the Open Learning Systems Education Trust's “English in Action” programme in South Africa from 1993 to the present. It describes the initial role of formative evaluation in highlighting the need for change in the programme's implementation theory from a model focused on enhancing learner involvement and learner gains, to a model of distance education and open learning focused on promoting teacher and learner gains through school, classroom and teacher support. It then documents the role of ongoing participatory multimethod evaluation in the programme's growth to scale over a twelve year period, both in assisting the development and sustainability of the programme, as well as in providing evidence supporting the renewed interest in radio learning which has taken place over recent years in developing countries, as well as more broadly internationally.
Improved Evaluation Designs for Educational Technology Projects
Presenter(s):
Michael Coe,  Northwest Regional Educational Laboratory,  coem@nwrel.org
Abstract: This presentation is a work-in-progress report on an NSF-funded project intended to develop improved models of evaluation and research for educational technologies. The premise is that the evaluation of educational technology applications in education is hampered by oversimplified, underspecified models of both the project designs and the project evaluations. Important aspects of the project designs are often left out of the evaluation designs, and often the relationships between project components are misrepresented in the evaluation designs. These issues may lead to unproductive evaluation questions and research methods. Many of these problems could be solved, at least in part, by applying program theory and basic causal modeling concepts. The paper will include the rationale for the project, brief examples of work we have done over the past few years, and preliminary findings from the current study.

Session Title: International Perspectives on Evaluation Part 2: Institutionalizing Evaluation in Government
Think Tank Session 130 to be held in Versailles Room on Wednesday, November 7, 4:30 PM to 6:00 PM
Sponsored by the AEA Conference Committee
Presenter(s):
Ross Conner,  University of California, Irvine,  rfconner@uci.edu
Abstract: Evaluators from outside the US face some similar and many different issues. Two of the currently-salient 'different issues' will be the focus of two consecutive parts of an IOCE-sponsored think tank session. The second session will focus on new initiatives to institutionalize monitoring and evaluation (or M&E as it is known in many developing countries) in government, in a more integral way and at a higher level than typically occurs in the US. The discussion will focus on recent developments in South Africa and in West Africa in Niger, where following the African Evaluation Association (AfrEA) 4th Conference there, Nigerien President Mamadou Tandja established a new Ministry of Monitoring and Evaluation (in the US context, this would be similar to establishing a new federal level department). Representatives from AfrEA will describe and discuss the background and implications for institutionalizing evaluation throughout Africa.

Session Title: International Perspectives on Evaluation Part 1: Evaluating the Impact of Development Projects
Think Tank Session 130 to be held in Versailles Room on Wednesday, November 7, 4:30 PM to 6:00 PM
Sponsored by the AEA Conference Committee
Presenter(s):
Ross Conner,  University of California, Irvine,  rfconner@uci.edu
Abstract: Evaluators from outside the US face some similar and many different issues in their work. Two of the currently-salient 'different issues' will be the focus of two consecutive parts of an IOCE-sponsored think tank session. The first session will build on a discussion now underway in developing countries about the best ways to assess the impacts and outcomes of these types of projects. This moderated Q&A session will feature three discussants from different stakeholder groups with a perspective on this issue: indigenous evaluators working in developing countries, donors and others from outside the developing world who support development projects, and evaluators from outside these countries doing international development work and who bridge the other two stakeholder groups. There will be no formal presentations; instead, the three discussants will answer questions about the issue posed by the moderator for about 30 minutes. The final 15 minutes will involve questions from the floor.

Search Results