Evaluation 2008 Banner

Return to search form  

Session Title: Evaluation Policy in the Changing United States Government Foreign Assistance Community
Panel Session 504 to be held in Capitol Ballroom Section 1 on Friday, Nov 7, 9:15 AM to 10:45 AM
Sponsored by the International and Cross-cultural Evaluation TIG
Chair(s):
Cynthia Clapp-Wincek,  Independent Consultant,  ccwincek@aol.com
Discussant(s):
Richard Blue,  Independent Consultant,  richardblue@earthlink.net
Abstract: In the 1970s and into the 1980s, USAID's evaluation system and practices were both innovative and well focused on impact and learning from experience. USAID was considered a leader in evaluation practices. In recent years, new actors have joined USAID in foreign assistance management. During this period, USAID has struggled with support to their evaluation system. This panel includes representatives from USAID, the Department of State's strategic planning office, the office of the Director of Foreign Assistance in the State Department and Millennium Challenge Corporation. Each will talk about evaluation policies and standards, how these are being developed, and how evaluation is playing out in practice in their agencies. After short presentations by each representative, the audience will be asked to contribute their experience and compare and contrast the different agency's approaches.
USAID Evaluation Policy and Practice
Gerry Britan,  United States Agency for International Development,  gerry@usaid.gov
The USAID experience in developing a stable and useful evaluation policy and implementation structure has had considerable ups and downs since the mid-1970s when the Office of Evaluation was created to strengthen the Agency's evaluation system and to carry out evaluations. By 1981, under President Ronald Reagan, USAID had formed the Center for Development Information and Evaluation, consolidating into one center the archival and evaluation leadership function within the Bureau for Policy and Program Coordination. The creation of the Director for Foreign Assistance shifted the evaluation staff and the focus of the work to the development of the FACTS monitoring and reporting system leaving no personnel or resources for evaluation for over a year. Evaluation seems to be back under discussion at both DFA and USAID. By the time of the Annual Meeting in November, USAID will have made progress in reinvigorating evaluation.
Bureau of the Director of Foreign Assistance Evaluation Policy and Practice
Donna Stauffer,  United States Department of State,  stauffer@state.gov
The bureau of the Director for Foreign Assistance was created in 2006 with a strong initial emphasis on input output monitoring within the context of the New Strategic Framework for foreign assistance. In 2007, DFA conducted an assessment of its own practices and made changes. DFA is currently working to define an evaluation policy.
United States Department of State Evaluation Policy and Practice
Melinda Crowley,  United States Department of State,  crowleym@state.gov
The U.S. Department of State implements foreign assistance programs as well but has not had a department-wide evaluation policy to support its strategic planning or results monitoring. The Bureau for Educational and Cultural affairs has had a strong evaluation unit and some other pockets of evaluation have developed as well. In the fall of 2008, the strategic planning office in the Resources Management Bureau (RM/SPP) began meeting with other bureaus to discuss evaluation issues and is working towards an evaluation policy.
Millennium Challenge Corporation Evaluation Policy and Practice
Harry Carr,  Millennium Challenge Corporation,  carrhc@mcc.gov
In 2004, Congress established the Millennium Challenge Corporation. MCC was 'purposely designed as a small corporation to promote, support and ensure accountability for the innovative foreign aid strategies it administers.' Due to that emphasis on accountability, MCC has used a set of performance indicators to determine countries' eligibility to receive programs. In 2006, MCC issued policy and guidance in monitoring and evaluation and the presentation will discuss what's working and what isn't.

Session Title: Communicating With Clients: An interactive demonstration of Do's and Don'ts for Evaluators
Demonstration Session 505 to be held in Capitol Ballroom Section 2 on Friday, Nov 7, 9:15 AM to 10:45 AM
Sponsored by the Independent Consulting TIG
Presenter(s):
LInda Thurston,  Kansas State University,  lpt@ksu.edu
Sally Bailey,  Kansas State University,  sdbailey@ksu.edu
Discussant(s):
Jan Middendorf,  Kansas State University,  jmiddend@ksu.edu
Christa A Smith,  Kansas State University,  christas@ksu.edu
Cindi Dunn,  Kansas State University,  ckdunn@ksu.edu
Sheryl A Hodge,  Kansas State University,  shodge@ksu.edu
Abstract: Successful evaluation necessitates good rapport between all stakeholders in the process. Communication between clients and evaluators is crucial for building collaborative and trusting relationships. This presentation will use interactive dramatization to examine critical issues in evaluator-client interactions. The audience will be part of the presentation; they will be encouraged to stop the action and point out roadblocks to successful communications in the scenarios that will be dramatized. Among the communication issues to be demonstrated are: the classic non-listener; the over-the-top expert; the anti-collaborator; cross-cultural partners; the parsimonious client; distrusting stakeholders. The presenters are professional evaluators and a professor of drama therapy and ethno theater.

Session Title: Policy and Practice Implications of a Multi-faceted Approach to Building Evaluation Capacity in Community Colleges
Multipaper Session 506 to be held in Capitol Ballroom Section 3 on Friday, Nov 7, 9:15 AM to 10:45 AM
Sponsored by the Assessment in Higher Education TIG
Chair(s):
William Rickards,  Alverno College,  william.rickards@alverno.edu
Discussant(s):
Mary Williams,  Lumina Foundation,  mwilliams@luminafoundation.org
Peggy Valdez-Fergason,  Independent Consultant,  pvaldezfergason@comcast.net
Abstract: This session brings together two independently developed approaches that are now coming together to create a complementary assessment strategy for community colleges. The strategy leverages both the organized and self-organizing system dynamics within community colleges. One approach focuses on the use of college-wide multiple data sources to identify college-wide issues about student learning and success that need to be addressed. The second approach uses a strategy of building Communities of Learning, Inquiry, and Practice (CLIPs). CLIPs are a structure for self-selected groups of faculty and staff to collaboratively investigate questions about their work. They undertake a three step evaluative inquiry process. The two processes work together to build a college-wide culture of inquiry. These papers address issues of policy and practice based on over three years of experience and research about the assessment and inquiry processes discussed in the papers.
Using Institutional Databases to Improve Student Achievement: The Role of Data Facilitators
Richard Voorhees,  Voorhees Group LLC,  rick@voorheesgroup.org
Over 80 community colleges nation-wide are involved in developing strategies to ensure student success through the 'Achieving the Dream' initiative of the Lumina Foundation. A vital part of the initiative is providing a 'data facilitator' for each college. The data facilitator plays a key role in helping a team (composed of many different stakeholders) analyze institutional level data to serve as the basis for policy and practice decisions about instructional and other changes to support student success. This paper describes lessons learned by a data facilitator as he helped colleges build their capacity to use data to make institutional level decisions. The paper addresses the broad policy issues both for colleges and for funders and leaders of large scale national initiatives. The paper looks at the complementary nature of the use of institutional level data with the use of self-organizing groups such as Communities of Learning, Inquiry, and Practice (CLIPs).
Faculty Engagement Through Communities of Learning, Inquiry, and Practice (CLIPs)
Beverly Parsons,  InSites,  beverlyaparsons@aol.com
Communities of Learning, Inquiry and Practice (CLIPs) are self-selected groups of faculty and staff who collaboratively investigate questions about their work. They undertake a three step evaluative inquiry process and operate under a set of Guiding Principles that integrate research on complex adaptive systems, inquiry processes, learning theory, communities of practice, appreciative inquiry, and change processes. The testing of the CLIP process at a California community college under a National Science Foundation grant showed that the process 1) builds collaborative relationships, 2) enhances evaluative inquiry skills, and 3) leads to changes in professional practice. This paper provides the theoretical basis for the CLIPs and the key findings from a three year action research study. It looks at how the CLIP process complements a more structured analysis of institutional level data to build a sustainable culture of inquiry and evidence based decision making.

Session Title: Shaping Policy and Practice: Evaluative Contributions to the delivery of Education Technology and Professional Development in Jordan
Multipaper Session 507 to be held in Capitol Ballroom Section 4 on Friday, Nov 7, 9:15 AM to 10:45 AM
Sponsored by the Distance Ed. & Other Educational Technologies TIG
Chair(s):
Audrey-Marie Moore,  Academy for Educational Development,  amoore@aed.org
Discussant(s):
John Gillies,  Academy for Educational Development,  jgillies@aed.org
Abstract: The Education Reform for the Knowledge Economy in Jordan has prompted the development and implementation of new approaches to professional development and the teaching of information management in classrooms. The USAID-funded Jordan ERfKE Support Program (ESP) has been instrumental in assisting the Ministry of Education to shape policy into practice through the use of a multi-step approach to evaluation. To understand the impact of ESP activities on teacher and student learning, and institutionalization of the reform process within the MoE itself, the Monitoring and Evaluation Team in Jordan developed a developmental approach, which included experimental design, case study and measurement of systems change, to show how MoE policies impacted teacher and student learning in Jordan while leading to a more systemic understanding policy implementation in the field. The approaches also contributed to MOE conceptualization of future activities that shaped on-going improvements of the interventions.
Contributions to Policy Decision-making: The Effectiveness of Blended, On-line learning strategies in Jordanian Secondary Education
John Rich,  Evaluvision,  drjohnrich@evaluvision.com
Audrey-Marie Moore,  Academy for Educational Development,  amoore@aed.org
In Jordan, the current approach to teaching the Management Information (MIS) stream curriculum is teacher-centered. The USAID-funded, ERfKE Support Program developed a series of online modules that introduced students to the same MIS concepts and facts through a student-centered, project-based learning approach. In this approach, students were guided through the process of self-selecting a profit-making business enterprise, and creating their own business plan, based on commonly used principles in the real-life business world. In the process of working through this business plan, required students to master concepts without reliance on textbooks and lecture alone. This paper will discuss the effectiveness of the approach when compared to the traditional approach; identify the characteristics of students, teachers, and/or schools that amplified, moderated, or suppressed the effect; and show how the approach and results of the evaluation contributed to policy decisions made by Ministry officials to institutionalize the approach in Jordan.
Evaluative Contributions to Teacher Professional Development: An Analysis of Demand-Driven Training in Jordan
David Balwanz,  Academy for Educational Development,  dbalwanz@aed.org
Razan Al-Hadidi,  Academy for Educational Development,  ralhadidi@esp-jordan.org
Ahmed Thawabieh,  Ministry of Education,  ahmadthawabieh@yahoo.com
This paper presentation will discuss evaluation findings that share teacher and student perceptions on the efficacy and impact of demand-driven training; the extent to which new content-knowledge and teaching strategies (co-planning, project-based learning, etc.) are used in the classroom; teacher and principal perceptions of learning team implementation in supporting teacher work and professional development, and all of the above in relation to the management and resource environments in which teachers work. The paper will also show how engagement of key Ministry personnel and the use of formative evaluation led to the evolution of the Ministry's professional development approach from a reform oriented intervention (to upgrade teacher skills for the introduction of a new curriculum) to one that could institutionalize a decentralized approach supporting the on-going professional development of teachers in Jordan.
Education Reform Support: Creating and Supporting Conditions for Success
Clive Cavanagh,  Knowledge Economy Support Program,  ccavanagh@esp-jordan.org
Audrey-Marie Moore,  Academy for Educational Development,  amoore@aed.org
This paper presentation will discuss institutionalization of reform components that include the strategies, approaches and capacities supporting the implementation of the management Information Stream (MIS) curriculum and the new professional development model for MIS teachers. To support discussion of sustainability and begin to measure the institutionalization of reform components, the ESP team used a set of rubrics to evaluate the development of institutions capable of providing technically and systematically sound support such as information systems, incentives, legal frameworks, and qualified staff members who have the ability to coordinate national priorities within the context of local and international cooperation. The framework allowed the project along with their Ministry counterparts to monitor changes in the conditions that allow for the integration of interventions ensure sustainable changes.

Session Title: The AEA Town Hall Meeting: Internal Scan Findings, Policy Governance, Actions, and Next Steps
Think Tank Session 508 to be held in Capitol Ballroom Section 5 on Friday, Nov 7, 9:15 AM to 10:45 AM
Sponsored by the AEA Conference Committee
Presenter(s):
Leslie Goodyear,  Education Development Center Inc,  lgoodyear@edc.org
Susan Kistler,  American Evaluation Association,  susan@eval.org
William Trochim,  Cornell University,  wmt1@cornell.edu
Discussant(s):
Thomas Chapel,  Centers for Disease Control and Prevention,  tchapel@cdc.gov
Thomas Schwandt,  University of Illinois Urbana-Champaign,  tschwand@uiuc.edu
Mary Stutzman,  Florida State University,  mstutzma@fsu.edu
Colleen Manning,  Goodman Research Group Inc,  manning@grginc.com
Abstract: Over 2000 members shared their background, input, concerns, and suggestions via the AEA Internal Scan, which included a membership survey, interviews and online discussion groups. The findings were useful and insightful and affected the association's strategic planning, programs, and services. We will highlight specific findings from the scan and tell how those findings have translated into plans and action for the association. We will also be sure that you know how to access the full scan, instruments, reports, and articles so that you may dig deeper into the scan results. Finally, we will examine our move to policy governance, share our plans for making AEA more engaged with the members, and more nimble and responsive, and open the floor for discussion and input as we plan for 2009 and beyond.

Session Title: The Science and Art of Creating Evaluation Standards That Stand the Test
Panel Session 509 to be held in Capitol Ballroom Section 6 on Friday, Nov 7, 9:15 AM to 10:45 AM
Sponsored by the AEA Conference Committee
Chair(s):
Hazel Symonette,  University of Wisconsin Madison,  symonette@bascom.wisc.edu
Abstract: Intended as a means to introduce the 3rd edition of The Program Evaluation Standards, this session goes well beyond simple descriptions of the product. Participants get an inside perspective on this extensive and difficult development process as well as the Joint Committee's focus on quality assurance. This session explores four facets of evaluation standards: (a) explanation of the operating procedures for developing standards why they are important, how they have changed, and why they changed; (b) the procedural changes employed by the Joint Committee and input received from the evaluation and practitioner professionals; (c) the fruits of the process that will serve as an American National Standard for the next decade or more, and (d) the steps the Joint Committee takes through an independent validation panel to assure quality in both the development process and the resulting product. Participant questions and comments are encouraged.
Key Elements of Developing Sound Evaluation and How the Joint Committee Works to Achieve Them
Arlen Gullickson,  Western Michigan University,  arlen.gullickson@wmich.edu
Since its inception in 1975, the Joint Committee has become the nation's preeminent developer of education evaluation standards. Its first and foremost product,' The Program Evaluation Standards.' is certified as an American National Standard by the American National Standards Institute (ANSI) and has guided evaluation planning and implementation both nationally and internationally since its first publication in 1981. This presentation focuses on the third edition of these standards. It describes the methods the Joint Committee employs to ensure high quality in its standards and the ways in which development methods have changed since the standards were first developed. This paper tracks the genesis of the standards development process, describes the forces that have shaped its development processes, and addresses the many practical issues that can derail best intentions to assure that the standards developed actually are principles commonly agreed to in the evaluation profession.
Stakeholder Contributions to the Development of The Program Evaluation Standards: 3rd Edition
Don Yarbrough,  University of Iowa,  d-yarbrough@uiowa.edu
Flora Caruthers,  National Legislative Program Evaluation Society,  caruthers.flora@oppaga.state.fl.us
Rodney Hopson,  Duquesne University,  hopson@duq.edu
Lyn Shulha,  Queen's University at Kingston,  shulhal@educ.queensu.ca
A primary goal of the Program Evaluation Standards development task force is that all interested stakeholders across the 16 sponsoring organizations as well as in the evaluation communities at-large have adequate opportunities to be involved, contribute suggestions, and make recommendations. This paper describes input received during development of The Program Evaluation Standards, 3rd Edition, including results of national and targeted surveys for needs assessment and format development, the 3-phased national and international review processes and results, the on-going hearings at national conferences, and the methodology and yields of the multi-stage field trials, some of which have been supported by the National Science Foundation. The full paper provides examples of feedback received via reviews, hearings, and field trials, and how that feedback has been put to good use. Remaining opportunities to contribute via field trials, reviews and national Web-based asynchronous 'hearings' will be discussed.
An Introduction to the Program Evaluation Standards, Third Edition
Lyn Shulha,  Queen's University at Kingston,  shulhal@educ.queensu.ca
Rodney Hopson,  Duquesne University,  hopson@duq.edu
Flora Caruthers,  National Legislative Program Evaluation Society,  caruthers.flora@oppaga.state.fl.us
Don Yarbrough,  University of Iowa,  d-yarbrough@uiowa.edu
The development process for the revised Program Evaluation Standards, 3rd Edition is nearing completion as the final reviews, hearings and field trials are completed following Joint Committee and ANSI guidelines. The near final draft is expected to be approved in September 2008 and is scheduled for publication in 2009. The 3rd edition is organized into individual chapters addressing the five key dimensions of quality in evaluations: Feasibility, Propriety, Accuracy, Utility, and Metaevaluation. In addition, it contains supporting chapters addressing the history of the standards development and development processes, what is 'new' to this edition and why a revision of the standards is needed now (and will be needed again within the next decade), how to use the standards in practical situations, and extended cases describing how multiple standards are applied in real life settings. This paper provides an introduction to the revised standards and to the 3rd edition book.
The New Program Evaluation Standards: A First Look
Marvin Alkin,  University of California Los Angeles,  alkin@gseis.ucla.edu
This presentation describes the validation panel's analysis of the Joint Committee's work in developing the standards and the quality of the standards that resulted. This panel, comprised of leading educators and evaluators and led by Dr. Alkin who is one of this nation's leading evaluation practitioners and theorists, is invested with determining whether the Joint Committee has met its own procedural expectations for development of standards and produced strong standards that truly are evaluation principles commonly agreed to by the education profession. The makeup of the panel, the procedures used, and the data employed are all described as factors important in panel determinations. This presentation directly addresses findings such as the congruence between procedures mandated for development of standards and actual procedures followed in the development process, as well as other criteria it applied to make in determining the Joint Committee's success in reaching the intended goals.

Session Title: Use of Indicators, Screeners, and Scaling Methods in Substance Use and Mental Health Evaluations
Multipaper Session 510 to be held in Capitol Ballroom Section 7 on Friday, Nov 7, 9:15 AM to 10:45 AM
Sponsored by the Alcohol, Drug Abuse, and Mental Health TIG
Chair(s):
Saumitra SenGupta,  APS Healthcare Inc,  ssengupta@apshealthcare.com
Development of Evaluation Indicators for a Policy Framework for Children and Youth Mental Health Services
Presenter(s):
Evangeline Danseco,  Center of Excellence for Child and Youth Mental Health,  edanseco@cheo.on.ca
Ian Manion,  Center of Excellence for Child and Youth Mental Health,  imanion@cheo.on.ca
Abstract: The Ontario Ministry of Children and Youth Services in Canada developed a policy framework for the children and youth mental health sector highlighting four broad goals: accessibility, integration, quality and evidence-based care, and accountability. Indicators to assess progress in achieving these goals were identified and selected using a modified Delphi approach. An expert panel of 21 stakeholders (parents, service providers and administrators, policy-makers, researchers, and advocates) reviewed candidate indicators based on relevance, validity, and feasibility. There were two rounds of online surveys and two panel meetings over six months. An initial 225 indicators were identified with seven of these being proposed for piloting and potential system-wide adoption. Through a series of online surveys and panel meetings over six months, the final list of recommended indicators included seven indicators based on validity, feasibility and relevance. This presentation will discuss the modified Delphi approach and the processes used to select the indicators.
Using Key Component Scaling to Evaluate the Implementation of a Community Crisis Bed Program
Presenter(s):
Purnima Sundar,  Carleton University,  purnimasundar@hotmail.com
John Sylvestre,  University of Ottawa,  jsylvest@uottawa.ca
Parastoo Jamshidi,  University of Ottawa,  pjams070@uottawa.ca
Matthew Manion,  University of Ottawa,  mmani069@uottawa.ca
Abstract: The Community Crisis Bed (CCB) program in Ottawa, Ontario, provides community-based alternatives to hospitalization for people experiencing a mental health crisis. Based on work by Cousins et al. (2004), a Key Component Scaling tool was developed to evaluate program implementation. From a program logic model, key programmatic activities were identified. For each activity, descriptions of “full” and “low” program implementation were written based on program documentation and feedback from an evaluation steering committee. Staff members from three program stakeholder groups, as well as clients and family members, participated in focus groups or interviews during which they rated and discussed implementation of the various program components. Scale data were used to categorize program components as achieving full implementation, inconsistent/partial implementation, or low level implementation. Qualitative data showed that implementation challenges were attributable to differences in expectation among the partners, as well as difficulties in communication, information sharing, and coordinating partners.
Theory of Planned Behavior and Propensity Scores: Applications for Longitudinal Evaluation of Underage Alcohol Use Prevention Programs
Presenter(s):
Robert Seufert,  Miami University at Middletown,  seuferrl@muohio.edu
Mark Carrozza,  University of Cincinnati,  mark.carrozza@uc.edu
Abstract: The research evaluates the utility of applying the Integrated Theory of Planned Behavior for understanding, predicting, and reducing underage alcohol use. The authors also assess the merit of using logistic regression-based propensity scores as a methodology to form statistically equivalent quasi-experimental groups during longitudinal program evaluations. Applications of other statistical procedures to evaluate program impact are assessed. Finally, we discuss and evaluate the policy implications of using media campaign messages to reduce underage drinking and related health and safety problems
Mental Health Performance Indicators to Measure Disparity: The California External Quality Review Experience
Presenter(s):
Saumitra SenGupta,  APS Healthcare Inc,  ssengupta@apshealthcare.com
Hui Zhang,  APS Healthcare Inc,  hzhang@apshealthcare.com
Sheila Baler,  APS Healthcare Inc,  sbaler@apshealthcare.com
Abstract: The role played by the External Quality Review Organizations (EQRO) in managed care waiver states in reviewing health and mental health plans are seldom discussed at the American Evaluation Association annual meetings. The proposed session will introduce the work of the California Mental Health EQRO in assessing statewide disparities in public mental health service access, expenditures and implications in policy initiatives and performance improvement at local and state levels.. The CA EQRO has utilized and improvised upon some commonly used performance indicators such as penetration rate, cost per beneficiary, and retention rate to examine racial and gender based disparities in the state. These indicators will be briefly described followed by some current findings and challenges of using this methodology in assessing disparity such as definitional issues, the EQRO role, as well as strengths of this approach including complementary data collection methods employed by the EQRO to enhance the review findings.

In a 90 minute Roundtable session, the first rotation uses the first 45 minutes and the second rotation uses the last 45 minutes.
Roundtable Rotation I: Blogging to the Beat of a New Drum: The Use of Blogs and Web Analytics in Evaluation
Roundtable Presentation 511 to be held in the Limestone Boardroom on Friday, Nov 7, 9:15 AM to 10:45 AM
Sponsored by the Integrating Technology Into Evaluation
Presenter(s):
Cary Johnson,  Brigham Young University,  cary_johnson@byu.edu
Stephen Hulme,  Brigham Young University,  byusnowboarder@yahoo.com
Abstract: Blogs are changing the face of marketing, business, and culture. How are they changing the practice of evaluation? This roundtable discussion will focus on how evaluators can use the combination of blogs and web analytics to assist in their evaluative processes both quantitatively and qualitatively. The facilitators will guide the participants in discussing some of the pros, cons, and potential hazards of using blogs in evaluations.
Roundtable Rotation II: Voicethread: A New Way to Evaluate
Roundtable Presentation 511 to be held in the Limestone Boardroom on Friday, Nov 7, 9:15 AM to 10:45 AM
Sponsored by the Integrating Technology Into Evaluation
Presenter(s):
Stephen Hulme,  Brigham Young University,  stephen_hulme@yahoo.com
Tonya Tripp,  Brigham Young University,  tonya.tripp@byu.edu
Abstract: Voicethread.com is a new website that allows users to comment on videos, images, documents, and presentations online. Not only can they make text comments, but they can also make audio and video comments. This is a rich resource for evaluators because they can gather far more data via a voicethread than a simple survey or interview. Because Voicethread.com is accessible to anyone with an internet connection, you can gather feedback from people outside your city, state, or even country. This roundtable will provide an open discussion about how this new technology could enhance evaluation in the future.

In a 90 minute Roundtable session, the first rotation uses the first 45 minutes and the second rotation uses the last 45 minutes.
Roundtable Rotation I: Painting a Picture of Evaluating Trainee Needs: Content and Experience With Training Methodologies
Roundtable Presentation 512 to be held in the Sandstone Boardroom on Friday, Nov 7, 9:15 AM to 10:45 AM
Sponsored by the Needs Assessment TIG
Presenter(s):
Sandra Naoom,  University of South Florida,  snaoom@fmhi.usf.edu
Adrienne Cadle,  University of South Florida,  adriennewoodleycadle@msn.com
Abstract: There is an intuitive benefit to evaluating a need for training as it identifies the motives for delivering (or not delivering) training, the method of delivery, the content to be trained and appropriate trainees. Despite the benefits of evaluating training, it is seldom reported in peer reviewed publications (Arthur et. al, 2003). This evaluation was aimed at assessing two different levels of school district leaderships’ need for training. The management-oriented approach was used to help trainers tailor the training they will deliver. The purpose of this formative evaluation was to identify training needs, past experiences with training others, and knowledge of training content. Training needs were assessed using mixed methodology, including paper and pencil surveys, as well as interviews. Organizational and person level analyses were conducted as part of the assessment of training needs. The benefits and limitations to conducting a needs assessment for training will be presented further in this paper.
Roundtable Rotation II: Measuring Impact Beyond the Training Room: Executive Leaders Training Program Evaluation
Roundtable Presentation 512 to be held in the Sandstone Boardroom on Friday, Nov 7, 9:15 AM to 10:45 AM
Sponsored by the Needs Assessment TIG
Presenter(s):
Jecky Misieng,  University of South Florida,  mjecky@gmail.com
Abstract: This evaluation was conducted to measure overall impact of executive leadership training on behavior changes and performance of district level administrators in a South Eastern school district. Approximately 100 participants, all new to their leadership positions, completed the training over a four year period (2004-2007). Kirkpatrick’s and Guskey’s evaluation models were used to provide insight on information at five different levels including reaction, learning, behavior, results, organizational support and change management. Several instruments were used to collect quantitative and qualitative data from participants, their supervisors and coaches in order to address the evaluation questions. These methods include post-training surveys, questionnaires, and interviews. Impacts at each of the above mentioned levels were examined based on the training evaluation criteria. Participant responses on the training, including logistic difficulties, expenses and limitations that provided useful information to improve the training will be discussed.

In a 90 minute Roundtable session, the first rotation uses the first 45 minutes and the second rotation uses the last 45 minutes.
Roundtable Rotation I: Measuring Changes in Teachers’ Mathematical Conceptual Understanding for Evaluation of Math-Science Partnership Professional Development Effectiveness
Roundtable Presentation 513 to be held in the Marble Boardroom on Friday, Nov 7, 9:15 AM to 10:45 AM
Sponsored by the Pre-K - 12 Educational Evaluation TIG
Presenter(s):
MaryLynn Quartaroli,  Northern Arizona University,  marylynn.quartaroli@nau.edu
Hollace Bristol,  Coconino County Education Services Agency,  hbristol@coconino.az.gov
Abstract: The Mathematics and Science Partnership (MSP) competitive grant program encourages partnerships between local education agencies and universities to collaboratively engage in professional development activities aimed at increasing teachers’ content knowledge and improving pedagogical practices. To facilitate planning and implementation of effective K-5 mathematics teacher training activities, one MSP grant team substantially modified the Frayer model graphic organizer into a new instrument, the Teachers’ Conceptual Understanding and Pedagogical Practice (TCUPP), as a pre-post assessment of teachers’ conceptual understanding of fundamental mathematical concepts and appropriate pedagogical practices. This instrument supplements data from standardized mathematics content tests and teacher self-assessments. In addition to serving as a pre/post measure, the new instrument can also be used to inform specific curricular decisions for summer institutes or other professional development sessions.
Roundtable Rotation II: Developing Higher Education Institution and University Partnerships in Reform of Science and Mathematics Education: A Three-Year Evaluation Study
Roundtable Presentation 513 to be held in the Marble Boardroom on Friday, Nov 7, 9:15 AM to 10:45 AM
Sponsored by the Pre-K - 12 Educational Evaluation TIG
Presenter(s):
Chad Ellett,  CDE Research Associates Inc,  cderesearch@charter.net
Judith Monsaas,  University System of Georgia,  judith.monsaas@usg.edu
Abstract: This round table discusses the findings of a recent qualitative evaluation (individual and focus group interviews) of factors that contribute to changes in organizational culture in four higher education institutions in a southeastern state and traditional policies and practices framing faculty tenure and promotion decisions and incentive/reward structures. The results were synthesized from extensive interviews with university/college level and department level administrators and key faculty informants. The primary focus of the study was to examine whether concerted (four-year) attempts to change higher education norms, values, and policies resulted in changes in organizational cultures supportive of new policy directions in faculty rewards and the enhancement of partnerships between K-16 science and mathematics education faculty and public school teachers and student achievement. The program evaluation results are interpreted and discussed in view of implications for the evaluation of statewide program innovation initiatives in mathematics and science education.

Session Title: Innovation, Complex Research Teams and Problems of Integration: The Missing Link
Panel Session 514 to be held in Centennial Section A on Friday, Nov 7, 9:15 AM to 10:45 AM
Sponsored by the Research, Technology, and Development Evaluation TIG
Chair(s):
Jerald Hage,  University of Maryland,  jerryhage@yahoo.com
Abstract: The management of innovation literature has a panacea that complex research teams with high rates of communication are responsible for innovation. The problems with this panacea are that it: (1) ignores the difficulties in creating high rates of communication in complex research teams; (2) does not consider the variety of complex research teams that exist; and (3) ignores the variety of linkages or ways in which teams can be integrated into networks of technical information. Our intent is to provide evaluators with check lists for how to determine the missing links of integration and how these might be measured.
Kinds of Complex Research Teams
Gretchen Jordan,  Sandia National Laboratories,  gbjorda@sandia.gov
The first problem and one that exists in all research teams is the amount of cognitive distance between the specialties within the complex team. As this distance grows, the possibility for radical innovation increases but communication tends to decline. To overcome this decline in communication, special mechanisms have to be used and policies instituted. The second problem of integration occurs when we consider the differences between small and large complex research teams. In the latter case, not only is the likely decline in communication augmented but the problem of maintaining a balance between research project autonomy and coordination occurs. The third problem of integration occurs when we consider the differences in the context of the small and large complex research teams. They are located in different kinds of idea innovation network contexts that can be distinguished by the kinds of gaps that are likely to occur with the growth in knowledge. The key is that the complex teams have to bridge these gaps.
Examples of Integrated and Non-integrated Research Teams in a Highly Innovative Research Organization' the Institut Pasteur
Jerald Hage,  University of Maryland,  jerryhage@yahoo.com
What are the organizational characteristics that encourage the formation of integration complex research teams? A historical analysis of the Institut Pasteur highlights the following structural characteristics that facilitated the creation of complex research teams: complex charter, terrific trio of visionary leaders, recruitment of scientists from very different channels, and multiple sources of funds. In addition, the Institut Pasteur illustrates multiple mechanisms for creating integration: common training course, duo-team leadership, joint research projects and publications, and a cultural sense of belonging to a family. The departments or laboratories in the Institut Pasteur that had these integrative characteristics achieved radical breakthroughs, while those lacking in these characteristics did not achieve similar breakthroughs. Likewise studies across time indicate how important it is to have complex teams that are integrated.
Critique of Current Network Studies: Not Measuring Complex Nodes, Project Integration and Gaps in the Idea Innovation Network
Jon Mote,  University of Maryland,  jmote@socy.umd.edu
A flurry of network studies has recently been completed but many of these have important limitations and are not connected to the management of innovation literature. In general, these studies fail to identify the network structures and processes that are most appropriate for achieving various types of innovation. One key issue is the need to study network ties in real time rather than on the basis of joint publications, joint research projects, etc. The research environment survey provides a series of questions that solve this, and examines the kinds and extent of various networks connected to the research project. In other words, we should measure links from the bottom up. Another key problem is the failure to measure gaps in the idea innovation network. It is not just a question of links, but which ones and are they in the right location.

Session Title: Core Quantitative Issues: New Developments in Outcome Measures
Panel Session 515 to be held in Centennial Section B on Friday, Nov 7, 9:15 AM to 10:45 AM
Sponsored by the Quantitative Methods: Theory and Design TIG
Chair(s):
Manuel C Voelkle,  University of Mannheim,  voelkle@rumms.uni-mannheim.de
Abstract: The session gives an overview of important new developments in constructing, analyzing and communicating outcome measures in evaluation research. It contains methodologically oriented presentations as well as examples from current evaluation studies. It is thus expected to be of interested to both the methodologically interested and applied researcher alike. The first presenter, Lee Sechrest, talks about the importance of properly calibrating outcome measures in order to make them meaningful for policy makers and practitioners. AndrTs Steffanowski proposes a new individualized outcome measure in rehabilitation research by combining information on status and change. Manuel Voelkle addresses the more general issue of emergent versus latent outcome measures in longitudinal designs. Mende Davis demonstrates the use of Rasch modeling to construct a measure of academic and professional success in mathematics, and illustrates its use in an outcome evaluation. Finally, Werner Wittmann discusses different approaches to synthesize outcome measures of large-scale evaluation studies.
Properly Calibrated Measures
Lee Sechrest,  University of Arizona,  sechrest@u.arizona.edu
The measures used in social science are often expressed in metrics that have no intrinsic meaning. Even measures of effect size are often not interpretable in any direct way. Persistent efforts should be exerted toward calibrating measures for meaning so that the effects of interventions can be described in ways that make sense to policy makers and practitioners. Unfortunately, very few such efforts have been mounted, and none systematically. It is easy, however, to exemplify the need and to illustrate the possibilities by reference to existing work. Examples also show why calibrated measures would be more persuasive and likely to result in implementation of effective interventions. Calibration research is an activity distinct from the development of basic measuring tools and need not interfere in any way with the production of evaluation measurement tools based on sound theory, state of the art methods, and pragmatic concerns for their implementation in the field.
Individualized Outcome Measures in Pre-Post-Studies: Combining Information on Status and Change
Andros Steffanowski,  University of Mannheim,  andres@steffanowski.de
Manuel C Voelkle,  University of Mannheim,  voelkle@rumms.uni-mannheim.de
Social interventions try to achieve two goals: First, to cure (or at least ameliorate) dysfunctional states (rehabilitation) and, second, to maintain desirable states (prevention). Regarding these two aspects, how is a fair outcome evaluation possible? The traditional approach has been to compute pre-post-difference scores. This, however, addresses only the first goal but not the second. In other words, maintaining a dysfunctional state, as well as maintaining a healthy state, would result in a zero-effect size (d = 0), with zero-effects being typically interpreted as 'no success'. Accordingly, using only pre-post-difference scores in outcome evaluation, the overall effect can be severely underestimated. To deal with this problem, an alternative measure has been developed by combining z-standardized pre-post-difference scores with z-standardized post-status scores for each item. The procedure is illustrated using a dataset of N = 858 psychosomatic inpatients with results indicating high reliability and good validity of the new outcome measure.
Emergent Versus Latent Outcome Measures in Longitudinal Analysis
Manuel C Voelkle,  University of Mannheim,  voelkle@rumms.uni-mannheim.de
Andros Steffanowski,  University of Mannheim,  andres@steffanowski.de
When designing and analyzing outcome criteria it is important to distinguish between emergent (i.e., formative) and latent (i.e., reflective) measures. While the former have typically been analyzed using methods of variance decomposition, the latter have a factor analytic tradition (Cole, Martin, & Steiger, 2005). In this presentation the distinction is reviewed for the analysis of longitudinal data. It is argued that approaching the distinction from the perspective of latent growth curve modeling as a general data analytic system for the analysis of change has several methodological, didactical and statistical advantages. All arguments are illustrated by our research on quality monitoring in ambulatory psychotherapy and results of a short Monte-Carlo simulation are presented to evaluate the underlying assumptions specific to the analysis of change.
Measuring the Educational Pipeline
Mende Davis,  University of Arizona,  mfd@u.arizona.edu
The measurement of change is often limited to single outcome variables, even when multiple measures have been collected. Relying on a single outcome measure can decrease power and reduce the likelihood of detecting the effect of an intervention. Combining multiple measures into a scale to measure change may result in greater sensitivity to intervention effects. This possibility is illustrated by the development and illustration of a scale to measure academic and professional progress in graduate school. An educational pipeline scale can incorporate multiple types of indicators, multiple sources of data, and even processes that play out over time. Rasch analyses are used in scale development, and the resulting scale is demonstrated in a program evaluation.
Outcome Measurement and Meta-Analysis, What are the Most Adequate Effect Sizes?
Werner Wittmann,  University Mannheim,  wittmann@tnt.psychologie.uni-mannheim.de
The majority of programs we evaluate are complex ones, i.e. intervention packages. Fair evaluations therefore need a set of different outcome measures. Applying meta-analysis to synthesize the effects of such programs we have several options. One is computing the effect sizes for each single outcome measure and average them or report them as an outcome profile to visualize level, scatter and shape what the program produced. Another strategy is considering the redundancy of the measures and reducing them via aggregation or factor analysis. The effect sizes resulting from the latter ones should be higher than the average of the single outcomes. Data from large-scale program evaluations using such multiple act outcome criteria are used to illustrate the differences. Meta-Analysis with no access to the intercorrelations of the outcomes used, will underestimate the effects of programs.

Session Title: Funder Evaluation Policy: Setting the Stage for Evaluation Practice and Use
Multipaper Session 516 to be held in Centennial Section C on Friday, Nov 7, 9:15 AM to 10:45 AM
Sponsored by the Presidential Strand and the Non-profit and Foundations Evaluation TIG
Chair(s):
Teresa R Behrens,  WK Kellogg Foundation,  tbehrens@wkkf.org
Evaluation as Organizational Strategy: Developing Evaluation Policy for Nonprofits and Foundations
Presenter(s):
Lester Baxter,  Pew Charitable Trusts,  lbaxter@pewtrusts.org
Glee Holton,  Pew Charitable Trusts,  gholton@pewtrusts.org
Scott Scrivner,  Pew Charitable Trusts,  sscrivner@pewtrusts.org
Abstract: This paper presents the results of a year-long process to prepare an evaluation policy at a national nonprofit. The founding premise for developing this policy is that evaluation should be positioned as a vital strategy for pursuing the organization’s core vision. The process included an analysis of the organization’s current status and trajectory and an assessment of the organization’s experience with evaluation. The revised policy integrates evaluation into the nonprofit’s core function—designing and implementing social-change strategies. Among the policy’s key elements are the following: (1) defining how evaluation will inform each major stage of the organization’s work; (2) specifying evaluation objectives and the primary client(s) in each stage; (3) proposing criteria to guide decisions about initiating evaluative work; and (4) addressing relevant operational issues (e.g., evaluation budgets, reporting relationships). The result is a policy designed to increase the value that evaluation delivers to the organization.
Evaluation Policy and Practice in Charity Accountability: The Better Business Bureau, Wise Giving Alliance
Presenter(s):
Nora Trochim,  Better Business Bureau,  ntrochim@council.bbb.org
Abstract: This paper presents the standards based evaluation framework of the BBB Wise Giving Alliance (the Alliance) and considers the value of such systems for establishing evaluation policies to enhance evaluation practice and encourage greater organizational effectiveness. The Alliance is a charitable affiliate of the Council of Better Business Bureaus, the national office of the Better Business Bureau system. Although most known for consumer complaint handling and business reporting activities, many BBBs also report on charities. The Alliance uses a standards driven approach to evaluate nationally soliciting charities to aid donors in making informed giving decisions and to encourage best practices among charitable organizations. The 20 Standards are used to assess charities and produce reports that describe their degree of compliance. The Standards also provide a framework for charities to conduct self-evaluation. This paper discusses the Standards and how they are applied, and critically assesses the practical implications of such approaches including how they encourage organizations to conduct consistent and regular effectiveness evaluations and how they advance the quality of evaluation policy and practice generally.
Potential, Process and Pitfalls of a Common Evaluation Format Among Funders
Presenter(s):
David Dobrowski,  First 5 Monterey County,  david@first5monterey.org
Abstract: A discussion about the process implemented in Monterey County to bring the five largest funders together to create a common evaluation format. The five funders (The United Way, The Community Foundation, Monterey Peninsula Foundation, Harden Foundation and First 5 Monterey County) worked on developing a common evaluation framework over the course of a year to use in conjunction with a common application form for funding. While a common framework was developed, its implementation has had varied success among the different funders. The way the process has affected policies of the funders and how the policies and approaches of the different funders effected the process will be discussed. This presentation explains the process taken, some of the challenges that were faced, how they were addressed and recommendations for other sets of funders looking to establish a common evaluation framework.

Session Title: Methodology and Tools for Assessing the Quality of Evidence
Panel Session 517 to be held in Centennial Section F on Friday, Nov 7, 9:15 AM to 10:45 AM
Sponsored by the Research on Evaluation TIG
Chair(s):
Nancy Kingsbury,  United States Government Accountability Office,  kingsburyn@gao.gov
Discussant(s):
Leslie J Cooksy,  University of Delaware,  ljcooksy@udel.edu
Abstract: Evaluators in all contexts are responsible for assuring that evaluation work follows good practice guidelines (e.g. AEA's Guiding Principles, Joint Committee Standards). Evaluators are also called upon to make judgments about the quality of others' studies. This may result in a metaevaluation of a single study, assessing study designs for their capacity to respond to particular policy questions, or, among other reasons, evaluating study quality for inclusion in a meta-analysis to determine the effectiveness of interventions from extant evaluations. Evaluators have devised heuristic tools such as checklists for making such judgments and scorecards to present summary snapshots of quality judgments to time-burdened decisionmakers. The first paper provides an overall framework within which the tools for assessing quality described in the other papers can be situated. The Discussant will focus on the challenges and opportunities evaluators' face when assessing quality through metaevaluation and its associated tasks.
Metaevaluation: A Conceptual Framework for Improving Practice
Valerie Caracelli,  United States Government Accountability Office,  caracelliv@gao.gov
Leslie J Cooksy,  University of Delaware,  ljcooksy@udel.edu
Metaevaluation, a mechanism for improving and informing stakeholders of an evaluation's strengths and weaknesses, is specifically called for in the Joint Committee's Program Evaluation Standards. A research program to clarify the current state of conceptual and technical development in metaevaluation was started in 2007. The Metaevaluation Project (Cooksy & Caracelli) has examined misunderstandings in terminology, purposes served, criteria of quality used, and its range of use in practice. The practice of metaevaluation as defined by the Joint Committee Standards remains limited, yet metaevaluative tasks requiring judgments of quality are frequently a part of other evaluation methodologies such as peer review, research synthesis, and meta-analysis. Drawing on the first phase of the project we will describe the theory and methodology associated with metaevaluation, including the criteria of quality (standards, principles, paradigm specific) used in practice. This paper sets the stage for the metaevaluative tools described in subsequent presentations.
Using Checklists to Assess Design Quality: Applications and Utility
Cindy K Gilbert,  United States Government Accountability Office,  gilbertc@gao.gov
Evaluators at the U.S. Government Accountability Office are often asked to assess the quality of evaluation designs - those of federal agencies as well as other research entities - at times making recommendations for improvement. For example, a recent report outlined the strengths and weaknesses of a Department of Defense (DoD) pilot program evaluation design, highlighting specific changes that could improve the validity of the study. Several types of evaluation checklists are available to guide such efforts, and are used to varying degrees at GAO. This paper will discuss available evaluation checklists, the characteristics these checklists have (and do not have) in common, the extent to which these and other frameworks are used at GAO, and their utility in a government auditing and evaluation setting.
Ensuring Quality in 'Score Card' Methodologies
Martin de Alteriis,  United States Government Accountability Office,  dealteriism@gao.gov
A number of US Federal Government agencies use 'score cards' or 'report cards' to assess key aspects of performance. For example, this methodology can be used to assess whether selected agencies have fully implemented, partially implemented, or taken no steps to implement a set of good practices. While 'score cards' allow for succinct presentations and definitive findings, there are also some potential limitations with this methodology, including possible oversimplification, and the 'shoehorning' of some issues into categories for which they are not a particularly good fit. Based on a review of reports that have used score cards, and input from evaluation specialists both inside and outside of government, this presentation will: 1) discuss the advantages and limitations of the score card methodology; 2) lay out key decisions that need to be made when constructing and using a score card; and 3) provide guidance on how quality can be ensured.
Shaping a Quality Product: A Balanced Approach to Assessing the Quality of Performance Audit in the United Kingdom
Jeremy Lonsdale,  National Audit Office United Kingdom,  jeremy.lonsdale@nao.gsi.gov.uk
Being seen to produce 'high quality' work is essential to the credibility of all evaluators. This paper will examine the quality assurance arrangements of the National Audit Office in the United Kingdom for its performance audit work, a form of evaluation that plays an important role in assessing the success of government programs. In particular, it will consider: How does the NAO define the quality of its products? Where do the criteria it uses come from? How does the NAO measure and monitor quality? And how is it trying to raise the quality of its work? The paper will assess how the NAO's approach to 'quality' has developed over the years, influenced in part by other disciplines. It will draw on a range of evidence, including 15 years of independent quality reviews, and highlight the current 'balance scorecard' arrangements designed to meet the expectations of different audiences.

Session Title: Excellence in Extension Evaluation: Views From Extension Education Evaluation Topical Interest Group Award Winners
Panel Session 518 to be held in Centennial Section G on Friday, Nov 7, 9:15 AM to 10:45 AM
Sponsored by the Extension Education Evaluation TIG
Chair(s):
Mary Arnold,  Oregon State University,  mary.arnold@oregonstate.edu
Discussant(s):
Mary Arnold,  Oregon State University,  mary.arnold@oregonstate.edu
Abstract: Panelists in this session will provide observation and reflection on Extension evaluation, from where it has been, to where things are currently, to where we are headed in the future. Each panelist will speak from the perspective of one of the five areas in which the EEE-TIG offers awards annually: 1) Sustaining excellence in Extension evaluation; 2) providing administrative leadership for evaluation; 3) promoting excellence in evaluation; 4) demonstrating excellence in evaluation teaching; and 5) providing outstanding leadership and service to the Extension evaluation profession. Presenters on this panel are past or current recipients of the EEE-TIG award in the category in which the panelist will speak. The collective wisdom and experience of this panel will provide great insight for any evaluator working as part of a complex organization such as Extension.
Building Evaluation Capacity: Excellence in Evaluation Training
Lisa Guion,  North Carolina State University,  lisa_guion@ncsu.edu
This award is given for evaluation training of outstanding quality. Criteria for the excellence in evaluation training include creativity, enablement of trainees in evaluation practice, and the use of training to further extension evaluation. The panelist is a past award winner with over 12 years of working in Extension evaluation. She has witnessed significant progress in evaluation capacity building efforts within Extension and within the two Extension systems in which she worked. Today, more in-depth, ongoing training is being provided that is being coupled with timely, user-friendly evaluation resources and tools. These efforts are producing Extension educators who have more positive/favorable attitudes towards evaluation, and are more knowledgeable and skilled in evaluation design and methods. This panelist will provide a framework and discussion of some of the critical components necessary for Extension organizations that want to engage in a holistic, comprehensive, multi-faceted evaluation capacity building effort within their state.
Sharing Extension's Impact: Excellence in Program Evaluation
Murari Suvedi,  Michigan State University,  suvedi@msu.edu
This award is given for an evaluation of outstanding quality. Criteria for this award are accuracy, feasibility, propriety, and utility, with particular emphasis placed on utility in improving Extension programs, policies and/or accountability. The panelist is a past award winner and professor of program evaluation at a major land-grand university. His professional background is varied, bringing a unique perspective to the evaluation of agriculture, environment and natural resources programs and projects. He conducts evaluation of Extension programs, teaches undergraduate and graduate courses in program development and evaluation, and offers short courses or workshops for development professionals focusing on evaluation. He has conducted evaluation of Extension programs and initiatives in Mexico, India, Nepal and the United States. The panelist will outline major factors influencing the quality of program evaluation in Extension and discuss ways to address them.
Supporting Evaluation Efforts: The Mary Nell Greenwood Extension Evaluation Award
Scott Reed,  Oregon State University,  scott.reed@oregonstate.edu
This award is given for administrative or program leadership toward outstanding organizational, policy and resource contributions to Extension evaluation. The panelist is currently a vice-provost of outreach and engagement at major land-grant university and will be nominated for this award in 2008. As vice-provost, this panelist also serves as the director of the Extension Service. The unique structure and emphasis of Extension in his state system has provided a rich context for the creative development and support of Extension Evaluation. The panelist will provide insight into the important ways that Extension administration can support and value evaluation efforts. And explore ways such efforts are instrumental in fulfilling the land grant mission of teaching, research and outreach.
Leading the Way: Leadership and Service to the Extension Evaluation Profession
Michael Lambur,  Virginia Polytechnic Institute and State University,  lamburmt@vt.edu
This award is given for leadership and service in creating networks and gaining visibility for Extension evaluators. A recent award winner in this category, this panelist will examine the evolution of Extension evaluation as a profession over the twenty-two years that Extension Education Evaluation Topical Interest Group (EEE TIG) has formally been in existence. The panelist will comment on the growth Extension evaluation as a profession through an exploration of the types and numbers of Extension evaluators, the role of the EEE TIG, professional development opportunities including AEA sessions, the role of our federal partner, and the structure for program evaluation within Extension organizations. Based on this, the presenter will provide perspectives on the future of Extension evaluation as a profession.
Building Evaluation Capacity That Lasts: Sustained Excellence in Extension Evaluation
Roger Rennekamp,  Oregon State University,  roger.rennekamp@oregonstate.edu
This award is given for conceptual, methodological and evaluation accomplishments making outstanding contributions to Extension evaluation. The panelist has a long career in Extension. His contributions to sustaining evaluation are well-recognized across the country, and he is frequently acknowledged by his peers as a leader in Extension evaluation. The panelist spent the majority of his career at one university, and during that time was instrumental in the evolution of evaluation throughout the organization. His work includes all aspects of developing a positive evaluation culture- from building individual capacity to implementing systems, processes and tools to enhance and sustain evaluation practice. Letters in support of this panelist's nomination for the sustained excellence award universally commented on the significant impact he made in changing the practice of evaluation. The panelist will share his knowledge and understanding of creating change within in a system that allows for evaluation success and sustainability over time.

Session Title: 'Expected Value of Perfect Information' in Health Economic Evaluation
Demonstration Session 519 to be held in Centennial Section H on Friday, Nov 7, 9:15 AM to 10:45 AM
Sponsored by the Costs, Effectiveness, Benefits, and Economics TIG
Presenter(s):
J Michael Menke,  University of Arizona,  menke@u.arizona.edu
Abstract: When do you have enough information or knowledge about a system, program, or treatment to make a sound decision? The expected value of perfect information (EVPI) when used with Bayesian methods in a decision analytic framework can address questions of whether resources should continue to be directed to system, program, or treatment effectiveness research. EVPI estimates the cost of uncertainty reduction in research and evaluation by indirectly estimating the cost of attaining 'perfect information', where the decision becomes error-free. This estimate is useful to inform health policy decisions, especially under resource constraints. Though effectiveness and cost effectiveness studies may help choose maximum health care or social welfare functions, they may still be poor guides, given the opportunity costs of overlooking alternative effective approaches. EVPI is a powerful tool for informing decisions by looking closely at uncertainty, and how much we are willing to spend to reduce it.

Session Title: Case Studies in Evaluation: United States Federal Agencies - Part 1
Multipaper Session 520 to be held in Mineral Hall Section A on Friday, Nov 7, 9:15 AM to 10:45 AM
Sponsored by the Government Evaluation TIG
Chair(s):
Joyce Keller,  St Edward's University,  joycek@stedwards.edu
Discussant(s):
Joyce Keller,  St Edward's University,  joycek@stedwards.edu
Improving Outcomes in Learning at the Defense Language Institute – How Program Evaluation is Contributing to the Increased Language Proficiency of Military Linguists
Presenter(s):
Melody Wall,  Defense Language Institute,  melody.wall@us.army.mil
Abstract: The Defense Language Institute Foreign Language Center annually hosts thousands of military members from all service branches learning more than 20 languages. The Institute has over the last several years implemented a robust plan consisting of numerous initiatives designed to improve the quality of learning, with almost $362 million of funding provided to support them. An effort of this size demands a vigorous evaluation plan to ensure that resources are being well spent. A team of evaluators is currently working at the Institute to accomplish this goal. The team has successfully used the participatory evaluation approach to enlist the interest and support of stakeholders, and has seen both formal and informal successes. This presentation will provide an overview of the participatory process at DLIFLC, the reactions to this approach from work groups, and how outcomes and behaviors have changed and improved as a result of the evaluation process.
The International Visitor Leadership Program (IVLP) Participants’ Perceptions of the Program: An Examination of the Influence of Major Events Covered By the Media on Program Evaluations and Assessments
Presenter(s):
Steven Gaither,  United States Department of State,  gaithersa@state.gov
Abstract: This paper examines the impact of media covered, world events on assessments and performance measurement in evaluations of Public Diplomacy (PD) programs. Using multiple regression and hierarchal linear modeling (HLM) the paper investigate the relationship between world events covered by the media and participant’s assessments of a frequently given PD program.
Evaluation Capacity Building Through Evaluating a Federal Leadership Development Program: Lessons Learned
Presenter(s):
Anna Dor,  United States Department of Homeland Security,  anna.dor@cgu.edu
Abstract: The pressure to retain quality workforce, has lead many federal agencies to focus on ways to develop the future leaders of the federal government. Over the next 10 years, 60% of people in the Federal workforce will be eligible to retire. More than likely, 40% of these individuals will retire when they are first eligible. Federal agencies are wondering if they are ready to fill if the leadership gap. I will describe a case study of a federal agency, the Federal Executive Board of Greater Los Angeles and its quest to maximize evaluation use and build evaluation capacity of its Leadership Associates Program. I will outline the benefits and lessons learned from the evaluation of the program. In my paper I outline lessons learned through practice, evaluation utilization and the potential establishment of evaluation policy with the Federal Executive Board.

Session Title: A Non-Technical Introduction to Analysis of Variance and Linear Regression
Demonstration Session 521 to be held in Mineral Hall Section B on Friday, Nov 7, 9:15 AM to 10:45 AM
Sponsored by the Quantitative Methods: Theory and Design TIG
Presenter(s):
Julius Najab,  George Mason University,  jnajab@gmu.edu
Caroline Wiley,  University of Arizona,  crhummel@u.arizona.edu
Susan Han,  George Mason University,  shan8@gmu.edu
Abstract: This demonstration is targeted toward evaluators unfamiliar with prevalent analyses (Analysis of Variance and linear regression) in evaluation work. The relative simplicity and ease of interpretation of the General Linear Model (GLM) makes it a common tool among program evaluators. However, failure to consider the role of and properly treat variance within the GLM in a program evaluation could potentially result in inaccurate inferences being made about program effectiveness. This demonstration will focus on variability and how variance is influential in conducting and interpreting ANOVA and linear regression (common GLM analyses). The session will linearly progress from data examination to proper statistical option selection, and end with interpretation of results. By the end of the demonstration, a non-technical audience will gain a better conceptual and technical understanding of how and why variance influences GLM analyses.

Session Title: Community Organizing Evaluation: Crossing the Next Frontier
Panel Session 522 to be held in Mineral Hall Section C on Friday, Nov 7, 9:15 AM to 10:45 AM
Sponsored by the Advocacy and Policy Change TIG
Chair(s):
Astrid Hendricks Smith,  The California Endowment,  ahendricks@calendow.org
Abstract: Community organizing is an area of advocacy work where there is considerable demand for more evaluation information and where there are new attempts to fill that void. This session is designed to provide the audience with an overview of numerous community organizing evaluation resources in existence, key themes in those resources, related tools that are available to the public and new models being tested. Case examples of recent evaluations done of community organizing work, as well as efforts to build organizations' capacities to do that work, will be presented and will bring several theories and models to life. The panel will explore the effectiveness of the tools and models used in the examples and the unresolved challenges faced. A group discussion with the audience will follow.
Resources and Themes for Community Organizing Evaluation
Susan Hoechstetter,  Alliance for Justice,  shoech@afj.org
There are a variety of different theories used and approaches taken by community organizing groups, evaluators, and funders in order to evaluate community organizing work. In an effort to identify resources for this work in a centrally available location, Alliance for Justice spent six months compiling electronic and hard copy evaluation models for an electronic compendium available to all. A review of resources demonstrated that some groups focus most heavily on empowering the community, others on achieving policy change, and yet others have a focus that combines these two areas and/or incorporate others. This presentation will provide an overview of the resources for informational purposes and will, at the same time, provide a basis for critiquing the evaluation examples in the ensuing presentations.
Evaluating Local and Statewide Community Organizing Efforts - What Does It Look Like?
Justin Louie,  Blueprint Research and Design Inc,  justin@blueprint.com
Moving beyond the theory, what does evaluation of community organizing look like on the ground? How do you engage organizers in developing their theories of change? What do you measure in community organizing and how do you measure it? How do you build organizers' own capacity to do evaluation? How do you feed back evaluation information to help improve organizing? We will present examples and lessons learned from our work with community organizers across the country, evaluating issues ranging from education reform to healthcare access to smart growth. We will discuss the local, regional, and statewide organizing efforts of individual organizing groups and organizing collaboratives, and provide real-world guidance to evaluators engaging or interested in evaluating community organizing.
Developing an Approach to Evaluating Community Engagement Advocacy Efforts
Jane Reisman,  Organizational Research Services,  jreisman@organizationalresearch.com
Sarah Stachowiak,  Organizational Research Services,  sstachowiak@orginaizationalresearch.com
Kendall Guthrie,  Bill and Melinda Gates Foundation,  kendall.guthrie@gatesfoundation.org
The Bill & Melinda Gates Foundation funds three community organizations to support a community engagement strategy in concert with high school redesign efforts in Austin, Texas. A key part of the learning for the Gates Foundation, as well as the Austin community partners and school district, has been an effort to develop theory of change models for each partner and for the effort overall based on the Foundation's comprehensive community engagement framework. Organizational Research Services (ORS) has been serving as an evaluation consultant to facilitate the theory of change development and build the capacity for designing tools, and collecting, analyzing and using data to evaluate the shorter-term advocacy changes identified in their models. This presentation will share the overall community engagement model and use the Austin experience as a case illustration to share models and tools and reflect on how this process has been working.

Session Title: Services, Quality, and Outcomes in Mental Health and Substance Use Evaluations
Multipaper Session 523 to be held in Mineral Hall Section D on Friday, Nov 7, 9:15 AM to 10:45 AM
Sponsored by the Alcohol, Drug Abuse, and Mental Health TIG
Chair(s):
Phillip Graham,  RTI International,  pgraham@rti.org
Does Quality Matter Too?: Evaluating A Tool to Assess Quality of Implementation in a Parenting-Based Program to Improve Children’s Postdivorce Mental Health
Presenter(s):
Cady Berkel,  Arizona State University,  cady.berkel@asu.edu
Sarah Jones,  Arizona State University,  sarahjp@asu.edu
Melissa Hagan,  Arizona State University,  melissa.hagan@asu.edu
Irwin Sandler,  Arizona State University,  irwin.sandler@asu.edu
Sharlene Wolchik,  Arizona State University,  wolchik@asu.edu
Tim Ayers,  Arizona State University,  tim.ayers@asu.edu
Erin Schoenfelder,  Arizona State University,  erin.schoenfelder@asu.edu
Anne Mauricio,  Arizona State University,  anne.mauricio@asu.edu
Abstract: While many contend program quality increases participants’ engagement, we lack standard measures for evaluating implementation, beyond fidelity checklists, which evaluate implementation quantity, but not quality. We present reliability and predictive utility of the QIM, an implementation quality measure rating seven domains of implementation (including fostering group cohesion, personalizing material). Data comes from video recordings of New Beginnings, a parenting program to improve children’s post divorce mental health, lead by two co-facilitators. For each session, we collect: 3 independently-rated QIMs (2 facilitators, 1 overall) 2 QIM self-assessments (1 per facilitator) Fidelity Engagement (participant attendance, homework completion) Using MLM with our nested, repeated measures data, we evaluate the following: Can adequate interrater reliability be achieved on the QIM? How do independent QIM ratings relate to fidelity, QIM self-assessments, and participant engagement? For independent ratings, how does an overall QIM score compare to a combination of the two facilitator QIM scores as predictors of engagement?
An Assessment of Substance Abuse Training and Technical Assistance Needs in Frontier States: Implications for the Adoption of Evidence-Based Practices
Presenter(s):
Joyce Hartje,  University of Nevada Reno,  jhartje@casat.org
Nancy Roget,  University of Nevada Reno,  nroget@casat.org
Wendy Woods,  University of Nevada Reno,  wwoods@casat.org
Joanne Brosh,  University of Nevada Reno,  jbrosh@casat.org
Abstract: The purpose of this project was to assess the training and technical assistance needs of substance abuse and allied health professionals in the six-state Mountain West Addiction Technology Transfer Center (MWATTC) region, and to examine their attitudes and level of interest in adopting specific evidence-based practices (EBPs). A web-based survey design and a snowball sampling procedure were used to facilitate collecting data from a representative sample in a short amount of time. Results from the needs assessment were used at the MWATTC Advisory Board meeting to guide state-level discussions regarding the training and technical assistance activities needed by substance abuse treatment professionals in their state. By getting stakeholder participation in this process, the MWATTC is able to more effectively implement stakeholder-driven training and technical assistance activities, thereby increasing the likelihood of events being relevant to participants, as well as enhancing event attendance and attitudes towards implementing EBPs.
Evaluating Findings from Nebraska's State Incentive Grant (SIG)
Presenter(s):
Phillip Graham,  RTI International,  pgraham@rti.org
Jason Williams,  RTI International,  jawilliams@rti.org
Abstract: The CSAP State Incentive Grant (SIG) program was the Center’s flagship initiative to reduce substance use among youth age 12-17 through the promotion and adoption of evidence-based prevention strategies. From 2004-2007, the State of Nebraska funded 27community coalitions to implement and evaluate comprehensive prevention strategies to reduce substance use among its youth. However, evaluation of this initiative has been hampered by several methodological challenges (e.g., lack of sound comparison data). Longitudinal Growth Modeling (LGM) will be applied to three waves of data collected in 2003, 2005, and 2007 to determine the overall impact of the SIG initiative. Results from latent profile analysis (LPA) will also be presented to identify key underlying characteristics of the most effective community coalitions. Findings will provide important insight regarding characteristics associated with program effectiveness in real world settings and will provide a critical test of federal policies requiring implementation of evidence-based strategies via community coalitions.
Linking Services to Outcomes in the Evaluation of San Francisco's Mental Health Services Act Full Service Partnerships
Presenter(s):
Thomas Bleecker,  Research, Evaluation, and Quality Management,  tom.bleecker@sfdph.org
Deborah Sherwood,  Research, Evaluation, and Quality Management,  deborah.sherwood@sfdph.org
Abstract: San Francisco’s implementation of Mental Health Services Act Full Service Partnerships (MHSA FSPs) targets underserved mental health and dual diagnosis clients. The treatment approach relies on an intensive case management model along with a commitment to do “whatever it takes” to support clients’ recovery and integration into the community. Client outcomes are tracked through the Data Collection and Reporting (DCR) system, a web-based application that captures initial assessments in life domains (e.g., residential, employment), changes in those domains (Key Events), and quarterly summaries. FSP mental health services are tracked in the county billing system. This presentation demonstrates our evaluation approach, which ties together the processes of service provision (from the billing system) to client outcomes (from the DCR). The presentation will demonstrate how differences in FSP client populations, treatment models, and outcomes were visually presented to program managers and staff, and how those results were used in ongoing program operations.

Session Title: Non-Profit Evaluation: Methods, Policies, and Practices
Multipaper Session 524 to be held in Mineral Hall Section E on Friday, Nov 7, 9:15 AM to 10:45 AM
Sponsored by the Non-profit and Foundations Evaluation TIG
Chair(s):
Ricardo Millett,  Millett & Associates,  ricardo@ricardomillett.com
Building Parent-Provider Shared Leadership to Improve Systems Serving Vulnerable Families
Presenter(s):
Arlene Andrews,  University of South Carolina,  arlene.andrews@sc.edu
Lisa Pion Berlin,  National Center for Shared Leadership Inc,  lpion-berlin@parentsanonymous.org
Margaret Polinsky,  Parents Anonymous Inc,  ppolinsky@parentsanonymous.org
Abstract: Shared Leadership in Action (SLA) was developed by Parents Anonymous® Inc., an international organization that helps vulnerable parents prevent child abuse through parental mutual support and parent-professional partnerships. Many parent leadership programs assume that if parents lead appropriately, service providers and systems will respond effectively. The SLA program aims to increase the probability of systems change by transforming the consumer-provider relationship. The SLA program involves joint parent leader-professional training followed by mentoring as the partners plan and enact reforms in community child and family service systems. This presentation will focus on the Shared Leadership Research Team. Using community-based participative research, parents, staff, and academicians participate in design and conduct of evaluation studies related to shared leadership. This presentation will review preliminary lessons learned about factors that facilitate and inhibit shared leadership in evaluation. Outcomes include factors related to parent empowerment, staff empowerment, partnership effectiveness, and targets of change.
Evaluation Policy and Planning within Nonprofits: A View from the Field
Presenter(s):
Kimberly A Fredericks,  Indiana State University,  kfredericks@indstate.edu
Joanne Carman,  University of North Carolina Charlotte,  jgcarman@uncc.edu
Abstract: In this paper, we explore how evaluation policy is formed within nonprofit organizations and by whom. The paper is based upon interview and self-assessment data gathered from 30 nonprofit organizations. The paper has four sections. First, we begin with a discussion about the important role that nonprofit organizations play in our society and the distinguishing characteristics that make them different from other organizations. Second, we describe four types of evaluation policies: 1) evaluation policies that are based upon the managing the requests and expectations from funders (i.e., government, the United Way, and foundations); 2) evaluation policies developed in tandem with strategic planning, which are integral to the nonprofit organization’s managerial processes and structures; 3) evaluation policies are formed as result of uncertainty in the funding environment; and 4) evaluation policies that are not well-developed. The paper concludes with recommendations connecting organizational theories to explain these practices.

Session Title: It's About the Relationship! Crosscutting Organizational policies That support Program Evaluation and Institutional Learning in Social Services.
Multipaper Session 525 to be held in Mineral Hall Section F on Friday, Nov 7, 9:15 AM to 10:45 AM
Sponsored by the Social Work TIG
Chair(s):
Benjamin Kerman,  Annie E Casey Foundation,  bkerman@caseyfamilyservices.org
Discussant(s):
Diane Purvin,  Annie E Casey Foundation,  dpurvin@caseyfamilyservices.org
Abstract: Social Services evaluation policy can be conceptualized as a series of formal and informal contracts developed among members of an organization to learn from its collective experience. The contract metaphor points to the intrinsic role of relationship in using evaluation effectively. Shared principles and goals underscore the contract parameters, such as the direct connection between evaluation and shared mission, valuing the commitment and capacity to apply feedback, the delineation of roles, confidentiality and the protection of participant rights. This session features three papers describing diverse evaluation activities with social services staff and managers illustrating how evaluation policies and activities manifest explicit and implicit contracts. The three projects involve the routine review of performance indicators, an implementation evaluation of a program redesign, and an outcome evaluation using a longitudinal follow-up. Evaluation policies need to adhere to common principles while remaining flexible enough to allow for adaptation to questions and context.
The Role of Evaluation Policies in Integrating Agency-wide Performance Measurement
Ellen Ross,  Annie E Casey Foundation,  eross@caseyfamilyservices.org
This paper examines how evaluation policy can facilitate agency wide performance measurement. In 2002, Casey Family Services adopted results-based accountability across the agency's core goals. This framework provides a simple, disciplined rubric, common language, and common ground for implementing new ideas. Operational tasks included how integrating evaluation into practice, recruiting and engaging social workers from the onset, and institutionalizing the framework. By involving all organizational levels in routine indicator review, leadership and staff were alerted to areas of concern and accomplishment. Critical in this process was the involvement and buy-in from senior management, as well as collaboration across research, field operations, information technology, and direct care staff. Challenges included: timing, funding, buy-in from stakeholders, and organizational culture. Strategies used to address these challenges included building departmental relationships, development of policies and infrastructure to embed the data collection and review in ongoing activities and structures, and attending to diverse stakeholder's needs.
Policies, Pressures, and Expectations: Managing Relationships During a Multifaceted Implementation Evaluation
Judy Lee,  Casey Family Services,  judymlee@msn.com
Benjamin Kerman,  Annie E Casey Foundation,  bkerman@caseyfamilyservices.org
Casey's Move to Greater Permanency initiative reflects an intensified focus on providing permanency planning and foster care services to youth in effort to increase legal and relational permanence. While some program refinements were similar across service sites (e.g., use of Permanency Teaming), additional refinements were site specific. As expectations for a swift and complete integration of Permanency Teaming practice increased among agency administrators, pressures also mounted among field leaders and staff who were experiencing increased uncertainty given the major shifts in practice and procedures. Amidst these changing demands and social worker anxiety, the evaluators used an array of existing policies to conduct a multifaceted implementation and short-term outcome evaluation. Close coordination between internal evaluators and the reliance on external consultants helped ensure that the relationships between evaluator and evaluand were appropriate and corresponded optimally to the particular evaluation question and component.
Follow-up Evaluation and the Long View: Preparation for the Casey Longitudinal Study of Foster Youth Development
Gretta Cushing,  Casey Family Services,  gcushing@caseyfamilyservices.org
Diane Purvin,  Annie E Casey Foundation,  dpurvin@caseyfamilyservices.org
Longitudinal and follow-up projects may require additional policies and structures that go beyond those of either performance measurement or concurrent program evaluation designs. This paper will discuss how researchers and social workers collaborated in the development of a longitudinal study aiming to provide the agency leadership with an unbiased view of outcomes, as well as to examine applied developmental questions concerning vulnerable youth and their families. All youth who were placed in a foster home after December 31 1993 and spent one year with Casey are included. Data collection includes record reviews and young adult interviews at ages 19, 22 and 25. A variety of agency policies and structures are essential to the conduct of longitudinal service outcome research, including location of former clients and response to service requests, internal communication and confidentiality, and handling reports of maltreatment, as well as the time investment inherent in longitudinal research.

Session Title: Collaborative Evaluations: Successes, Challenges, and Lessons Learned
Multipaper Session 526 to be held in Mineral Hall Section G on Friday, Nov 7, 9:15 AM to 10:45 AM
Sponsored by the Collaborative, Participatory & Empowerment Evaluation TIG
Chair(s):
Susan Hibbard,  University of South Florida,  hibbard@coedu.usf.edu
An Accountability Evaluation Supportive of Stakeholder Learning
Presenter(s):
Cheryl Poth,  University of Alberta,  2cnp@queensu.ca
Abstract: Using the two-year evaluation of the Queen’s University Inter-Professional Patient-centred Education Direction (QUIPPED) program as a case study, this paper describes a participatory approach supportive of learning and ongoing program development while meeting the evaluation’s primary purpose for accountability to government funders. Our current understandings of the participatory approach describe it as useful to support individual and organizational learning. However, studies of evaluation use have yet to examine specifically how evaluators create opportunities for stakeholders to learn and at the same time, how evaluators produce outcomes that are useful for external accountability purposes. This paper reports the analysis of evaluator behaviour as it endorsed, modified, and extended current notions of the participatory approach.
Empowerment evaluation: Did The Approach Work with a National School Breakfast Program In Australia?
Presenter(s):
Wayne Miller,  Avondale College,  wayne.miller@avondale.edu.au
Heather Yeatman,  University of Wollongong,  hyeatman@uow.edu.au
Abstract: The empowerment evaluation approach was used in an evaluation of a national school breakfast program in Australia. Preliminary findings have been reported previously (Miller and Lennie, 2005). The purpose of this presentation is to report key evaluation outcomes associated with employing the empowerment evaluation approach rather than program outcomes as a result of the evaluation. Evaluation outcomes were found to be of two types - those that are clearly linked to the steps of the empowerment evaluation approach, and others linked more generally to the principles of empowerment evaluation. The three steps provided a simple, progressive framework for the evaluation. This built early confidence in the approach which contributed to the achievement of significant evaluation and program outcomes. However, participants reported both alignment and misalignment with the ten principles. The presentation will discuss findings in relation to the application of the empowerment evaluation approach. Strengths and limitations with respect to its use with a community-based, dispersed and diverse program will be highlighted. Some signposts will be provided for those looking to use empowerment evaluation in similar applications.
Strategies for Using Stakeholder Input to Improve Evaluation Designs
Presenter(s):
Miles McNall,  Michigan State University,  mcnall@msu.edu
Lauren Lichty,  Michigan State University,  lichtyla@msu.edu
Abstract: Participatory and collaborative evaluation approaches call for meaningful involvement of stakeholders in nearly every aspect of evaluation, including evaluation design. When stakeholders participate in decisions about evaluation design, they impart critical information about the context in which the evaluation will be implemented and their information needs. In the absence of such information, evaluations are unlikely to meet standards of feasibility, utility, or accuracy. In our experience, discussions with stakeholders about evaluation design surface issues such as the political and logistical feasibility of various methods and the kinds of evidence that will have credibility for various stakeholder groups. In this paper, we present a framework to assist evaluators and stakeholders in considering these issues and discussing their implications for evaluation design. The framework is designed to promote evaluation designs that meet stakeholders information needs (utility) and are responsive to local realities (feasibility) while upholding methodological integrity (accuracy).
Walking a Fine Line: The Role of the Evaluator in Monitoring Program Implementation
Presenter(s):
Rebeca Diaz,  WestEd,  rdiaz@wested.org
Abstract: This presentation will discuss the role of evaluators in monitoring the implementation of programs in school districts. The impetus for this topic comes from an evaluation study of a federal Smaller Learning Communities (SLC) grant implemented in a school district in northern California. The implementation of the grant was virtually void of district leadership, and as a result, the program contributed to negative student outcomes. The evaluator, who was subcontracted by the district, found this experience to be particularly challenging as she witnessed low socio-economic, ethnic minority students at a low performing high school expressing feelings of inferiority as a result of their participation in the program. She began to question her own responsibility in the situation. This presentation will explore questions such as: What is the responsibility of the evaluator in the implementation of school programs? How does one conduct a fair and meaningful evaluation while respecting both the client and the goals of the program?
Collaborative Evaluation Strategies that Promote Culturally-Responsive Evaluations
Presenter(s):
Monifa Beverly,  University of Central Florida,  mbeverly@mail.ucf.edu
Karyl Askew,  University of North Carolina Chapel Hill,  karyls@email.unc.edu
Michelle Jay,  University of South Carolina,  jaym@gwm.sc.edu
Abstract: Collaborative evaluation is an orientation to evaluation that promotes culturally-responsive practice. The goal of both collaborative evaluation and culturally-responsive approaches is to use an understanding of the cultural context to develop and execute evaluations that accurately capture program impact and provide maximum benefit to stakeholders. This presentation provides three examples of evaluations that demonstrate the use of four collaborative evaluation techniques resulting in practices that were culturally-responsive: collaborative planning, evaluation technical assistance, evaluation capacity building, and evaluation fairs. The evaluands include single-site and multi-site programs focused on delivery of services to ethnically diverse populations. The presenters, five individuals trained in the collaborative evaluation approach, offer critical reflection on using collaborative evaluation methods to promote culturally responsive practices. The evaluators will present successes, challenges, and practical recommendations for using collaborative evaluation techniques that can be widely applied to enhance the cultural-responsiveness of evaluations and evaluators.

In a 90 minute Roundtable session, the first rotation uses the first 45 minutes and the second rotation uses the last 45 minutes.
Roundtable Rotation I: Issues, Challenges and Solutions of Evaluating International Scholars/Participants Training in the United States
Roundtable Presentation 527 to be held in the Slate Room on Friday, Nov 7, 9:15 AM to 10:45 AM
Sponsored by the Multiethnic Issues in Evaluation TIG and the International and Cross-cultural Evaluation TIG
Presenter(s):
Deborah Oh,  California State University at Los Angeles,  doh2@calstatela.edu
Abstract: Twenty-six (26) university senior administrators from Guangdong University of Foreign Studies completed the Guangdong University of Foreign Studies Senior Administrators Training Program, a comprehensive institute at California State University, Los Angeles (CSULA). The scope of this three week institute held from July 30, 2007 to August 18, 2007 was two-fold: focused training on various topics, conducted primarily by faculty at CSULA, and fieldwork visits to various university campuses here in Southern California and in the East Coast, thereby featuring a hands-on intensive orientation, and familiarization of current organizational structures and governance, curriculum, student affairs, development, infrastructure support, and partnerships of institutions of higher learning in the US. The issues and challenges and possible solutions related to evaluating participants in a cultural sensitive manner, namely, the manner in which the surveys were conducted, the survey results are understood, and the manner in which the surveys uncovered unanticipated answers which were not part of the program or evaluation design will be discussed.
Roundtable Rotation II: Evaluating the Impacts of Bolsa Escola Programs on Student Performance in Brazil
Roundtable Presentation 527 to be held in the Slate Room on Friday, Nov 7, 9:15 AM to 10:45 AM
Sponsored by the Multiethnic Issues in Evaluation TIG and the International and Cross-cultural Evaluation TIG
Presenter(s):
Amy Lora,  Florida International University,  alora001@fiu.edu
Abbas Tashakkori,  Florida International University,  tashakko@fiu.edu
Abstract: Beginning in 1997, conditional cash transfer programs started throughout Brazil, giving children the opportunity to choose school over work. Currently, these programs (known as Bolsa Escola) support over eight million children through the distribution of R$127.2 million to 5,512 municipalities per month. Thus far, evaluations and investigations of Bolsa Escola programs have focused on enrollment, attendance, and dropout rates, indicating that the programs have been largely successful in achieving these goals. However, studies have largely neglected to investigate, or have failed to demonstrate, impacts on students' academic performance. This paper presents the results of an evaluation project, examining the possible effects of the program on student performance in Mathematics and Language Arts in grades 4 and 8.

Session Title: Conducting Evaluation With a Focus on Stakeholders
Multipaper Session 528 to be held in the Agate Room Section B on Friday, Nov 7, 9:15 AM to 10:45 AM
Sponsored by the Graduate Student and New Evaluator TIG
Chair(s):
S Lory Hovsepian,  University of Montreal,  sarine.lory.hovsepian@umontreal.ca
Crossing Country and Cultural Boundaries for Work
Presenter(s):
Lennise Baptiste,  Kent State University,  lbaptist@kent.edu
Abstract: The development co-operation strategy adopted in 1996 by the Organization for Economic Cooperation and Development, an international organization focused on economic and social issues. This strategy shaped two partnerships formed between one European and two Caribbean non-governmental organizations. The lessons learned from the evaluation of the partnerships centered on, cross-cultural relations, access, project plan logistics, and of negotiation of the Terms of Reference which required qualitative methodology. One key finding was that routine tasks can become monumental when evaluators are in unfamiliar terrain even though language was not a barrier. This presentation can inform evaluators who undertake projects spanning cultural and country boundaries.
Exploring Connections Between Programs, Their Evaluation and Those Involved: An Example of the Implementation of a Community-based Pilot Program for Women Living with HIV
Presenter(s):
S Lory Hovsepian,  University of Montreal,  sarine.lory.hovsepian@umontreal.ca
Astrid Brousselle,  University of Montreal,  astrid.brousselle@umontreal.ca
Joanne Otis,  University of Quebec Montreal,  otis.joanne@uqam.ca
Abstract: Background: Community-university collaborators developed and implemented “Pouvoir Partager/Pouvoirs Partagés”, a community-based program (CBP) helping women living with HIV in Montreal, Canada to manage the disclosure of their HIV status. Process and outcome evaluations were undertaken in the context of a formative evaluation. Objective: We explore the interconnections between the CBP, its evaluation and participatory processes. Methods: Interviews with stakeholders, focus groups with program participants and committee meeting observations were undertaken. Results: Extent of participatory evaluation varied according to stakeholder expertise and interests, evaluation type and other practical considerations. The evaluation process seemed to facilitate participation in program development and the program, and aid its continuation. Although evaluation contributed to improving the program and its implementation, some perceived it as strain, restricting implementation flexibility and limiting program impact. Conclusion: Planning, implementing and evaluating CBPs require that stakeholders address issues emerging from the interrelations between the program, its evaluation and participatory processes.
Evaluation Decision Making as Authoritative Allocation of Value
Presenter(s):
A Rae Clementz,  University of Illinois,  clementz@uiuc.edu
Jeremiah Johnson,  University of Illinois Urbana-Champaign,  jeremiahmatthewjohnson@yahoo.com
Abstract: Evaluation resources are limited. There is rarely enough time, funding, people, etc., to accomplish everything we might need and want to. David Easton (1954) defined policy as "an authoritative allocation of value." In this same way, the decisions made during the course of evaluation regarding evaluation design and implementation reflect value stances and are actualized through the allocation of the limited evaluation resources at our disposal. Thus, our practical and functional decisions in evaluation practice carry value implications. Evaluation decisions are never made solely by the evaluator. They are made in context and reciprocally with the group of individuals involved in conducting the evaluation. This session discusses these value decisions and their role in evaluation practice. Separate from method or approach, this framework is intended to engage the practical implications of holding various value stances in the conduct of an evaluation.
Evaluability Assessment: What Makes a Program Ready for Evaluation?
Presenter(s):
Gargi Bhattacharya,  Southern Illinois University at Carbondale,  gargi@siu.edu
Meghan Lowery,  Southern Illinois University at Carbondale,  meghanlowery@gmail.com
Alen Avdic,  Southern Illinois University at Carbondale,  alen@siu.edu
Abstract: Evaluation of program impact should guide the development of programs; however, in real-life organizational situations, clients are often unaware of this essential program development process. Professional evaluators often encounter clients who are unsure of how they want to evaluate a program or how they want to measure program impact. Applied Research Consultants (ARC), a vertical practicum built within the doctoral program of Applied Psychology at Southern Illinois University at Carbondale, is a graduate student-run consulting firm. ARC conducts professional evaluations and provides other services (e.g., job analysis, survey research, data analysis etc.) to clients in and outside the University community. Different programs evaluated by ARC were analyzed and evaluators were interviewed to identify components of programs that made them easier to evaluate or become evaluation-ready.

Session Title: Evaluation Policy and Practice: Its Impact on African American Women's Sexual and Reproductive Health
Multipaper Session 529 to be held in the Agate Room Section C on Friday, Nov 7, 9:15 AM to 10:45 AM
Sponsored by the Feminist Issues in Evaluation TIG
Chair(s):
Jenny L Jones,  Virginia Commonwealth University,  jljones2@vcu.edu
Abstract: This multipaper presentation will include three studies that highlight the need for sound evaluation policies which influence practice. Each paper focuses on African American women and health disparities as they relate to HIV/AIDS as well as adverse birth outcomes. Quantitative and qualitative findings will be presented for Young Black College Women: HIV/AIDS Secreted 'At Risk' Group and An Exploration of Factors That Influence the Sexual Decisions of Young Black Women: What Are They Really Thinking respectively. Exploring Adverse Birth Outcomes Among African American Women with Means: Revealing a New Population will include a conceptual discussion regarding the adverse birth outcomes of African American women. Topics related to black feminist as well as gender and power theories will be discussed.
HIV Risk Behaviors among a Sample of Young Black College Women
Binta Alleyne,  University of Illinois Chicago,  balleyne@uic.edu
The relationship between certain factors associated with the theory of gender and power including: sexual relationships, condom use self-efficacy, and substance use, to young Black college women's condom use was studied. A sample of 189 young Black women between the ages of 18 and 24 from a Historically Black University in the South was selected to complete a HIV/AIDS Risk Behavior Survey. An ordinal logistic regression revealed that self-efficacy and type of sexual relationship were positively associated with condom use, whereas substance use was negatively associated with condom use. This discussion will focus on the need to identify young Black women in college as a non-disenfranchised 'at risk' group, understand how the type of sexual relationships in which young Black women are involved influence their condom usage, and the need to evaluate current HIV prevention practice and policy with this population.
An Exploration of Factors That Influence the Sexual Decisions of Young Black Women: What are they Really Thinking?
Aisha Tucker-Brown,  University of Georgia,  aktb@bellsouth.net
This study explored the factors that influence the decision making of 25 African American women between the ages of 18 and 25. They were purposefully selected and interviewed. Interviews and focus groups served as the sole source of data for this study. This was a basic interpretive study and a qualitative research design was used to explore the factors that influence sexual decision making among these participants. Data were analyzed using the constant comparative method. An analysis of the data revealed that participants were in fact quite knowledgeable about HIV/AIDS and prevention strategies. This raises the question, what are we evaluating regarding HIV prevention programs? Are we allowing practice to dictate policy or vice versus. Conclusions also address the intersection of race, class, and gender and its relevance to the development of sound evaluation policy and practice with regards to HIV prevention initiatives.
Exploring Adverse Birth Outcomes Among African American Women with Means: Revealing a New Population
Kimberly Farris,  University of Illinois at Chicago,  kimberlydfarris@gmail.com
Health researchers have long noted racial disparities in birth outcomes, including pre-term delivery, low birth weight, and infant mortality. Recent literature has shown that African American women in higher SES categories are more likely to have adverse birth outcomes including low birth weight, pre-term deliveries, and fetal death. While researchers have focused on socioeconomic and behavioral explanations, few have examined psychosocial factors among this population. The issue warrants further investigation through the initiation of dialogue to identify potential methods for researchers/evaluators to reach women not traditionally considered as disenfranchised in terms of SES factors but still experience adverse outcomes. Furthermore, the discussion will focus on conceptualizing potential psychosocial factors that may predict adverse outcomes among this population, developing a model examining the extent to which these factors affect birth outcomes, and exploring methods on recruitment strategies of participants and issues related to understanding the role of evaluation practice and policy.

Session Title: Evaluating Early Childhood Programs: Methodological Challenges and Lessons Learned
Multipaper Session 530 to be held in the Granite Room Section A on Friday, Nov 7, 9:15 AM to 10:45 AM
Sponsored by the Pre-K - 12 Educational Evaluation TIG
Chair(s):
Barbara J Helms,  Education Development Center Inc,  bhelms@edc.org
Avoiding the Ethical and Methodological Deficiencies Found in Widely Acclaimed Early Childhood Education Studies
Presenter(s):
Morris Lai,  University of Hawaii Manoa,  lai@hawaii.edu
Susan York,  University of Hawaii Manoa,  yorks@hawaii.edu
Abstract: Although many have praised the High/Scope Perry Preschool and the Abecedarian Preschool studies, we seriously question whether such studies strongly generalize today and meet prevailing ethical, evaluation, and other professional standards. We are disturbed by the devastating fate of the control-group children; we have also noted other types of inappropriateness in published studies on the effects of prekindergarten early childhood education (ECE). After delineating our concerns about several research or evaluation studies in the ECE literature, we develop recommendations that address the numerous inadequacies. Evaluations of ECE initiatives should (a) accommodate allowing all children to participate in the main treatment if their families so choose, (b) include nonformal/noncenter-based programs, (c) use data-collection instruments designed to specifically measure children’s readiness for kindergarten, (d) regard participant feedback as primary evaluation data, (e) analyze the data focusing on effect size rather than on statistical significance, and (f) ensure cultural appropriateness.
Pre-Kindergarten Teachers' Literacy Knowledge: Instrument Development and Lessons Learned
Presenter(s):
Melissa Chapman,  University of Iowa,  melissa-chapman@uiowa.edu
Abstract: This proposal is contextualized in this evaluator’s experiences working as an external consultant for an Early Reading First project and the problems encountered when federal site visitors requested that project teachers be evaluated on their knowledge of the mandated early reading areas: oral language (expressive and receptive), phonological awareness, print awareness, and alphabet knowledge (Whitehurst & Lonigan, 1998; USDOE, 2006). Given the absence of widely available, technically adequate assessments that would serve this purpose, an open-ended instrument was first created and used. This proposal is focused on improving the technical adequacy of a Language and Literacy Assessment (LLA) by conducting cognitive interviews (Willis, 1999, 2005) with expert teachers and researchers and then assessing a large sample of pre-K teachers in a Midwestern state. Further item level statistics and other technical adequacy issues will be examined following the statewide distribution.
The Use of Parent and Teacher Task Scenarios in the Evaluation of a Literacy Website
Presenter(s):
Margaret Vaughan,  Goodman Research Group Inc,  vaughan@grginc.com
Margaret Tiedemann,  Goodman Research Group Inc,  tiedemann@grginc.com
Abstract: The paper is based on a comparative study, one component of a comprehensive website evaluation. The study involved recruiting parent and teacher participants to examine the strengths and weaknesses of a targeted early childhood literacy site in comparison to its primary competitors. A unique aspect of the evaluation was the inclusion of hypothetical task scenarios. A task scenario, often part of usability testing, is a representation of work a user would likely perform while visiting a website. These task scenarios were designed to gain feedback from the website’s primary audiences and to be meaningful to their perspectives and roles. The scenarios provided participants with a child-based situation to consider rather than having them follow a step-by-step approach. The paper will report on methodology and findings with a focus on parents and teachers overall impressions of the website as well as specific details about their experiences while completing the task.
Development of an Interview Protocol for Families of English Language Learners in Head Start Programs
Presenter(s):
Marc Winokur,  Colorado State University,  marc.winokur@colostate.edu
Victoria Buchan,  Colorado State University,  buchan@cahs.colostate.edu
Abstract: The Head Start English Language Learner Project (HELLP) is one of eight Head Start Innovation and Improvement Projects funded by the U.S. Department of Health and Human Services to strengthen Head Start programs nationwide. Specifically, HELLP is focused on supporting Head Start programs to better serve English language learner (ELL) children and their families. Through evidence-based curricula, hands-on training, and programmatic assistance, HELLP is designed to build the capacity for Head Start personnel and families to assist ELL children in their literacy development. To evaluate this ambitious initiative, a comprehensive program evaluation was fashioned. A major component of the evaluation was a parent-family interview designed to explore the level of engagement of ELL parents in supporting their children’s literacy development in the home and at school. The proposed paper will present the unique approach used in the protocol development and administration process, while reporting on findings from the interviews.

Session Title: HIV/AIDS Prevention at the Community Level: Innovative Evaluation Approaches
Multipaper Session 531 to be held in the Granite Room Section B on Friday, Nov 7, 9:15 AM to 10:45 AM
Sponsored by the Health Evaluation TIG
Chair(s):
Ann Zukoski,  Oregon State University,  ann.zukoski@oregonstate.edu
Evaluating the Effectiveness of the Baby and Me Tobacco Free Program
Presenter(s):
Anne Gadomski,  Bassett Research Institute Computing Center,  anne.gadomski@bassett.org
Laurie Adams,  Tri County Tobacco Control Program,  lauriea@windstream.net
Nicole Krupa,  Bassett Research Institute Computing Center,  nicole.krupa@bassett.org
Patrick Burdick,  Bassett Research Institute Computing Center,  patrick.burdick@bassett.org
Nancy Tallman,  Bassett Research Institute Computing Center,  nancy.tallman@bassett.org
Nancy Weissflog,  Bassett Research Institute Computing Center,  nancy.weissflog@bassett.org
Paul Jenkins,  Bassett Research Institute Computing Center,  paul.jenkins@bassett.org
Abstract: The Baby & Me Tobacco Free Program is a multi-pronged intervention that includes cessation counseling, biomarker feedback and an incentive program to facilitate prenatal and post-partum smoking cessation. The intervention includes four face-to-face contacts prenatally with a cessation specialist who does counseling and carbon monoxide (CO) testing. For 1 year post-partum, mothers are CO tested every 3 to 4 weeks and, if negative, receive a diaper voucher. Outcomes include biochemically validated monthly abstinence rates and self reported quit rate at 12 months post-partum ascertained by telephone interview. To date, 22 sites (WIC offices and obstetric clinics) in 9 counties in New York State have enrolled 400 women. Preliminary results include 289 women who have completed, on average, between 5 and 6 (mean=5.63) follow-up tests (CO and/or saliva), with a mean passing/quit rate of 93.8%. Among the 45 women who have completed 12 follow-up tests, the mean passing/quit rate is 96.9%.
Examination of Peer Advocates in HIV/AIDS Care and Treatment
Presenter(s):
Marcia Dutcher,  Kansas City Free Health Clinic,  marciad@kcfree.org
Brenda Loscher Hudson,  Kansas City Free Health Clinic,  brendal@kcfree.org
Abstract: The Peer Education Training Site (PETS) project is a Health Resources and Services Administration (HRSA) multiyear funded project to increase the number of trained HIV peer educators and the number of organizations with the capacity to utilize peers in their HIV service delivery. Three sites located in California, Missouri, and New York have developed peer and organization training programs tailored to the assets, needs, and preferences of their individual target populations. Training sessions have been offered since 2006. A cross-site quantitative longitudinal design is used to measure research questions for peers and organizations while a qualitative longitudinal design is used to address research questions for clients of peer advocates. The paper will discuss the research methodologies to evaluate the emerging role of peer advocates in HIV/AIDS care and treatment and how the implementation of a multi-dimensional evaluation design utilized in non-profit health care settings impact evaluation practice and theory.
Evaluating a Community Level HIV and Substance Use Prevention Program: The Fortaleciendo la Comunidad Project
Presenter(s):
Kathryn DeRoche,  Mental Health Center of Denver,  kathryn.deroche@mhcd.org
Antonio Olmos,  Mental Health Center of Denver,  antonio.olmos@mhcd.org
Lydia Prado,  Mental Health Center of Denver,  lydia.prado@mhcd.org
Hollie Granato,  University of Colorado Denver,  hollie.granato@email.cudenver.edu
Shara Shaw,  University of Northern Colorado,  sshaw_dance@yahoo.com
Abstract: The Fortaleciendo la Comunidad project is a HIV, sexual transmitted infections, and substance abuse prevention program designed for and by the Latino community in West Denver. The project includes four programs designed for selected sub-populations in the community deem to have the highest need for prevention including: adolescents, parents, heterosexual adult females, and a community outreach program for injecting drug users and men who have sex with men. In the evaluation, we have determined the progress of the four individual prevention programs, while also evaluating the cumulative affects of the project within the community. The presentation will discuss the evaluation results, the process of increasing evaluation capacity of service organizations within the community, and the process of conducting a community-wide, culturally competent evaluation.
Innovations in HIV/AIDS Prevention and Education: Reining in the Evaluation After We Got Creative
Presenter(s):
Jennifer Camacho Catrambone,  Ruth M Rothstein CORE Center,  jcamacho@corecenter.org
Rebecca Goldberg,  Ruth M Rothstein CORE Center,  rgoldberg@corecenter.org
Jackie Boone,  Ruth M Rothstein CORE Center,  jboone@corecenter.org
Chester J Kelly,  Ruth M Rothstein CORE Center,  ckelly@corecenter.org
Peter McLoyd,  Ruth M Rothstein CORE Center,  pmcloyd@corecenter.org
Abstract: A multidisciplinary team based at a midwestern infectious disease clinic developed this innovative presentation designed to educate its audience (primarily community college students in Chicago) by shocking them into paying attention and taking HIV/AIDS prevention seriously. This approach was developed to combat “prevention fatigue,” a phenomenon in which people have heard HIV/AIDS prevention in the background so often that it seems to have become easier to ignore. This session will first describe the multifaceted and innovative approach our team took and contrast it with more traditional HIV/AIDS prevention messages. Subsequently, presenters will discuss the problems the unique presentation style created for keeping the evaluation methodologically sound, missed evaluation opportunities and their causes, as well as two years worth of evaluation data. Finally, the team will list future directions of the evaluation.

Session Title: Evaluation Apprenticeship: Teaching Evaluation by Doing
Panel Session 532 to be held in the Granite Room Section C on Friday, Nov 7, 9:15 AM to 10:45 AM
Sponsored by the Teaching of Evaluation TIG
Chair(s):
Cynthia Tananis,  University of Pittsburgh,  tananis@education.pitt.edu
Abstract: The Collaborative for Evaluation and Assessment Capacity (CEAC) at the University of Pittsburgh's School of Education provides collaborative workspace for evaluation work in education (preK-16), child advocacy and support, and related initiatives through faculty and graduate students. Graduate students come to CEAC with no or little evaluation experience and often faculty serve as content-specialist consultants rather than bringing any particular evaluation expertise. This panel will explore the ways in which apprenticeship (specifically for graduate students) can provide for their own field-based training in evaluation to augment their primary studies in other areas of education such as comparative studies, international development, educational leadership, and research methodology and simultaneously provide high-quality evaluation services for CEAC clients. The panel will consist of two faculty members (one who serves as the Director of CEAC and one who works in collaboration with CEAC) and three graduate student evaluators, one who serves as the full-time Project Manager for CEAC and two others who serve as graduate assistants. Lessons learned and dilemmas from practice will be explored from each perspective.
From the Evaluation Director Perspective
Cynthia Tananis,  University of Pittsburgh,  tananis@education.pitt.edu
As the director of the evaluation collaborative, I am ultimately responsible for contracts and grants, evaluation planning and development, hiring and supervision of staff and graduate students and all management and leadership aspects of the Collaborative. As such, I offer this perspective on the issues identified in the panel description and rationale. My role on the panel is both to manage the conversation (to assure ample coverage from each of the presenters' perspectives and the audience) and to present the issues and dilemmas of managing a large evaluation group with staff and students who have no formal evaluation training or experience.
From the Perspective of Full-time Staff: The Project Manager
Cara Ciminillo,  University of Pittsburgh,  ciminill@education.pitt.edu
As the Project Manager for CEAC and doctoral student in educational leadership (with the Director as my academic advisor), I wear numerous hats. I am the first-line supervisor for the other graduate students who work for CEAC while also being a peer student and evaluator with them on projects. I serve as a peer evaluator with my academic advisor and employer/CEAC Director. I supervise and hire undergraduate students who work as support staff for our unit. While I had no formal evaluation experience prior to working with my advisor and CEAC, I find my continuing academic pursuits and interests being influenced by my work and my work being equally influenced by my academic interests. My perspective is varied and unique among CEAC staff and students.
From the Perspective of a Doctoral Student: Graduate Assistant
Keith Trahan,  University of Pittsburgh,  kwt2@pitt.edu
I serve as a graduate assistant at CEAC and am a full-time doctoral student in social and comparative analysis in education. I am lead evaluator on a number of projects and also work as a team member on a number of large projects as well. My academic interests include history and philosophy and their interplay in how educational issues are framed and studied. I have worked with CEAC for two years now and offer a perspective of how my work with CEAC has influenced my academic work and how my academic focus has also brought certain features to the evaluation work I do individually and as a team member.
From the Perspective of a Masters Student: Graduate Assistant
Tracy Pelkowski,  University of Pittsburgh,  tlp26@pitt.edu
I have worked as a graduate assistant for CEAC for one year and am completing a masters degree in social and comparative analysis in education. I have been a K-12 teacher in an urban district for a number of years as well. As both a novice graduate student (hoping to move directly into the doctoral program at the conclusion of my masters degree) and a novice evaluator, my first year at CEAC has been all about learning. Often I come to projects that are well underway and, as such, need to catch up quickly with both an understanding of context as well as evaluation. I bring the perspective of a true novice to CEAC as well as the panel.
From the Perspective of a Collaborating Faculty Member
Junlei Li,  University of Pittsburgh,  jul@pitt.edu
I am an Assistant Professor in the School of Education and also serve as the Evaluation Director for the Office of Child Development, a department in the School. I have worked with CEAC to consider avenues of professional collaboration and development for our staff to share expertise and learning. Additionally, we have collaborated on a grant proposal and a number of shared projects across our units. I offer the perspective of a faculty colleague as well as a manager of evaluation services.

Session Title: Evaluation: What Stakeholders Don't Know Can Hurt Evaluation
Demonstration Session 533 to be held in the Quartz Room Section A on Friday, Nov 7, 9:15 AM to 10:45 AM
Sponsored by the Collaborative, Participatory & Empowerment Evaluation TIG
Presenter(s):
L Shon Bunkley,  Community Research Partners,  sbunkley@communityresearchpartners.org
Gary Timko,  Community Research Partners,  gtimko@communityresearchpartners.org
Abstract: This 90-minute demonstration will provide attendees with practical guidelines for engaging a broad range of stakeholders in the evaluation process, from conceptualizing the evaluation plan to making use of evaluation findings. Session attendees will be given specific and tested examples of: - Gauging and broadening stakeholders' understanding of the purpose, value, and process of evaluation - Actively involving stakeholders in the identification of conclusions and recommendations - Increasing the readability of reports

Roundtable: Portfolio Planning and Evaluation: Advancing Knowledge and Practice
Roundtable Presentation 534 to be held in the Quartz Room Section B on Friday, Nov 7, 9:15 AM to 10:45 AM
Sponsored by the Organizational Learning and Evaluation Capacity Building TIG
Presenter(s):
Jennifer Martineau,  Center for Creative Leadership,  martineauj@ccl.org
Erika Fulmer,  RTI International,  fulmer@rti.org
Djimo D Adoum,  United States Department of Agriculture,  dadoum@csrees.usda.gov
Abstract: A quick search of the online resource Google Scholar reveals over 1,700 references to 'portfolio evaluation', many of which refer to work that was done in the teaching and financial investment arenas. But rarely do the references point to organizations examining the strength of their portfolios in meeting the needs of their clients and constituents. Members of the evaluation field have begun to bring the best of the evaluation practice to the assessment of a portfolio - a set of programs, products, and services - an effort that we believe will add significant value to organizations in ensuring they are able to achieve their goals and meet their missions. This session will first present three examples of portfolio evaluation and then actively engage the audience in examining these cases in an effort to make further improvements to the practice.

Session Title: Reviving the Evaluability Assessment (EA) Method: EAs of Interventions to Prevent Childhood Obesity
Multipaper Session 535 to be held in Room 102 in the Convention Center on Friday, Nov 7, 9:15 AM to 10:45 AM
Sponsored by the Health Evaluation TIG
Chair(s):
David Cotton,  Macro International Inc,  david.a.cotton@macrointernational.com
Discussant(s):
Debra Rog,  Rockville Institute,  debrarog@rockvilleinstitute.org
Laura Kettel Khan,  Centers for Disease Control and Prevention,  ldk7@cdc.gov)
Nicola Dawkins,  Macro International Inc,  nicola.u.dawkins@macrointernational.com
Abstract: In 2007, the Robert Wood Johnson Foundation funded the Centers for Disease Control and Prevention (CDC) Foundation, which provided funds to the CDC and Macro International to conduct 25 evaluability assessments (EAs) on selected initiatives to improve the eating habits and physical activity levels of children. EAs provide a systematic way to determine if an initiative is ready for rigorous evaluation. Developed by Joseph Wholey and colleagues, EAs received significant attention 30 years ago but later fell into disuse (Rog 1985; Trevisan 2007). Today the method is seeing renewed interest, particularly for quickly and cheaply gauging whether an initiative is plausible and feasible, and whether investment of evaluation resources is appropriate. This session will share the method for selecting initiatives and implementing the EA process, process for identifying and training 33 professionals to conduct EAs, and lessons and products of the process. Discussants will reflect on the project to date.
Evaluability Assessments of Interventions to Prevent Childhood Obesity: What, Why, and How
Holly Wethington,  Centers for Disease Control and Prevention,  eyk5@cdc.gov
Nicola Dawkins,  Macro International Inc,  nicola.u.dawkins@macrointernational.com
Seraphine Pitt-Barnes,  Centers for Disease Control and Prevention,  spe6@cdc.gov
Diane Dunet,  Centers for Disease Control and Prevention,  dol0@cdc.gov
David Cotton,  Macro International Inc,  david.a.cotton@macrointernational.com
Leah Robin,  Centers for Disease Control and Prevention,  ler7@cdc.gov
Jo Anne Grunbaum,  Centers for Disease Control and Prevention,  jpg9@cdc.gov
Laura Leviton,  Robert Wood Johnson Foundation,  llevito@rwjf.org
Laura Kettel Khan,  Centers for Disease Control and Prevention,  ldk7@cdc.gov)
Evaluability assessments (EAs) provide a systematic way to determine if an intervention is ready for rigorous evaluation and identify refinements that may be needed prior to formal evaluation. This method gauges an initiative's plausibility and feasibility, stakeholder agreement on objectives, data systems and capacity to collect data, and consistency of key elements with program theory. In 2007, the Early Assessment of Programs and Policies to Prevent Childhood Obesity project conducted 25 EAs on programs and policies that address the physical, economic, or social environment to prevent childhood obesity. Three focus areas were selected: programs or policies in the after school/daycare settings; programs or policies to increase access to healthier foods in urban, low-income communities; and school district local wellness policies. This presentation will describe the EA methodology and will demonstrate the value of using a systematic process to conduct rapid assessments to identify those most suited for full-scale evaluation.
Teaching Evaluability Assessment Methodology
Thearis Osuji,  Macro International Inc,  thearis.a.osuji@macrointernational.com
Nicola Dawkins,  Macro International Inc,  nicola.u.dawkins@macrointernational.com
Holly Wethington,  Centers for Disease Control and Prevention,  eyk5@cdc.gov
Seraphine Pitt-Barnes,  Centers for Disease Control and Prevention,  spe6@cdc.gov
Diane Dunet,  Centers for Disease Control and Prevention,  dol0@cdc.gov
David Cotton,  Macro International Inc,  david.a.cotton@macrointernational.com
Leah Robin,  Centers for Disease Control and Prevention,  ler7@cdc.gov
Jo Anne Grunbaum,  Centers for Disease Control and Prevention,  jpg9@cdc.gov
Laura Leviton,  Robert Wood Johnson Foundation,  llevito@rwjf.org
Laura Kettel Khan,  Centers for Disease Control and Prevention,  ldk7@cdc.gov)
In 2007, the Early Assessment of Programs and Policies to Prevent Childhood Obesity project conducted evaluability assessments (EAs) of programs and policies with the potential to improve the eating habits and physical activity levels of children. One goal was to develop a network of evaluation professionals with the skills to conduct evaluability assessments. This presentation will focus on the process of identifying and training 33 professionals in the EA methodology, results from a post EA survey of the evaluators, and insights from the process of coordinating a large number of evaluation professionals in 3 months. The training process highlighted: the benefit of a diverse team approach; the distinction between EA and full evaluation; the critical nature of strong qualitative interview skills; the significant time commitment for evaluators; and the need for centralized oversight for methodologic integrity among evaluators.
Lessons Learned: Practical Tips for Conducting Evaluability Assessments
Seraphine Pitt-Barnes,  Centers for Disease Control and Prevention,  spe6@cdc.gov
Holly Wethington,  Centers for Disease Control and Prevention,  eyk5@cdc.gov
Karen Cheung,  Macro International Inc,  karen.cheung@macrointernational.com
Nicola Dawkins,  Macro International Inc,  nicola.u.dawkins@macrointernational.com
Diane Dunet,  Centers for Disease Control and Prevention,  dol0@cdc.gov
David Cotton,  Macro International Inc,  david.a.cotton@macrointernational.com
Leah Robin,  Centers for Disease Control and Prevention,  ler7@cdc.gov
Jo Anne Grunbaum,  Centers for Disease Control and Prevention,  jpg9@cdc.gov
Laura Leviton,  Robert Wood Johnson Foundation,  llevito@rwjf.org
Laura Kettel Khan,  Centers for Disease Control and Prevention,  ldk7@cdc.gov)
The Early Assessment of Programs and Polices to Prevent Childhood Obesity project offers key lessons learned from quickly conducting large numbers (25-30 annually) of evaluability assessments (EAs). These lessons are offered to inform others interested in or using this methodology. Among the lessons learned are: (1) clear communication is critical with evaluators and sites about roles, expectations, and the EA site visit process; (2) site documents often do not represent the program accurately; (3) involving the site in the selection of stakeholder interviewees is key to ensure comprehensive perspectives; (4) logic models are an appropriate and enlightening method for examining program goals and activities; (5) standard reporting formats provide clarity and consistency by evaluators; and (6) brief technical assistance calls and formal site reports are highly valued, low cost incentives for participation and useful tools for program/policy improvement.
Impact of the Evaluability Assessment Process: Early Returns
Laura Leviton,  Robert Wood Johnson Foundation,  llevito@rwjf.org
Holly Wethington,  Centers for Disease Control and Prevention,  eyk5@cdc.gov
Seraphine Pitt-Barnes,  Centers for Disease Control and Prevention,  spe6@cdc.gov
Diane Dunet,  Centers for Disease Control and Prevention,  dol0@cdc.gov
Leah Robin,  Centers for Disease Control and Prevention,  ler7@cdc.gov
Jo Anne Grunbaum,  Centers for Disease Control and Prevention,  jpg9@cdc.gov
David Cotton,  Macro International Inc,  david.a.cotton@macrointernational.com
Nicola Dawkins,  Macro International Inc,  nicola.u.dawkins@macrointernational.com
Laura Kettel Khan,  Centers for Disease Control and Prevention,  ldk7@cdc.gov)
The Early Assessment of Programs and Polices to Prevent Childhood Obesity project had four notable consequences in its first year of operation. (1) The process revealed more than ten policies and programs worthy of more rigorous evaluation. (2) Three such policies and programs rose to the top of the list of priorities: a day care policy for prevention of childhood obesity, a program designed to bring supermarkets back to underserved inner city environments, and an incentive program for poor families to buy fruit and vegetables at farmers' markets. (3) Most sites visited found the feedback helpful for program improvement. (4) Based on site visitor reports and expert panel discussions, themes and research issues emerged for all policy and program types that were assessed. This process might easily be emulated for other policies and programs to find out "what works."

Session Title: Evaluation Policy and Training Evaluation in Child Welfare
Panel Session 536 to be held in Room 104 in the Convention Center on Friday, Nov 7, 9:15 AM to 10:45 AM
Sponsored by the Human Services Evaluation TIG
Chair(s):
Henry Ilian,  New York City Children's Services,  henry.ilian@dfa.state.ny.us
Abstract: This panel will address evaluation policy's influence on evaluation practice within child welfare systems. It posits that the absence of evaluation policy creates structural barriers to conducting training evaluation. In the absence of policy, training and training evaluation compete with all other organizational priorities for resources and managerial attention. This competition is governed by such factors as organizational structure, type of leadership, leadership turnover, organizational culture, and caseload size. Thereby the ability of evaluators to get data and to show how training is or is not influencing practice is frequently impeded. This conclusion is based on two sources of evidence: (1) an overview of the training evaluation literature; (2) lessons drawn from training evaluation within child welfare agencies in three localities, the state of Kentucky and the cities of Los Angeles and New York.
Effects of Leadership Change on Evaluation Efforts in Kentucky
Anita Barbee,  University of Louisville,  anita.barbee@louisville.edu
Dr. Barbee was selected to participate in this panel because she has worked with one state child welfare agency conducting training evaluation over the past 16 years under four different governors, three different configurations of the agency and eight different commissioners. These changes have had an impact on organizational structure and culture, leadership style and expertise as well as case load size. All of these variables have affected access to decision makers, whether or not research and evaluation is valued, the evaluator's ability to obtain data from stressed and overworked employees, the quality of that data, the usefulness of the data to the agency, and the willingness of the agency to make changes based on the reported results. One solution to the problems would be the establishment of a formal evaluation policy so that such changes would not undermine the evaluation enterprise. This and other possible solutions will be discussed.
Evaluation Policy and Child Welfare Training Evaluation: Lessons from Los Angeles
Todd Franke,  University of California Los Angeles,  tfranke@ucla.edu
Dr. Franke was selected for this panel as a result of his extensive work at the intersection of evaluation and child welfare issues. He has been involved with agencies that serve thousands of families representing unique geographic and cultural communities in California, particularly southern California counties. He is the PI for the evaluation of the First 5 Los Angeles funded Partnership for Families Initiative. Partnerships for Families is a five year secondary prevention initiative that is designed to prevent child maltreatment by creating opportunities for families, communities, and governmental entities to meaningfully partner in the prevention of child abuse and neglect throughout Los Angeles County. It is designed to assist at-risk families including pregnant and parenting teens. Dr. Franke also is the PI for the Small County Initiative for the past six years, which is designed to systematically evaluate the State of California's efforts to build and enhance child abuse and neglect prevention efforts in 11 rural counties in northern California. This effort required extensive collaboration with multiple county agencies and employed a mixed model approach to data collection. For the past several years he has been working with the Department of Children and Family Services (DCFS-Los Angeles) in evaluating their pre-service and in-service training for Child Protective Service (CPS) workers. In this capacity he has had the opportunity to work collaboratively with the Training Director and supervisors.
Training Evaluation and Operational Imperatives at New York City's Children's Services
Henry Ilian,  New York City Children's Services,  henry.ilian@dfa.state.ny.us
Dr. Ilian was selected to participate on this panel because of his work over twenty years in evaluating training at Children's Services, the New York City child welfare agency. He is strongly interested in the organizational dynamics--including change and persistence in organizational culture and the role of leadership--that permit or inhibit the evaluation of training. He will address lessons learned from efforts undertaken between 2002 and 2005 to conduct a multidimensional evaluation of a large-scale supervisory training program. The evaluation ultimately had to be abandoned, but the experience proved to be rich in insights into the conditions under which training evaluations are and are not successfully conducted within large child welfare organizations. Implications point to an evaluation policy that mediates between the competing imperatives of conducting an agency's ongoing work and ensuring the success of efforts to instill a mode of practice that meets nationally recognized standards.

Session Title: Evaluating Programs in Times of Complexity and Chaos
Panel Session 537 to be held in Room 106 in the Convention Center on Friday, Nov 7, 9:15 AM to 10:45 AM
Sponsored by the Systems in Evaluation TIG
Chair(s):
Daniel Folkman,  University of Wisconsin Milwaukee,  folkman@uwm.edu
Abstract: How do we evaluate programs that operate in complex, rapidly changing, and indeterminate environments? Join our panel presentation to learn how this one evaluation strategy incorporates principles from complexity theory along with more traditional approaches. The utility of our evaluation strategy is demonstrated through a case study of the Healthier Wisconsin Partnership Program. This program brings together a group of community partners including an urban middle school, a school based health clinic, a community based agency, and a family leadership development program. The goal is to address the academic, health, safety and wellness issues facing low-income, high-risk, middle school girls. The challenge is that each partner operates within highly complex environments where implementing strategies are continuously changing and outcomes emerge that were not anticipated by the program model. Learn how the evaluation design documents the emerging outcomes and facilitates stakeholder decision making while operating in complexity and near chaos.
Evaluating Programs That Operate in Complex and Chaotic Environments
Daniel Folkman,  University of Wisconsin Milwaukee,  folkman@uwm.edu
My presentation is an overview of the evaluation design and its implementation. A central component involves creating a learning community among the partners, which meets monthly to review the progress and challenges encountered as the program unfolds. One tool that is used to document the implementation process is called 'Headlines.' Here each partner records the major accomplishments, challenges, and critical learning that has occurred over the previous month. These headlines provide the basis for constructing a narrative on how the program is evolving as it is happening. A second tool involves creating an outcome database called 'good stuff' This is literally a database of unique accomplishments that the girls are producing as they participate in the program but could never be anticipated. This tool is highly sensitive to the emergent patterns and outcomes that the girls are producing. These introductory comments set the stage for the next two presentations.
Creating Partnership in the Face of Complexity
Devarati Syam,  University of Wisconsin Milwaukee,  devasyam@uwm.edu
Creating a community-based partnership model for addressing the health, safety and wellness issues facing high-risk, African American middle school girls is a central goal of this project. Therefore, the collaborative endeavor requires multiple partners to coordinate their services. The partners belong to different complex systems (both institutional and situational) and are attempting to develop a program by managing different expectations, values, roles and responsibilities that are part of their institutional commitments. My presentation will report the progress we have made in this partnership building attempt and how the partners have experienced their participation in this process so far - their learning as well as their struggles and challenges. The partners meet once a month to problem solve around the different issues that crop up in the implementation. This presentation will provide the highlights from these conversations and draw implications of evaluating the partnership building project in this context.
Facilitating Teen Support Groups Where Complexity is the Norm and Chaos is Just Around The Corner
Yvette Dobson,  PEARLS for Teen Girls Inc,  yvette@pearlsforteengirls.com
This presentation is focused on implementing the PEARLS program. The school serves some of the most 'high-risk' youth in the district making the environment highly charged and ready to erupt at any moment. Therefore, my programming context is characterized by a high degree of uncertainty and unpredictability. My primary responsibility is to work closely with the teen girls and develop a relationship with them. But it also entails working with teachers, nursing staff, parents, and community residents. The discussion will include critical incidents that demonstrate the complexity of working with teen girls within a middle school setting. The incidents will also show the connections with other partners including school nurses, parents, and community influences. The school principal captures my challenge nicely when she described by role as being like the cartoon character Plastic Man. To be successful I must bend and shape to the changing conditions within my environment.

Session Title: Evaluation Strategies and Assessment in the Arts
Multipaper Session 538 to be held in Room 108 in the Convention Center on Friday, Nov 7, 9:15 AM to 10:45 AM
Sponsored by the Evaluating the Arts and Culture TIG
Chair(s):
Ching Ching Yap,  University of South Carolina,  ccyap@gwm.sc.edu
Discussant(s):
Rob Horowitz,  Columbia University,  artsresearch@aol.com
Developing a Theory-Based Evaluation Strategy Within a University-Based Arts Education Program
Presenter(s):
Mark Hansen,  University of California Los Angeles,  hansen.mark@gmail.com
Noelle Rivera,  University of California Los Angeles,  novari320@hotmail.com
Abstract: ArtsBridge at the UCLA School for the Arts and Architecture provides high-quality arts education experiences for youth and teachers in local schools who would otherwise have little or no access to such opportunities. Recently, the Social Research Methodology Evaluation Group of the UCLA Graduate School of Education and Information Studies began consultation with ArtsBridge. Our purpose was to assess the program’s current approach to evaluation and to identify opportunities for enhancing these efforts. By engaging a broad base of stakeholders, the assessment process clarified objectives related to multiple target audiences. Logic models were developed to illustrate hypothesized links between these objectives and program activities. Using these models and input from stakeholders concerning areas of inquiry that should be prioritized, draft questionnaires were created and pilot-tested. This paper will describe the consultation process, share preliminary findings and demonstrate the appropriateness of theory-based approaches to evaluation within the field of arts education.
Exploring the Nonformal Education Frontier: Using Cluster Analysis to Investigate Children’s Museum Programming for Special Learner Groups
Presenter(s):
Deborah Chapin,  University at Albany - State University of New York,  debchapin50@juno.com
Dianna Newman,  University at Albany - State University of New York,  dnewman@uamail.albany.edu
Abstract: This study presents the findings of a study on the perceptions of eighty education directors of children’s museums responding to a paper-pencil survey examining patterns of planning, designing, and implementing nonformal education programs for minority youth and children with disabilities. K-means cluster analysis, applied as an exploratory tool, demonstrated that three major children’s museum types exist, modest-sized urban “dedicated” museums, small suburban “elite” museums, and large urban “generic” museums. Dedicated and elite museums planned their curriculum using State Standards, teachers, and community cultural leaders more than generic museums. Dedicated and generic museums followed Universal Design strategies when serving children with disabilities; however, only dedicated museums used elements of culturally relevant education with minority youth. “Dedicated” children’s museums charged the least per admission, devoting fewer resources to out-reach to special learner groups and offering less variety in their current programming, but had staff holding more specialized degrees creating specific science content.
Examining Arts Educators Use of Arts Assessment Results
Presenter(s):
Tara Pearsall,  University of South Carolina,  tmcpearsall@yahoo.com
Kate Halladay,  University of South Carolina,  kate.halladay.art@gmail.com
Ashlee Lewis,  University of South Carolina,  scaap@gwm.sc.edu
Ching Ching Yap,  University of South Carolina,  ccyap@gwm.sc.edu
Abstract: In 1999, the South Carolina Department of Education, University of South Carolina, and South Carolina arts agencies collaborated to develop a comprehensive program that encourage standards-based arts instruction and measure standards-based arts achievement. To achieve those goals, collaborators created professional development programs for arts educators and developed standardized arts assessments in all four arts areas. The arts assessments provide school-level information regarding student arts achievement and the professional development programs enable arts educators to use the results to inform their instruction. Although numerous arts educators involved in this comprehensive program have informally communicated accounted their experiences with the program, including their use of the assessment results, no formal attempts has been made to collect that information. The purpose of this study is to collect information regarding teachers’ experiences with this comprehensive program.

Session Title: Modeling the Logic and Measuring the Performance of Government Programs
Multipaper Session 539 to be held in Room 110 in the Convention Center on Friday, Nov 7, 9:15 AM to 10:45 AM
Sponsored by the Government Evaluation TIG
Chair(s):
David J Bernstein,  Westat,  davidbernstein@westat.com
Discussant(s):
David J Bernstein,  Westat,  davidbernstein@westat.com
Assisting the Local Program Level with Performance Measurement
Presenter(s):
Leanne Kallemeyn,  Loyola University Chicago,  lkallemeyn@luc.edu
Abstract: A critical evaluation policy, the Government Performance and Results Act of 1993, has contributed to the growth of performance measurement. While some evaluators have criticized performance measurement, others have contributed to its development and incorporated it into their evaluations. The purpose of this paper is to further contribute to this conversation about how evaluators can engage performance measurement. I use a case example of how I, as an evaluator, assisted a local Head Start program in using mandated school readiness assessments. Based on the literature and case, I identified five areas for an evaluator to work within and challenge performance measurement: What is assessed, how it is assessed, how results are interpreted and used, what is the purpose, and what relationships are facilitated. I also provided example activities that I completed with the local program to illustrate how evaluators can assist local programs with performance measurement.
Practical Limitations to Gathering Data in Support of GPRA and PART
Presenter(s):
William Scarbrough,  Macro International Inc,  william.h.scarbrough.iii@macrointernational.com
Herbert Baum,  Macro International Inc,  herbert.m.baum@macrointernational.com
Renee Bradley,  United States Department of Education,  renee.bradley@ed.gov
Andrew Gluck,  Macro International Inc,  andrew.gluck@macrointernational.com
Abstract: Providing the Office of Management and Budget with data to support a program’s Program Rating Assessment Tool (PART) represents an unfunded mandate for the program. As such, the program often can only gather data on a “shoestring” budget. Macro International has worked for 4 years with the Research to Practices Program, Office of Special Education Programs, US Department of Education, to develop program measures and collect data in support of these. In this paper we briefly describe how we have used expert panels to review material, and the challenges associated with this approach.
Partnership Contribution to Program Outcomes: Logic Models as Tools for Developing Measures for Partnerships
Presenter(s):
Valerie Williams,  RAND Corporation,  valerie@rand.org
Lauren Honess-Morreale,  RAND Corporation,  laurenhm@rand.org
Abstract: Partnerships are a common strategy for many programs. Government programs, in particular, rely on partnerships because they offer the potential to deliver more effective programs as well as increased efficiency in producing and delivering outputs. Despite the ubiquitous nature of partnerships, there is very little information on developing measures to assess the value of partnerships, particularly as it relates to the contribution of partnerships to achieving program outcomes. In this paper, the authors describe the use of a logic model template as a means for developing measures that can be applied to partnerships. The logic model provides a framework for describing partnership inputs, activities, outputs and outcomes, and as such, serves as a basis for developing measures to assess 1) specific elements of partnerships; 2) the cost efficiency and effectiveness of the partnership; and 3) the extent to which “partnership,” rather than simply two entities working together, contribute to program outcomes.

Session Title: Alternative Approaches to Strategic Management and Ex Ante Assessment of Portfolios of Research Programs
Multipaper Session 540 to be held in Room 112 in the Convention Center on Friday, Nov 7, 9:15 AM to 10:45 AM
Sponsored by the Research, Technology, and Development Evaluation TIG
Chair(s):
Cheryl J Oros,  United States Department of Veterans Affairs,  cheryl.oros@va.gov
Evaluation of Tekes Funding for Research Institutes and Universities
Presenter(s):
Jari Hyvarinen,  Tekes,  jari.hyvarinen@tekes.fi
Abstract: A focus of Tekes (Finnish Funding Agency for Innovation and Technology) funding for research institutes and universities is the renewing and collaborative technology, service and business oriented research projects. This paper presents recent evaluation results of the Tekes projects for research institutes and universities that were carried out in the early 2000s. Tekes database includes evaluation results of around 2500 Tekes funded projects and 100 research organizations. First, it shows a survey of recent findings of those projects using the four-step model of evaluation and impact assessment implemented in Tekes. This model concentrates on additionality effects including R&D investments, results, direct effects and impact on national economy and society. Second, it emphasizes several improvements how by using ex post evaluation and impact analysis it can be found more precise and accurate evaluation results, and how to make more precise ex ante decisions when planning new funding for research institutes and universities.
Cost-Benefit Analysis for Research and Development Program Evaluation
Presenter(s):
Jiyoung Park,  Korea Institute of Science and Technology Evaluation and Planning,  jypark@kistep.re.kr
Seung Kyu Yi,  Korea Institute of Science & Technology Evaluation and Planning,  skyist@kistep.re.kr
Abstract: Feasibility analysis system for national R&D programs has been legislated since Dec. 2006 in Korea. It is performed to decide governmental budget distribution. Technical, policy, and economic/impact analysis is carried out to decide approval or rejection of the newly proposed R&D programs. The economic evaluation is performed to measure economic benefits and appropriateness of the program. Cost-benefit analysis is often conducted as a part of ex ante program evaluation, and should provide information on the full costs and expected benefits. However, identifying all of the benefits and costs of the program prior to their initiation is difficult for R&D programs. The assessment should include direct and indirect benefits and costs and has to always contain a “do nothing” alternatives to provide a baseline. Tangible or intangible benefits of R&D programs should be measured through different methodologies. In this paper, case studies on B/C analysis for the R&D programs are introduced.
Content and Values in the Strategic Management of Research and Development Portfolios
Presenter(s):
Juan Rogers,  Georgia Institute of Technology,  jdrogers@gatech.edu
Abstract: Most R&D programs in government agencies consist in portfolios of projects that have become more collaborative and multi- or interdisciplinary. Each field represented in a multidisciplinary or interdisciplinary arrangement has different field specific cognitive values that are often difficult to negotiate and reconcile within collaborative teams. There are also new networks of relevant stakeholders that implicitly or explicitly adjudicate whether desirable or undesirable outcomes have occurred. This paper will propose a theoretical framework for the evaluation of such portfolios that includes the empirical elucidation of the content-value structures for strategic management. It will apply the framework to the case of government agencies in the US and a South American country to exemplify the expected results and the how the approach can be used.

Session Title: Process Use-Methods and Measures
Multipaper Session 542 to be held in Room 105 in the Convention Center on Friday, Nov 7, 9:15 AM to 10:45 AM
Sponsored by the Evaluation Use TIG
Chair(s):
Sandra Ortega,  National Data Evaluation Center,  ortegas@ndec.us
Promoting the Use of Evaluation Information: Early Reading First in Chicago Charter Schools
Presenter(s):
Tania Rempert,  University of Illinois Urbana-Champaign,  trempert@uiuc.edu
Lizanne DeStefano,  University of Illinois Urbana-Champaign,  destefan@ad.uiuc.edu
William Teale,  University of Illinois Chicago,  wteale@uic.edu
Abstract: This presentation is organized via alignment of evaluative questions, measurement tools, and reporting strategies. Comments from program staff highlight the methods, strategies, and tools that have been most useful to them: observations, teacher portfolios, assessment data, and PD questionnaire. This evaluation has been designed to be longitudinal in nature in order to capture changes that occur over time and across settings. Formatively, the evaluation routinely monitors progress toward specific objectives, assesses the quality of implementation, and gauges the short-term impacts of the project. This information is reported to program staff regularly at project management meetings so that it could be used to guide program development and improvement. Summatively, the evaluation employs a quasi-experimental pre-test, post-test, non-equivalent comparison group design to document the impact of the project on teaching, learning and family. Finally, the evaluation is designed to provide information on effective strategies for replication of the program in other settings.
Initial Results from a Planned Approach to Process Use: Using Evaluation to Influence Public School Policy and the Classroom Practice of Formative Assessment
Presenter(s):
Debra Heath,  Albuquerque Public Schools,  heath_d@aps.edu
Nancy Carrillo,  Albuquerque Public Schools,  carrillo_n@aps.edu
River Dunavin,  Albuquerque Public Schools,  dunavin_r@aps.edu
Abstract: This paper describes how evaluators in an urban school district employed strategies to promote evaluation use, foster organizational alignment and influence evaluation policy. Findings from an evaluation of district-wide formative assessment practices suggest that teachers under-utilized formative assessment techniques and that district support systems offered inconsistent formative assessment definitions, strategies and resources. We wanted to do more than provide the standard written report and set of presentations; we wanted to help optimize the district’s support and practice of proven formative assessment strategies. We used a collaborative evaluation approach with the intention of solidifying commitment to formative assessment, increasing communication between stakeholders, creating shared understandings regarding formative assessment, and fostering collaborations between departments. We outline and critique our methods for achieving process use, and describe the techniques and tools we used to measure process use, including group processes, questionnaires, observations and interviews. Finally we share lessons learned.
Evaluation as Change Agent: The Precarious See-Saw Between Helping and Judging and How it Affects Process Use
Presenter(s):
Gale Mentzer,  University of Toledo,  gmentze@utnet.utoledo.edu
Abstract: This paper tracks the precarious path of program improvement that began with a required program evaluation for a federally funded project. It illustrates the complexity of working with “reluctant” stakeholders through the example of the evolution of a summer program designed to introduce secondary school students to teaching. A mixed methods evaluation was used to measure stated outcomes and to discover whether other, unanticipated factors were affecting program goal attainment. Application of the evaluation model to the program revealed not only that students’ misconceptions about teaching were being cleared up but also discovered a major unanticipated factor—the program instructors (university faculty) held misconceptions as to the goals of the program. This paper shows how, over the course of four years, group process techniques were used to engage the stakeholders in the evaluation in order to implement necessary programmatic changes as indicated by program evaluation results.

Session Title: Communities in Schools (CIS) Evaluations
Multipaper Session 543 to be held in Room 107 in the Convention Center on Friday, Nov 7, 9:15 AM to 10:45 AM
Sponsored by the Cluster, Multi-site and Multi-level Evaluation TIG
Chair(s):
Pamela Bishop,  Oak Ridge Institute for Science and Education,  pbaird@utk.edu
Determining the Impact of CIS Programs: Student-Level Randomized Controlled Trials
Presenter(s):
Clawson Heather,  ICF International,  hclawson@icfi.com
Allan Porowski,  ICF International,  aporowski@icfi.com
Felix Fernandez,  ICF International,  ffernandez@icfi.com
Christine Leicht,  ICF International,  cleicht@icfi.com
Abstract: Communities In Schools, Inc. (CIS) is a nationwide initiative to connect community resources with schools to help at-risk students successfully learn, stay in school, and prepare for life. CIS is currently in the midst of a comprehensive, rigorous three-year national evaluation, culminating in a two-year multi-site randomized controlled trial (RCT) to ascertain program effectiveness. In this presentation, we will draw from our experience working with Austin, TX and Jacksonville, FL public schools, present our overall study design, and the process involved in conducting a student-level RCT. Preliminary results from Year One will also be discussed.
The Communities in Schools Inc (CIS) National Evaluation: Conducting Case Studies of Large, Diverse Community-Based Organizations
Presenter(s):
Sarah Decker,  ICF International,  sdecker@icfi.com
Kelle Basta,  ICF International,  kbasta@icfi.com
Susan Siegel,  Communities In Schools Inc,  ssiegel@cisnet.org
Abstract: Communities In Schools, Inc. (CIS) is a nationwide initiative to connect community resources with schools to help students in need. As part of the national student- and school-level evaluation, we have conducted eight case studies in an effort to gain an in-depth understanding of the processes that contribute to successful CIS affiliate- and site-level outcomes. This presentation will provide a two-fold description of case study implementation within diverse, community-based organizations. First, we will examine the methodology behind our case study of CIS sites. We will include a discussion of challenges encountered while conducting the studies in elementary, middle, and high schools located in rural, suburban, and urban areas around the country, and the solutions we developed to circumvent these issues. Second, we will use this presentation as an opportunity to present our findings from the case studies, and will discuss their value-added to the overall national evaluation framework.
Findings from the Communities In Schools (CIS) School-Level Quasi-Experimental Study
Presenter(s):
Allan Porowski,  ICF International,  aporowski@icfi.com
Julie Gdula,  ICF International,  jgdula@icfi.com
Susan Siegel,  Communities In Schools Inc,  ssiegel@cisnet.org
Abstract: Communities in Schools, Inc. (CIS) is a nationwide initiative to connect community resources with schools to help at-risk students successfully learn, stay in school, and prepare for life. As part of a five-year comprehensive national evaluation of CIS, a school-level quasi-experimental study was conducted to determine whether CIS sites effected school-level change on a range of outcomes, including dropout rates, graduation rates, attendance, and academic performance. A brief description of our matching procedures, which included a technique called propensity score matching, will be included, along with a general overview of how this component fits into our comprehensive national evaluation plan. The final results from this study will also be presented, which covers a sample of 604 CIS sites and 604 matched comparison sites. The presentation will conclude with lessons learned about the effectiveness of community-based integrated student supports (CBISS) in general, and the CIS program in particular.
Typology Results from the National Evaluation of Communities In Schools Inc (CIS)
Presenter(s):
Jing Sun,  ICF International,  jsun@icfi.com
Katerina Passa,  ICF International,  kpassa@icfi.com
Allan Porowski,  ICF International,  aporowski@icfi.com
Susan Siegel,  Communities In Schools Inc,  ssiegel@cisnet.org
Abstract: Communities In Schools, Inc. (CIS) is a nationwide initiative to connect community resources with schools to help at-risk students successfully learn, stay in school, and prepare for life. To capture the diversity of the CIS network as part of our multi-component, national evaluation of the program, threshold analysis was employed to develop a typology, which identifies “high implementers” and “partial implementers” according to an ideal implementation model from the CIS Total Quality System (TQS). By comparing program outcomes (including dropout indicators and academic achievements) across typologies, we found that CIS “high implementers” outperformed their similar non-CIS comparison schools in most outcomes, while “low implementers” did not. We used this methodology to gain a deep understanding of how and why specific program models work. In this presentation, we will describe our methodology and results, as well as the processes that can be linked to positive outcomes. Implications for CIS will also be discussed.

Session Title: Strengthening Effective Implementation of Evidence-Based Interventions through Innovative Evaluation Tools and Approaches
Demonstration Session 544 to be held in Room 109 in the Convention Center on Friday, Nov 7, 9:15 AM to 10:45 AM
Sponsored by the Organizational Learning and Evaluation Capacity Building TIG
Presenter(s):
Aisha Gilliam,  Centers for Disease Control and Prevention,  agilliam@cdc.gov
Winifred King,  Centers for Disease Control and Prevention,  wking@cdc.gov
Jeanette Nu'Man,  Macro International Inc,  jeanette.h.numansheppard@orcmacro.com]
Abstract: It is necessary to provide capacity building assistance to help community-based organizations enhance and sustain their capacity to monitor and evaluate behavioral interventions. In this workshop, participants will be able to: - Understand the use of evaluation field guides as a tool to build evaluation capacity - Map behavioral risk factors to inform logic models used for evaluation planning - Learn the constructs and tools to implement evaluation of an evidence-based intervention We will provide an overview of an evaluation field guide and its use in building evaluation capacity; and illustrate the process for mapping risk behaviors and developing a logic model that can be used for planning, implementation and evaluation. We will then share an example of how these constructs are used to implement evaluation in one of CDC's evidenced- based interventions. This demonstration workshop will include a structured power-point presentation, demonstration of tools and group discussion.

Session Title: The Challenges of Conducting Experimental and Quasi-Experimental Small-Budget Evaluations of Teacher Professional Development Programs: The Case of Teaching American History Grants
Multipaper Session 545 to be held in Room 111 in the Convention Center on Friday, Nov 7, 9:15 AM to 10:45 AM
Sponsored by the Pre-K - 12 Educational Evaluation TIG
Chair(s):
Sally Bond,  The Program Evaluation Group LLC,  usbond@mindspring.com
Abstract: Experimental and quasi-experimental designs are becoming customary in RFPs from federal agencies including the U.S. Department of Education, National Science Foundation, and others. This session will explore the application of those designs in small-budget evaluations. Presenters will draw upon their years of experience using such designs in Teaching American History (TAH) grant project evaluations to illustrate key features of design and implementation. Of particular emphasis will be negotiating program and evaluation design with program staff, matching assessment to content (including design and validation of locally developed instrumentation), and challenges of implementing experimental/quasi-experimental evaluations (including factors necessitating changes in evaluation design and strategies for recruiting control or comparison groups). Presenters will discuss their own experiences and will engage attendees in discussing lessons learned in this and other evaluation contexts.
Negotiating Project Design to Permit Experimental and Quasi-Experimental Evaluation Designs
Sally Bond,  The Program Evaluation Group LLC,  usbond@mindspring.com
Negotiating program design elements to accommodate a robust experimental or quasi-experimental design is often overlooked when developing proposals or designing programs. In such cases where experimental or quasi-experimental evaluation designs are desirable and/or required, it is the responsibility of the evaluator to collaborate with program staff to (1) educate them about the demands of the desired evaluation design and (2) design a program that is consistent with the aims of the project and the needs of the evaluation design. The presenter will illustrate the iterative process of program and evaluation design in a TAH program, including selection of the linchpin assessment of teacher learning which provided the best fit with the resulting design and content to be conveyed.
The Critical Role of Assessment in Experimental and Quasi-Experimental Designs
Michael Herrick,  Herrick Research LLC,  herrickresearch@aol.com
Teaching American History grants are required to be evaluated with experimental or quasi-experimental designs. The expected outcomes of these projects are often expressed as gains in student achievement. This session will explore two different TAH evaluations, each using different types of assessments for measuring student achievement. One project used nationally validated NAEP items, while the other used locally developed and locally validated test items. The pre to post test gains in the experimental group using NAEP items were not statistically different than those of the control group. For the project using locally development and validated items, the gains of the experimental group over the control group were statistically significant at the .01 level. The primary conclusion was that the NAEP items were not strongly aligned to the content of the project, whereas the locally developed assessments were aligned and, therefore, sensitive to instruction.
The Challenges of Implementing Experimental and Quasi-Experimental Designs in Teaching American History Grant Evaluations
Kristine Chadwick,  Edvantia Inc,  kristine.chadwick@edvantia.org
Kim Cowley,  Edvantia Inc,  kim.cowley@edvantia.org
Georgia Hughes-Webb,  Edvantia Inc,  georgia.hughes-webb@edvantia.org
Through six projects, TAH program evaluators at Edvantia have taken on the challenge of implementing mandated experimental or quasi-experimental designs on shoestring budgets. Working with slightly different designs and methods, these evaluators have learned a few lessons about how to adhere to the rigorous evaluation mandate in the program solicitation, soothe client fears, partner with clients to implement these designs, collect data for both formative and summative purposes, stay within budget, and successfully manage at least a few threats to validity. Three main challenges will be discussed. Evaluations were often downgraded from true experimental designs to quasi-experimental either in the initial design phase or later in the project implementation. Negotiating and recruiting teacher comparison groups required different solutions in every project. Comparison group data collection has required evaluator flexibility and extensive use of persuasion skills, thus highlighting both the technical and people skills evaluators must have to be effective.

Session Title: Evaluation Planning Incorporating Context (EPIC): A Model and Case Examples of Practice
Panel Session 546 to be held in Room 113 in the Convention Center on Friday, Nov 7, 9:15 AM to 10:45 AM
Sponsored by the Theories of Evaluation TIG
Chair(s):
Debra Holden,  RTI International,  debra@rti.org
Abstract: This session will present an overview of a Sage book (release date: 10/08), entitled A Practical Guide to Program Evaluation Planning. The book provides a step-by-step process to guide evaluators as they begin developing a program evaluation with a number of settings. Our five step conceptual framework, called Evaluation Planning Incorporating Context (EPIC), includes 1) assessing context (e.g., stating the purpose of the evaluation); 2) gathering reconnaissance (e.g., determining the evaluation uses); 3) engaging stakeholders (e.g., ensuring stakeholders' buy-in); 4) describing the program (e.g., stating the theoretical underpinnings of the program); and 5) focusing the evaluation (e.g., assessing feasibility of proposed measures). We will first provide an overview of our model, including an explanation of each step and corresponding issues to anticipate for each, and then invited presenters, who are authors of chapters in the book, will present case examples of their planning process for evaluations conducted in differing contexts.
Evaluation Planning Incorporating Context (EPIC) Model Overview
Debra Holden,  RTI International,  debra@rti.org
Marc A Zimmerman,  University of Michigan,  marcz@umich.edu
In this presentation (and book), we introduce an evaluation planning process model we call the Evaluation Planning Incorporating Context (EPIC) model. We define evaluation planning as the initial or background processes that go into the final design and implementation of a program evaluation. We describe five general steps in this process assess context, gather reconnaissance, engage stakeholders, describe the program, and focus the evaluation that provide a guideline for evaluators as they develop an evaluation plan. The EPIC model provides a heuristic for evaluation planning rather than a specified set of steps that are required for all evaluations. Some parts of the model may be more or less applicable depending on such issues as the type of evaluation, the setting of the evaluation, the outcomes of interest, and the sponsor's interests. Thus, the EPIC model can be used as a kind of instruction guide to prepare for a program evaluation.
Planning for an Education Evaluation
Julie Marshall,  University of Colorado Denver,  julie.marshall@uchsc.edu
In this presentation, contextual factors identified in planning the evaluation of a school-based nutrition curriculum in a rural, low-income community will be described. The curriculum delivery included fun, hands-on food preparation, cooperative learning, and activities that were tied to content standards in math, literacy and science. The evaluation focused on understanding long term curriculum effectiveness and factors that influence curriculum adoption and delivery. Evaluation planning considered local and state stakeholders; the stress surrounding high-stakes testing and the burden placed on schools for health related activities. Including teachers as part of the evaluation team was critical for informing the evaluation questions, for informing the context within which teachers and students operate that may modify curriculum delivery and impact, and in developing evaluation tools. Strategies for planning meaningful evaluations in a school setting and lessons learned will be highlighted.
Evaluation Planning for a Service Agency Program
Mari Millery,  Columbia University,  mm994@columbia.edu
This case study represents an evaluation planning process in a service program context. The EPIC model is applied to describe the planning for a study at the Leukemia & Lymphoma Society's Information Resource Center (LLS IRC), which responds to nearly 80,000 annual telephone inquiries from cancer patients and their family members. The study focused on a patient navigation intervention consisting of follow-up calls LLS IRC made to its clients. The author, who was contracted as an external evaluator, worked closely with LLS IRC managers and staff throughout the 3-month planning process which resulted in a fairly rigorous study design. The presentation will describe each planning step and discuss lessons learned. Issues of particular importance to service programs will be highlighted, including the complexity of the context, importance of stakeholders, process vs. outcome evaluation, and use of tools, conceptual frameworks, and evaluation research concepts while working with service program stakeholders.
Evaluation Planning for a Community-Based Program
Thomas Reischl,  University of Michigan,  reischl@umich.edu
Susan P Franzen,  University of Michigan,  sfranzen@umich.edu
Planning an evaluation of a new community-based program required successful partnerships with the project's coordinating agency and other community-based organizations. In the presentation we describe these relationships and the role the evaluators played in developing and evaluating the new program. We adopted a "responsive predisposition" (Stake, 2004) to focus on key issues, problems, and concerns experienced by the program's stakeholders in plan development. We also adopted principles that engaged key stakeholders in the evaluation planning and in the implementation of the evaluation study. We describe the development of the evaluation plan for a new telephone information and referral service focused on serving African American families and reducing infant mortality among African American mothers. Finally, we discuss the utility of using an Evaluation Planning Matrix to help focus the evaluation. The evaluation plan focused on process evaluation goals and attempted an outcome evaluation study using baseline data from a previous study. Stake, R. E. (2004). Standards-based and responsive evaluation. Thousand Oaks, CA: Sage.
Planning for a Media Evaluation: Case Example of the National Truth Campaign
W Douglas Evans,  The George Washington University,  sphwde@gwumc.edu
Media evaluation is an overarching subject area that includes the study of marketing campaigns intended to promote or change consumer behavior, as well as assessments of educational and entertainment media and the effects of news media on public discourse and policy. In this presentation, we describe evaluation planning strategies and research methods for health communication and marketing campaigns designed to affect consumer health behavior. Media evaluation is distinct from other forms of program evaluation. It focuses on media effects on healthy behaviors or avoidance of unhealthy behaviors, as opposed to broad evaluation strategies that cross-cut multiple venues and approaches. Media evaluations measure four key process and outcome dimensions of campaign effectiveness: 1) exposure and recall, 2) message reactions and receptivity, 3) behavioral determinants, and 4) behavioral outcomes. After describing media evaluation methods, we describe the truth campaign evaluation, the largest antitobacco media campaign conducted in the United States.

Return to Evaluation 2008
Search Results for All Sessions