Evaluating Outputs and Outcomes
What is it?
Simply put, evaluation assesses whether or to what degree a given effort is achieving its intended outcome. For hunting and shooting sports R3, the ultimate intended outcome is to maintain or increase participation in and support for hunting and shooting sports. For more details and background about R3 outcomes, see the chapters on R3 Foundations and Definitions and Planning and Strategy.
In the Planning & Strategy Chapter, we learn how to design R3 efforts to help participants through the Outdoor Recreation Adoption Model. In this chapter, we learn how to evaluate whether or not those programs meet their objectives.
Evaluation is a systematic process determining if your programs achieve their intended purpose.
Evaluation results can be used to:
- Show successful outcomes to your stakeholders.
- Build additional support for R3 programming.
- Identify specific ways to improve implementation.
- Make strategic decisions about which programs to continue, modify, or discontinue.
It is common for practitioners and administrators to lose sight of evaluation as they focus on program development and implementation. However, if applied early and consistently, evaluation will help you collect the data you need to:
- Stop conducting programs/efforts that aren’t achieving the outcomes you seek.
- Do more programs/efforts that are working (or try new efforts that might work).
- Make adjustments to programs/efforts that show promise.
Why is it important?
The R3 community has been implementing programs with no systematic evaluation for decades. That is, the R3 community has been doing all manner of “stuff” with little or no idea whether it is working. With hunter numbers declining, the community can’t afford to continue with this approach.
Recent research1 shows that hunting license sales will continue to decline as the license-buying population declines due to the effects of aging and other social, cultural, and demographic changes.
The situation for shooting sports is slightly different. The number of people purchasing and using firearms is growing, not declining. However, there is no obvious metric like license sales to measure overall participation. The need to carefully identify R3 objectives and measure progress against them is even more acute.
How can your organization help its practitioners?
- Insist that R3 efforts be evaluated against specific objectives to ensure effectiveness.
- Give the R3 Team the resources and authority to conduct meaningful evaluation.
- Ensure staff have access to data needed to measure outcomes (dashboards, CRMS, license sales data, customer demographic information, marketing campaign evaluation results etc.).
- Work toward implementing a full Customer Relationship Management (CRM) system (or at least collaboration among the agency divisions that touch customers).
- Ensure your state agency is participating in the Real-Time Hunting and Fishing License Data Dashboard and your state agency or organization uses an internal dashboard to track sales/engagement at the state/organization level.
Other types of Evaluation
There are specific evaluation methods and metrics for many of the disciplines that combine to form R3. Each of these chapters of the guide has an evaluation section of its own. Here are links to those other sections of the guide if you are looking for that specific assistance:
Create a Plan Format and Process
Revising your plan over time
Marketing Evaluation
How to Evaluate Partnership Engagement
How to Evaluate and Refine Your Education and Outreach Initiatives
The New Future of Hunting and Fishing Technical Report
Nationwide license sales data are used to quantitatively predict the future volume of license sales over time.
What are you talking about? – Definitions
A big challenge the R3 community has grappled with is identifying what, specifically, constitutes an R3 effort. Too often, R3 practitioners inherit or are directed to implement programs or efforts that aren’t really R3 at all. There will always be political, economic, and other realities that practitioners must deal with, i.e., you may have to complete tasks that are not R3. Identifying the programs and efforts that are and are not R3 is the first step in prioritizing where you can most effectively spend your evaluation resources.
Additionally, there are terms and concepts used in the R3 community without common definitions. Some of them are measured in different ways. This makes it very difficult to evaluate efforts effectively and/or compare one effort’s results to another.
Working with R3 practitioners from around the country, the Western Association of Fish and Wildlife Agencies is leading an effort to define R3 terms and concepts more clearly2. Practitioners would do well to read the paper “R3 Metrics” in its entirety. Key elements follow.
Overall, the purpose of R3 efforts is to maintain or increase participation in and support for hunting and shooting sports. This is most often done by influencing or modifying human behavior, which can be specifically defined according to the three Rs.
Recruitment – A behavioral influence (from an R3 effort or other external influence) resulting in the initial choice to participate in a target activity.
Retention – A behavioral influence (from an R3 effort or other external influence) resulting in continued year-to-year participation in a target activity.
Reactivation – A behavioral influence (from an R3 effort or other external influence) resulting in the renewed participation in an abandoned target activity.
Thus, most R3 efforts should be viewed as interventions or actions needed to initiate, support, or re-establish an individual’s adoption of desired behaviors.
An individual’s or group’s adoption of a new behavior is likely dependent upon two things. One is how well the behavior aligns with the individual’s existing values, attitudes, or social norms. The second is how well an R3 effort addresses the unique barriers (real and perceived) that restrict a particular audience from participating in the behavior3. Ideally, R3 efforts should be designed and implemented to address the specific barriers and subjective learning needs of a particular group or demographic. If the effort is designed without an understanding of the audience’s values, beliefs, and barriers to participation, it may fail to motivate individuals to participate in the future.
More to the point, given the diversity of those likely to engage in an outdoor activity, it is imperative that R3 efforts be evaluated for effectiveness. R3 practitioners need data illuminating how and when to modify their R3 efforts to increase efficacy and efficiency. This underscores the need for adaptive program design and implementation of R3 efforts that incorporate an evaluation system capable of documenting participants’ behavioral changes directly attributable to the R3 effort.
Evolution and Current Use of the Outdoor Recreation Adoption Model
This resource helps define and outline the ORAM and its stages.
Common R3 definitions
Without common definitions, it is nearly impossible to coordinate a national R3 effort. It is also very difficult to train new practitioners. Many of the definitions below are excerpts from the “R3 Metrics” white paper developed by the Western Association of Fish and Wildlife Agencies’ R3 Evaluation Criteria Working Group or from the National Dashboard. The latter are marked with an asterisk to denote their source.
Agency – For this discussion of R3, an agency is defined as a federal, tribal, state, municipal, or other governmental entity in charge of the conservation and management of fish and wildlife.
R3 – An acronym for the words “recruitment, retention and reactivation”; it refers to a series of efforts intended to modify participant behaviors to cause increased or maintained participation in target activities (e.g., boating, fishing, hunting, recreational shooting, etc.).
Recruitment – A behavioral influence (from an R3 effort or other external influence) resulting in the initial choice to participate in a target activity.
Retention – A behavioral influence (from an R3 effort or other external influence) resulting in continued year-to-year participation in a target activity.
Reactivation – A behavioral influence (from an R3 effort or other external influence) resulting in the renewed participation in an abandoned target activity
Churn rate – The proportion of a total participant population who do not participate in a target activity (as indicated by license purchase patterns or other participation metrics) in a given year or years.
*Individual organizations could define the amount of time a person has to spend without a license to consider them in your churn rate. The national dashboard calculates churn rate as the percentage of prior year license holders that were not license holders in the reporting year.
*License Holder/Participant – an individual who holds one or more licenses granting primary privileges for that activity. The National Dashboard counts any individual hold his/her license for at least one day as a license holder.
*License Rate: the percentage of the population for the reporting state or region that were license holders for the reporting period. The National Dashboard counts any individual hold his/her license for at least one day as a license holder.
Lapsed participant – An individual who does not participate in a target activity in the year, or years, after they previously participated.
*Individual organizations could define the amount of time a person has to be without a license to consider them in lapsed. The national dashboard calculates an individual as lapsed if she/he had a license for at least one day of the prior licensing period and does not have a license in the current licensing period. (based on churn def)
Reactivated participant – A lapsed participant who did not participate in a target activity in the previous year or years, and then resumes participation in the current year.
Reactivation Rate – The proportion of lapsed participants who did not participate in a target activity in the previous year or years, but then resumed participation in the current year.
Recruited participant – An individual who participates in a target activity for the first time. The National Dashboard considers a person who was a license holder in the reporting year, but was not a license holder in the 5 years prior a new participant/new recruit.
Recruitment Rate – The proportion of the participant population who participates in the target activity for the first time.
Retained participant – An individual who participates in a target activity in both the previous year and the current year.
Retention Rate – The proportion of individuals in a participant population who participated in a target activity in both the previous year and the current year.
Evolution and Current Use of the Outdoor Recreation Adoption Model
This resource helps identify and define words used in the ORAM.
Plan for Evaluation From the Beginning
One of the primary reasons so few R3 efforts have been evaluated in any meaningful way is the fact that very few organizations think about evaluation until the effort is well underway or completed. To be effective, every R3 program or action should have evaluation woven into it from the very beginning. For every program, ask what is this program’s objective—what, specifically, is it trying to achieve? How does this program help us meet the overall R3 outcomes we seek?
Review the chapter Planning and Strategy for guidance on setting meaningful objectives. Defining your evaluation objectives up front establishes guidance for what questions about the program need to be answered. It helps to decide how to collect data to track metrics, how these data are analyzed, and how the results will be used and shared.
Even if your agency or organization has been doing a program for many years, if it doesn’t have a specific objective identified, stop everything, and assign an objective(s). Otherwise, you will have no way of knowing whether that program or effort is worth the time and effort you are putting into it. Sometimes, when you attempt to attach a specific objective to a program, you realize that it is not really R3 after all. Or, maybe it makes people feel good about your agency/organization or “makes us more visible in the community” or something along those lines. Perhaps the public relations objective is essential, and you may decide to continue doing it for that reason. If a program is not achieving R3 outcomes, recognizing that and calling attention to it is the first step toward making your true R3 efforts effective.
Once you have suitable objectives, the next step is to identify which metric you will measure. This determines whether you have been successful, to what degree you have been successful, and how often it will be measured. After you set objectives and metrics, the real work of evaluation can happen. This means looking closely at results over time and acting on the results. Adjust or eliminate the things that aren’t working, so you can do more things that make a difference.
Often, it is difficult to determine when enough is enough. As the plan develops, the amount of information wanted tends to grow larger and larger. This can lead to data collection instruments that are too long or demand staffing and fiscal resources you don’t have. Asking yourself who this evaluation is intended for and what information will be most useful to them will make the task of narrowing your focus more manageable. Is the evaluation intended to provide information to R3 practitioners, funding partners, agency directors, or commissions? An evaluation for program staff may look very different than one intended for funding partners or directors. What information is most meaningful to them? Different types and quantities of information will be more or less valuable in different situations.
Outputs and Outcomes
The difference between outputs and outcomes is important in evaluating R3 efforts. Simply stated, outcomes are the ultimate ends we seek from our efforts, and outputs are what we accomplish to achieve those outcomes. Most agencies and organizations select some variation of the following ultimate outcomes for their R3 programs:
- Maintain/increase participation in hunting/shooting activity and/or license purchases.
- Increase public acceptance/support of hunting/shooting activity.
Nearly all workshops, activities, communications, and other efforts you do to get people to do these things are outputs. Far too often, when agencies and organizations set out to evaluate their programs, they only measure outputs. Things like the number of programs delivered in a year, the number of participants going through programs, the number of training videos produced, clicks on websites, online impressions, and so on. Very few of these efforts lead directly to the outcomes we seek. For instance, it usually takes many interactions over time and across a variety of topics to create an independent hunter or shooter. While measuring outputs is very important, it is even more important that you do not stop there. It is insufficient to have dozens or even hundreds of attractive outputs if it doesn’t achieve the ultimate outcome. These are all part of effective R3 program evaluation. Ideally, you should measure as many outputs as possible to be sure they are tracking in the right direction. Then measure your outcomes to be sure the rest of it makes a difference.
Set yourself up for success. Are your outcomes and outputs realistic–can you achieve them in the amount of time you have? Are your outcomes about change–how will the participants be different because of the program? Can you think of existing or new ways to measure these outcomes?
Results Chains
Results chains are foundational to evaluation. They provide a framework for designing evaluation activities. A results chain visualizes the step-by-step process leading from a program’s inputs to the outcomes. It shows how the effort is intended to work. It will help you ensure that your activities align with your intended outcomes. It also helps identify appropriate metrics for measuring the success of your efforts.
Inputs are what we invest in our efforts. They include staff and volunteer time, money, equipment, and facilities. Outputs are what you use the inputs to create. They are the products and services you deliver and who you reach with these products and services. Finally, outcomes are the results of your outputs. You will primarily be interested in two types of outcomes: effort/program-specific outcomes and R3 outcomes. Effort- or program-specific outcomes look at how the output affects participants’ awareness, knowledge, attitudes, skills, intentions, or motivations. R3 outcomes are the ultimate goal. They are how the effort or program results in R3 behaviors (recruitment, retention, reactivation).
Suppose your program or effort already has a results chain developed. In that case, it should be the starting point for planning your evaluation. If you don’t already have one, creating a results chain involves visually representing the sequence of events that leads from program inputs to outcomes.
Start by defining the inputs, representing the resources (personnel, equipment, etc.) that go into a program or effort. Layout your planned activities that will be implemented. From here, you will identify these activities’ direct and tangible results. These should be specific and measurable. Incorporate key indicators at each stage of the results chain. These will help you measure progress and success. You will need to clearly articulate the effort-specific outcomes and your R3 outcomes. Remember, your effort-specific outcomes may not represent the desired R3 outcome but can reflect progress toward desired R3 outcomes. Typically, several effort-specific outcomes must be achieved before achieving your R3 outcome. You may find that participants will need additional interactions beyond a single program or effort. When building your results chain, clearly demonstrate how each step contributes to the next stage of the chain.
As you lay out the results chain, ensure a logical flow from inputs to outcomes. Each step in the chain should make sense in the context of the program’s theory of change and goals. A theory of change is a systematic description of how and why a desired change is expected to happen in a particular situation (e.g., the Outdoor Recreation Adoption Model). It articulates the underlying logic and assumptions of a program.
From your results chain, you should be able to fill in the blanks below:
If we do (proposed activity), then based on (theory of change), we can logically expect (result) to happen as evidenced by (indicator).
The following is an example of what a results chain for a shooting sports R3 program might look like.
If you would like a more in-depth discussion about results chains, the Dunfee and Pedder 2017 webinar on R3 evaluations is a great place to start.
Don’t try this alone!
Program evaluation often requires a range of skills and knowledge areas. It includes research design, data collection and analysis, subject matter expertise related to the program, and understanding the context in which the program operates. Just as most marketing specialists shouldn’t assume they could conduct a research project on turkey population ecology, most R3 practitioners should not assume they can design and implement effective program evaluation. Specific expertise is required. Establishing an evaluation team is important for conducting a thorough, objective, and credible program evaluation. Include staff from outside the R3 program (e.g., licensing, social science, marketing/communications, GIS, etc.) who can critically support the evaluation process. The combination of diverse skills, perspectives, and expertise enhances the quality of the evaluation process and its outcomes.
Part of the work of an evaluation team is to determine the staff resources, data systems, abilities, and fiscal resources needed and available for the evaluation. Your team might be able to support all your evaluation needs, but if you are stuck or need an outside perspective, seek input as early as possible. Find a social scientist, an experienced evaluator, or a subject matter expert who can give advice and notice problems early in the process. More and more fish and wildlife agencies and conservation organizations have social scientists on staff. But they are often spread thin and working across many divisions and projects. Engaging them early in the process and articulating your program’s objectives and outcomes will help them understand how to provide help. These steps are critical. Depending on your budget, you may be able to hire a consultant to assist with some or all of the process. They may help frame the research questions, develop evaluation instruments, assist implementation, and/or perform data analysis and reporting.
Diving into Evaluations (90 min.)
This webinar discusses the basics of both improvement and impact evaluations.
Hunting, Fishing, Sport Shooting, and Archery Recruitment, Retention and Reactivation: A Practitioner’s Guide
This resource helps identify areas and methods for formal evaluation. Pg. 77-94.
Open Standards for the Practice of Conservation
This PDF/PPT deck presents information on the theory of change and how to build a results chain.
Open Standards for the Practice of Conservation
This resource covers all aspects of using results chains to design, implement, and evaluate conservation initiatives.
Outcomes versus Outputs (5 min.)
This quick webinar discusses the differences between outcomes and outputs. Important, foundational knowledge for preparing programs for evaluations.
Frameworks and Standards for R3 Effort and Strategy Evaluation
An in-depth study and resource for program evaluation from planning to identifying useful metrics.
Common R3 Metrics
The ultimate R3 outcome most organizations seek is participation in hunting and the shooting sports. For hunting, the obvious metric is license sales. Not everyone who buys a license participates, and not everyone who participates is required to buy a license. But the correlation is high, and license sales are a well-accepted proxy for hunting participation. License sales (or sales resulting from specific interventions) are one of the primary ways agencies and organizations can assess the success of hunting R3 efforts.
Following are some metrics commonly used to assess changes in license sales over time. Some are straight counts; others are percent change, where percent change is calculated:
Recruitment – A behavioral influence (such as from an R3 effort) resulting in the initial choice to participate in a target activity.
- Attain X number of unique new permit holders by [date].
- Increase by X percent the number of unique new permit holders by [date].
- Increase participant recruitment rate X percent by [date].
To be considered recruited, an individual should never have had a previous license (or range permit) issued by your organization.
Retention – A behavioral influence (such as from an R3 effort) resulting in continued year-to-year participation in a target activity.
- Increase by X percent the number of unique new permit holders who maintain those permits for Y years by [date].
- Reduce by X percent the number of participants who have lapsed for Y years by [date].
- Increase participant retention rate X percent by [date].
Individual organizations could define the number of years a person has to renew their license (or range permit) to consider them in your retention rate. The national dashboard calculates the retention rate as the percentage of prior year license/permit holders that renewed their license/permit in the reporting year.
Reactivation – A behavioral influence (such as from an R3 effort) resulting in the renewed participation in an abandoned target activity.
- Increase by X percent the number of lapsed participants who purchase a hunting license/range permit in the current year.
- Increase participant reactivation rate X percent by [date].
Individual organizations could define the amount of time a person has to be without their license (or range permit) to consider them in your reactivated when they purchase a new license/permit. The national dashboard considers individuals reactivated if they missed at least one entire year and then renewed their license/range permit.
Although the above metrics measure the primary R3 outcomes most agencies seek, they don’t tell us whether the outputs (our R3 programs and efforts) positively contribute to these outcomes. For that, we need additional proxy metrics. For instance, if we have a program designed to recruit new hunters (and we collect information that will allow us to match them to our license database), we could calculate a program conversion rate, that is, the proportion of first-time license sales purchased by our program participants (X unique, new participants in R3 programs account for Y percent of first-time license sales by [date]).
This still isn’t a direct metric because we can’t know whether our program or R3 effort was the cause of the recruitment. But it is reasonable that if this proportion is high, either our R3 effort is having some effect, and/or people participating in R3 programs are already more likely to purchase a license for the first time. We could use self-reported data (e.g., post-event survey) to learn more about this possible effect.
Although it isn’t perfect, this kind of evaluation can tell us a lot about whether or to what degree our R3 efforts result in the behavior change we seek.
Diving into Evaluations (90 min.)
This webinar discusses the basics of both improvement and impact evaluations.
Hunting, Fishing, Sport Shooting, and Archery Recruitment, Retention and Reactivation: A Practitioner’s Guide
An expansive compilation of R3 survey/assessment results. These sections identify potential R3 metrics. Pg. 77-94, 328.
Outcomes versus Outputs (5 min.)
This quick webinar discusses the differences between outcomes and outputs. Defines common metrics and ideas.
Recommendations and Strategic Tools for Effective Angler Recruitment, Retention and Reactivation (R3) Efforts
This resource was designed for aquatic educators. Much is readily transferable to any R3 effort. This guide identifies potential metrics depending on your situation.
Try, Test, Learn, Adapt: The Massachusetts Approach to R3 (55 min.)
This webinar includes an in-depth description of one state’s R3 program evolution and evaluation of their Learn to Hunt clinics. Provides an example of program evaluation.
Frameworks and Standards for R3 Effort and Strategy Evaluation
An in-depth study and resource for program evaluation from planning to identifying useful metrics.
License Dashboard
Thanks to a Multistate Conservation Grant sponsored by the Association of Fish and Wildlife Agencies, state agencies can now see their license sales trends at a higher scale. License dashboards were created to allow state agencies to view and share real-time license sales information in a standard format. Since license sales are one commonly used indicator of participation, understanding regional and national license sales trends helps agencies better understand their operating environment, and whether these trends exceed or underperform against set goals or regional averages. With these insights, states can identify when changes are needed or if current R3 efforts are on track. These kinds of real-time data are what the industry uses to make marketing and sales decisions on a daily basis. Now that this technology is in place for tracking license sales, state agencies should do the same.
Primary reasons agencies are encouraged to participate:
- Data Confidentiality: Participation in the real-time dashboard does NOT require the sharing of license buyers’ personal identifying information (PII).
- Frequency: Daily license sales are collected and reported in historical (ten-year), cumulative (year-to-date), monthly, and daily visualizations. Sales data undergo a standard, rigorous normalization process to ensure accurate and reliable hunting and fishing participation insights.
- With the shift to API technology, demands on state agency staff are minimal.
- Cost: The dashboards are funded by an AFWA multistate grant and FREE to all state agencies.
Click here to see the latest results from participating states. Go here for more information. If your state is not currently participating, contact Lisa Parks ([email protected]) for additional details or to sign up.
Hunting Licenses, Holders, and Costs by Apportionment Year
These dashboards present licenses issued as reported by the states.
R3 Dashboard – Why Your Agency Should Participate
Data dashboards provide quick, easy-to-interpret R3 insights. This guide encourages participation in the dashboards.
How to Identify Meaningful (and Reasonable) Evaluation Techniques
There are many different ways that R3 efforts can be evaluated (assuming specific objectives have been set). The trick is selecting techniques that give meaningful results at a reasonable cost. Data come in a variety of forms. Regardless of the manner of collection, data must be gathered and organized before analysis can begin. Before you start, ask yourself what the objectives of the evaluation are, what type of data you are collecting, and what methods you will use to gather, organize, and analyze the data. Answering these questions will help you avoid a very common problem—being unable to answer your evaluation questions. Accurate data are necessary to draw correct conclusions and maintain research integrity. If you are unsure if a technique suits your needs, seek help from a social scientist or experienced evaluator.
Ethical Considerations
Data collection involves gathering information from your program participants. This makes ethical considerations paramount to ensure data privacy, informed consent, and minimizing the potential impact on participants. There are online resources that can help you ensure that your evaluation is conducted with integrity and sensitivity. If your evaluation team includes a social scientist or experienced evaluator, they should be well-versed in this area as well.
Administrative Data – License Sales
If your evaluation efforts have license sales or range permit sales as a metric, external data may exist for measuring impact. License sales data and program-related records often include numbers and statistics that you can use to track programs and participants.
The points below are a general guide on using administrative data to evaluate an outcome:
- Define the outcome you seek and how you will measure it.
- Identify relevant administrative data sources, including your agency’s license sales database, license dashboard, a CRM system, and other program-specific records.
- Prepare and clean the data – This involves handling missing or duplicate data, correcting errors, and standardizing formats. This step is crucial for the reliability of your analysis.
- Link data sources – If your outcome metric requires information from multiple sources, you may need to establish a unique identifier to link data from different sources.
At this point, you can analyze, interpret, and share the results.
Primary Data Collection
If your evaluation efforts do not have license/permit sales as a metric and no relevant administrative data are available, you will need to collect data directly from participants. Following is a general overview of quantitative and qualitative methods. If your evaluation team lacks the expertise for social science primary data collection, consult with a social scientist or experienced evaluator. Poor design often leads to bad data, and bad data are often worse than no data at all.
You need to collect quantitative data for evaluations looking to precisely measure results or gain large-scale statistical insights. If the hope is to explore ideas, understand experiences, or gain detailed insights in a specific context, collect qualitative data. Often, we hope to do both, in which case we can use a mixed methods approach and collect both quantitative and qualitative data.
Original R3 Action Plan Appendices
The original R3 plan appendices. These help guide R3 program development and evaluation.
R3 Specific Evaluation & Social Science Training and Resources for the Modern R3 Practitioner
Evaluation Training Resources designed specifically for the R3 Practitioner.
How to Evaluate and Refine Your Education and Outreach Initiatives
Periodically evaluating your education and outreach initiatives will help identify gaps, improve implementation, and increase program effectiveness.
Evaluation is how we prove and improve program methodology. Periodic evaluation of your programs will help identify strengths and weaknesses. They can be consistently improved upon using both short- and long-term changes. When your evaluation methods are consistent, logical, and goal-oriented, they bring great power and leverage to refining your education and outreach initiatives.
Using a program mapping worksheet (See How to Identify Program Needs), you can identify the gaps in our programming.
However, we also need to understand if the programs that we are running are having an impact.
In the short term, evaluating specific programs starts with establishing good objectives. It includes pre-, post-, and follow-up assessments for each program. Then, we can measure our progress against our objectives to evaluate their success. Suppose our objective for a squirrel hunting class is to have class members buy a small game hunting license. In that case, we can set up assessment stages that tell us how well we met that objective.
Pre-assessment:
A pre-assessment gives us a baseline. The pre-assessment stage can occur before a program is offered to help us better understand our audience. Suppose our pre-assessment reveals that folks in our squirrel hunting class have never been hunting. In that case, we know that the course should focus on knowledge and skills gaps that are common for new hunters.
Post-assessment:
In the post-assessment stage, we can poll the class participants about how confident and prepared they feel about participating in small game hunting. This lets us know how the program is doing and whether adjustments are needed to increase participants’ likelihood of actually going hunting after the program.
Follow-up assessment:
Finally, our follow-up assessment can identify the folks who took the class and whether or not they purchased a small game license to hunt squirrels. This lets us know if we met our overall objective. Utilize the follow-up assessment to identify effective programs or strategies to inform future educational practices.
The things we most often evaluate in R3 programming are the demand for a specific program, what participants learned from a program, whether participants did the activity immediately following the program, and whether they subsequently purchased a license to continue the activity on their own.
Regarding hunting, this could look like:
- An assessment of interest in a deer hunting program (and current knowledge of deer hunting).
- An assessment of what participants learned in the deer hunting program.
- An assessment of participants who went on a mentored deer hunt after the program.
- Checking the license database to see if participants purchased a deer tag and went deer hunting in the following year(s).
Regarding shooting sports, this could look like:
- An assessment of interest in a target shooting program (and current knowledge of shooting).
- An assessment of what participants learned in the target shooting program.
- An assessment of participants who went target shooting as part of the program or immediately following it.
- A follow-up assessment of participants to see how many regularly go target shooting six months or a year after the program.
Every stage of the evaluation process allows us to make changes and improve the R3 effort so more people take the class, more specific barriers and needs are met, and more R3-specific behaviors are encouraged. For much more specific information on effective evaluation, see the chapter on Evaluating Outputs and Outcomes.
2015 MN DNR Adult Learn to Hunt Firearms Deer Pre-Program Survey
This resource is a sample pre-assessment survey for an Adult Learn to Firearms Deer Hunt program from the MN DNR.
Georgia Hunt and Learn Program Satisfaction Survey
This resource is a sample post-assessment post-program satisfaction survey from the Georgia DNR.
Learn to Hunt Turkey 2016 Pre-program Survey
This resource is a sample pre-program pre-assessment for a Learn to Hunt Turkey program.
Locavore.guide: Locavore Program Evaluation
This lesson lays out the basics of evaluation for a hunter/angler training program.
Pre-Event Questionnaire Template
This resource is a generic pre-event questionnaire that can used for many different types of programs.
Qualitative Versus Quantitative Research
Quantitative – Quantitative research results can be compared numerically. The advantage of quantitative methods is that they provide results that can be easily compared.
Quantitative methods are best used to establish cause-and-effect relationships, test hypotheses, and determine a large population’s opinions, attitudes, and practices.
Quantitative methods include:
- Surveys
- Experiments
- Analysis of existing data
Qualitative – Qualitative research attempts to understand human thoughts, experiences, and motivations. The complex nature of this information can’t be quantified and compared as easily as quantitative research, but qualitative analysis is necessary to understand the human experience.
Qualitative methods are best for developing hypotheses and theories and describing processes such as decision-making or communication.
Qualitative methods include:
- Focus groups
- Interviews
- Open-ended questionnaires
- Content analysis
It’s impossible to determine which research technique is best for you without understanding your situation, available resources, and objectives. Social science research requires a complex skillset. This chapter of the Practitioner’s Guide is designed to give you enough of an understanding of the process to coordinate with your organization’s social science specialists or hire a consultant rather than doing the research yourself. However, if you have the background to do research yourself, the following sections will help you with the specifics. If you aren’t going to do your own research, you can probably skip these sections:
Quantitative Methods
A primary advantage of quantitative methods is that they can provide objective, reliable, and generalizable results. Quantitative research encompasses the techniques used to collect and analyze numerical data. These techniques most commonly use random sampling to describe patterns in data, make predictions, test relationships, and generalize findings to wider populations. One of the main limitations of quantitative research is it may not capture the richness, depth, and diversity of people’s experiences, meanings, and perspectives.
To ensure generalizability of your results, samples need to be carefully designed. But even the best designed studies have error associated with them. There are two broad categories of error: random error and systematic error. Random error is caused by the natural variability in the measurement process and is due to chance. Random error can be reduced by taking repeated measures using their average and increasing your sample size. Systematic errors are predictable and create bias in your data. While both errors can be problematic, systematic errors are typically considered worse and can be tricky to identify and fix. You can reduce these types of errors by using multiple data collection techniques, doing cognitive testing on questionnaires, and using probability sampling methods like random sampling.
Two quantitative techniques you are most likely to use in R3 program evaluation are analysis of external data and questionnaires. External data such as administrative data for programs often include numbers and statistics that you can use to track programs and participants. This type of data is already collected for other purposes. It may come from surveys or as part of your agency’s normal business. This type of data may be sufficient for meeting many of your information needs.
Some advantages to using external data sources are:
- It can include comprehensive data over a long period of time.
- Data collection time is minimal since the data has already been collected by someone else.
Some limitations to this type of data are:
- Often only numerical data and does not provide the reason or cause behind the data.
- Cleaning poorly structured records can be time-consuming.
- Incomplete or inaccurate records often cannot be fixed.
Surveys are the most common way to collect quantitative data. Surveys comprise a series of questions designed to collect data to help you answer your evaluation questions. An important consideration in collecting quality data with a survey is to know and avoid four sources of error.
- Coverage error is the difference between the target population (the group of people about which we want to generalize) and the sample population (the group of people from which a sample is drawn). A survey can inform us about the sample population and no more. For example, you would not survey bow hunters to find out how all hunters think. Current and accurate sampling lists can be difficult to build and maintain. Missing, duplicate, and ineligible entries are important considerations when developing sample lists. The amount of coverage error depends on how different these entries are from the target population.4
- Sampling error is a reality for all who conduct sample surveys. It is the extent to which a sample is limited to perfectly describe a population. Conducting a full census is the only way to completely avoid sampling error; however, time, staffing, and financial resources usually make this impractical. Sampling error is, however, fairly easy to quantify and can be minimized by increasing sample size. There are tables to help estimate appropriate sample sizes you can access if you do not wish to do the math behind the formula yourself. It is important to note, though, that the sample sizes indicated are the number of surveys completed and not the number of people you asked to complete the survey. The latter will be much larger.5
- Measurement error happens when a response to a question is inaccurate, imprecise, or cannot be compared to other responses. While coverage and sampling error occur before data are collected, measurement error and nonresponse error occur during collection. There are typically four sources of measurement error: the survey mode, questionnaire wording and construction, influence of the interviewer, and the respondent’s behavior. The type of survey or mode (e.g., telephone, mail, online, telephone, face-to-face) place different demands on respondents. For some questions, different types of surveys may yield different results than other types. Questions that deal with abstract concepts or sensitive issues tend to be affected more so by this source of measurement error. Errors can also occur if the questionnaire wording is not clearly understood in the same way by all respondents or response choices are not clearly defined. Question structure may confuse respondents, making it impossible to answer. Asking double-barreled questions (two questions in one) and not providing mutually exclusive answer choices are common ways question structure can lead to errors. Interviewers in telephone and face-to-face surveys may also influence responses. Even with training, the way an interviewer asks a question can influence or lead respondents. Finally, respondents must be able and willing to provide accurate responses. Measurement error from respondents can occur if they deliberately provide inaccurate answers or inadvertently if they misunderstand the question.6
- Nonresponse error happens when a significant number of people in the sample do not respond to the survey and are different in meaningful ways from those who do. Even if you have a complete sample list, draw a large enough sample, and have accurate measurement, you will still have to deal with nonresponse error. One way to combat this is to aim for a high response rate, and if one is not received to conduct a nonresponse bias check to compare respondents with nonrespondents.5
It is important to remember that your surveys are open to four distinct sources of errors and that you cannot focus on one to the exclusion of others. Sampling error and nonresponse error often get specific call out in final reports, but that does not make coverage error and measurement error any less important.
The following tips for choosing a survey mode, selecting a sample, and writing good questions take into consideration these four sources of error and how to overcome them. There are multiple modes for delivering surveys, including telephone, mail, in-person, online, drop-off, and mixed modes.
Things you should consider when selecting a survey mode include:
- Questionnaire length
- Completion time
- Complexity of questionnaire
- Available resources
- Availability of sample contact information
Comparing Survey Modes:
- Mail and online survey formats tend to be more suitable for longer surveys.
- In-person and telephone surveys are best suited for surveys with shorter completion times.
- Mail and online surveys also allow for greater question complexity than in-person or telephone.
- Telephone surveys tend to have greater success on completing open-ended questions, screening questions, controlling the order questions are presented and answered, and avoiding item nonresponse. Telephone surveys also tend to be less sensitive to the design layout relative to mail, in-person, and online surveys.
- Costs vary depending on specific situations; however, online surveys tend to have lower costs while in-person tend to have higher costs associated with them.
- Mixed-mode surveys use two or more types of surveys to build on the strengths of each type. For example, you might choose to do an online survey for those in your sample for whom you have email addresses, and then send a mail survey to the sample members who you don’t have emails for.
- Another example could be using a telephone survey for screener questions and then providing a mail survey based on the screener results.7
There are three main things you need to do when sampling: identify your target population, create your sample frame, and draw the sample. You want to identify your target population as precisely as you can and in a way that makes sense for the objectives of your evaluation. Finding a good list can sometimes be a challenge. In a random sample every member of your target population must have an equal chance of selection in the sample. Very few sample lists are perfect. The more you know about your sample lists, the better you can identify the limitations of your survey results and generalizations. Finally, the most common method of sampling is the simple random sample, which can be accomplished with a random numbers table or a computer-generated list of random numbers. You may run into more complex sampling designs. These often require more sophisticated statistical techniques for data analysis and weighting procedures.
Survey Question Design
The key to having useful information is taking the time to turn the ideas that motivate your evaluation into good questions. Typically, your questions will pertain to what your respondents do or their characteristics (behaviors and attributes) or what they think or say they want (attitudes and beliefs). Below are the most common types of questions you will use. You can find examples of R3-related questions in the table below.
- Open-ended – This type of question does not provide answer choices.
- Close-ended – In this type of question, the complete range of possible answers is provided.
- Scale questions – This type of question is used to gauge respondents’ attitudes and opinions. You commonly see them set up to measure the extent someone agrees or disagrees, or the likelihood they will do something. These can be either unidirectional or bipolar in construction.
- Partially close-ended – Between the open-ended and close-ended questions are the partially close-ended. These questions provide answer choices, but respondents have the option to also provide their own response.
Questions relating to attitudes and beliefs, in some cases, can be difficult to measure because they are imprecise, change, and may not be well-formed in advance of a survey. To help minimize measurement errors avoid abstract issues, use more than one question type so you don’t rely on a single question.
Writing good questions usually takes more than one or two attempts. Below are some tips to keep in mind when wording questions8.
- Avoid vaguely worded questions and questions with slanted or leading introductions.
- Avoid unequal comparisons or unbalanced response choices.
- Strive for a neutral and objective tone.
- Don’t ask respondents to calculate numbers.
- Include both positive and negative side in question stem (i.e., how strongly do you agree or disagree….).
- Answer categories should be exhaustive and mutually exclusive. If you don’t know all the possible choices, you can provide an “other, please specify” option or use an open-ended question.
Avoid scales that have more than one zero point. Use unipolar scales when a construct naturally starts at zero and negative frequencies don’t make sense.
- Use bipolar scales when it is appropriate to include both the negative and positive side of a scale. In this case the zero point should be in the middle of the scale and there should be an equal number of negative and positive responses. If you choose to not include the middle option in your answer choice, the zero point is still implied to be in the middle of the positive and negative choices.
- Whether you start with the negative end or the positive end of the scale, stay consistent with this order throughout your survey.
- Answer choices of undecided or don’t know are different from a neutral or neither/nor, and should be placed at the end of a scale and not in the middle.
- Fully label all scale points.
If you do not have access to a social scientist and would like to learn more about writing good questions, Internet, Phone, Mail, and Mixed-Mode Surveys: The Tailored Design Method by Don Dillman et. al. and Survey Research and Analysis: Applications in Parks, Recreation and Human Dimensions by Jerry Vaske are both excellent books that provide a wealth of information about survey research from start to finish.
Qualitative Research Methods
Qualitative research (Page 56 of the doc) is a method of inquiry that looks to understand, interpret, and describe the complex nature of human experiences and social processes. It is a research approach that focuses on the richness and depth of data, often involving non-numerical information such as words, images, and narratives9. It is important to note this research methodology makes no attempt to identify or interview a representative sample of the target audience, nor do researchers use statistical techniques to generalize results across a large population. Rather, researchers look for general themes and broad insights into issues, concerns, and problems that may affect participants’ perceptions and experiences.
Qualitative research uses purposeful sampling to intentionally select participants who have experienced the key concepts being explored in the research or belong to key target audiences. The intention here is not to generalize the information across broader populations but to gain insight into the research topics. Qualitative study samples are typically made up of a small number of participants who provide in-depth information about the areas of interest. It is common for the final sample size to not be determined until the study is conducted to ensure sufficient data have been collected and the process reaches saturation–the point where any new data collected does not add to understanding of the research topic.
Qualitative research techniques are particularly well suited for exploring complex topics in depth. The research design is often flexible and adaptable, allowing researchers to modify their approach based on emerging findings. This type of research often seeks to capture the perspectives and experiences of the participants directly. In qualitative research, context is crucial. Studies aim to understand the context in which events or behaviors occur, recognizing that meaning is often context-dependent.9
Common methods used in qualitative research include:
Interviews: The most common format of data collection in qualitative research. This method allows for the collection of in-depth information that will help you explain, better understand, and explore participants’ perceptions and experiences. Interviews are typically semi-structured or unstructured, allowing you to explore topics in depth and follow the natural flow of conversation.
Focus Groups: Group discussions focusing on a specific topic or set of topics under the guidance of a facilitator. This method aims to gather in-depth insights into people’s attitudes, perceptions, opinions, and experiences through group interaction and discussion. Focus groups typically consist of six to ten participants. Encouraging group interaction is crucial, as participants can react to and build upon each other’s responses.
Case Studies: In-depth analysis that focuses on a single unit (individual, group, event, etc.) of study. This method is particularly useful for exploring complex activities in real-life contexts and gaining a deeper understanding of the dynamics involved.
Content Analysis: A systematic examination of the content of documents, media, or other communication artifacts to identify patterns, themes, and trends. Content analysis is a versatile method that can be adapted to various research questions and contexts.
Observations: Systematic watching and recording of behaviors, events, or activities as they naturally occur in their real-world settings.
Like quantitative research, qualitative research is susceptible to various biases that can impact the validity and reliability of findings. It’s essential to be aware of potential biases and take steps to minimize their influence. Positioning refers to the process by which researchers critically examine and acknowledge their own perspectives, biases, and influence on the research process. It involves reflecting on how the researcher’s background, experiences, values, and assumptions may shape the research design, data collection, analysis, and interpretation of findings. This is a key element in ensuring transparency, rigor, and ethical conduct in qualitative research.
Below are some common issues9
- Researcher subjectivity – We all bring our own perspectives, beliefs, and values to our research, potentially influencing data collection, analysis, and interpretation. To address this bias, researchers can use positioning, acknowledging their own biases, and employing techniques such as peer debriefing to ensure a more objective analysis.
- Interviewer effect – The presence and behavior of the researcher can influence participant responses. Participants may alter their behavior or responses to align with perceived expectations. You can minimize this effect by establishing rapport with participants, being nonjudgmental, and using standardized protocols.
- Social desirability bias – Participants may provide responses that they perceive as socially desirable rather than expressing their true thoughts or behaviors. This can be addressed by creating a comfortable and nonjudgmental environment, ensuring confidentiality, and using multiple sources of data to validate findings.
- Cultural bias – Cultural differences between researchers and participants can lead to misunderstandings or misinterpretations of data. It’s crucial for researchers to be culturally sensitive, conduct pilot studies to ensure cultural appropriateness of instruments, and consider the impact of culture on participants’ perspectives.
Developing effective focus group topic guides and interview scripts is crucial for facilitating meaningful discussions and gathering valuable insights from participants. A well-structured guide or script helps ensure that the research objectives are addressed, and that the discussion flows smoothly. Here are some guidelines for development:
- Clearly define research objectives – This will guide the development of relevant topics.
- Introduction – Begin with a warm introduction and icebreaker questions to help participants feel comfortable and encourage open communication. These initial questions should be easy and non-threatening.
- Organize topics logically – Start with broad and introductory topics before moving into more specific and detailed ones.
- Use open-ended questions – Frame questions in an open-ended format to encourage participants to express themselves freely. Avoid leading questions that might bias responses. Question starters such as who, what, where, when, why, and how often lead to more elaborate responses. Be careful when crafting questions that start with why, as these may come across as accusatory.
- Probe for details – Include probing questions to encourage participants to elaborate on their responses. This helps uncover deeper insights and ensures a comprehensive exploration of each topic.
- Consider context – Be mindful of cultural, social, or environmental factors that may influence participant responses.
- Use neutral language – Avoid influencing participants’ responses by using neutral and non-biased language. Be aware of the potential impact of wording on participant reactions.
- Participants’ language – Frame questions using language familiar to participants.
- Sensitive topics – If addressing sensitive topics, approach them with care and consider placing them later in the process after rapport has been established.
- Visual aids – If applicable, incorporate visual aids to prompt discussion. This can be especially helpful in marketing research.
- Pilot test – Before conducting the focus group or interview, pilot test the guide or script with a small group to identify any issues with question wording, sequencing, or potential misunderstandings.
- Be flexible – Be prepared to adapt during the discussions. Participants may introduce unexpected perspectives or issues, and flexibility allows you to explore emerging themes.
- Closings – Reserve time at the end for participants to provide any final thoughts or comments.
- Ethical considerations – Include ethical considerations in your guide or script, such as obtaining informed consent and ensuring participant confidentiality.
A Multi-state Approach to Evaluate the Success of Programs (43 min.)
Seminar discussing the results of surveys to evaluate multiple, different programs held in several states.
Mixed Methods Research
Mixed methods combine qualitative and quantitative approaches into a single or multiphase study. These types of studies draw on the potential strengths of both qualitative and quantitative methods. They can be used to understand connections or contradictions better, provide participants with a strong voice and share their experiences across the research process, and facilitate different avenues of exploration. All of this tends to enrich results and enable questions to be answered more deeply. Mixed methods research comprises different design categories. The two most common of these are explanatory and exploratory designs. An explanatory design is a two-phase mixed methods approach. The first phase of the study collects and analyzes quantitative data. The second phase is designed to follow up the quantitative results with an in-depth qualitative study to explain why these results occurred and expand upon them. In an exploratory design, the research begins with the qualitative research phase, and then, building off the qualitative results, moves into the quantitative research phase to assess the overall prevalence of results found during the qualitative phase. It is strongly recommended that researchers have experience with both quantitative and qualitative research independently before undertaking mixed methods research.10
Best Practices Guide for Program Evaluation for Aquatic Educators 2012
This program evaluation guide was designed for aquatic educators. However, much is readily transferable to any R3 effort.
Original R3 Action Plan Appendices
The original R3 plan appendices. These help guide R3 program development and evaluation.
R3 Specific Evaluation & Social Science Training and Resources for the Modern R3 Practitioner
Evaluation Training Resources designed specifically for the R3 Practitioner.
Try, Test, Learn, Adapt: The Massachusetts Approach to R3 (55 min.)
This webinar includes an in-depth description of one state’s R3 program evolution and evaluation of their Learn to Hunt clinics. Provides an example of program evaluation.
Data Analysis
Evaluating programs is more than just collecting data. It involves analyzing the data and interpreting the results within the context of the program and its objectives. What story is the data telling, and how can you use it to make decisions about your program?
Conducting quantitative analysis involves applying statistical techniques to analyze numerical data and draw meaningful conclusions. These techniques allow researchers to quantify relationships, patterns, and trends that you can use to assess the effectiveness of programs and interventions. It helps measure outcomes, determine impact, and guide adjustments to improve program efficiency.
Here is a step-by-step guide for conducting a quantitative data analysis:
- Prepare and clean the data by addressing missing data and outliers and verifying that variables are correctly coded and formatted.
- Clearly articulate the research questions or hypotheses that guide the analysis, and specify the variables of interest and relationships you want to explore.
- Begin with descriptive statistics to summarize and describe the main features of your data. This might include measures such as mean, median, mode, standard deviation, and percentiles.
- Choose the significance level for your statistical tests (commonly set at 0.05). This represents the threshold for deciding whether results are statistically significant.
- Pick the appropriate statistical test based on the nature of your data and research questions. Common tests include t-tests for comparing means, chi-square tests for categorical data, ANOVA for comparing multiple groups, and regression analysis for exploring relationships.
- Perform the chosen statistical tests and ensure that the tests’ assumptions are met. See video here.
- Examine the results of statistical tests, paying attention to p-values, effect sizes, and confidence intervals.
- Interpret results in the context of your research questions. Based on your analysis, draw conclusions about the relationships or differences in your data, and discuss the practical significance of your findings.
- Acknowledge any limitations in your study and consider the generalizability of your findings to your target population.
Conducting quantitative data analysis using more advanced statistical tests requires understanding statistical methods and the specific context of your research. If you are unsure about the appropriate methods or interpretation of results, seek guidance from someone with background/experience in statistics.
Qualitative data analysis involves systematically analyzing and interpreting non-numerical data, such as text, audio, images, or video, to identify patterns, themes, and insights.
Here’s a step-by-step guide for conducting qualitative analysis11:
- Transcribe interviews or focus group discussions, if applicable, and organize your data systematically.
- Read or listen to your data multiple times to become familiar with the content, taking notes on initial impressions, recurring themes, and noteworthy patterns.
- Start the coding process by assigning descriptive labels (codes) to segments of your data. It is entirely possible and acceptable to code qualitative data by hand. However, if you have access to and are familiar with qualitative data analysis software, this can decrease the time spent coding.
- Create a coding scheme that includes a set of categories or themes. You will refine and expand the coding scheme as you proceed with the analysis.
- Systematically apply the codes to relevant sections of the data. You can ensure consistency in coding by establishing clear coding rules. Continuously compare new data with existing codes and categories. This iterative process helps refine categories and identify nuances. Group related codes into broader themes. Regularly review and refine your identified themes.
- Look for relationships and patterns within the coded data using visual representations to explore connections.
- Seek input from colleagues or peers to enhance the credibility of your analysis.
- Consider how the themes contribute to answering your research questions.
- Keep a detailed record of your analysis process, including decisions made, coding choices, and changes to the coding scheme. This record enhances transparency and allows for verification.
- Be open to alternative explanations and interpretation of your data. Think critically about your own biases and assumptions.
- Write up your findings, including quotes and examples. Clearly articulate the themes and patterns identified in your analysis.
Qualitative data analysis is often iterative. Be prepared to revisit and revise your analysis, especially in response to new insights or feedback. A team approach to qualitative data analysis offers several benefits. It can enrich the analysis by ensuring a more comprehensive and nuanced understanding of the data. The risk of bias is reduced by multiple team members independently reviewing and coding data, increasing the reliability of your findings. A team approach also allows for cross-checking, helping ensure consistency in coding and interpretation. Engaging in reflective conversations with team members can help you become more aware of your perspectives and potential influences on the analysis.
R3 Specific Evaluation & Social Science Training and Resources for the Modern R3 Practitioner
Evaluation Training Resources designed specifically for the R3 Practitioner.
Understanding Evaluation Results
Sometimes, it is “easy” to evaluate an R3 effort. A quick look at the metrics you assigned shows that your R3 effort significantly outperformed or underperformed your objective for it. Let’s say your goal was to reduce range permit churn by 5%, and the effort you implemented reduced it by 10%. Your program was a success, and you can continue as planned. On the other hand, if your effort has not reduced churn, you need to revise it and try again or cancel the effort and try something different.
The most important thing to learn about evaluation is that failure lies not in canceling an effort that is not working but in not evaluating a failing effort and allowing it to continue.
Sometimes, the results of an evaluation are not obvious. In those cases, determining the meaningfulness of evaluation results involves assessing the findings’ relevance, significance, and validity within the context of the evaluation’s objectives. Interpretation goes beyond just reporting the data. It adds context, meaning, and an understanding of the practical significance of the findings. Ask yourself: what do the results mean? Why did they turn out this way? Do they have implications for the individual program or the overall R3 initiative? Your interpretation should cover how the results address your evaluation objectives and outcomes. Acknowledge if there are any unexpected results. These can offer insights and inform future planning and improvements. In addition to the data interpretation, a part of evaluation is determining the degree to which the results provide insights into whether the R3 effort performed as expected or how it compares to established criteria.
By their very design, many of the steps involved in qualitative data analysis help determine if the results are meaningful. However, for many quantitative statistical tests, the statistical significance alone cannot tell you if the results are meaningful. In some cases, large sample sizes can impact the statistical tests, making it easier to find statistical significance. It is important to not just focus on the statistical significance but to also look at effect size or measures of association. These can help you determine if there are not only statistically significant differences, but also if those differences are meaningful. The many different measures of association and the choice of which to use depends on the variables being studied and the assumptions of the statistical test being used. You should select measures appropriate for your data type and research questions.
R3 Specific Evaluation & Social Science Training and Resources for the Modern R3 Practitioner
Evaluation Training Resources designed specifically for the R3 Practitioner.
Communicating Evaluation Results
Communicating your evaluation results effectively is critical for ensuring that others understand the findings and can make informed decisions based on your evaluation. With the diversity in audiences who may be interested in your results, it is difficult to produce a single final report that will meet the needs of all the users. For example, less is more for executives, legislators, and media. You will want to use many visuals, minimal text, and very few pages. Distill it down as much as possible. For a technical audience, you can provide more details. You must understand your audience(s) and tailor your communication to their level of expertise and interest.
The following tips can help with communicating your findings across a wide range of audiences:
- Narratives help make data more engaging. Tell the story of your program’s impact. Connect your findings with your program’s objectives and intended outcomes. Start by giving the context for your evaluation, why you conducted it, why this evaluation is important to your program and organization, and why your audience should be interested in the results.
- Depending on their interest, you may want to tell your audience how you collected the data and from whom, what kinds of analyses you performed and why, and clearly describe any data limitations.
- Identify and highlight the most important findings.
- Present your findings in an accessible manner. Use clear and straightforward language, avoiding jargon and technical language.
- Provide clear and actionable recommendations to help illustrate how the findings can be used to improve the program and the R3 community.
- Adjust your approach based on the needs of your audience. Present your results using a variety of communication channels, such as presentations, reports, and interactive discussions, to reach different audiences most effectively.
- Verbal presentations help communicate research results to audiences and have the advantage of communication being more interactive.
- Reports, executive summaries, handouts, or slides used in verbal presentations benefit from accompanying visual aids.
- Visual aids like charts, graphs, and tables make data more accessible and easier to interpret. Visuals can often convey complex information more effectively than text alone. They can be used to illustrate trends, patterns, and key findings.
Good charts and graphs are easy to understand, present findings clearly, summarize information and need little interpretation. Remember that the goal of a chart or graph is to simplify complex information and ease understanding. Know your audience and tailor your charts and graphs to meet their needs and preferences. Some tips for creating charts and graphs are:
- Choose a chart that best represents the type of data you have and can tell the data’s story. Common types include bar charts, line charts, pie charts, and scatter plots.
- Charts and graphs should contribute to your audience’s understanding of the information and not be redundant to the accompanying text. All information needed to interpret a graph or chart must be included in its caption.
- Keep your charts and graphs simple and uncluttered. Avoid unnecessary gridlines, three-dimensional graphs, or labels that don’t add to the main message. Focus on the essential elements.
- Provide a clear and descriptive title that summarizes the main point of the chart or graph. Clearly label the axes with descriptive and easy-to-understand titles. If appropriate, include units of measure in the axes’ titles. Ensure labels and tick marks are legible.
- Use color purposefully to highlight important information. Too many colors may be hard to differentiate between and confuse your viewers.
- Ensure charts and graphs are accessible to diverse audiences. Use accessible color combinations, provide alternative text for images, and use patterns or textures in addition to color.
While figures present findings directly, tables require the viewer to analyze the components to understand the message. Tables are suitable for presenting certain types of data, but it’s important to use them judiciously. Consider the complexity of the data and your audience’s needs when deciding whether a table is needed. Tables can be appropriate when displaying structured information, comparisons, or specific details. Some tips for creating tables are:
- Organize the data logically, making it easy for readers to follow.
- Include a concise and informative title that summarizes its contents. The title should clearly communicate what information the table presents.
- Provide clear and descriptive headings for each column and row that reflect the contents in the corresponding cells.
- Consider using alternate row shading to make it easier to follow rows across tables.
- Minimize merging cells, which can complicate the reading process and create confusion.
- Include units of measurement for numerical data to avoid ambiguity.
- Limit the number of decimal places to avoid unnecessary precision.
- Include totals and subtotals for numerical data to provide a quick summary if applicable.
- Consider breaking the data into smaller, related tables if the data is extensive.
- Design tables with accessibility in mind. This ensures the table structure is logical when read by screen readers and provide alternative text for tables.
Good visuals can significantly improve the communication of results. Slides should complement your presentation, not replace it. Use slides to help convey your message effectively and keep your audience engaged.
- Keep it simple. Simplicity makes it easier for your audience to grasp your message. Limit each slide to one main point.
- Use a uniform color scheme, font, and layout to maintain.
- Avoid using busy backgrounds that can distract. Use light text on a dark background or vice versa.
- Guide your viewers’ attention with visual cues. Make important points more prominent through font size or contrast.
- If using images, ensure they are high-quality and relevant to your content.
- Use animations and transitions sparingly. These should enhance your presentation and not detract from the content.
You can find several online resources geared toward improving presentations. One resource often referenced is a TEDx video How to Avoid Death by PowerPoint by David Phillips.
A Multi-state Approach to Evaluate the Success of Programs (43 min.)
Seminar discussing the results of surveys to evaluate multiple, different programs held in several states.
Best Practices Guide for Program Evaluation for Aquatic Educators 2011
This program evaluation guide was designed for aquatic educators. However, much is readily transferable to any R3 effort. Sections 5.1-5-13 guide communicating your evaluation results
How to Avoid Death by PowerPoint
This TEDx talk discusses ways to make your presentations more engaging and professional.
Hunting, Fishing, Sport Shooting, and Archery Recruitment, Retention and Reactivation: A Practitioner’s Guide
An expansive compilation of R3 survey/assessment results. These sections will help practitioners communicate evaluation results. PG. 66-94, 325-338
RBFF-Georgia New Angler Retention Pilot Program
This project report provides a simple yet detailed evaluation of a campaign to increase renewal rates for first-time anglers.
- Chase, L. & Dunfee, M. (2022). The new future of hunting and fishing. A Multi-State Conservation Grant. Technical Report presented to the Association of Fish and Wildlife Agencies, https://find.nationalr3community.org/media/?mediaId=F580F7AF-F93A-4861-9B49A85DD3A5A93B [↩]
- Currently in draft format. Scott Lavin, WAFWA R3 Committee Chair & Matt Dunfee, WMI are the document keepers at this point in time. [↩]
- Byrne and Dunfee, 2018. “Evolution and Current Use of the Outdoor Recreation Adoption Model. [↩]
- Groves, R. M. (2005). Survey errors and survey costs. John Wiley & Sons; Salant, P., & Dillman, D. A. (1994). How to conduct your own survey. John Wiley & Sons; Babbie, E. (2004). The Practice of Social Research. 10th edition. Wadsworth, Thomson Learning; Vaske, J. J. (2008). Survey research and analysis: Applications in parks, recreation and human dimensions. Venture Publishing; Dillman, D. A., Smyth, J. D., & Christian, L. M. (2014). Internet, phone, mail, and mixed-mode surveys: The tailored design method. John Wiley & Sons. [↩]
- Groves, R. M. (2005). Survey errors and survey costs. John Wiley & Sons; Salant, P., & Dillman, D. A. (1994). How to conduct your own survey. John Wiley & Sons; Vaske, J. J. (2008). Survey research and analysis: Applications in parks, recreation and human dimensions. Venture Publishing; Dillman, D. A., Smyth, J. D., & Christian, L. M. (2014). Internet, phone, mail, and mixed-mode surveys: The tailored design method. John Wiley & Sons. [↩] [↩]
- Groves, R. M. (2005). Survey errors and survey costs. John Wiley & Sons; Salant, P., & Dillman, D. A. (1994). How to conduct your own survey. John Wiley & Sons; Vaske, J. J. (2008). Survey research and analysis: Applications in parks, recreation and human dimensions. Venture Publishing; Dillman, D. A., Smyth, J. D., & Christian, L. M. (2014). Internet, phone, mail, and mixed-mode surveys: The tailored design method. John Wiley & Sons. [↩]
- Vaske, J. J. (2008). Survey research and analysis: Applications in parks, recreation and human dimensions. Venture Publishing; Dillman, D. A., Smyth, J. D., & Christian, L. M. (2014). Internet, phone, mail, and mixed-mode surveys: The tailored design method. John Wiley & Sons. [↩]
-
Salant, P., & Dillman, D. A. (1994). How to conduct your own survey. John Wiley & Sons;
Vaske, J. J. (2008). Survey research and analysis: Applications in parks, recreation and human dimensions. Venture Publishing;
Dillman, D. A., Smyth, J. D., & Christian, L. M. (2014). Internet, phone, mail, and mixed-mode surveys: The tailored design method. John Wiley & Sons. [↩]
-
Creswell, J. W. (2007). Qualitative inquiry and research design: Choosing among five approaches. 2nd edition. Sage Publication;
Creswell, J. W. and Clark, V. L. P. (2018) Designing and Conducting Mixed Methods Research. 3rd edition. Sage Publications. [↩] [↩] [↩]
-
Tashakkori, A., & Teddlie, C. (1998). Mixed methodology: Combining qualitative and quantitative approaches (Vol. 46). sage;
Creswell, J. W. and Clark, V. L. P. (2018) Designing and Conducting Mixed Methods Research. 3rd edition. Sage Publications. [↩]
- Creswell, J. W. (2007). Qualitative inquiry and research design: Choosing among five approaches. 2nd edition. Sage Publication; Richards, L. (2014). Handling qualitative data: A practical guide. Sage Publications: Creswell, J. W. and Clark, V. L. P. (2018) Designing and Conducting Mixed Methods Research. 3rd edition. Sage Publications. [↩]