Americi Org Capacity Assessment- Culture tips
Americi Org Capacity Assessment- Culture tips
Americi Org Capacity Assessment- Culture tips
Assessment Tool
AmeriCorps Office of Research and Evaluation
October 2017
This tool was developed for the Corporation for National and
Community Service (2020, DBA AmeriCorps) by ICF under
contract #CNSHQ16T0073.
Contributing Authors:
Adrienne DiTommaso
Bethany Slater
Joe Raymond
Venessa Marks
Nanette Antwi-Donkor
Trevor Hoffberger
Suggested citation: AmeriCorps. (2017). Organizational Capacity Assessment Tool. Washington, DC:
Author.
AmeriCorps thanks the following people for serving on a technical working group and advising the
authors in the development of this tool: Isaac Castillo, Robert Cox, Meghan Duffy, and Chukwuemeka
Umeh. AmeriCorps would also like to thank the following staff for their contribution to the development
of this tool: Jennifer Kerner, Lily Zandniapour, Anthony Nerino, Carla Ganiel, Molly Pelzer, and Rob Cox.
AmeriCorps also thanks the many organizations that participated in the pilot testing and validation of
this tool.
Office of Research & Evaluation
Table of Contents
Introduction ........................................................................................................... 2
Key Domains of Organizational Capacity ....................................................................... 3
Using This Tool ................................................................................................................ 4
Leadership Capacity .......................................................................................................... 6
Vision and Mission ........................................................................................................... 7
Leadership and Governance ........................................................................................... 7
Strategy and Planning ..................................................................................................... 8
Culture and Values .......................................................................................................... 8
Management and Operations Capacity ........................................................................ 10
Financial Management .................................................................................................. 11
Human Resources .......................................................................................................... 11
Infrastructure and Information Technology.................................................................. 12
Community Engagement Capacity ................................................................................ 13
Fund Development ........................................................................................................ 14
Communications and Advocacy ................................................................................... 14
Volunteer Management ................................................................................................ 14
Community Partnerships ............................................................................................... 15
Service Capacity ............................................................................................................... 17
Program Design ............................................................................................................. 18
Program Implementation .............................................................................................. 18
Performance Management ........................................................................................... 19
Evaluative Capacity.......................................................................................................... 21
Evaluation Planning ....................................................................................................... 22
Data Collection .............................................................................................................. 22
Measuring Outcomes and Impact ................................................................................ 22
Learning and Continuous Improvement ....................................................................... 23
Appendix: Scoring Rubric ............................................................................................... 25
References ........................................................................................................................ 28
1
Office of Research & Evaluation
Introduction
AmeriCorps, its State Service Commissions, and intermediaries work with thousands of direct
service providers each year to improve the lives of American citizens. This tool was created for
AmeriCorps grantees and the broader field to assess organizational capacity to deliver
effective services. High-performing organizations typically have a strong understanding of
their organization’s strengths and challenges. This tool provides a practical method of
organizational self-assessment that can be used to acknowledge strengths, clarify different
perceptions, and plan strategies to enhance capacity in identified areas. This introduction
explains the intended use of this tool and highlights the five domains of organizational
capacity assessed by the instrument. For each domain, we offer a brief synopsis of the
research literature on effective practice followed by a series of capacity assessment questions.
Each domain also includes suggested reading and research to build capacity in that area.
Adapted from Grantmakers for Effective Organizations. (2016). Strengthening nonprofit capacity: Core concepts
in capacity building. Washington, DC: Author.
The goal of this tool is to provide AmeriCorps grantees and the broader field with a
research-based instrument to promote organizational capacity self-assessment. State
commissions and other intermediaries may find this tool particularly helpful in working with
subrecipients to identify capacity strengths and areas for support. The tool is designed to be a
conversation- starter within an organization and between organizations engaged in a technical
assistance relationship.
2
Office of Research & Evaluation
Leadership Capacity
Vision/Mission, Governance, Strategy/Planning, Culture/Values
Service Capacity
Program Design
Program Implementation
Performance Management
Management & Operations
External Focus
Internal Focus
Capacity
Financial Management
Human Resources Community
IT & Infrastructure Engagement Capacity
Fund Development
Communications & Advocacy
Volunteer Management
Community Partnerships
Evaluative Capacity
Evaluation Planning, Data Collection, Measuring Impact, Learning &
Continuous Improvement
3
Office of Research & Evaluation
Prepare
• Identify at least 2-3 individuals within your organization to independently take this
assessment. This could include the CEO, members of the board of directors,
leadership team members, or even well-informed external stakeholders.
• Set a deadline for respondents to complete the assessment and schedule a
meeting to debrief your findings and conclusions.
Assess
• Take the assessment. Check off any statements that are true for your organization
from your perspective. Note any questions that are not applicable to your
organization.
• Tally the number of checked statements at the end of each domain and the
number of skipped questions if any were not applicable.
• Complete the scoring rubric in Appendix A. Identify the domains where your
organization has the strongest capacity and domains that could benefit from
capacity building.
Reflect
• Discuss your conclusions with the larger respondent team. Dig into any areas of
disagreement, seeking to understand and address differences of opinion.
• Share your group’s key findings, conclusions, and any action items with the larger
staff and board.
4
Office of Research & Evaluation
Appendix A offers a scoring rubric to help you identify domains and subdomains of capacity
that might particularly benefit from capacity-building efforts. To simplify and streamline
scoring, all questions are framed negatively – requiring you to simply check off whether a
specific capacity is a challenge or a gap for your organization.
This tool has been validated for use with a wide variety of organization types: national and
local nonprofits; state, local and tribal governments; institutes of higher education; and
funders and intermediary organizations. If a question is not appropriate for your organization,
simply skip that question and note its exclusion in your scoring calculation.
The tool was also designed to help organizations assess changes to capacity over time.
Consider taking and retaking this assessment on an annual or biannual basis to track how
organizational capacity strengths and needs change over time.
Capacity building takes time and effort. This capacity tool can be a critical first step toward
increasing basic understanding about capacity and prioritizing potential capacity-building
efforts. The suggested resources at the end of each domain section provide a helpful starting
place to learn more about effective practices for organizational development.
5
Office of Research & Evaluation
Leadership Capacity
This domain focuses on capacity functions that are typically the responsibility of senior
leadership and executive board members (in the case of nonprofits) to guide or execute.
Vision and Mission: An organization’s vision and mission statements articulate its sense
of purpose and direction (McKinsey & Company, 2001). Effective vision and mission
statements set parameters for what the organization will and will not do; inspire staff,
volunteers, and donors; and set the basis for strategy (McKinsey & Company, 2001;
Paynter & Berner, 2014; Smith, Howard, & Harrington, 2005).
Leadership and Governance: An organization’s governance model is a critical
component for organizational functioning and sustainability (Liket & Mass, 2015). For
nonprofits with executive boards, clear separation between the board and the
organization’s leadership is important, as are documented roles and responsibilities
(Liket & Mass, 2015). Research suggests that professional diversity, the ability to raise
funds, and the size of the board can affect nonprofit effectiveness. Note: Organizations
that do not have an executive board or suitable proxy should mark that question as not
applicable (N/A).
Strategy and Planning: An organization’s vision and mission establish its aspirations,
but its strategy articulates the means for achieving those goals (McKinsey & Company,
2001). Research has shown that strategic planning – the process of mission review,
stakeholder analysis, and visioning, coupled with the development of resource
allocation strategies, boosts organizational capacity (Bryson, Gibbons, & Shaye, 2001;
Paynter & Berner, 2014).
Culture and Values: An organization’s culture affects every aspect of its functioning –
from how leaders interact with the board and staff to how staff members respond to
external or internal challenges. Building a strong values-based culture is a strategic and
often difficult process that must be led and modeled by organizational leadership.
Organizational culture is typically divided into three interrelated components: core
values, beliefs, and behavior norms (McKinsey & Company, 2001). Cultural
competency, diversity, equity, and inclusion are critical components of a strong
organizational culture.
6
Office of Research & Evaluation
Instructions: Read each statement and check the True box to the left if it is true or mostly
true for your organization. If the statement does not apply to your organization (e.g., it
refers to a governing board practice, and you are a school and do not have a traditional
governing board or suitable proxy), check the N/A box to the right. If the statement is not
true for your organization, check the Not True box and proceed to the next statement.
Note that selecting all N/A boxes within a domain will produce a form error.
Not
True N/A True
1.2 Our mission statement does not clearly define what we want to achieve
and for whom.
1.3 Not all staff fully embrace or could clearly describe our vision and
mission to individuals who have never heard of our organization.
1.4 Organizational decisions are sometimes not reflective of the mission and
vision of the organization and detract from its fulfillment.
1.6 Our board does not adopt and regularly review an annual set of
organizational strategic goals and measurable outcomes.
1.7 Our board does not adopt an annual budget aligned with its strategic goals
and measurable outcomes.
1.8 Our board does not regularly update and adopt a set of policies to govern
the organization in the areas of finance, human resources, fund
development, and communications.
1.9 Our board does not evaluate the performance of its CEO on regular basis.
1.10 Our board does not evaluate its performance on a regular basis.
1.11 Our board does not have the right mix of skills and expertise to govern the
organization and routinely consider diverse points of view from internal and
external stakeholders.
1.12 The composition of our board does not reflect the community we serve.
1.13 Board members do not have enough knowledge about our organization and
current issues relevant to our organization to make effective policy decisions.
7
Office of Research & Evaluation
1.14 Few or none of the board members are effective at getting others in the
community to invest time, money, or other resources in our organization.
1.16 Our organization does not formally share progress on the strategic plan’s
goals and objectives with board and staff members on a regular basis.
1.17 Our organization either did not solicit or did not use external stakeholder
input as it developed its strategic plan.
1.18 Our board either has not reviewed or has not approved the existing
strategic plan in the past 12 months.
1.19 Our organization has too many priorities, and our capacity is insufficient or
stretched too thin to achieve all of our goals.
1.21 Our overall strategy is not broadly known and has limited influence over day-
to-day behavior.
1.22 There is a lack of clarity on how to make decisions when priorities come into
conflict with each other.
1.25 Our organization does not provide regular opportunities for staff to express
constructive feedback or concerns to leadership.
1.26 Many staff members are not culturally sensitive with respect to internal
management or delivery of services.
1
A strategic plan is a documented framework that communicates an organization’s goals, sets priorities,
and focuses energy on actions that accomplish those goals (Balanced Scorecard Institute, n.d.).
2
Strategic goals are the realistic and clearly defined outcomes that guide implementation of a program
or intervention (The NCJA Center for Justice Planning, n.d.).
3
Strategic objectives are concrete explanations of how goals will be accomplished and the necessary
steps to reach that end (The NCJA Center for Justice Planning, n.d.).
8
Office of Research & Evaluation
1.28 Our organization does not openly embrace diversity of race, ethnicity, class,
gender, sexuality, ability, and other facets of human identity.
1.29 The demographics of our staff do not represent the population it serves.
Leadership Capacity 0 0
• The strategic plan is dead. Long live strategy, by Dana O’Donovan and Noah Rimland
Flower. Stanford Social Innovation Review. January 10, 2013.
• Boards that make a difference: A new design for leadership in nonprofit and public
organizations, by John Carver. December 10, 2007.
• Trying hard is not good enough: How to produce measurable improvements for
customers and communities, by Mark Friedman. March 8, 2015.
9
Office of Research & Evaluation
10
Office of Research & Evaluation
Instructions: Read each statement and check the True box to the left if it is true or mostly
true for your organization. If the statement does not apply to your organization (e.g., it
refers to a governing board practice, and you are a school and do not have a traditional
governing board or suitable proxy), check the N/A box to the right. If the statement is not
true for your organization, check the Not True box and proceed to the next statement.
Note that selecting all N/A boxes within a domain will produce a form error.
Not
True N/A True
Financial Management
2.1 Our organization does not have an up-to-date fiscal policy and procedures
manual.
2.2 Our organization does not compare actual with budgeted expenses each
month.
2.3 Our operations plan and annual budget do not align with our current
strategic plan.
2.4 Our organization rarely reforecasts year-end revenue and expenses to assist
management decision-making.
2.5 Our organization does not effectively manage its finances (e.g., it does not
have balanced books, appropriate internal controls, on-time accounts
payable, or an adequate reserve fund, or it has year-over-year deficits).
Human Resources
2.6 Our organization does not have written human resource policies that have
been approved by the board and explained to staff.
2.7 Staff members are not given constructive feedback from managers or
supervisors on a regular basis.
2.8 Our organization does not routinely assess workloads to ensure adequate
resources are available to meet performance objectives.
2.9 Our organization does not have an adequate total compensation system4,
including salary standards, retirement benefits, health care benefits, and
systems for bonuses, awards, or recognition of high performance, that is on
par with similar organizations.
2.10 Our organization does not fill open positions with highly qualified applicants
in a timely manner.
4
Total compensation is a holistic model of employee payment that incorporates both monetary
compensation (such as base pay, performance-based pay, and bonuses) and nonmonetary
compensation (such as health care benefits, trainings, and retirement benefits) (Sharpe, 2016).
11
Office of Research & Evaluation
2.12 Our organization does not have sufficient expertise (on staff or through
volunteers or consultants) to effectively and efficiently run and manage our
technology systems.
2.13 Our staff members do not have the necessary hardware (e.g., computers)
and software (e.g., word processing systems and database systems) to do
their jobs consistently, efficiently, and effectively.
2.14 Important data and files are not backed up at least once a month.
• Managing to change the world: The nonprofit manager’s guide to getting results, by
Alison Green and Jerry Hauser. 2012.
• An executive director’s guide to financial leadership, by Kate Barr and Jeanne Bell. The
Nonprofit Quarterly. Fall/Winter 2011.
• Financial management for human service administrators, by Lawrence Martin. May 5,
2016.
12
Office of Research & Evaluation
Fund Development: The lack of core, stable, long-term funding is often noted as the
greatest challenge to the development of organizational capacity (Hall et al., 2003).
Uncertainties about future funding and constraints on how funds can be used can have
a significant impact on the ability of an organization to plan strategically – or to
execute those plans (Misener & Doherty, 2009). Organizations that are mature in their
fund development capacity have provisions for covering overhead costs and routine or
formal fundraising activities (such as annual campaigns or events) and have a diverse or
strategic array of funding sources (Schuh & Leviton, 2006).
Communications and Advocacy: Increasingly in the digital age, effective and
transparent communications are considered essential to nonprofit effectiveness (Liket
& Mass, 2015). Communications capacity includes marketing skill, online presence,
media relations, and use of social media to raise awareness, advocate, and attract
resources to the organization or issue (Grantmakers for Effective Organizations,
2016b). Transparency is often judged by posting the organization’s strategic plan and
annual and financial reports online and by providing a list of executive board members
on the organization’s website (Liket & Mass, 2015).
Volunteer Management: Many small community-based nonprofits, as well as larger
organizations, rely on volunteers to deliver services or cover other essential staff
functions. For some small community-based organizations, the commitment of
volunteers can be more important than other capacity areas, such as infrastructure
(Paynter & Berner, 2014). Effective volunteer management requires the development
and execution of effective recruitment, screening, training, and retention strategies.
Community Partnerships: Partnership capacity includes the skills and mindset to
create and sustain relationships with peer organizations, government, corporations,
and other key stakeholders to advance the organization’s mission (Grantmakers for
Effective Organizations, 2016b). Many direct service providers rely on organizations
with complementary services to meet the needs of their clients. Volunteer-based
organizations often heavily rely on corporations or religious organizations to help
recruit volunteers or provide in-kind donations.
13
Office of Research & Evaluation
Instructions: Read each statement and check the True box to the left if it is true or mostly
true for your organization. If the statement does not apply to your organization (e.g., it
refers to a governing board practice, and you are a school and do not have a traditional
governing board or suitable proxy), check the N/A box to the right. If the statement is not
true for your organization, check the Not True box and proceed to the next statement.
Note that selecting all N/A boxes within a domain will produce a form error.
Not
True N/A True
Fund Development
3.1 Our organization would shut down or dramatically reduce services if it lost
1-2 key funders.
3.4 Our organization does not have a viable fundraising plan that was
developed within the past 12 months.
3.7 Our materials or website do not reflect the quality of our organization.
3.9 Our organization leaders are rarely asked by other community or nonprofit
leaders to provide leadership, knowledge, or advice on community-level
issues.
Volunteer Management
3.10 Our organization does not have a written volunteer recruitment and
management plan.
3.11 Our organization often fails to recruit the volunteers it needs to provide
essential services.
5
A communications strategy is a document that establishes the objectives, audiences, messages,
resources, responsibilities, and measures for an organization’s outreach. The objectives in a
communication strategy should be segmented by target audience (Hovland, 2005).
14
Office of Research & Evaluation
3.15 Volunteers do not always receive the resources, support, and training they
need to do their jobs.
3.16 Our organization often struggles to recruit the right mix of volunteers (e.g.,
with the right skills, availability or with backgrounds that reflect the
community).
Community Partnerships
3.17 Our organization spends insufficient time meeting, interacting, and
collaborating with community members, program participants, and leaders
for the purpose of learning about what is going on in the community.
3.19 Our organization has spent so much time on partnership work that it
interferes with our ability to implement important goals.
3.20 Our organization has focused efforts on partnership work or networking that
is not aligned with our mission.
3.21 Our organization has not assessed the results of key partnerships, alliances,
or participation in networks.
15
Office of Research & Evaluation
• Ten nonprofit funding models, by William Foster, Peter Kim, and Barbara Christiansen.
Stanford Social Innovation Review. Spring 2009.
• Twenty-first-century communications versus the illusion of control: An epic battle, by
Ruth McCambridge. Nonprofit Quarterly. August 27, 2014.
• Working better together: Building nonprofit collaborative capacity, by Grantmakers for
Effective Organizations. 2013.
• Management of human service programs, by Judith A. Lewis, Thomas R. Packard, and
Michael D. Lewis. August 15, 2011.
16
Office of Research & Evaluation
Service Capacity
This domain focuses on the capacity of the organization to design research-informed
programs, monitor and support quality implementation, and make course corrections as
needed.
Program Design: Programs are more likely to produce reliable, positive outcomes for
their clients if they use evidence-based practices and have a clearly articulated logic
model or theory of change (Easterling & Metz, 2016). A critical element in strong
program design includes taking steps to understand and document relevant
community and individual-level needs and assets.
Community needs assessment, asset mapping, and focus groups with potential clients
and key stakeholders are all strategies that can assist organizations in designing (or
refining) programs that are responsive to client needs and the larger community
environment (Sharpe, Greaney, Lee, & Royce, 2000).
Program Implementation: Program implementation is more effective and sustainable
if it is documented, monitored, and well-coordinated with other program or
organizational functions. Policy and procedure manuals provide evidence of a
structured, step-by-step approach to programming and are an essential knowledge
and risk management tool (Paynter & Berner, 2014). Coordination across functional
teams or other interagency programs can keep programs from operating in isolation
and reduce inefficiencies. Finally, monitoring fidelity to policies and practices or to
evidence-based programs (if applicable) is essential to ensure that programs provide
the intended services (Easterling & Metz, 2016).
Performance Management: Similar to evaluative capacity, performance management
capacity focuses on the organization’s ability to identify, collect, and monitor key
performance indicators (KPIs) directly related to service provision. These KPIs are
typically program activities and outputs that provide real-time input on program
implementation and client participation (Parmenter, 2015).
Note: if your organization runs multiple programs, be sure to align with your colleagues also
taking the assessment on which program(s) you are focusing on as you complete the
questions.
17
Office of Research & Evaluation
Instructions: Read each statement and check the True box to the left if it is true or mostly
true for your organization. If the statement does not apply to your organization (e.g., it
refers to a governing board practice, and you are a school and do not have a traditional
governing board or suitable proxy), check the N/A box to the right. If the statement is not
true for your organization, check the Not True box and proceed to the next statement.
Note that selecting all N/A boxes within a domain will produce a form error.
Not
True N/A True
Program Design
4.1 Our organization does not have a clear understanding of how our resources
and strategies will result in our intended outcomes.
4.2 Our program design is not grounded in the best and most recent research
literature available.
4.6 Our organization does not conduct regular assessments of client needs.
4.7 Our new programs are created largely in response to funding availability
rather than client needs or community service gaps.
Program Implementation
4.8 Policy and procedure6 documents are out-of-date or insufficient to provide
staff guidance on current program practices.
4.10 Not all of our program staff has the required knowledge, experience, or skills
to implement our program in a manner that will achieve the greatest positive
effect.
4.11 Staff members with different roles rarely have the time to meet and share
their work, coordinate their work, or develop ideas for working together.
6
Policy and procedure documents define how an organization operates and provide guidance on
program-specific practice (NCVO Knowhow Nonprofit, 2016).
18
Office of Research & Evaluation
4.12 Program leadership does not regularly monitor fidelity to program design7 or
adaptations 8 made to implementation.
4.13 Staff members do not have a clear understanding of the program logic
model 9 or the relationship between implementation and expected
outcomes.
Performance Management
4.14 Our program does not have clearly defined key performance indicators.10
4.16 Internal performance data are rarely used to improve the program or
organization.
4.17 Our organization rarely or never compares our program performance with
comparable programs run by other organizations.
Service Capacity 0 0
7
Fidelity is the “extent to which delivery of an intervention adheres to the protocol or program model
originally developed” (Mowbray, Holter, Teague, & Bybee, 2003). Providing consistent services is
important for evaluating impact and making adjustments.
8
Program adaptations are data-driven changes to implementation to ensure sustainability and
effectiveness (Center for Public Health Systems Science, n.d.).
9
A logic model is a visual and written depiction of the inputs and activities that will result in the desired
outputs and outcomes (W.K. Kellogg Foundation, 2004).
10
A key performance indicator is a quantifiable performance measurement that indicates the
effectiveness of a program or organization in achieving its goals (Jackson, 2015).
19
Office of Research & Evaluation
• Within our reach: Breaking the cycle of disadvantage, by Lisbeth B. Schorr. March 23,
2011.
• Designing and managing programs: An effectiveness-based approach, by Peter Kettner,
Robert Moroney, and Lawrence Martin. January 20, 2016.
20
Office of Research & Evaluation
Evaluative Capacity
This domain focuses on the capacity of an organization to gather data, measure impact, and
assess lessons learned to strengthen the organization’s work over time.
21
Office of Research & Evaluation
Instructions: Read each statement and check the True box to the left if it is true or mostly
true for your organization. If the statement does not apply to your organization (e.g., it
refers to a governing board practice, and you are a school and do not have a traditional
governing board or suitable proxy), check the N/A box to the right. If the statement is not
true for your organization, check the Not True box and proceed to the next statement.
Note that selecting all N/A boxes within a domain will produce a form error.
Not
True N/A True
Evaluation Planning
5.1 Within the past three years, our organization has not developed or not
revisited a systematic plan that defines the purpose of our evaluation efforts
and our methodology, outlines our evaluation activities, and establishes clear
responsibilities.
5.2 Our senior leadership does not prioritize evaluation and does not routinely
dedicate resources to it.
Data Collection
5.5 Our organization does not have clear protocols 11 for data collection.
5.6 Our organization does not provide regular staff training on how to use data
collection protocols.
5.7 Our organization does not have sufficient or effective data collection
systems. 12
5.9 The questions in our evaluation instruments 13 are not clearly stated.
11
Data collection protocol is the systematic procedure through which individuals and organizations
collect, maintain, secure, and use data. Protocols ensure that evaluations are effective and valid (Faculty
Development, 2005).
12
Data collection systems, typically using computer-based software, aggregate and analyze sets of data
in an efficient manner (Techopedia, n.d.).
13
An evaluation instrument is a questionnaire or survey that assesses knowledge gain or behavior
change in a group of program participants (Rutgers University, n.d.).
22
Office of Research & Evaluation
5.10 The questions in our evaluation instruments are not in-line with our
proposed methods of evaluation and program design.
5.13 Our organization does not openly and widely share evaluation findings with
key stakeholders. 17
5.14 Our organization makes limited use of internal evaluation data to make
decisions regarding organization strategy or fiscal allocations.
5.15 Our organization makes limited use of external research to make decisions
regarding organization strategy or fiscal allocations.
Evaluative Capacity 0 0
Use this space for thoughts and reflections about this domain.
What is your organization doing well?
14
A quasi-experimental study compares outcomes for individuals receiving an intervention with
outcomes for comparable individuals not receiving that intervention (Moore, 2008).
15
A randomized control trial randomly assigns individual participants to either a control or treatment
group to measure the impact of an intervention on specific outcomes (Himmelfarb Health Sciences
Library, n.d.).
16
For results to be attributed to program interventions, a causal relationship must exist between them,
effectively ruling out other variables as the primary cause (Australian Bureau of Statistics, 2013).
17
Key stakeholders are individuals or organizations that share an interest in the program’s success.
Stakeholders can be funders, partners, community members, participants, board members, or
volunteers (The Denver Foundation, n.d.).
23
Office of Research & Evaluation
• The challenge of organizational learning, by Katie Smith Milway and Amy Saxton.
Stanford Social Innovation Review. Summer 2011.
• Building a strategic learning and evaluation system for your organization, by Hallie
Preskill and Katelyn Mack. 2013.
• Collective genius, by Linda Hill, Greg Brandeau, Emily Truelove, and Kant Lineback.
Harvard Business Review. June 2014.
• Building evaluation capacity: Activities for teaching and training, by Hallie Preskill and
Darlene Russ-Eft. September 15, 2015.
24
Office of Research & Evaluation
Leadership Capacity 0 29 0 29 0%
Management and
0 14 0 14 0%
Operations Capacity
Community
0 21 0 21 0%
Engagement Capacity
Service Capacity 0 17 0 17 0%
Evaluative Capacity 0 16 0 16 0%
Scoring Instructions:
Figure: AmeriCorps Organizational Capacity Assessment Tool Domains
1. Tally the number of boxes marked
Not True within each domain and Leadership Capacity
record it in the Number of Vision/Mission, Governance,
Statements Marked Not True Strategy/Planning, Culture/Values
column.
Operations
3. Subtract the Number of Statements Capacity
Marked N/A from the Total Number Financial Management Community
of Statements column to determine Human Resources Engagement Capacity
IT & Infrastructure
your Total Applicable Statements Fund Development
count for each domain. Communications & Advocacy
Volunteer Management
4. Divide the Number of Statements Community Partnerships
Marked Not True by the Total
Applicable Statements.
Evaluative Capacity
5. Convert the answer into a Evaluation Planning, Data Collection, Measuring Impact,
percentage and write that number in Learning & Continuous Improvement
the Percentage column.
25
Office of Research & Evaluation
After completing the table, briefly reflect on your results in the space provided. By identifying
your strongest domains and the areas of greatest need, you will be better equipped to
prioritize capacity building efforts.
Which domain(s) within your organization do you feel are strongest, based on your
assessment results?
1. Domain:
Use this space for thoughts and reflections about this domain.
What is your organization doing well?
2. Domain:
Use this space for thoughts and reflections about this domain.
Where specifically could your organization grow, and what resources may be helpful?
Which domain(s) show the greatest need for capacity building? These gaps can
represent possible focal points for strategic planning, technical assistance, or staff
training. The Resources to build capacity section at the end of each domain can
support your exploration.
1. Domain:
Use this space for thoughts and reflections about this domain.
Where specifically could your organization grow, and what resources may be helpful?
2. Domain:
Use this space for thoughts and reflections about this domain.
Where specifically could your organization grow, and what resources may be helpful?
26
Office of Research & Evaluation
Be sure to discuss your initial conclusions with colleagues and board members who have also
completed the assessment to explore areas of alignment and differences in perspective. The
team should be prepared to summarize the group’s conclusions and share recommendations
on next steps with the larger staff and board. If your organization is working with a technical
assistance provider or consultant, share your findings and work together to identify capacity
building priorities and next steps.
27
Office of Research & Evaluation
References
Australian Bureau of Statistics. (2013, July 3). Correlation and causation. Retrieved from
http://www.abs.gov.au/websitedbs/a3121120.nsf/home/statistical+language+correlation+and+
causation
Authenticity Consulting. (n.d.). Nonprofit organizational assessment. Retrieved from
https://www.surveymonkey.com/r/SVF38MM?sm=nqYAfItd5pCME8J7VJjBQpxt%2b7TXVQBxd
Zt6z7IiPZg%3d
Balanced Scorecard Institute. (n.d.). The basics of strategic planning, strategic management and
strategy execution. Retrieved from http://www.balancedscorecard.org/Resources/Strategic-
Planning-Basics
Bourgeois, I., & Cousins, J. B. (2013). Understanding dimensions of organizational evaluative capacity.
American Journal of Evaluation, 34(3), 299–319. doi:10.1177/1098214013477235
Bryson, J. M., Gibbons, M. J., & Shaye, G. (2001). Enterprise schemes for nonprofit survival, growth, and
effectiveness. Nonprofit Management & Leadership, 11(3), 271–288.
Center for Public Health Systems Science, Washington University in St. Louis. (n.d.). Program
sustainability assessment tool. Program adaptation. Retrieved from
https://sustaintool.org/understand/program-adaptation
The Denver Foundation. (n.d.). Identifying internal and external stakeholders. Retrieved from
http://www.nonprofitinclusiveness.org/identifying-internal-and-external-stakeholders
Easterling, D., & Metz, A. (2016). Getting real with strategy: Insights from implementation science. The
Foundation Review, 8(2), 97–115.
Faculty Development and Instructional Design Center, Northern Illinois University. (2005). Data
collection. Retrieved from
https://ori.hhs.gov/education/products/n_illinois_u/datamanagement/dctopic.html
Grantmakers for Effective Organizations. (2016a). Shaping culture through key moments. Washington,
DC: Author. Retrieved from https://www.geofunders.org/resources/708
Grantmakers for Effective Organizations. (2016b). Strengthening nonprofit capacity: Core concepts in
capacity building. Washington, DC: Author. Retrieved from
https://www.geofunders.org/resources/710
Hall, M., Andrukow, A., Barr, C., Brock, K., de Wit, M., Embuldeniya, D., Vaillancourt, Y. (2003). The
capacity to serve: A qualitative study of the challenges facing Canada’s nonprofit and voluntary
organizations. Toronto: Canadian Centre for Philanthropy. Retrieved from http://www.vsi-
isbc.org/eng/knowledge/pdf/capacity_to_serve.pdf
Himmelfarb Health Sciences Library, George Washington University. (n.d.). Study design 101:
Randomized controlled trial. Retrieved from
https://himmelfarb.gwu.edu/tutorials/studydesign101/rcts.html
Hovland, I. (2005, January). Planning tools: How to write a communications strategy. Overseas
Development Institute: Shaping policy for development. Retrieved from
https://www.odi.org/publications/5186-planning-tools-how-write-communications-strategy
Jackson, T. (2015, March 5). 18 key performance indicator examples defined for managers. ClearPoint
Strategy. Retrieved from https://www.clearpointstrategy.com/18-key-performance-indicators
28
Office of Research & Evaluation
Liket, K. C., & Mass, K. (2015). Nonprofit organizational effectiveness: Analysis of best practices.
Nonprofit and Voluntary Sector Quarterly, 44(2), 268–296. doi:10.1177/0899764013510064
McKinsey & Company. (2001). Effective capacity building in nonprofit organizations. Report for Venture
Philanthropy Partners. Retrieved from
https://www.neh.gov/files/divisions/fedstate/vppartnersfull_rpt_1.pdf
McKinsey & Company. (2016, April). Organizational capacity assessment tool (OCAT) 2.0. Retrieved
from http://mckinseyonsociety.com/ocat/
Misener, K., & Doherty, K. (2009). A case study of organizational capacity in nonprofit community sport.
Journal of Sport Management, 23(4), 457–482. Retrieved from doi:10.1123/jsm.23.4.457
Moore, K. A. (2008). Quasi-experimental evaluations: Part 6. Washington, DC: Child Trends. Retrieved
from https://www.childtrends.org/wp-content/uploads/2008/01/Child_Trends-
2008_01_16_Evaluation6.pdf
Mowbray, C. T., Holter, M. C., Teague, G. B., & Bybee, D. (2003). Fidelity criteria: Development,
measurement, and validation. American Journal of Evaluation, 24(3), 315–340.
doi:10.1177/109821400302400303
The NCJA Center for Justice Planning. (n.d.). Goals and objectives. Retrieved from
http://www.ncjp.org/strategic-planning/overview/where-do-we-want-be/goals-objectives
NCVO Knowhow Nonprofit. (2016, July 1). Employment policies and procedures. Retrieved from
https://knowhownonprofit.org/people/employment-law-and-hr/law-and-hr-basics/policies
Parmenter, D. (2015). Key performance indicators: Developing, implementing, and using winning KPIs
(3rd ed.). Hoboken, NJ: John Wiley & Sons.
Paynter, S., & Berner, M. (2014). Organizational capacity of nonprofit social service agencies. Journal of
Health and Human Services Administration, 37(1), 111–145.
Rutgers University. (n.d.). Developing a survey instrument. Retrieved from
http://njaes.rutgers.edu/evaluation/resources/survey-instrument.asp
Schuh, R. G., & Leviton, L. C. (2006). A framework to assess the development and capacity of non-profit
agencies. Evaluation and Program Planning, 29(2), 171–179.
doi:10.1016/j.evalprogplan.2005.12.001
S. D. Bechtel, Jr. Foundation. (2016, March). Resiliency guide 2.0. Retrieved from
http://sdbjrfoundation.org/wp-content/uploads/2016/03/ResiliencyGuide.pdf
Sharpe, B. (2016, January 26). How to define total compensation: A quick guide [blog]. Retrieved from
https://hrsoft.com/blog/how-to-define-total-compensation-a-quick-guide
Sharpe, P. A., Greaney, M. L., Lee, P. R., & Royce, S. W. (2000). Assets-oriented community assessment.
Public Health Reports, 115(2–3), 205–211. Retrieved from
https://www.ncbi.nlm.nih.gov/pubmed/10968755
Smith, W. J., Howard, J. T., & Harrington, K. V. (2005). Essential formal mentor characteristics and
functions in governmental and non-governmental organizations from the program
administrator’s and the mentor’s perspective. Personnel Administration, 34(1), 31–58.
doi:10.1177/009102600503400103
Taylor-Ritzer, T., Suarez-Balcazar, Y., Garcia-Iriarte, E., Henry, D. B., & Balcazar, F. E. (2013).
Understanding and measuring evaluation capacity: A model and instrument validation study.
American Journal of Evaluation, 34(2), 190–206. doi:10.1177/1098214012471421
29
Office of Research & Evaluation
30