Skip to main content
Erschienen in:
Buchtitelbild

Open Access 2017 | OriginalPaper | Buchkapitel

13. Learning Lessons with Knowledge Audits

verfasst von : Olivier Serrat

Erschienen in: Knowledge Solutions

Verlag: Springer Singapore

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Knowledge from evaluations will not be used effectively if the specific organizational context, knowledge, and relationships of evaluation agencies, and the external environment they face, are not dealt with in an integrated and coherent manner. Knowledge management can shed light on this and related initiatives can catalyze and facilitate identification, creation, storage, sharing, and use of lessons.
Hinweise
These Knowledge Solutions abridge a paper presented at the Malaysian Evaluation Society’s Third International Evaluation Conference held from 31 March to 4 April 2008 in Kuala Lumpur, Malaysia.
In a Word Knowledge from evaluations will not be used effectively if the specific organizational context, knowledge, and relationships of evaluation agencies, and the external environment they face, are not dealt with in an integrated and coherent manner. Knowledge management can shed light on this and related initiatives can catalyze and facilitate identification, creation, storage, sharing, and use of lessons.
https://static-content.springer.com/image/chp%3A10.1007%2F978-981-10-0983-9_13/MediaObjects/372422_1_En_13_Figa_HTML.gif

Introduction

Most development agencies have committed to become learning organizations. But the use of evaluation for learning may be less important than that of other inputs, such as self-evaluation and training, and evaluation results may only marginally support policy, strategy, and operational changes. The Independent Evaluation Department of the Asian Development Bank (2006) determined to apply knowledge management to lesson learning. In 2007, it formulated a strategic framework to improve the organizational culture, management system, business processes, information technology solutions, community of practice, and external relations and networking for that. These Knowledge Solutions explain the strategic framework. They also describe the knowledge audit methodology developed to tie in with the department’s audiences. The online, questionnaire-based survey of perceptions conducted as a first exercise that year provided ready and multiple entry points against which the department can take measures to that intent, as well as a comprehensive baseline assessment against which to judge progress. Fundamentally, these Knowledge Solutions contend that evaluation agencies should move from “make and sell,” at the simplest level, to “sense and respond” in ways that are increasingly satisfying to stakeholders. Knowledge from evaluations will not be used effectively if the specific organizational context, knowledge, and relationships of evaluation agencies, and the external environment they face, are not dealt with in an integrated and coherent manner. Knowledge management can shed light on possible operating frameworks for this and knowledge management initiatives can be applied to catalyze and facilitate identification, creation, storage, sharing, and use of lessons. That would be knowledge utilization indeed.

Knowledge, Relationships, Context, and External Environment

Knowledge must not be seen as something supplied from one person to another or from better-off countries to developing countries, but as something that can flow back and forth and be continually improved, adapted, and refreshed using knowledge management tools. What is more, knowledge management tools are more effective where the specific knowledge, relationships, and context of organizations and the external environment they face are dealt with in an integrated and coherent manner.

Audiences

Evaluations are conducted to find out what results are being achieved, what improvements should be considered, and what is being learned. In ADB, this is done with systematic and impartial assessment of policies, strategies, partnerships, programs and projects, including their design, implementation, and results. Sharing lessons1 also demonstrates good governance and advances understanding of what an organization aims to accomplish, thereby generating support for it. The principal audiences for evaluations, using ADB as an example, include the Board of Directors, Management, the operations departments, ADB’s developing member countries, the international evaluation community, and of course ADB’s Independent Evaluation Department itself.

Interfaces

In the case of ADB, inter- and intra-organizational relationships encompass ADB’s Independent Evaluation Department, other departments,2 developing member countries, and the international evaluation community. Figure 13.1 shows these primary interfaces with the specific organizational context, knowledge, and relationships of ADB’s Independent Evaluation Department and the external environment it faces to structure entry points for lesson learning.

Architecture

Knowledge management must be embedded into an organization’s business processes. It is not an activity delivered exclusively by a distinct business unit or a particular process. An architecture must be built to initiate and implement organization-wide knowledge management initiatives. Here, four pillars are critical to success. They are (i) leadership, (ii) organization, (iii) technology, and (iv) learning. The below table outlines the core functions, typical activities, and implementation elements of a stable architecture for lesson learning.
Table. Architecture for lesson learning
Pillar
Function
Typical activity
Illustrative implementation element
Leadership
Drive values for knowledge management
• Identify knowledge critical to learning lessons in ADB
• Conduct work-centered analysis
• Plan high-level strategic approach
• Establish goal and prioritize objectives
• Define requirements and develop measurement program
• Promote values and norms.
• Implement strategy
• Strategic planning
• Vision sharing
• Definition of goal and objectives
• Executive commitment
• Knowledge management programs tied to metrics
• Formal knowledge management roles in existence
• Tangible rewards for use of knowledge management
• Encouragement, recognition, and reward for knowledge sharing
• Communications
Organization
Organize to support values for knowledge management
• Identify critical knowledge gaps, opportunities, and risks
• Develop business process model
• Engage key audiences with incentives
• Organizational structure
• Organizational culture
• Business process workflows
• Business process reengineering
• Management by objectives
• Total quality management
• Operating procedures for knowledge sharing
• Knowledge performance metrics
• Communications
Technology
Collect and connect knowledge
• Enhance system integration and access
• Deploy intelligent agents for people
• Exploit semantic technologies
• Reuse existing capabilities in new ways
• Monitor, measure, and report knowledge performance metrics
• E-mail
• Data warehousing
• Data management software
• Multimedia repositories
• Groupware
• Decision support systems
• Intranet
• Search engines
• Business modeling systems
• Intelligent agents
• Neural networks
• Lessons learned systems
• Videoconferencing
• Communications
Learning
Cultivate and use virtual teams and exchange forums for knowledge management
• Enliven collaboration
• Facilitate communities of practice
• Encourage storytelling
• Recognize and reward knowledge sharing
• Tacit and explicit knowledge
• Capturing, organizing, and disseminating knowledge
• Team learning
• Management support for continuous learning
• Virtual teams
• Exchange forums
• Communities of practice
• Encouragement, recognition, and reward for innovation
• Communications
Source ADB (2007)

Knowledge Management Tools

Learning lessons is contingent on improving organizational performance in five areas of competence. They are (i) strategy development, (ii) management techniques, (iii) collaboration mechanisms, (iv) knowledge sharing and learning, and (v) knowledge capture and storage (Collison and Parcell 2001).3 Sundry knowledge management tools can support endeavors in each area, including, for example, knowledge audits, activity-based knowledge mapping, action learning sets, peer assists, and exit interviews. Conspicuously, the advent of the Internet has brought information technologies that complement and supplement the knowledge management tools at hand to make knowledge flow more effectively around and across organizations. The technologies include e-learning, web conferencing, collaborative software, content management systems, Yellow Pages, e-mail lists, wikis, and blogs. Where an organization might aim to be in specified time and the priority areas of competence that it might therefore decide to focus on can be investigated by means of such diagnostic tools.

Putting It All Together: The Strategic Framework

Drawing the elements of knowledge, relationships, context, and external environment; audiences; interfaces; architecture; and knowledge management tools in a conceptual structure generates the operating framework within which decisions on knowledge management initiatives can be taken and implemented. Figure 13.2 depicts the operating framework within which knowledge management tools were leveraged by ADB’s Independent Evaluation Department for lesson learning in ADB.

Business Planning

Organizations looking to knowledge management develop business plans that are aligned with their goals and objectives. To raise knowledge vigilance to the point where attitudes are realistic and automatic, and tacit knowledge is internalized, such plans usually identify needs and issues within the organization and are couched against a framework for addressing these. Needs and issues, as well as the business processes associated with them, are typically determined by (i) the external environment; (ii) the mandate, vision, goal, and objectives of the organization; (iii) the overall strategic direction; (iv) the size and spread of the organization; (v) organizational history and culture; (vi) staff skills and experience; and (vii) available resources.
The elemental steps of business planning are (i) identify key staff groups in the organization; (ii) conduct comprehensive and holistic analyses with the key staff groups to identify needs and issues and barriers to organizational performance; (iii) supplement the analyses with inputs from managers and organizational strategy documents to determine an overall strategic focus; (iv) develop findings and recommendations to address the needs and issues and to tackle the barriers identified; and (v) implement a series of knowledge management pilots based on the findings and recommendations, leveraged by suitable knowledge management tools, and with concern for measuring the effectiveness of outreach. Figure 13.3 illustrates the process commonly followed to develop a business plan for knowledge management.
Learning is a process, not an attainment. Hence, in ADB, the Independent Evaluation Department’s knowledge management business plans are aligned against ADB’s to set in train the drive for continuous improvement that is at the heart of strategic frameworks. Moreover, the annual business planning process specifies that regular annual knowledge audits linked to annual business plans will deliver outputs steadily against each interface based on operational needs and priorities but also resources, with flexibility and adaptability.

Auditing Knowledge

Knowledge audits help organizations identify their knowledge-based assets and develop strategies to manage them.
  • Definition and Purpose Developing a knowledge-sharing culture is a change process on the way to better organizational performance. To achieve that change, an organization needs a vision of where it wants to be and an accurate picture of where it is now—that is, its current reality. A knowledge audit is one way of taking that picture. What is a knowledge audit? The traditional concept of an audit is an evaluation of a person, business, system, process, project, or product by an independent third party. Financial audits are well understood. They examine the financial statements of a company to check performance against standards. A knowledge audit works differently, and some demystification is called for. It is by and large—granted differing objects, breadth of coverage, and levels of sophistication—a qualitative review (or inventory, survey, check) of an organization’s knowledge health at both the macro and micro levels. The defining feature of a knowledge audit is that it places people at the center of concerns: it purports to find out what people know, and what they do with the knowledge they have. It can be described as an investigation of the knowledge needs of an organization and the interconnectivity among leadership, organization, technology, and learning in meeting these. Put in a different way, a knowledge audit is an investigation of the strengths and weaknesses of an organization’s knowledge, and of the opportunities and threats that face it.
    A knowledge audit can have multiple purposes, but the most common is to provide tangible evidence of what knowledge an organization needs, where that knowledge is, how it is being used, what problems and difficulties exist, and what improvements can be made. Although there can be no blueprint, a typical knowledge audit will—not necessarily at the same time or level of detail4—query the following:
    • What are an organization’s knowledge needs?
    • What tacit and explicit knowledge assets does it have and where are they?
    • How does knowledge flow within the organization, formally and informally, and to and from clients and relevant organizations?
    • How is that knowledge identified, created, stored, shared, and used?
    • What obstacles are there to knowledge flows, e.g., to what extent do its people, business processes, and technology currently support or hamper the effective movement of knowledge?
    • What gaps and duplications exist in the organization’s knowledge?
  • Constituents of Knowledge Audits The typical constituents of knowledge audits, each of which can be conducted at different levels of complexity using various tools,5 are shown in the Fig. 13.4.6 They are preferably, but not necessarily, in the following order: (i) knowledge needs analysis, (ii) knowledge inventory analysis, (ii) knowledge flow analysis, and (iv) knowledge mapping. Throughout investigations, elements of knowledge, relationships, context, and external environment should be borne in mind, together with the fact that about 80% of an organization’s knowledge is tacit—the greatest challenge lies in the audit of that.

Knowledge Audit Methodology

In 2007, IED particularized a knowledge audit methodology, its principal means, and associated time frame, to be applied in four phases spanning about 5 months. The methodology draws on the elements of knowledge, relationships, context, and external environment; interfaces; and architecture deemed most relevant to the department. The four phases are (i) knowledge audit preparations, (ii) knowledge audit analysis, (iii) knowledge audit review, and (iv) business planning. Box 1 enumerates possible related steps and activities and Box 2 sketches an indicative time frame for implementation. Since knowledge management is a process for optimizing and leveraging the stores of knowledge in an organization, the accent placed (concurrently or in turn) on each constituent of a knowledge audit will depend on where an organization is and where it wants to be. Boxes 1 and 2 should be interpreted in view of that. A second important caveat is that the following section on the survey of perceptions conducted by IED in 2007, which emphasized identification of knowledge needs, should not be taken as all that a knowledge audit can be.
Box 1: Knowledge Audit Methodology—Suggested Steps and Activities
Phase 1
1. Plan Knowledge Audit
• Identify objectives
• Conduct background investigations
• Hold preliminary discussions
2. Assimilate Core Knowledge Activities
• Identify
• Create
• Store
• Share
• Use
3. Delineate Interface Characteristics
• IED–IED
• IED–ADB
• IED–developing member countries
• IED–international evaluation community
4. Identify and Liaise with Key Audiences
• Agree on interface representatives
• Make initial contact
5. Select and Design Audit Forms
• Consider interface characteristics
• Formulate audit deliverables
Phase 2
6. Identify Knowledge Needs
• Investigate what important knowledge the interfaces need to meet goals and objectives
• Determine what important knowledge is available and what is missing
• Consider, with attention to people, business processes, and technology, how faster access to important knowledge might be secured
7. Draw up Knowledge Inventory
• Track down explicit knowledge products and services, their locations, purposes, relevance, and accessibility
• Make out tacit knowledge about who the key audiences are, where they are, what they do, what they know, and what they learn
• Identify gaps in tacit and explicit knowledge
8. Analyze Knowledge Flows
• Examine how knowledge products and services flow in IED, and to and from its interfaces, with attention to people, business processes, and technology
• Characterize stock-based and flow-based knowledge, trends and patterns, and efficiency and effectiveness
9. Create Knowledge Maps
• Locate knowledge products and services and map out flows, constraints, and sinks
• Map knowledge gaps
• Analyze social networks
Phase 3
10. Assess Knowledge Audit Findings
• Analyze evidence
• Suggest courses and means of action
• Devise improvements
11. Discuss Knowledge Audit
• Carry out after-action reviews and retrospects
• Conduct in-house workshops
12. Close Knowledge Audit
• Incorporate suggestions for improvement
• Identify matters for follow-up
Phase 4
13. Decide on Knowledge Management Initiatives
• Prioritize knowledge management initiatives
• Design knowledge management initiatives
14. Formulate Business Plans
• Propose capital, operational, administrative, and recurrent expenditures
• Submit annual budget document
Source ADB (2008)
Box 2:
Indicative Knowledge Audit Time Frame
https://static-content.springer.com/image/chp%3A10.1007%2F978-981-10-0983-9_13/MediaObjects/372422_1_En_13_Figb_HTML.gif
Source ADB (2008)
To underpin future knowledge audits, IED formulated in 2007 survey questionnaires that drew out perceptions of the performance of independent evaluation across the four interfaces. The questionnaires were designed against the Five Competencies Framework. The framework of organizational competence for knowledge management comprises (i) strategy development, (ii) management techniques, (iii) collaboration mechanisms, (iv) knowledge sharing and learning, and (v) knowledge capture and storage.7 The questionnaires were comprehensive, organized, systematic, and inclusive; they provide the framework within which IED can search for continual opportunities to ameliorate the independent evaluation function and its feedback mechanisms. The responses to the questionnaires also revealed rich seams of “as-is,” baseline information, which will be mined vigorously. Box 3 shows for each interface the area of competence on which the questionnaires centered.
Box 3:
Perceptions Survey Questionnaires—Interface and Areas of Competence
Interface
Strategy development
Management techniques
Collaboration mechanisms
Knowledge sharing and learning
Knowledge capture and storage
IED–IED
IED–ADB
 
IED–DMCs
  
IED–IEC
  
ADB Asian Development Bank, DMC Developing member country, IEC International evaluation community, IED Independent Evaluation Department
Source ADB (2008)

The Survey of Perceptions

Knowledge surveys The survey that opened IED’s first knowledge audit aimed to gain insight into how people within the four interfaces perceive the department’s knowledge management activities. From the results, IED measured awareness of and identified gaps in the department’s knowledge products and services. The Five Competencies Framework was used to assess the department’s organizational performance. This framework identifies these areas of organizational competence for knowledge management as (i) strategy development, wherein tools are used to help an organization achieve a particular goal in knowledge management through a long-term plan of action; (ii) management techniques, which cover a range of practices from assessing the forces for and against desired organizational changes to assessing managerial approaches to mistakes, in order to do things right; (iii) collaboration mechanisms, which pertain to facilitating effective practices in working with others; (iv) knowledge learning and sharing, which means using techniques to learn from and improve future activities; and (v) knowledge capture and storage, wherein routines are applied to ensure that an organization retains essential knowledge. For each interface, survey questionnaires covered only the areas of competence deemed most relevant.
The survey adopted a variant of the Likert scale8 to evaluate the perceived level of organizational performance per competence area, wherein respondents specify the extent of their agreement with a statement. Five choices were given per question to gauge perception of IED’s competences: (i) never, (ii) seldom the case, (iii) sometimes the case, (iv) often the case, and (v) always the case. Two methods were used to determine overall perception of organizational performance in each area of competence. The first was based on the responses of the majority per question. The second established an objective measure by computing the weighted average score to account for the perception of the entire sample.

Survey Results

Box 4 gives a snapshot of the perception of the performance of IED in each area of competence by respondents from each interface.9 Respondents from IED thought that the department is doing well in the areas of strategy development, collaboration mechanisms, and knowledge capture and storage. But the department is “on the fence” in knowledge sharing and learning, and its competence with management techniques must get better. Respondents from other departments felt that IED does well only in strategy development. They were ambivalent with regard to collaboration mechanisms. They recommended that the department should deploy more efforts in the areas of knowledge sharing and learning, and knowledge capture and storage. Respondents from the international evaluation community felt that the department is doing well in all three areas of competence regarding which their opinions were sought.
Box 4:
Summary of Perceptions by Area of Competence
Interface
Strategy development
Management techniques
Collaboration mechanisms
Knowledge sharing and learning
Knowledge capture and storage
IED–IED
X
?
IED–ADB
 
?
X
X
IED–IEC
  
ADB Asian Development Bank, IEC International evaluation community, IED Independent Evaluation Department
Note ✓ = More than half of the items in the questionnaire were rated as “often the case” to “always the case” by most of the respondents; ? Half of the items in the questionnaire were rated as “often the case” to “always the case” by most of the respondents, while the other half were rated as “sometimes the case” to “never;” X = More than half of the items in the questionnaire were rated as “sometimes the case” to “never” by most respondents
Source ADB (2008)

Associated Initiatives

The survey of perceptions substantiated the basis of the knowledge management initiatives that IED introduced from 2007. It clarified the need for others. Among the new knowledge products and services developed, Learning Curves are handy, two-paged quick references designed to feed findings and recommendations from evaluation to a broader range of clients. Success Rates present condensed information on successful ADB projects. The Evaluation Information System is an online database of lessons, recommendations, and ADB Management responses. The department hosts the secretariat of the Evaluation Cooperation Group.10 It has also overhauled ECGnet, the group’s communication tool. Evaluation Alerts are targeted information bytes delivered to personal mailboxes. Methods and guidelines for using plain English, disseminating findings and recommendations, and conducting exit interviews have been prepared. The evaluation pages were refurbished from top to bottom. They are updated daily and are now one of the most accessed first-level directories in adb.org. IED formulated regional technical assistance for capacity development for monitoring and evaluation, expected also to suggest a strategy for evaluation capacity development.11 The department advertises its knowledge products and services on ADB Today 12 and adb.org (and through other channels and at several venues) with one-time, near-term, and continuous efforts. The survey of perceptions suggested other opportunities. They included Evaluation Chats, a communication tool that would facilitate the establishment of an evaluation community of practice focused on the conduct and dissemination of strategic evaluations, harmonization of performance indicators and evaluation methodologies, and development of capacity in evaluation and evaluative thinking. Evaluation News and Evaluation Presentations were introduced too. They offer, respectively, reports on events in monitoring and evaluation and short photographic or PowerPoint displays on evaluation topics. IED’s knowledge management initiatives are framed deliberately to increase value-added from operations evaluation, and are managed with knowledge performance metrics. Client feedback is sought regularly by various means.
Box 5 identifies the that might be leveraged to fill the remaining knowledge management gaps identified, and emphasizes with shading the areas of organizational competence found wanting at the time of the survey.
Box 5:
Knowledge Management Tools Prioritized for Leverage
Interface
Strategy development
Management techniques
Collaboration mechanisms
Knowledge sharing and learning
Knowledge capture and storage
IED–IED
• Knowledge Audit
• Most Significant Change
• Outcome Mapping
• Scenario Testing and Visioning
• Five Competencies Framework
• Communities of Practice
• Action Learning Sets
• Social Technologies
• Peer Assists
• Challenge Sessions
• After-Action Reviews and Retrospects
• Intranet Strategies
• Taxonomies for Documents and Folders
• Exit Interviews
• Shared Network Drives
IED–ADB
• Scenario Testing and Visioning
 
• Communities of Practice
• Social Technologies
• Peer Assists
• Challenge Sessions
• After-Action Reviews and Retrospects
• Intranet Strategies
• Staff Profile Pages
• Blogs
IED–IEC
   
• Stories
• Peer Assists
• After-Action Reviews and Retrospects
• Staff Profile Pages
• Blogs
ADB Asian Development Bank, IEC International evaluation community, IED Independent Evaluation Department
Source ADB (2008)
The opinions expressed in this chapter are those of the author(s) and do not necessarily reflect the views of the Asian Development Bank, its Board of Directors, or the countries they represent.
Open Access This chapter is licensed under the terms of the Creative Commons Attribution-NonCommercial 3.0 IGO license (http://​creativecommons.​org/​licenses/​by-nc/​3.​0/​igo/​) which permits any noncommercial use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the Asian Development Bank, provide a link to the Creative Commons license and indicate if changes were made.
Any dispute related to the use of the works of the Asian Development Bank that cannot be settled amicably shall be submitted to arbitration pursuant to the UNCITRAL rules. The use of the Asian Development Bank’s name for any purpose other than for attribution, and the use of the Asian Development Bank’s logo, shall be subject to a separate written license agreement between the Asian Development Bank and the user and is not authorized as part of this CC-IGO license. Note that the link provided above includes additional terms and conditions of the license.
The images or other third party material in this chapter are included in the chapter’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the chapter’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder.
Fußnoten
1
Lessons are of two types: operational and developmental. Operational lessons relate, among others, to performance measurement, aid coordination, resource requirements, team building and coordination, procurement practices, delivery and reporting systems, and logistics. Developmental lessons pertain to realization of development results, improvement of developmental practice, and delivery on priorities.
 
2
In large measure, these are operations departments. But ADB’s Independent Evaluation Department also interacts with nonoperations departments and offices including the Asian Development Bank Institute, the (then) Economics and Research Department, the (then) Regional and Sustainable Development Department, and the Strategy and Policy Department.
 
3
The Five Competencies Framework helps determine priorities for immediate action by selecting the area that will yield the greatest benefits if improved.
 
4
The audit could span the whole organization, but preferably cover constituent parts of it. For the same reason that opinion polls do not sample the entire population, marginal returns diminish as the scale of related exercises increases. The same consideration applies to the number of questions that might be posed.
 
5
The common tools used for knowledge audits are face-to-face and telephone interviews; structured, semi-structured, and unstructured questionnaires; workshops; focus group discussions; and online consultations. Other data and information can be gathered by referring to the documentation of the organization, conducting direct inspections, and examining the information and communications technology infrastructure, including the organization’s website.
 
6
Naturally, in a large and diverse organization, the dimensions and conduct of a knowledge audit will differ radically from that applicable to a small, less complex one.
 
7
A competency approach befits organizational learning. It offers safeguards against drain of knowledge, inappropriate use of knowledge, and accumulation of poor knowledge.
 
8
A Likert scale is usually composed of an odd number of points measuring positive or negative responses to a statement.
 
9
Sadly, no responses were received from evaluation agencies in developing member countries, with implications for the tools that can be applied to that interface in the future.
 
10
The Evaluation Cooperation Group was established by the heads of evaluation in multilateral development banks in 1996. Its membership comprises the African Development Bank, ADB, European Bank for Reconstruction and Development, European Investment Bank, Inter-American Development Bank, International Monetary Fund, and the World Bank Group. The United Nations Evaluation Group and the Evaluation Network of the Development Assistance Committee of the Organisation for Economic Co-operation and Development are observer members.
 
11
Progressively more, evaluation ownership must move from ADB to its developing member countries.
 
12
ADB Today is a daily e-information resource for all ADB staff in headquarters, resident missions, and representative offices. It is the main internal communication vehicle to keep ADB staff abreast of events and activities of ADB-wide interest. It is produced and edited each working day by the Department of External Relations with inputs from other departments.
 
Literatur
Zurück zum Zitat ADB (2007) Learning lessons in ADB. Manila ADB (2007) Learning lessons in ADB. Manila
Zurück zum Zitat ADB (2008) Auditing the lessons architecture. Manila ADB (2008) Auditing the lessons architecture. Manila
Zurück zum Zitat Collison C, Parcell G (2001) Learning to fly: practical knowledge management from leading and learning organizations. Capstone Publishing Collison C, Parcell G (2001) Learning to fly: practical knowledge management from leading and learning organizations. Capstone Publishing
Zurück zum Zitat Independent Evaluation Group—World Bank (2006) Annual report on operations evaluation. World Bank, Washington, DC Independent Evaluation Group—World Bank (2006) Annual report on operations evaluation. World Bank, Washington, DC
Zurück zum Zitat Ramalingam B (2005) Implementing knowledge strategies: lessons from international development agencies. Working Paper 244. Overseas Development Institute Ramalingam B (2005) Implementing knowledge strategies: lessons from international development agencies. Working Paper 244. Overseas Development Institute
Zurück zum Zitat Senge P (1990) The fifth discipline: the art and practice of the learning organization. Currency Double Day, New York Senge P (1990) The fifth discipline: the art and practice of the learning organization. Currency Double Day, New York
Metadaten
Titel
Learning Lessons with Knowledge Audits
verfasst von
Olivier Serrat
Copyright-Jahr
2017
Verlag
Springer Singapore
DOI
https://doi.org/10.1007/978-981-10-0983-9_13