This section presents the experiences of users with selected evaluation tools.
The users assessed their experience with the tools using a standardized methodology, which can be found
here: CoEvalAMR evaluation matrix. The updated version consists of a brief introduction to the case study and to the tool separately. Moreover, 12 functional aspects and nine content themes are scored. Results are illustrated using radar diagrams. The answers are scored from 1-4; where 1 = not satisfactory at all, 2 = need major improvements, 3 = almost perfect but minor improvements are needed, 4 = completely satisfactory. With each numeric score (1-4) a comment is requested explaining the score.
The 12 functional aspects to score are:
- User-friendliness related to wording, guidance and layout of the tool or framework
- Analysis and interpretation of evaluation data
- Analysis and interpretation of evaluation data
- Amount and complexity of data required, where complexity is defined as different kinds of data from multiple sources in different formats or as primary data collection required
- Ability to address the stated evaluation objectives
- Evaluation of One Health (OH) aspects (collaboration across sectors/disciplines, knowledge integration, added value of OH approach, etc.)
- Required level of knowledge of users regarding surveillance, epidemiology and evaluation
- Training to get acquainted with the tool
- Costs related to the access and use of the tool
- Number of people in the evaluation team
- Duration of the evaluation process
- Generation of actionable evaluation outputs
In addition, users are asked to answer four questions using a SWOT-like approach: 1) The strengths of this tool are, 2) The weaknesses of this tool are, 3) The added value(s) of using this tool is, and 4) This tool might be criticized because of.
Please note that these reports are subjective and do not constitute the endorsement of any tool.