Skip to content

NBISweden/elixir_evaluation_evaluation

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

4 Commits
 
 
 
 
 
 

Repository files navigation

elixir_evaluation_evaluation

An evaluation of the ELIXIR course evaluation.

Goals

  • To evaluate the ELIXIR course evaluation
  • If possible, propose a better one

Introduction

How it started

On 2025-01-20 the NBIS Training Steering Group had a meeting on course evaluations. The first question was 'How are evaluations evaluated?'.

History of the NBIS short-term evaluation

The NBIS short-term evaluation questions are based on the ones by ELIXIR. ELIXIR developed these evaluation questions to assess the quality and impact of bioinformatics training [Gurwitz et al., 2020]. and the resulting metrics can be found at https://training-metrics-dev.elixir-europe.org/all-reports.

How good are these questions?

The question is: how good are these questions? To determine this, I'll first describe how these questions are developed. After that, I will describe the written-down goal of the questions and each of these in detail.

How were these questions developed?

To get a first idea of how good these evaluation questions are can be obtained by reading how these were developed. We can read in [Gurwitz et al., 2020] how this process went: 'These metrics were developed out of those already collected by ELIXIR training providers, as well as from discussions with stakeholders, external training providers, and literature review', with references to two papers. The first paper described the demograpics and short-term (i.e. directly after the course) questions used by DataCarpentries [Jordan et al., 2023], where the second is about long-term impact [Brazas & Ouellette, 2016]. At a first glance, this looks like a proper procedure.

What are the goals of these questions?

flowchart TD
  goal_1[ELIXIR: Determine training quality]
  goal_2[NBIS: Find out how participants have used the skills and knowledge they gained]
  goal_3[NBIS: Improve the course and the materials we deliver]
  goal_1 <-.-> |similarish| goal_3
  goal_2 <-.-> |dissimilar| goal_3
Loading

To get a second idea of how good these evaluation questions are can be obtained by reading the goal of the questions.

Here is the ELIXIR goal [Gurwitz et al., 2020]:

We were interested in participant satisfaction as a reflection on training quality in order to be able to inform best practice for ELIXIR training.

Here the intention of the NBIS short-term evaluation form is quoted:

The intention of the STF survey is to find out how participants have used the skills and knowledge they gained through participating in the NBIS course.

To me, this seems like a copy-paste error from the long-term survey by NBIS...? Section 1 seems to be more in line with a short term evaluation:

It is really important to us in order to continually improve the course and the materials we deliver

So it seems the goal of the STF is to improve the course and its materials.

What are thequestions?

With the goal of the SFT ('to improve the course and its materials') in mind, here we go through the questions that resulted from the process described in the previous paragraph. Here is Section 3 - Quality Metrics of the NBIS short-term evaluation:

5. Have you used the tools/resource(s) covered in the course before? 

- Never - Unaware of them
- Never - Used other service
- Occasionally
- Frequently

6. Will you use the tools/resource(s) covered in the course again?

- Yes
- No
- Maybe

7. Would you recommend the course?

- Yes
- No
- Maybe

8. What is your overall rating for the course

- Poor (1)
- Satisfactory (2)
- Good (3)
- Very Good (4)
- Excellent (5)

9. A. May we contact you by email in the future for more feedback?

- Yes
- No

Question 5

Question 5 is an interesting way to evaluate the quality of a course, because it is about something learners have done before the course took place.

These are the metrics collected at 2025-01-24 7:04 Stockholm time (https://training-metrics-dev.elixir-europe.org/feedback-report):

Reponse n Frequency (%)
Never - Unaware of them 4350 23.5
Never - aware of them 3838 20.8
Never - Used other service 1803 9.7
Occasionally 6974 37.7
Frequently 1528 8.3

Does this indicate good or bad courses? Are the right people reached? It would be interesting to know, how these values are used to determine the quality of a course.

Question 6

Question 6 is another interesting way to evaluate the quality of a course, because it is about the usefulness of the topic being taught, combined with predicting the future.

These are the metrics collected at 2025-01-24 7:16 Stockholm time (https://training-metrics-dev.elixir-europe.org/feedback-report):

Reponse n Frequency (%)
Maybe 2822 15.1
No 105 0.6
Yes 15792 84.4

Also here, does this indicate good or bad courses? It would be interesting to know, how these values are used to determine the quality of a course.

Question 7

Question 7 attempt to measure course quality by asking the learner if he/she would recommend the course. The (little) research on this practice shows that this may be true [Ang et al., 2018].

These are the metrics collected at 2025-01-24 8:27 Stockholm time (https://training-metrics-dev.elixir-europe.org/feedback-report):

Reponse n Frequency (%)
Maybe 19597 89.5
No 1790 8.2
Yes 519 2.4

Question 8

Question 8 too attempt to measure course quality by asking the learner to rate it. There seems to be overlap between this and previous question, hinged on the assumption that if a course is recommended that is likely to be rated positively.

Asking learners for their course satifaction however, is sketch. This is already acknowledged by ELIXIR:

We acknowledge that training quality is more complex than solely participant satisfaction and that the community would benefit from future work to obtain a fuller picture on training quality [Gurwitz et al., 2020]

There is, however, according to a meta-analysis, no relation between training quality and participant satisfaction [Uttl et al., 2017] and this meta-analysis gives some examples how problematic this metric is.

These are the metrics collected at 2025-01-24 8:28 Stockholm time (https://training-metrics-dev.elixir-europe.org/feedback-report):

Reponse n Frequency (%)
Excellent 7736 37
Very good 8437 40.4
Good 3543 17
Satisfactory 993 4.8
Poor 192 0.9

Question 9

Question 9 is an interesting way to measure the course quality, based on the learner being willing to answer questions on the future. It seems more likely that question should be placed outside of the section Section 3 - Quality Metrics.

These are the metrics collected at 2025-01-24 8:32 Stockholm time (https://training-metrics-dev.elixir-europe.org/feedback-report):

Reponse n Frequency (%)
No 8756 49.7
Yes 8860 50.3

Missing questions

There are some questions that were removed from the Data Carpentries evaluations.

How to evaluate these questions?

The goal is this exercise is:

  • To evaluate the ELIXIR course evaluation
  • If possible, propose a better one

To achieve this goal, these are the steps:

  • Collect evaluation questions that were thought of as 'good'
  • Of these and the NBIS SFT evaluation questions, collect reasons why this would be a 'good' or 'bad' question
  • Of the questions and reasonings, assign a rating per question

Epilogue

We know that teachers reflecting on their work is one of the best ways to increase his/her teaching quality. Or: 'student ratings can only become a tool for enhancement when they feed reflective conversations about improving the learning process and when these conversations are informed by the scholarship of teaching and learning [Roxå et al., 2021]. The other best way for teachers to improve is to do peer observations. Note that neither practice needs an evaluations.

If we really care about teaching quality, shouldn't we encourage doing the things that actually work?

Notes

I think this is a useful evaluation using learning objectives:

I think this is a useful evaluation for improving as a teacher:

Here is an example of how useful questions can be determined:

References

  • [Gurwitz et al., 2020] Gurwitz, Kim T., et al. "A framework to assess the quality and impact of bioinformatics training across ELIXIR." PLoS computational biology 16.7 (2020): e1007976.

  • [Jordan et al., 2023] Jordan, Kari, François Michonneau, and Belinda Weaver. "Analysis of Software and Data Carpentry’s pre-and post-workshop surveys." Software Carpentry. Retrieved April 13 (2018): 2023.

  • [Brazas & Ouellette, 2016] Brazas, Michelle D., and BF Francis Ouellette. "Continuing education workshops in bioinformatics positively impact research and careers." PLoS computational biology 12.6 (2016): e1004916.

  • [Ang et al., 2018] Ang, Lawrence, Yvonne Alexandra Breyer, and Joseph Pitt. "Course recommendation as a construct in student evaluations: will students recommend your course?." Studies in Higher Education 43.6 (2018): 944-959.

  • [Roxå et al., 2021] Roxå, Torgny, et al. "Reconceptualizing student ratings of teaching to support quality discourse on student learning: a systems perspective." Higher Education (2021): 1-21.

About

An evaluation of the ELIXIR course evaluation

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published