The Final average final rating across all metrics was 4.38 / 5.
Conclusion: Overall score not bad yay! Will continue to manage expectations carefully. Need to look into worst scores, distribution and time ->
The worst score given for any of the metrics on any part of the training
Conclusion: Both too easy and too hard; difficulty scores are also the most spread out. Will focus on managing varying skill levels: Will add more time for preliminary reading to get everyone to the same base; will add ‘bonus’ tasks; will push for multi-level training conducted in parallel (as originally planned). Overall intersting, so will keep the general topics covered.
Having ‘Too Hard’ outliers should have been anticipated. Need to identify them early on and provide extra support. ‘Mentoring’ plan to make sure everyone gets up to speed (Maybe assign mentors based on self-assessed skill levels).
nice that feedback was implemented almost instantly Action: Will update the slides for more coherent structure.To keep the instant feedback implementation but have better structure, I will add “blank” timeslots throughout the training. That way we can stick to the structure and still add in the feedback. Any ‘digressing’ topics I will note down and cover in those timeslots rather than jumping around within a session.
wide range of prior knowledge means training is too difficult and too hard at the same time
Action: To accomodate for varying skill levels, will add ‘bonus’ tasks for more advanced users to all practical exercises. Will send out prerequisite reading earlier & push for reading time berforehand to make sure skill levels are harmonised
Action: The next training will be organised around a real-world example project around which all exercises are built. Will try out having theory and training sessions in mini-intervals, so that each piece of theory can be tried out immediately (slowly building up the example project analysis)
Actions: As I am transferring the training also to Eliora, we will make all materials available, and put together a “trainer tool” on how to run the training. To the training plan, I will add a ‘follow up’ plan on how the shared ideas will be executed / implemented together with the participants. Will convert the skype group into an R-User group, to continue the discussion Have regular “micro events”; for example regular “show & tell” call of what people have been working on
I’m making a single page linking all data unit resources. Adding the above to that list.
Overview over existing tools will add
non-statistical sampling and how to correctly report findings this is a big topic, it probably needs to be a separate training track. It is not the priority because this is more manageable with ‘oldschool’ tools.
Brainstorming / Sharing etc.: I will add this under the umbrella term ‘support community growth’ as a declared goal of the trainings and other data unit activities; exact changes not clear yet, but of course more time and more structure for brainstorming, discussion etc. during trainings.
Listed below is the individual full text feedback sorted by agenda item it relates to. Please note that for the responses below, it was asked specifically for negative feedback, and to be as critical (but constructive) as possible.
The idea is nice and there is nothing bad about exchanging ideas. However, the actual coding part of the hackathon will just put pressure on those who can code already and leave the beginners where they are.
Before doing something like this we should bring everybody to an acceptable level where they can effectively contributing to the designing and building of the tool.
Overall fine
All good but next time more exercises!
Today’s sessions were good, I am leaving with more knowledge than I came with
I feel like there is quite a big difference in the level of previous experience for this, which might be frustrating for some of the people in the training (i.e. those with a lot of experience). Considering the goal to have people with the skills to build tools in each mission, some people might need more training than others (in the future, after the RCM). However, as a complete novice, this applied approach is very useful for me.
A bit too slow
Discussion on analysis guidelines would be good
The first day was very informative regarding to we work with data/research in generally. Maybe (1) program outline on the srceen (what do we do now and what comes next) could help to stay on track for people with different speed of progress, as well as learning-friendly dataset
All the ideas presented were really good, hope we can move to actions not like what happened during Geneva workshop. (Brainstorming but then nothing happened :( )
The first day overall was good but more examples/exercises with actual datasets would have been useful. I think it would also be good to get a better overview of what kind of tools currently exist, what they are doing and at what development stage they are.
All good, shame we couldn’t play around with hypegrammaR! I enjoyed the brainstorming session.
It is VERY intersting! It would be nice though to have a tiny bitesize practical example for each item, so that we can try things out as we learn them. Or maybe semihomework?
Very good! I am quite interested to see how Markdown can improve the way we structure our data analysis
Hey, do you suggest purchasing a book/guide of any sort like the “R for Data Science”, (or any other guide that helped you in getting where you are)?
Being able to use hypegrammaR was great! I can’t wait to see it progress!
I found the hands-on exercises this morning very useful, and good to apply the basics I learnt during the last days. also, I think it’s a good foundation for using R in the future. However, I think we need to make sure that we’ll have enough time to also use the things we learnt, so the tools actually benefit our daily work. would be good if that’s taken into consideration of people’s TOR.
Would be interesting to have a bit of a group discussion on what people are doing in terms of development/automation/programming
We should create cross-country project-specific teams (WASH / MSNA / Market Monitoring, Migration) and make them work together to find solution that can be generalizable.
I am wondering how we can transfer the knowledge we aquire here into our missions. I am sure people learn a lot and gain some insights on how to make their own work more efficient/analysis better/etc., but if it’s only one person per mission learning this, the gap between this group and the others will just increase. I think we should think of a way on how to bring other AOs/coordination-focused people to a certain level so that they can actually use the tools. At the moment, I can’t really see how we could do that (specifically in our mission, where the basic technical level in not bad, but there’s no real experts that could push to get everyone further). (Rating below is for the idea of incorporating a training for everyone in the mission)
Great training, good to see the tools developed at Impact. Very useful for people to share a common level. Very interesting brainstorming part.
Overall the training was super smooth and clear. It was not “technically” advanced (simply it could not be that way knowing that we are all coming from different background and not necessarily having the basic skills for that). Could you please upload the new version of your presentation. It is super helpful thanks :) :) :)
This is a very valuable training! Some things about it: -it was not too task- specific, so just allowed to conceptualize, explore guidelines, understand general principles of how to look for solution and how to debug the bug -it was an amazing opportunity to meet people from the other missions, learn from them, enrich own perspective -i think it is very hardcore for the trainer to be the only trainer during all three days, expecially taking into account that this is a pilot -it was nice that feedback was implemented almost instantly, so that finally we can do/try everything we wanted to -it was very inspirational! Bravo
I think overall this has been a good training. I liked the practical exercises particularly. Overall, it was a good introduction to existing R tools and how to generally load/run/troubleshoot R scripts etc. and should enable participants to use existing tools at least (which I guess was the intention of this, so, goal achieved). I think it could’ve been a bit more structured at times (the general structure was probably useful but we did jump a bit between different things on the first day). I also would have liked more practical exercises and personally felt it could have been a bit more challenging. Maybe it could have been good to have a specific output as a desired result of a multiple-step process involving multiple tools and eventually leading to a brief analysis of a few hypotheses or indicators. Maybe also could’ve been good to have a few small tasks for people to try out with their own data, e.g. aggregate data at a higher admin level, group data by population group or region, create a visualisation for one of these etc. (just with a few lines of code everyone could do on their own). I get that we had different skill levels and limited time so maybe not completely feasible but just wanted to bring it up as it could have helped people to have a few more experiences of success/progression and maybe a bit more insight into coding.
Overall the training was awesome! Thank you for putting this together. I feel much more comfortable now that I know where to find useful packages, read the documentation, and ask for help if needed. Getting to know GitHub was great and I will start implementing it in all my R projects now. Finally, for future reference, I’d like to see something regarding non-statistical sampling and how to correctly report findings.
It would be beneficial if, as an organization, we compiled useful training materials, tutorials, packages, etc. (including everything from this training) that could be accessible for anyone looking to learn or improve their skills.
Each day, this fully anonymous feedback form was filled out by the training participants ad-hoc between sessions, as well as after the workshop was finished. Feedback was then implemented in the subsequent sessions, which is why some of the feedback may be “contradictory” between sessions (i.e.: “X is missing from the training”, and later “it was good that we did X”.
At the end of the last day, the form was filled out by all participants. All ratings are based on 25 responses overall.
| Zeitstempel | I.have.. | My.feedback.question.relates.to… | My.feedback.question..s.. | Rate..from.0…not.at.all..to.10…completely….This.part.of.the.training.was…..Too.easy. | Rate..from.0…not.at.all..to.10…completely….This.part.of.the.training.was…..Too.hard. | Rate..from.0…not.at.all..to.10…completely….This.part.of.the.training.was…..Too.detailed. | Rate..from.0…not.at.all..to.10…completely….This.part.of.the.training.was…..Interesting. | Rate..from.0…not.at.all..to.10…completely….This.part.of.the.training.was…..Boring. | Rate..from.0…not.at.all..to.10…completely….This.part.of.the.training.was…..frustating. |
|---|---|---|---|---|---|---|---|---|---|
| 11.02.2019 16:31:30 | feedback | not session specific | Would be interesting to have a bit of a group discussion on what people are doing in terms of development/automation/programming | 2 | 0 | 1 | 4 | 0 | 0 |
| 11.02.2019 17:24:08 | feedback | day 1 afternoon | All good but next time more exercises! | 4 | 0 | 1 | 5 | 1 | 1 |
| 11.02.2019 17:26:46 | feedback | day 1 afternoon | Today’s sessions were good, I am leaving with more knowledge than I came with | 1 | 1 | 1 | 4 | 1 | 2 |
| 11.02.2019 17:26:46 | feedback | day 1 afternoon | I feel like there is quite a big difference in the level of previous experience for this, which might be frustrating for some of the people in the training (i.e. those with a lot of experience). Considering the goal to have people with the skills to build tools in each mission, some people might need more training than others (in the future, after the RCM). However, as a complete novice, this applied approach is very useful for me. | 0 | 2 | 1 | 4 | 0 | 2 |
| 12.02.2019 08:33:33 | feedback | day 1 afternoon | A bit too slow | 4 | 0 | 2 | 5 | 0 | 0 |
| 12.02.2019 08:34:59 | feedback | Day 1 | Overall fine | 4 | 1 | 2 | 3 | 2 | 2 |
| 12.02.2019 08:35:09 | feedback | day 1 afternoon | Discussion on analysis guidelines would be good | 3 | 0 | 0 | 4 | 1 | 1 |
| 12.02.2019 08:35:18 | feedback | Reading material | It would be beneficial if, as an organization, we compiled useful training materials, tutorials, packages, etc. (including everything from this training) that could be accessible for anyone looking to learn or improve their skills. | 1 | 1 | 2 | 4 | 1 | 1 |
| 12.02.2019 08:36:24 | feedback | Brainstorming/Hackathon | The idea is nice and there is nothing bad about exchanging ideas. However, the actual coding part of the hackathon will just put pressure on those who can code already and leave the beginners where they are. — — Before doing something like this we should bring everybody to an acceptable level where they can effectively contributing to the designing and building of the tool. | 3 | 0 | 3 | 5 | 1 | 1 |
| 12.02.2019 08:36:57 | feedback | day 1 afternoon | The first day was very informative regarding to we work with data/research in generally. Maybe (1) program outline on the srceen (what do we do now and what comes next) could help to stay on track for people with different speed of progress, as well as learning-friendly dataset | 2 | 0 | 0 | 5 | 0 | 1 |
| 12.02.2019 08:37:24 | feedback | day 1 morning | The first day overall was good but more examples/exercises with actual datasets would have been useful. I think it would also be good to get a better overview of what kind of tools currently exist, what they are doing and at what development stage they are. | 3 | 0 | 2 | 3 | 1 | 2 |
| 12.02.2019 08:37:37 | feedback | day 1 afternoon | All the ideas presented were really good, hope we can move to actions not like what happened during Geneva workshop. (Brainstorming but then nothing happened :( ) | 4 | 1 | 3 | 4 | 0 | 0 |
| 12.02.2019 08:38:34 | Idea | not session specific | We should create cross-country project-specific teams (WASH / MSNA / Market Monitoring, Migration) and make them work together to find solution that can be generalizable. | 4 | 0 | 2 | 5 | 0 | 0 |
| 12.02.2019 08:39:15 | a question | not session specific | I am wondering how we can transfer the knowledge we aquire here into our missions. I am sure people learn a lot and gain some insights on how to make their own work more efficient/analysis better/etc., but if it’s only one person per mission learning this, the gap between this group and the others will just increase. I think we should think of a way on how to bring other AOs/coordination-focused people to a certain level so that they can actually use the tools. At the moment, I can’t really see how we could do that (specifically in our mission, where the basic technical level in not bad, but there’s no real experts that could push to get everyone further). (Rating below is for the idea of incorporating a training for everyone in the mission) | 1 | 4 | 3 | 4 | 0 | 3 |
| 12.02.2019 11:55:07 | feedback | day 2 morning | It is VERY intersting! — It would be nice though to have a tiny bitesize practical example for each item, so that we can try things out as we learn them. Or maybe semihomework? | 3 | 0 | 1 | 5 | 0 | 1 |
| 12.02.2019 13:04:45 | feedback | day 2 morning | Very good! I am quite interested to see how Markdown can improve the way we structure our data analysis | 1 | 3 | 2 | 5 | 0 | 0 |
| 12.02.2019 13:18:09 | a question | day 2 morning | Hey, do you suggest purchasing a book/guide of any sort like the “R for Data Science”, (or any other guide that helped you in getting where you are)? | 0 | 0 | 0 | 5 | 0 | 0 |
| 13.02.2019 08:05:02 | feedback | day 2 afternoon | All good, shame we couldn’t play around with hypegrammaR! I enjoyed the brainstorming session. | 0 | 0 | 0 | 5 | 0 | 0 |
| 13.02.2019 11:22:37 | feedback | day 3 morning | Being able to use hypegrammaR was great! I can’t wait to see it progress! | 0 | 0 | 0 | 5 | 0 | 1 |
| 13.02.2019 11:23:37 | feedback | not session specific | Great training, good to see the tools developed at Impact. — Very useful for people to share a common level. — Very interesting brainstorming part. | 4 | 1 | 1 | 5 | 0 | 0 |
| 13.02.2019 11:25:00 | feedback | day 3 morning | I found the hands-on exercises this morning very useful, and good to apply the basics I learnt during the last days. also, I think it’s a good foundation for using R in the future. However, I think we need to make sure that we’ll have enough time to also use the things we learnt, so the tools actually benefit our daily work. would be good if that’s taken into consideration of people’s TOR. | 0 | 0 | 0 | 4 | 0 | 0 |
| 13.02.2019 11:25:26 | Overall | Overall the training was awesome! Thank you for putting this together. I feel much more comfortable now that I know where to find useful packages, read the documentation, and ask for help if needed. — Getting to know GitHub was great and I will start implementing it in all my R projects now. — Finally, for future reference, I’d like to see something regarding non-statistical sampling and how to correctly report findings. | 2 | 0 | 0 | 5 | 0 | 1 | |
| 13.02.2019 11:29:33 | feedback | not session specific | Overall the training was super smooth and clear. It was not “technically” advanced (simply it could not be that way knowing that we are all coming from different background and not necessarily having the basic skills for that). Could you please upload the new version of your presentation. It is super helpful thanks :) :) :) | 3 | 1 | 3 | 4 | 0 | 0 |
| 13.02.2019 11:30:24 | feedback | not session specific | This is a very valuable training! Some things about it: — -it was not too task- specific, so just allowed to conceptualize, explore guidelines, understand general principles of how to look for solution and how to debug the bug — -it was an amazing opportunity to meet people from the other missions, learn from them, enrich own perspective — -i think it is very hardcore for the trainer to be the only trainer during all three days, expecially taking into account that this is a pilot — -it was nice that feedback was implemented almost instantly, so that finally we can do/try everything we wanted to — -it was very inspirational! Bravo — | 0 | 0 | 0 | 4 | 0 | 1 |
| 13.02.2019 11:34:17 | feedback | not session specific | I think overall this has been a good training. I liked the practical exercises particularly. Overall, it was a good introduction to existing R tools and how to generally load/run/troubleshoot R scripts etc. and should enable participants to use existing tools at least (which I guess was the intention of this, so, goal achieved). I think it could’ve been a bit more structured at times (the general structure was probably useful but we did jump a bit between different things on the first day). I also would have liked more practical exercises and personally felt it could have been a bit more challenging. Maybe it could have been good to have a specific output as a desired result of a multiple-step process involving multiple tools and eventually leading to a brief analysis of a few hypotheses or indicators. Maybe also could’ve been good to have a few small tasks for people to try out with their own data, e.g. aggregate data at a higher admin level, group data by population group or region, create a visualisation for one of these etc. (just with a few lines of code everyone could do on their own). I get that we had different skill levels and limited time so maybe not completely feasible but just wanted to bring it up as it could have helped people to have a few more experiences of success/progression and maybe a bit more insight into coding. | 3 | 0 | 0 | 4 | 2 | 0 |