Polish language e-learning system

 
 
Mockup_topphoto_polishsystem.jpg

objective

The objective of this project was to design, prototype, and test an e-learning system utilizing the principles of learner-centered design. I developed a set of testable learning objectives for the course, designed the course content, and evaluated how well the system helped people meet those intended objectives.

Context

This E-learning system was created for HCI520: Learner-Centered Design at DePaul University

METHODS & TOOLS:

Axure RP 9

role:

UI/UX Designer, researcher, course content creator

 

 

Course description

I began the project by outlining the course goals and deciding what the course would cover. The course length was limited to roughly a 30 minute course, including both the pre and post tests.

Scope:

For my project, I created an e-learning system featuring an introductory Polish language lesson designed for travelers visiting Poland. The course covers vocabulary and phrases separated into the following modules:

• Basic Greetings

• Simple Conversation and Pleasantries

• Transportation and Getting Around

• Accommodation

• Shopping and Dining

Content Source:

For this project, I relied on my own personal knowledge of the Polish language, as I am fluent in Polish. I used Quizlet to create a matching game for vocabulary practice.

 

 

Course Goals

After I narrowed the scope of the lesson, I created 5 specific, testable learning goals, which were measured by relevant questions on the pre and post tests.

By the end of the course students should be able to:

  • remember and use simple greetings and pleasantries in the correct context

  • match destination/transportation vocabulary and phrases with the correct context and translation

  • ask for directions to a destination (bathroom, museum, bus etc.) as well as understand some basic directional responses

  • ask the price of something and ask for an item they would like at a store, restaurant, or hotel

  • use and match learned travel phrases and vocabulary to specific contexts of use

 


 

Application of Learner-Centered Design principles

Contiguity Principle

While I did not have images and text to be shown near each other, I made sure to apply the contiguity principle in my practice question feedback. I placed the feedback directly beneath each question, right near the “Check My Answer” button, so that learners would see it quickly and easily. Here is an example:

 
 
contiguitysnap1.PNG
 
 
 
 

Modality Principle

Rather than presenting purely textual information, I presented the information in videos. The videos consisted of audio over slides of the vocabulary and phrases in textual form:

 
 
moduleexample snapshot1.JPG
 
 
 
 

Redundancy

Since this is foreign language learning, I purposefully violated the redundancy principle and showed the information in textual form while simultaneously saying the same words utilizing audio. I made sure to highlight the words as I went along to aid learners in following along:

 
 
redundancysnap.PNG
 
 

Segmenting

I broke up the course into 5 modules to segment the information into manageable pieces:

 
 
segmentingsnap.PNG
 
 

Practice

I utilized a lot of varied practice in the system. For some modules, I included a Quizlet matching game for vocabulary, and for other modules, I included other custom practice questions, with feedback given when the learner clicked “Check My Answer”:

 
 
practicesnap1.PNG
practicesnap2.PNG
contiguitysnap1.PNG
 
 

Learner Control

Since my system is designed for beginners, I allowed very little learner control. It was mostly system controlled, with the “Next” buttons disabled until the learner watched the lesson and then clicked the practice. I also locked the subsequent modules until the previous ones were done. Additionally, learners could control the lesson videos and replay or pause if necessary.


For more details on all the learner-centered design principles applied in this system, see the full report here.

 

 

User Testing and system evaluation

Participants

A total of 14 participants, consisting of friends, friends of friends, and DePaul University students. While not all of them plan to travel to Poland, all of them were beginners at Polish and had not studied the language previously at all.

Testing

Each participants was sent a prototype link via email, in which I briefly described the project topic and gave instructions to take the Pre-test, then follow the instructions on the system to complete the course, and then take the Post-test. The testing was 100% remote, done at the participants’ leisure. In order to affiliate pre and post test scores, I asked each participant to create a unique ID which they then inputted on both pre and post tests. I manually checked each submission and recorded all scores on an Excel spreadsheet. Each question was worth 1 point, and I graded question #8 as X/11, as there were 11 words to categorize (see full report link above to view pre and post test questions). In order to have an accurate comparison, Pre and Post test questions remained identical, but in a different order.

 

 

Statistical Outcomes

I conducted a paired t-test to compare the pre and post test scores of 14 participants. I found a significant difference (t = -7.40, p < .05), such that participants after completing the e-learning system scored higher (M = 8.36, SD = 1.25) than before completing the system (M = 4.23, SD =2.12).

t(13)= -7.40, p < .05

 
 
statssnap.PNG
 
 

Effect Size:

Calculated Effect Size: 2.4

The calculated effect size indicates a very strong relationship between test results and completing the e-learning course.

 

 

Overall Evaluation & System Limitations

Overall, based on the statistical tests, the system was effective. The results were statistically significant, with a very large effect size. However, as always, there are some limitations including:

  • a small sample size, and testing within a similar age range, as I recruited friends, friends of friends, and other students, but did not test with any older users.

  • I wonder if I should have presented different questions for the post-test, with more of a “knowledge transfer” element involved, or designed the assessment questions in a different manner than multiple choice

  • I originally wanted to include audio within the assessments, but found it too difficult to implement with the tools available and time restriction. That might be something worth considering in further developments, as hearing and speaking the language is such a big part of learning it

  • Finally, there is always the factor of uncertainty in results when conducting an unmoderated remote test, as you cannot see the participants and must trust that they’ve followed the system instructions, and not looked up answers etc.

Further Research & Development

In addition to trying out use of audio clips in assessment, in future iterations, I might want to research other methods of conducting lessons utilizing the multimedia principle and images more, despite the fact that they would probably be more representational than anything else. Additionally, I would recommend external links for the pre and post tests, or an addition Axure page built in, stopping participants from moving onto the lessons if they had not yet taken the pre-test. I had a few participants miss the pre-test because it was an embedded Google form and didn’t load immediately, thus making their post-test scores unusable. I was trying to make the system uniform and streamlined, but in this case it created more trouble than cohesiveness. I might also consider comparing groups of users, as unintentionally, some of my participants were fluent in a different language already, and I wonder if that didn’t increase scores due to similarities they may have noticed between their language and Polish.

Finally, a mobile version of the system would be a more viable option for people to learn as they travel- the desktop version is fine if they are learning before their trip, but a mobile e-learning system would be a more convenient and useful option for the user.