Chen Li, Bai Bin, Xu Yang, Shen XinYi
Beijing Normal University, China
Abstract:
Most of the quality assurance (QA) criteria developed in various settings are mainly concerns of the Online and Distance Learning institutions, instructors, assessors and funding bodies. To promoting the quality in Online and Distance Learning, it is critical to understand learners’ views since learning outcome relies to a greater extent on learners’ motivation and engagement. This research mainly focus on the factors which are more important than others in assuring the quality of Online and Distance Learning from the perspective of China learners and differences in learners’ perception of the quality according to age groups, learning methods, and learning experience. What’s more, we compare the difference between learners and teachers. An online survey was developed to gather empirical evidence about a set of 59 items in the ten dimensions of QA in Online and Distance Learning. 653 learners who were enrolled in three institutions in mainland China and 91 teachers including senior managers, instructors and tutors from eight institutions in mainland China were surveyed. We hope the findings from this survey will supply valuable references for developing QA framework for China. Readers can get deep understanding QA situation in China.
Keyword: Quality assurance,Student Survey, Critical Factors, ODL
Background
Since China kicked off the ICT-supported distance education pilot project in 1999, online education had achieved remarkable development in China. However, as a newborn thing, modern distance education had some problems and challenges naturally. Among the last twelve years for the pilot project, the government, the society (the third party) and the school all had made a lot of efforts to build the quality assurance system of distance education. The Ministry of Education had gradually carried out more than 100 policy documents about the approval and management, recruitment and employment, certificates and electronic registration, examination, annual inspection, assessment and teaching standards for online education college, public service system and learning center, etc, and had established an information-based platform and mechanism for quality control of online education. The distance education providers had actively formulated appropriate rules and regulations in all aspects of distance education, such as student management, teacher management, test management and other aspects, to ensure that online education is rule-based, and to achieve the scientific, standardized management of online education. Some distance education providers even took attempt to introduce the ISO9000 to make the quality management of distance education normative.
Although the government and distance education providers had attempted to improve the quality, the current situation was still not satisfactory. Generally speaking, besides objective factors, we should put more emphasis on the area that government, society and provider was working on .For instance, there was not any national quality standard, any third party, any unified quality values and quality management. In a word, we concluded that the quality assurance system of distance education in China had not been formed.
Many other countries like United State, Britain, Malaysia, Australia and the European region had issued relevant quality standards documents. The United States and Australia's distance education had its specific quality standard system. American Distance Education and Training Council issued Distance education institutions accreditation manual and Institute for Higher Education Policy in USA released online education quality standard, while Australasian Council on Open, Distance and E-Learning was non-governmental organization and aimed to provide with policy and practice guidance of open, remote, flexible and digital learning. The quality management organization of Britain and Malaysia belonged to national ministry of education, and their distance education quality standards and the traditional university quality standard was unified.
Two major production procedures are applied in developing the quality standards in these countries. One is up-down model and the government takes the same evaluation framework on the assessment of the quality of DE/e-learning as well as that of traditional education. The other is third party model, which means the standard framework is developed by the third party with wide range of applications. Nonetheless, no matter which procedure is applied, the existing quality standards are mainly based on the perspective of the managers. Learners are core stakeholders in distance education, which indicates their views should be crucial in drafting the quality standard. There had been a few studies which examined learners’ views of the quality of distance education.
Cashion and Palmieri (2002) investigated Australian learners’ and educators’ views of the quality of online learning in the vocational education and training context. They found that the view of learners was different from view of providers. Flexibility was the most important factor for learners but not for the providers. And learners said they didn’t need initial support to study online which was emphasized by providers. Besides, skill in using technology rated by learners was lower than providers.
Chin-Wen Chang(2008) carried out a survey about learners’ view of online learning in Taiwan. They found that students were positive to online learning and they hold the view that course management, communication and feedback, course flexibility, course design are the important factors for online learning.
Salvador(2008) selected Master of Engineering in Professional Practice (MEPP) distance education program and Sloan Consortium in University of Wisconsin-Madison as research case. He interviewed the learners with the topic of ‘what are the main characteristic quality online education program’. He found that learners attached great important to the diversity of learners and participation, experienced and responsible professional personnel, project design and flexibility of the structure, the contact level of curriculum content and the practice, advanced technology and technical support. Besides, interaction between teachers and students was the core factor in the quality of online education pregame.
Yi Yang(2006)selected three items from American Institute for Higher Education Policy consisted of seven items, that were the process of teaching, course organization, learner support, and formed a questionnaire about the evaluation of online education quality. He found that the main factors were peer interaction level, teachers’ feedback and online course structure.
While these studies contribute to our understanding of learners’ perspectives on quality in DE/e-learning, they also point to the fact that there can be some significant differences across different contexts and depending on learner variables such as gender and learning experience. In order to provide useful references to develop QA framework for DE/e-learning in China, we need to identify Chinese learners’ perspective of DE/e-learning quality and understand personal differences in Chinese learners’ perceptions, concerns and experiences which affect their participation in DE/e-learning.
The present study aims to provide practical references to the development of QA frameworks for DE/e-Learning by investigating Chinese learners’ perceptions of the quality of DE/e-learning and attending to their differences with teachers in such dimensions.
Research Question
In examining Chinese learners’ perspective of the quality of DE/e-learning, specific research questions are formulated as follows:
Methodology
Participants
653 learners who were enrolled in three different DE/e-learning institutions in China (School of Distance Learning and Continuing Education Beijing Jiaotong University, Online Education of Renmin University of China, Beijing Open University) participated in this study in February 2010 and July 2011 separately. Questionnaires are distributed through mails and online survey. At last, there are 404 valid questionnaires in all. Table 1 shows the demographic features of participant learners.
Table 1 Demographics of participants (N=404) |
||||||||
Label |
N |
% |
Label |
N |
% |
|||
Gender |
||||||||
Female |
191 |
47.28 |
Male |
209 |
52.73 |
|||
Missing |
4 |
0.99 |
||||||
DE/e-learning experience |
Age |
|||||||
Extensive |
19 |
4.73 |
-20 |
112 |
27.86 |
|||
Quite a lot |
122 |
30.35 |
21-30 |
174 |
43.28 |
|||
Some experience |
131 |
32.59 |
31-40 |
96 |
23.88 |
|||
Very limited |
50 |
12.44 |
41-50 |
18 |
4.48 |
|||
None |
75 |
18.66 |
51-60 |
2 |
0.50 |
|||
Missing |
5 |
1.24 |
61-70 |
0 |
0.00 |
|||
71- |
0 |
0.00 |
||||||
Missing |
0 |
0.00 |
||||||
Learning place |
Learning time |
|||||||
At home |
255 |
56.82 |
Weekdays morning |
7 |
2.35 |
|||
In a library |
6 |
1.67 |
Weekdays afternoon |
11 |
3.69 |
|||
At my workplace |
79 |
7.80 |
Weekdays evening |
127 |
42.62 |
|||
In a café |
0 |
14.21 |
Weekdays late night |
15 |
5.03 |
|||
Outside in a park |
2 |
0.00 |
Weekends morning |
13 |
4.36 |
|||
No particular place |
27 |
0.56 |
Weekends afternoon |
9 |
3.02 |
|||
Other |
0 |
56.82 |
Weekends evening |
17 |
5.70 |
|||
Missing |
41 |
1.67 |
Weekends late night |
13 |
4.36 |
|||
Other |
9 |
3.02 |
||||||
Missing |
7 |
25.84 |
||||||
Major learning method |
||||||||
Reading textbooks/printed learning materials |
134 |
34.99 |
||||||
Using broadcast programs |
8 |
2.09 |
||||||
Content-based online study |
154 |
40.21 |
||||||
Interactive online study |
23 |
6.01 |
||||||
Face to face tutorials |
64 |
16.71 |
||||||
Others |
0 |
1.31 |
||||||
Missing |
5 |
34.99 |
||||||
Biggest difficulty |
||||||||
Financial difficulties |
14 |
3.81 |
||||||
Living far from study centers |
49 |
13.03 |
||||||
Lack of study time |
85 |
22.61 |
||||||
Conflict with other responsibilities at work |
126 |
33.51 |
||||||
Conflict with family obligations |
12 |
3.19 |
||||||
Inadequate technology environment |
5 |
1.33 |
||||||
Lack of self-motivation to study |
37 |
9.84 |
||||||
Lack of prior knowledge on subjects |
21 |
5.59 |
||||||
Lack of distance learning skills |
16 |
4.26 |
||||||
Other (specify) |
2 |
0.53 |
||||||
Missing |
9 |
2.39 |
In addition, the research team collects 91 teachers’ perspectives with the same QA framework as students’ questionnaires. There are senior managers, instructors and tutors from eight different DE institutions and 77 valid questionnaires are taken into further data analysis.
Instruments
An online survey was conducted to gather empirical evidence about a set of 59 items in the ten dimensions of QA in DE/e-learning from a perspective of adult learners in China. The purpose of the survey was to determine the level of importance of the items across ten dimensions so as to identify quality dimensions and criteria as perceived by learners in a DE/e-learning environment. In order to develop a valid and reliable survey questionnaire, the following steps were undertaken.
Sufficient common ground has been found in previous studies to enable the research team to develop an initial list of quality dimensions for use in the study. As a team member of Pandora project-Quality Assurance (QA) Models, Standards and Key Performance Indicators for ICT-supported Distance Education (DE) in Asia,which kicks off in 2009, we carry out similar student survey, draft QA framework for Asian DE/e-learning institutions, and organize several international workshops around QA framework for Asian DE/e-learning institutions. Due to that research experience in Pandora project, we are able to develop the initial list which included 55 items across ten dimensions (Institutional Credibility, Course Development, Information & Publicity, Evaluation & Assessment, Faculty Support, Interactive Tasks, Teaching and Learning, Infrastructure, Internal Quality Assurance System, and Student Support).
Following that, around 20 distance learners and 6 experts from four different DE providers and research centers are invited to review the draft. As a result, three were deleted because they were redundant, nine were revised for terminology clarification, and seven new items were added to reflect diverse DE/e-learning situations in China. In all, 59 items were included in the final survey.
In the survey questionnaire, respondents were asked to rate each item’s importance in assessing and assuring the ‘QUALITY’ of DE/e-learning (1 being lowest 5 being highest).
Data analysis
Confirmatory Factor Analysis (CFA) via Structural Equation Modeling (SEM) was carried out with AMOS 18.0 to answer the research questions 1,2 and 4. Independent sample t-test, one-way ANOVA and post hoc, multiple comparisons were carried out to answer the question 3.
DE/e-learning quality dimensions
It appeared that all ten dimensions were important in assuring the quality of DE/e-learning from the perspective of Chinese learners and teachers. In students’ questionnaires, factor loadings ranged from .656 (Institutional credibility) to .911 (Teaching and learning). Teachers’ questionnaires’ results about these ten dimensions is a bit different from students’. Table 2 has detailed information about the difference between students and teachers.
Table 2 Comparison of the importance of ten dimensions between students and teachers
Order | Students N=404(Factor loading) |
Teachers N=77(Factor loading) |
1 | Teaching and learning(.911) |
Infrastructure(.952) |
2 | Student support(.907) |
Information and publicity(.932) |
3 | Evaluation and assessment(.883) |
Evaluation and assessment(.877) |
4 | Interactive tasks(.855) |
Faculty support(.822) |
5 | Course development(.849) |
Interactive tasks(.822) |
6 | Faculty support(.848) |
Course development(.807) |
7 | Information and publicity(.837) |
Teaching and learning(.779) |
8 | Infrastructure(.826) |
Student support(.750) |
9 | Internal quality assurance(.816) |
Institutional credibility(.657) |
10 | Institutional credibility(.656) |
Internal quality assurance(.649) |
Regards to 59 items in those ten dimensions, the results appeared that most items were important in assuring the quality of DE/e-learning from the perspective of Chinese learners and teachers. In students’ questionnaires, factor loadings ranged from .656 (Institutional credibility) to .911 (Teaching and learning). Teachers’ questionnaires’ results about these ten dimensions is a bit different from students’. Table 3 has detailed information about the difference between students and teachers.
Table 3 Comparison of the importance of 59 items between students and teachers
Note: Items with * indicates the factor loading is lower than 0.6 which is not a good indicator. |
||||||||
Items |
Factor Loading |
|||||||
Students |
Teachers |
|||||||
Institutional credibility |
.656 |
.657 |
||||||
1 |
External accreditation at the national level |
.604 |
.400* |
|||||
2 |
International accreditation |
.549* |
.391* |
|||||
3 |
Strong leadership |
.870 |
.578* |
|||||
4 |
Clear lines of authority in decision making |
.856 |
.703 |
|||||
5 |
Qualified faculty/staff |
.611 |
.470* |
|||||
Course development |
.849 |
.807 |
||||||
6 |
Clear guidelines for course development |
.676 |
.612 |
|||||
7 |
Inclusion of multimedia components in courses |
.668 |
.621 |
|||||
8 |
Well structured course materials |
.675 |
.635 |
|||||
9 |
Inclusion of video-recorded lectures |
.639 |
.358* |
|||||
10 |
Course content adaptability to students’ needs |
.747 |
.598* |
|||||
11 |
Course content adaptability to students’ levels |
.759 |
.592* |
|||||
Information and publicity |
.837 |
.932 |
||||||
12 |
Providing course information (course objectives, assignments, timelines, study requirements, resources, learning outcomes, etc. for each courses) |
.778 |
.774 |
|||||
13 |
Clear indication of requirements for assignments (due dates, evlautation criteria, etc.) |
.747 |
.793 |
|||||
14 |
Guidelines about how to finish assignments |
.766 |
.810 |
|||||
15 |
Easy channels to know own learning progress |
.647 |
.372* |
|||||
16 |
Make good use of online platform to inform learners |
.646 |
.328* |
|||||
Evaluation and assessment |
.883 |
.877 |
||||||
17 |
Timely feedback to student assignments and questions |
.797 |
.919 |
|||||
18 |
Fair rubrics for learning assessment |
.874 |
.894 |
|||||
19 |
Periodic student evaluation of teaching and learning |
.780 |
.748 |
|||||
20 |
Periodic institutional review of lectures’/tutors’ performance |
.750 |
.681 |
|||||
21 |
Feedback from graduates |
.786 |
.603 |
|||||
22 |
Feedback from employers |
.722 |
.464* |
|||||
Faculty support |
.848 |
.822 |
||||||
23 |
Continuous assistance for faculty/tutors/staff in course development, delivery and management |
.793 |
.694 |
|||||
24 |
Periodic training for faculty/tutors/staff |
.824 |
.744 |
|||||
25 |
Policy and procedures for faculty/tutor/staff selection |
.801 |
.679 |
|||||
26 |
Faculty/tutor/staff welfare (e.g. financial aid and health care for faculty/tutor/staff) |
.558* |
.280* |
|||||
Interactive tasks |
.855 |
.822 |
||||||
27 |
Inclusion of collaborative learning activities in courses |
.801 |
.614 |
|||||
28 |
Inclusion of individualized learning activities in courses |
.830 |
.916 |
|||||
29 |
Inclusion of problem/case-based learning activities in courses |
.829 |
.904 |
|||||
Teaching and learning |
.911 |
.779 |
||||||
30 |
Student interaction with instructors |
.758 |
.782 |
|||||
31 |
Student interaction with other students |
.800 |
.726 |
|||||
32 |
Asynchronous online interaction (discussion boards, emails, etc.) |
.769 |
.740 |
|||||
33 |
Synchronous interaction (video conferencing, fixed line telephone etc.) |
.747 |
.733 |
|||||
34 |
Flexibility in learning methods |
.752 |
.543* |
|||||
35 |
Flexibility in learning pace |
.770 |
.523* |
|||||
36 |
Face-to-face tutorials |
.719 |
.475* |
|||||
37 |
Online tutorials |
.762 |
.646 |
|||||
38 |
Access to online library resources |
.742 |
.664 |
|||||
39 |
Access to physical library resources |
.634 |
.418* |
|||||
40 |
Informal face-to-face meeting with instructors/tutors |
.723 |
.473* |
|||||
41 |
Informal face-to-face meeting with other students |
.676 |
.607 |
|||||
42 |
Instructors’ teaching skills |
.482* |
.176* |
|||||
Infrastructure |
.826 |
.952 |
||||||
43 |
Reliable media/technology infrastructure |
.710 |
.645 |
|||||
44 |
Reliable learning management systems |
.263* |
.672 |
|||||
45 |
Physical classrooms |
.520* |
.354* |
|||||
46 |
Media/technology production facilities |
.745 |
.551* |
|||||
47 |
Security of student data system |
.643 |
.625 |
|||||
Internal quality assurance system |
.816 |
.649 |
||||||
48 |
Existence of quality standards specifically for distance education |
.811 |
.478* |
|||||
49 |
Periodic internal evaluation by a distance education institution |
.790 |
.873 |
|||||
50 |
Periodic evaluation by external experts |
.798 |
.560* |
|||||
51 |
Clear guidelines for quality assurance |
.782 |
.800 |
|||||
Student support |
.907 |
.750 |
||||||
52 |
Providing program/course administration information (admission requirements, tuition fees, technical and assessment requirements, students support services, etc) |
.617 |
.386* |
|||||
53 |
Distance learning skills training for students(e.g. how to succeed in DE) |
.792 |
.678 |
|||||
54 |
Media/technology support for students(e.g. operating 24/7 help desk) |
.788 |
.802 |
|||||
55 |
Social support for students (e.g. encouragement, interpersonal communication with faculty and tutors) |
.798 |
.713 |
|||||
56 |
Psychological support for students(e.g. counseling services) |
.617 |
.666 |
|||||
57 |
Administrative support for students(e.g. enrollment and admission services) |
.733 |
.606 |
|||||
58 |
An established appeal mechanism |
.653 |
.458* |
|||||
59 |
Flexible payment method |
.660 |
.286* |
Age group differences in rating DE/e-learning quality dimensions
The participants were grouped in two categories to see age group differences in the learners’ perception of DE/e-learning quality: ‘Younger than 31’ and ‘31 or older’ groups. About 71% belonged to the first group and 29% in the second group. An independent-sample t-test disclosed that there were mean differences between these two groups in rating the dimensions of DE/e-learning quality. Table 4 shows that the older group perceived these 10 dimensions more important in assessing the quality of DE/e-learning than the younger group, which is quiet significant in the dimension of teaching and learning.
Table 4 Age Group Differences in Rating DE/e-learning Quality Dimensions
Younger than 31 |
31 or older- |
||||||||||||||||||||||
DE/e-learning quality |
M |
SD |
n |
M |
SD |
n |
df |
t Sig. |
|||||||||||||||
Credibility |
4.24 |
.7944 |
286 |
4.29 |
.6751 |
116 |
648.8 |
-0.60 |
|||||||||||||||
Course development |
4.12 |
.755 |
286 |
4.27 |
.651 |
116 |
645.12 |
-1.92 |
|||||||||||||||
Information & publicity |
4.21 |
.647 |
286 |
4.31 |
.727 |
116 |
592.53 |
-1.29 |
|||||||||||||||
Evaluation & assessment |
4.12 |
.85 |
286 |
4.16 |
.77 |
116 |
633.89 |
-0.48 |
|||||||||||||||
Faculty support |
4.13 |
.897 |
286 |
4.30 |
.729 |
116 |
660.03 |
-1.75 |
|||||||||||||||
Interactive tasks |
4.10 |
.856 |
286 |
4.23 |
.741 |
116 |
644.13 |
-1.44 |
|||||||||||||||
Infrastructure |
4.07 |
.738 |
286 |
4.25 |
.663 |
116 |
635.69 |
-2.36 |
* |
||||||||||||||
Teaching & learning |
3.98 |
.784 |
286 |
4.31 |
1.011 |
116 |
573.72 |
-3.56 |
*** |
||||||||||||||
Internal QA system |
3.96 |
.914 |
286 |
4.04 |
.882 |
116 |
620.03 |
-0.89 |
|||||||||||||||
Student support |
4.06 |
.729 |
286 |
4.18 |
.716 |
116 |
616.65 |
-1.48 |
|||||||||||||||
* p<.05,*** p<.001 |
Effects of learning methods on rating DE/e-learning quality dimensions
One-way ANOVA revealed that there were differences in learners’ perception of each quality dimensions according to learning methods (see Table 5). A post hoc test using pair-wise comparison with LSD’s method showed that in assessing the quality of DE/e-learning, a group whose main method was ‘interactive online study’ perceived most quality dimensions more important compared with other groups, especially groups using reading textbooks/printed learning materials and broadcast programs. A group using content-based online study felt that ‘Faculty support’, ‘Infrastructure’ , ‘Teaching & learning’ and‘Student support’ were especially important in measuring the quality when compared with groups using reading textbooks/printed learning materials. Students who study online appeared to view ‘Course development’, ‘Infrastructure’, ‘Internal QA system’ and ‘Student support’ as more important dimensions of quality e-learning and list ‘Evaluation and assessment’ as the most important one compared with other students who conventional DE methods such as prints and broadcast programs.
Table 5 Effects of Learning Methods on DE/e-learning Quality Dimensions
Reading textbooks/printed learning materials (1) |
Using broadcast programs (2) |
Content-based online study (3) N=157 |
Interactive online study (4) N=124 |
Face to face tutorials (5) N=64 |
||||||||||||||||||
DE/e-learning quality |
M |
SD |
M |
SD |
M |
SD |
M |
SD |
M |
SD |
F |
Sig. |
Post hoc |
|||||||||
Credibility |
4.22 |
0.06 |
4.40 |
0.26 |
4.24 |
0.61 |
4.42 |
0.18 |
4.29 |
0.09 |
.723 |
a) |
||||||||||
Course development |
4.09 |
0.06 |
4.06 |
0.35 |
4.19 |
0.06 |
4.48 |
0.18 |
4.17 |
0.08 |
1.633 |
1>4 |
||||||||||
Information & publicity |
4.13 |
0.06 |
4.35 |
0.17 |
4.28 |
0.05 |
4.57 |
0.11 |
4.23 |
0.07 |
2.208 |
3 >1 |
||||||||||
Evaluation & assessment |
4.00 |
0.07 |
4.08 |
0.34 |
4.19 |
0.06 |
4.61 |
0.12 |
4.14 |
0.11 |
2.857 |
* |
||||||||||
Faculty support |
4.11 |
0.07 |
4.38 |
0.22 |
4.26 |
0.07 |
4.47 |
0.15 |
4.08 |
0.12 |
1.579 |
3, 4>1 |
||||||||||
Interactive tasks |
4.05 |
0.07 |
4.38 |
0.16 |
4.23 |
0.06 |
4.50 |
0.18 |
4.03 |
0.11 |
2.670 |
* |
a) |
|||||||||
Infrastructure |
4.02 |
0.06 |
4.17 |
0.26 |
4.19 |
0.06 |
4.49 |
0.13 |
4.13 |
0.09 |
2.786 |
* |
4,3>1 |
|||||||||
Teaching & learning |
3.96 |
0.06 |
4.15 |
0.22 |
4.17 |
0.08 |
4.30 |
0.20 |
4.07 |
0.09 |
1.816 |
3>1 |
||||||||||
Internal QA system |
3.94 |
0.07 |
4.16 |
0.22 |
4.06 |
0.07 |
4.39 |
0.17 |
3.80 |
0.14 |
2.970 |
* |
4>1,5 |
|||||||||
Student support |
3.98 |
0.06 |
4.19 |
0.31 |
4.17 |
0.06 |
4.62 |
0.11 |
4.03 |
0.10 |
4.303 |
*** |
3,4>1 |
|||||||||
* p<.05,*** p<.001 |
Effects of DE/e-learning experience on rating DE/e-learning quality dimensions
One-way ANOVA disclosed that DE/e-learning experience affected learners’ perception of several quality dimensions (see Table 6). Dimensions such as ‘Teaching and learning’, ‘Student support’ , ‘Evaluation and assessment’ and ‘Faculty support’ were perceived important regardless learners’ level of DE/e-learning experience. Students with more experience in DE/e-learning viewed some dimensions such as ‘Course development’ and ‘Faculty support’ more important than less experienced learners.
Table 6 Effects of Experience in DE/e-learning on DE/e-learning Quality Dimensions
Extensive (1) |
Quite a lot(2) |
Some experience (3) |
Very limited (4) |
None (5) N=75 |
|
|
|
|||||||||||||||
DE/e-learning quality |
M |
SD |
M |
SD |
M |
SD |
M |
SD |
M |
SD |
F |
Sig. |
Post hoc |
|||||||||
Credibility |
4.69 |
0.75 |
4.35 |
0.58 |
4.12 |
0.86 |
4.32 |
0.73 |
4.11 |
0.82 |
3.426 |
** |
|
1,2>3,5 |
||||||||
Course development |
4.81 |
0.37 |
4.28 |
0.61 |
3.98 |
0.80 |
4.20 |
0.62 |
4.04 |
0.81 |
6.450 |
*** |
|
1>2,3,4,5 |
||||||||
Information & publicity |
4.77 |
0.40 |
4.31 |
0.59 |
4.12 |
0.70 |
4.25 |
0.58 |
4.22 |
0.78 |
4.161 |
*** |
|
1 >2,3,4,5 |
||||||||
Evaluation & assessment |
4.76 |
0.51 |
4.25 |
0.66 |
3.98 |
0.89 |
4.04 |
0.96 |
4.11 |
0.70 |
4.332 |
*** |
|
1>2,3,4,5 |
||||||||
Faculty support |
4.86 |
0.33 |
4.36 |
0.64 |
3.97 |
0.96 |
4.18 |
0.97 |
4.10 |
0.84 |
5.501 |
*** |
|
1>2,3,4,5 |
||||||||
Interactive tasks |
4.79 |
0.37 |
4.33 |
0.68 |
3.89 |
0.91 |
4.18 |
0.77 |
4.09 |
0.87 |
6.490 |
*** |
|
1>2,3,4,5 |
||||||||
Infrastructure |
4.69 |
0.38 |
4.28 |
0.63 |
3.93 |
0.80 |
4.08 |
0.63 |
4.11 |
0.71 |
5.818 |
*** |
|
1>2,3,4,5 |
||||||||
Teaching & learning |
4.78 |
0.38 |
4.17 |
0.73 |
3.87 |
1.07 |
4.15 |
0.73 |
4.01 |
0.75 |
4.747 |
*** |
|
1>2,3,4,5 |
||||||||
Internal QA system |
4.73 |
0.33 |
4.18 |
0.76 |
3.81 |
0.95 |
4.07 |
0.89 |
3.72 |
1.02 |
6.458 |
*** |
|
1>2,3,4,5 |
||||||||
Student support |
4.78 |
0.31 |
4.18 |
0.70 |
3.90 |
0.77 |
4.18 |
0.62 |
4.11 |
0.71 |
6.254 |
*** |
|
1>2,3,4,5 |
||||||||
** p<.01,*** p<.001 |
Findings
As Table 2 shows, ‘teaching and learning ’, ‘student support’ and ‘evaluation and assessment’ were viewed as top three important indicators of the quality of DE/e-learning from the perspective of Chinese learners. In addition, Chinese teachers point out ‘Infrastructure’, ‘Information and publicity’ and ‘Evaluation and assessment’ are the most influential indicators.
In the ten dimension, Chinese learners regards the following ten criteria as the most important items for quality assurance in DE/e-learning, which includes Strong leadership, Course content adaptability to students’ levels, Providing course information, Fair rubrics for learning assessment, Periodic training for faculty/tutors/staff, Inclusion of individualized learning activities in courses, Student interaction with other students, Media/technology production facilities, Existence of quality standards specifically for distance education, Social support for students.
From Table 4 to Table 6 present the analysis about learners’ characteristics effects on their view of the quality of DE/e-learning. Generally speaking, learning experience has a greater impact on learners’ perspectives than age and learning methods.
The data result in Table 3 reveals the big difference between learners teachers in the perception of the quality of DE/e-learning. Three major difference lies in the dimensions of institutional credibility, teaching and learning and course development.
Conclusions
This study has explored Chinese learners’ perception of the quality of DE/e-learning and compare the difference between learners’ and teachers’ by dividing the concept of the quality of DE/e-learning into ten dimensions, which includes Institutional credibility,Course development,Information and publicity,Evaluation and assessment,Faculty support,Interactive tasks,Infrastructure,Teaching and learning,Internal quality assurance,and Student support. The importance of each criteria in each dimension viewed by learners and teachers was empirically examined using Confirmatory Factor Analysis (CFA) via Structural Equation Modeling (SEM). The results revealed considerable support for policies makers especially quality benchmarks developer in distance education. Besides, the difference of learners’ and teachers’ perception and parts of learners’ view according age, learning method and experience were found to be significant, providing significant insight into developing proper quality standards.
As with all empirical research, this study has its limitations. First, this study aimed to investigate learners’ view of quality of DE/e-learning in higher education. As a result, the findings of this study may not be generalized to the quality assurance in different contexts, such as elementary schools and high schools. More research that aims to examine the proposed quality framework using a variety of samples in similar and different contexts is needed to further validate and refine the investigation instrument. Second, this study has investigated the perception of quality of DE/e-learning by incorporating ten variables from various perspectives, and the variables in the proposed investigation theoretical framework were validated important by Confirmatory Factor Analysis (CFA) via Structural Equation Modeling (SEM). However, there is still room for improvement. To further our understanding of the perception of quality of DE/e-learning in various contexts, we strongly encourage future research that deletes less important variables, adjust less important criteria, and combine parts of dimensions into a broader concept to enhance the usability of this framework.
Acknowledgements
The authors thank the participants for their valuable feedback on this investigation. This study was funded by International Council of Distance Education, ICDE, Pandora project-Quality Assurance (QA) Models, Standards and Key Performance Indicators for ICT-supported Distance Education (DE) in Asia.
References