Engaging online learners: The impact of Web-based learning technology on college student engagement

Engaging online learners: The impact of Web-based learning technology on college student engagement

Computers & Education 54 (2010) 1222–1232 Contents lists available at ScienceDirect Computers & Education journal homepage: www.elsevier.com/locate/...

207KB Sizes 1 Downloads 55 Views

Computers & Education 54 (2010) 1222–1232

Contents lists available at ScienceDirect

Computers & Education journal homepage: www.elsevier.com/locate/compedu

Engaging online learners: The impact of Web-based learning technology on college student engagement Pu-Shih Daniel Chen a,*, Amber D. Lambert b, Kevin R. Guidry b a b

Department of Counseling and Higher Education, University of North Texas, 1155 Union Circle #310829, Denton, TX 76203-5017, USA Center for Postsecondary Research, Indiana University Bloomington, USA

a r t i c l e

i n f o

Article history: Received 31 July 2009 Received in revised form 30 October 2009 Accepted 16 November 2009

Keywords: Online learning Engagement College University NSSE Web-based Deep learning

a b s t r a c t Widespread use of the Web and other Internet technologies in postsecondary education has exploded in the last 15 years. Using a set of items developed by the National Survey of Student Engagement (NSSE), the researchers utilized the hierarchical linear model (HLM) and multiple regressions to investigate the impact of Web-based learning technology on student engagement and self-reported learning outcomes in face-to-face and online learning environments. The results show a general positive relationship between the use the learning technology and student engagement and learning outcomes. We also discuss the possible impact on minority and part-time students as they are more likely to enroll in online courses. Ó 2009 Elsevier Ltd. All rights reserved.

1. Introduction The Internet and other digital technologies have become thoroughly integrated in the lives of today’s college student. A recent study by EDUCAUSE (Hawkins & Rudy, 2008) found that the vast majority of US students at baccalaureate degree-granting institutions own and use their own computers. Online learning management systems (LMS) such as Blackboard, D2L, or Sakai are nearly ubiquitous on American colleges and universities, and wireless Internet access permeates most college classrooms (Green, 2007; Hawkins & Rudy, 2008). Outside the classroom, Internet connections are available in virtually all on-campus residence halls (Hawkins & Rudy, 2008) and an estimated 79– 95% of all American College students use Facebook and MySpace (Ellison, 2007). Most first-year college students now arrive on campus with their own personal computer, digital music player, cell phone, and other digital devices (Salaway & Caruso, 2008). As technology becomes a part of modern life and fuel price remains high, more and more college students opt to take online or hybrid courses using readily-available computers and information technologies (Allen & Seaman, 2008). Moreover, many students expect instructors to integrate Internet technologies, such as online learning management systems and collaborative Internet technologies, into traditional face-to-face classes to enhance learning experience, believing those tools make the educational experience more convenient and educationally effective (Salaway & Caruso, 2008). Since the early 2000s, Web-based applications have become the de facto standard platform for distance education courses and learning management systems (Parsad & Lewis, 2008). The widespread adaptation of digital technologies and online courses has caused many researchers (Bråten & Streømsø, 2006; Kuh & Hu, 2001; Robinson & Hullinger, 2008; Zhou & Zhang, 2008) to question the impact of the Internet and Web-based learning technology on student’s educational engagement and learning outcomes. The concept of student engagement is not new to educators. Years of research has shown that what students do during college counts more in terms of learning outcomes than who they are or even where they go to college (Austin, 1993; Kuh, 2004; Pace, 1980; Pascarella & Terenzini, 2005). In the Seven principles for good practice in undergraduate education, Chickering and Gamson (1987) argued that good college education should promote student-faculty interaction, cooperation among students, active learning, prompt feedback, time on task, high expectations, and respect for diverse talents and ways of learning. In a follow-up article published in 1996, Chickering and Ehrmann (1996) stated that new

* Corresponding author. Tel.: +1 940 369 8062; fax: +1 940 369 7177. E-mail addresses: [email protected] (Pu-Shih Daniel Chen), [email protected] (A.D. Lambert), [email protected] (K.R. Guidry). 0360-1315/$ - see front matter Ó 2009 Elsevier Ltd. All rights reserved. doi:10.1016/j.compedu.2009.11.008

Pu-Shih Daniel Chen et al. / Computers & Education 54 (2010) 1222–1232

1223

communication and information technology alone will not lead to student success. Instead, educators must utilize technology as a lever to promote student engagement in order to maximize the power of computers and information technology as a catalyst for student success in college (Ehrmann, 2004). Most studies on the topic of technology and student engagement have affirmed the utility of computers and information technology on promoting student engagement (Hu & Kuh, 2001; Nelson Laird & Kuh, 2005; Robinson & Hullinger, 2008). For example, Robinson and Hullinger found that asynchronous instructional technology allows learners more time to think critically and reflectively, which in turns stimulates higher order thinking such as analysis, synthesis, judgment, and application of knowledge. Duderstadt, Atkins, and Houweling (2002) stated, ‘‘When implemented through active, inquiry based learning pedagogies, online learning can stimulate students to use higher order skills such as problem solving, collaboration, and stimulation” (p. 75). Furthermore, students taking online courses are expected to work collaboratively, which is an important component of student engagement, plus that collaborative components have been integrated into most Web-based course designs (Thurmond & Wambach, 2004). Other than promoting student engagement, research focused on the connection between technology and learning outcomes has been mixed. George Kuh and his associates have published several articles related to this issue using the National Survey of Student Engagement (NSSE) data. In Kuh and Hu (2001), the authors suggested a positive relationship between a student’s use of computers and other information technologies and self-reported gains in science and technology, vocational preparation, and intellectual development. Hu and Kuh (2001) also found that students attending more ‘‘wired” institutions reported more frequently use computing and information technology and higher levels of engagement in good educational practices than their counterparts at less wired institutions. A similar study conducted by Kuh and Vesper (2001) concluded that increased familiarity with computers was positively related to developing other important skills and competencies, including social skills. Studies conducted by other researchers, however, have mixed outcomes that have often not been as positive as those reported by George Kuh and his associates. A meta-analysis commissioned by the US Department of Education examined empirical evidence of the impact of online and hybrid courses on learning outcomes. The authors found that both online and hybrid courses have a significant positive impact on learning outcomes, with hybrid courses having a greater impact. However, the authors caution that the ‘‘positive effects associated with blended learning should not be attributed to the media, per se” (p. ix) (Means, Toyama, Murphy, Bakia, & Jones, 2009). This reflects long-standing findings that, contrary to many naïve beliefs, media do not have a significant impact on learning outcomes (Clark, 2009). Other meta-analyses of distance education impacts on learning outcomes have supported these mixed findings (Bernard et al., 2004; Sitzmann, Kraiger, Stewart, & Wisher, 2006). While it is unclear if students learn more in online courses, it does seem clear that there is an increase in students’ information literacy. For example, Robinson and Hullinger (2008) found a correlation between taking online courses and the improvement of students’ computer skills. Though most online courses do not require students to have high level computer skills in order to complete the courses, they nevertheless require students to become familiar with essential information technological skills such as using e-mail, participating in online chatting, posting to a Web-based discussion board, and using word processing, presentation, and spreadsheet software. Even though there are many educational benefits associated with using computer technologies, there are also downsides. Critics have argued that online learning and the use of information technology may put certain student populations in disadvantage. Echoing Jenkins’ ‘‘participation gap” idea (Jenkins, 2006), some researchers have suggested that characteristics such as socioeconomic status (Gladieux & Swail, 1999) and institutional resources (Hu & Kuh, 2001) play a significant role in students’ use of and the impact of computers and the Internet. In addition, some researchers asserted that the lack of face-to-face interactions in online learning may reduce instructional effectiveness for students of certain learning styles (Bullen, 1998; Terrell & Dringus, 2000; Ward & Newlands, 1998). Sanders (2006) argued that no communication technology can replace the physical presence and the serendipitous moments of learning such as the spontaneous discussion or the overheard remarks during class break that so often occurred in a face-to-face environment. 1.1. Purpose of study and research questions Although studies have found positive connections between the use of computers and information technology and student engagement and learning outcomes, most of them studied the general use of information technology instead of the specific use of instructional and learning management systems. This study investigates the nature of student engagement in the online learning environment to find out if student and institutional characteristics affect the use of the learning technologies and their impact on student engagement. Specifically, the following research questions were addressed: 1. How often do college students in different types of courses use the Web and Internet technologies for course-related tasks? 2. Do individual and institutional characteristics affect the likelihood of taking online courses? 3. Does the relative amount of technology employed in a course have a relationship with student engagement, learning approaches, and student self-reported learning outcomes? 2. Methods 2.1. Instrument and data source The data for this study come from the 2008 administration of the National Survey of Student Engagement (NSSE). NSSE is an annual survey created and administered by the Indiana University Center for Postsecondary Research. Since the inception of the NSSE in 2000, more than a million first-year students and seniors at more than 1300 baccalaureate degree-granting colleges and universities in the United States and Canada have reported the time and energy that they devote to the educationally purposeful activities measured by this annual survey (Indiana University Center for Postsecondary Research, 2008b). Participating institutions use their student engagement results to identify areas where teaching and learning can be improved. NSSE results have been found to positively correlate with desired learning outcomes, such as critical thinking ability and grades (Carini, Kuh, & Klein, 2006; Kuh, 2004; Ouimet, Bunnage, Carini, Kuh, & Kennedy,

1224

Pu-Shih Daniel Chen et al. / Computers & Education 54 (2010) 1222–1232

2004; Pike, 2006). The conceptual framework and psychometric properties of the NSSE and the development of NSSE scales have been amply documented (Kuh, 2004; Nelson Laird, Shoup, & Kuh, 2005). In 2007, researchers at NSSE developed a set of questions to investigate the nature of student engagement in the online learning environment. The original set of questions includes 22 questions. After pilot testing and expert review, the items were revised and the numbers were reduced to 13 (see Appendix for the list of items). The final set of 13 items asks respondents to identify the number of classes in which they were enrolled in the last academic year and how many of those courses were conducted entirely online or face-to-face with a significant online component. Survey respondents also reported on specific behaviors related to their collegiate experiences, including in- and out-of-class behaviors, time usage, and learning approaches that are known to contribute to desirable learning outcomes. 2.2. Sample The NSSE online learning questions were attached to the end of the NSSE online survey and sent to participating students at 45 US baccalaureate degree-granting institution. The 45 institutions were randomly selected from the pool of 763 institutions participated in the 2008 NSSE administration. The institutions include 14 (31%) public and 31 (69%) private institutions; 8 (19%) of them were classified by the Carnegie Foundation for the Advancement of Teaching (2009) as doctoral institutions, 16 (38%) were master’s institutions, and 18 (43%) were baccalaureate institutions. Detailed institutional characteristics of the 45 participating institutions and their comparison with all 2008 NSSE participating institutions can be found in Table 1. The survey was sent to 77,714 first-year and senior college students and approximately 23,706 students responded to this set of questions, yielding a response rate of 30.5%. However, about 4500 students who were purposely sampled by the institutions were excluded from analysis, which leaves only students who were randomly sampled. Additionally, one institution that offers online courses only was removed from the dataset because no comparison among different course delivery methods can be made at this online institution. Removing this online institution did not greatly affect the general characteristics of the sample. Finally, 1825 students, who accounted for 7.7% of the total respondents, were excluded as their responses indicated that they may not understand these questions in the manner intended by the researchers (when summed, their responses indicated that over 100% of their classes were online or hybrid classes); this indicates a likely data reliability issue with these new questions that will be addressed when discussing this study’s limitations. The final data set for this study has 17,819 respondents, in which 8065 (45%) were first-year students and the remaining 9754 (55%) seniors. Nearly 7000 respondents (35%) were male and 13,000 (65%) female. The majority (97% for first-year students and 87% for senior students) of the surveyed students were enrolled full-time at their institution. Detailed student characteristics including gender, enrollment status, and race and ethnicity can be found in Table 2.

Table 1 Institutional characteristics.

a b

Institutions participated in this study (n = 45)

All NSSE 2008 institutions (n = 763)a

All US institutionsb

Count

Percentage (%)

Count

Percentage (%)

Percentage (%)

Control

Public Private

14 31

31 69

320 443

42 58

35 65

Carnegie classifications

Doctoral Master’s Baccalaureate

8 16 18

19 38 43

103 303 244

16 47 38

18 41 41

Urbanicity

City Suburban Town Rural

27 6 7 5

60 13 16 11

333 154 173 53

47 22 24 7

46 22 21 9

Not all NSSE participating institutions are classified by the Carnegie Foundation for the Advancement of Teaching. US percentages are based on data from the 2007 IPEDS institutional characteristics file as reported in Indiana University Center for Postsecondary Research (2008a).

Table 2 Respondent demographics. First-year

Senior

Count

Percentage (%)

Count

Percentage (%)

Gender

Male Female

2771 5274

34 66

3351 6375

35 65

Enrollment status

Part-time Full-time

259 7789

3 97

1175 8562

13 87

Race or ethnicity

African American or Black American Indian or other Native American Asian, Asian American, or Pacific Islander White (non-Hispanic) Hispanic, Mexican or Mexican American, Puerto Rican Other Multiracial No response

676 40 483 5753 279 124 208 502

8 1 6 71 4 2 3 6

881 60 437 7132 273 111 194 666

9 1 5 73 3 1 2 7

1225

Pu-Shih Daniel Chen et al. / Computers & Education 54 (2010) 1222–1232

2.3. Variables and data analysis For the purposes of this study, a Web or online course is defined as a course that is conducted entirely through the Internet without any face-to-face contact among instructor(s) and students. In contrast, a face-to-face course is a course that conducted entirely in a physical classroom without using any Internet technology for course management or instructional purpose. Although there are many definitions for hybrid learning, or so-called blended learning (Bersin, 2004; Driscoll, 2002; Reay, 2001; Rossett, 2001; Sands, 2002; Ward & LaBranche, 2003), Graham (2006) indicated that blended learning can be sorted into three categories: enabling blends, enhancing blends, and transforming blends. Enabling blends focus primarily on improving student access and convenience. Enhancing blends allow for incremental changes to the pedagogy while transforming blends carry radical transformation of the pedagogy. Learning management systems and technology equipped classrooms are two examples of enhancing blends. For the purpose of this study, the researchers adopted enhancing blends as the definition of hybrid courses. Therefore, a hybrid course is defined as one that blends both Web and face-to-face components in the same course. A hybrid course must include both face-to-face contacts among instructor(s) and students and the use of the Internet or Web technology for course management or instructional purpose. If the only utilization of the Internet or Web technology in a face-to-face course is for non-instructive or routine communication, the course is considered a face-to-face course rather than a hybrid course. To answer the first research question, descriptive statistics including means and standard deviations were reported for all of the survey items. The Kruskal Wallis Test (Siegel & Castellan, 1988), a nonparametric equivalent of the analysis of variance (ANOVA), was conducted to examine if statistically significant differences exist in student’s technology use among different course delivery methods. Hierarchical linear modeling (HLM) was utilized to answer the second research question (Raudenbush & Bryk, 2002). The assumption underlying the HLM analysis is that institutions have a differential impact on student’s course taking behaviors and technology usage. The benefit of using HLM is that it allowed the researchers to partition the variance attributable to the individual and the variance attributable to the institution. The dependent variables for the HLM analysis are the ratio of classes taken online. The independent variables include individual (level 1) variables such as the student’s gender, enrollment status (part-/full-time), ethnicity, major, and parental education. The institutional level variables (level 2 variables) are dummy-coded 2005 Carnegie basic classification, control (public/private), and urbanicity or locale. The third research question, which addresses the impact of learning technologies on student engagement and outcomes, was answered using Ordinary Least Squares (OLS) multiple regression analysis. A regression analysis is a statistical technique that allows the researcher to investigate the relationship between one dependent variable and several independent variables (Tabachnick & Fidell, 2007). The dependent variables for this analysis include four of the five NSSE Benchmarks of Effective Educational Practice (Kuh, 2004; LaNasa, Cabrera, & Trangsrud, 2009; Pascarella & Seifert, 2008) – level of academic challenge (LAC), active and collaborative learning (ACL), student-faculty interaction (SFI), and supportive campus environment (SCE), the three student self-reported Gain Scales (Chen, Ted, & Davis, 2007; Pike, 2006) – gain in general education, gain in personal and social development, and gain in practical competence, and the three deep learning scales (Nelson Laird et al., 2005) – higher order thinking, reflective learning, and integrative learning. One of the NSSE Benchmarks – enriching educational experiences (EEE) – is excluded from the analysis because technology use is part of the benchmark. The independent variables include the percentage of classes taken online, the percentage of classes that were hybrid classes, a composite score of course-related technology use, and controls for student and institutional characteristics. 3. Results 3.1. Descriptive statistics The first three questions of the survey asked students how many courses they took in the current academic year, how many of those courses used the Web or Internet as the primary method to delivery course content, and how many of those courses were hybrid courses. Using those responses, we were able to classify course delivery methods into three categories: Web or Internet-only, face-to-face, and hybrid. As a result of this classification, students can take courses in seven different patterns: Web-only, face-to-face-only, hybrid-only, some Web and hybrid, Web and face-to-face, some face-to-face and hybrid, and all three delivery methods. As shown in Table 3, very few (2.1%) of the 17,819 students who adequately completed the survey took all their courses in Web-only mode. A larger percentage of students took some Web courses and some hybrid courses (5.2%) while a similar percentage enrolled in both Web and face-to-face courses (7.6%). The majority (84.8%) took classes with at least some face-to-face component. Although some of those students were also enrolled in Web (7.6%), hybrid (21.5%), or both Web and hybrid (34.9%) courses, one-fifth (20.8%) of the respondents were enrolled only in face-to-face classes with no significant Web or Internet component. These seven groups were collapsed into five groups for later analyses: Web-only, hybrid-only, some Web, face-to-face and hybrid, and face-to-face-only. As shown in Tables 4 and 5, students whom one would expect to use technology more often – those enrolled in Web and hybrid classes – indeed used online learning tools and technologies more frequently than students who only took face-to-face courses. More specifically, Table 3 Distribution of course options. Course delivery method

First-year students Frequency

Senior students Percentage (%)

Frequency

Combined Percentage (%)

Web-only Hybrid-only Face-to-face-only Web and hybrid Web and face-to-face Face-to-face and hybrid All three delivery methods

90 628 1718 362 573 1699 2995

1.1 7.8 21.3 4.5 7.1 21.1 37.1

281 789 1988 561 776 2139 3220

2.9 8.1 20.4 5.8 8 21.9 33

Total

8065

100.00

9754

100.00

Frequency 371 1417 3706 923 1349 3838 6215 17,819

Percentage (%) 2.1 8 20.8 5.2 7.6 21.5 34.9 100.00

1226

Pu-Shih Daniel Chen et al. / Computers & Education 54 (2010) 1222–1232

Table 4 First-year student engagement in online learning activities. Web-only

How often: discussed or completed an assignment using a synchronous tool like instant messaging, online chat room, video conference, etc. How often: discussed or completed an assignment using an asynchronous tool like e-mail, discussion board, listserv, etc. How often: used your institution’s Web-based library resources in completing class assignments How often: used the Internet to discuss with an instructor topics you would not feel comfortable discussing face-to-face or in a classroom How often: used an electronic medium (listserv, chat group, Internet, instant messaging, etc.) to discuss or complete an assignment How often: used e-mail to communicate with an instructor To what extent does your institution emphasize using computers in academic work?

Hybrid-only

Some Web

SD

Mean

Hybrid and face-to-face

SD

Mean

SD

Face-to-faceonly

Mean

SD

Mean

1.91

1.174

1.72

.961

1.62

.886

1.50

.810

1.45

Mean

SD .824

3.12

1.091

2.62

.974

2.46

.931

2.39

.893

2.00

.928

2.40

.997

2.60

.910

2.45

.900

2.44

.861

2.29

.919

1.70

.993

1.87

.989

1.78

.940

1.69

.874

1.62

.882

3.07

1.095

2.66

1.044

2.66

1.037

2.61

1.001

2.33

1.047

3.40 3.56

.761 .781

3.25 3.42

.790 .744

3.25 3.33

.781 .780

3.17 3.30

.778 .753

3.04 3.15

.824 .821

Table 5 Senior student engagement in online learning activities. Web-only

How often: discussed or completed an assignment using a synchronous tool like instant messaging, online chat room, video conference, etc. How often: discussed or completed an assignment using an asynchronous tool like e-mail, discussion board, listserv, etc. How often: used your institution’s Web-based library resources in completing class assignments How often: used the Internet to discuss with an instructor topics you would not feel comfortable discussing face-to-face or in a classroom How often: used an electronic medium (listserv, chat group, Internet, instant messaging, etc.) to discuss or complete an assignment How often: used e-mail to communicate with an instructor To what extent does your institution emphasize using computers in academic work?

Hybrid-only SD

Some Web

Hybrid and face-to-face

Face-to-faceonly

Mean

SD

Mean

Mean

SD

Mean

SD

Mean

2.05

1.160

1.62

.921

1.64

.889

1.51

.812

1.34

SD .734

3.29

1.032

2.82

.986

2.69

.942

2.58

.915

2.07

.979

2.72

1.042

2.81

.964

2.75

.933

2.77

.939

2.52

1.020

1.77

1.086

1.82

.990

1.74

.933

1.61

.850

1.48

.819

3.25

1.018

2.99

1.009

2.91

.991

2.81

.979

2.47

1.067

3.67 3.72

.604 .594

3.53 3.64

.687 .613

3.47 3.49

.691 .716

3.43 3.48

.707 .711

3.28 3.37

.788 .799

respondents who were enrolled in online courses more frequently used both synchronous and asynchronous communication tools for instructional or learning purposes. Compared with students in traditional face-to-face setting, online students also more frequently used electronic media to discuss or complete assignments, and these differences were consistent for both first-year and senior students. One interesting finding is that students who took hybrid courses more frequently utilized the institutional Web-based library resources in completing class assignment than students who only had online courses or those only had face-to-face courses. A probable explanation is that students who took hybrid courses are more familiar with doing research online than students who took only face-to-face courses. On the other hand, students who only took online courses may feel comfortable with the Internet technologies but may not receive sufficient instruction on how to conducting research using Web-based library resources. We attempted to perform an analysis of variance (ANOVA) on the mean scores for these seven questions for both first-year and senior students to determine which, if any, of the apparent differences are statistically significant. These tests were abandoned as the assumptions of ANOVA, particularly homoscedacity, were only met in two of the 14 tests. A nonparametric test, the Kruskal Wallis Test, indicated that there are significant differences in the mean scores for each question among at least some of the groups of students. However, the very large number of respondents makes it difficult to make much meaning of the significant results of those tests given the sensitivity of the tests to the high number of respondents. 3.1.1. HLM one-way ANOVA model To answer the second research question, a hierarchical linear model (HLM) was built to investigate the impacts of individual and institutional variables on students’ course taking behaviors. Before estimating the full, two-level HLM to examine the effects of individual and institutional variables in the student’s likelihood of taking online courses, we used the one-way ANOVA model or so-called ‘‘null model” to estimate the proportion of variance that exists between and within colleges. The proportion of variance between institutions ranges from 0.033 for first-year students to 0.157 for seniors (Table 6). The result indicates that institutional variables have more influence on seniors than first-year students in their decision to take online courses. This result also warrants further investigation into what individual and institutional variables may affect student’s decision to take online courses. 3.1.2. HLM random coefficient regression and intercept- and slopes-as-outcomes models The second step of the modeling procedure is the creation of the random coefficient regression model, also known as the level 1 model or the individual level model. This procedure tests and establishes the individual-level independent variables before estimating the full, intercept- and slopes-as-outcomes model. Table 7 presents the descriptive statistics of the independent variables included in the analysis. The level 1 independent variables include student’s gender (0 = male, 1 = female), enrollment status (0 = full-time, 1 = part-time), ethnicity

1227

Pu-Shih Daniel Chen et al. / Computers & Education 54 (2010) 1222–1232 Table 6 Variance components of dependent variable. Ratio of online courses taken by the student

Total variance Variance within institutions Variance between institutions Proportion between institutions

First-year students

Seniors

.05929 .05731 .00198 .033

.08028 .06767 .01261 .157

Table 7 Descriptive statistics for independent variables included in models. First-year students

Individual characteristics First generation college student Female Part-time enrollment Ethnical minority STEM Arts, Humanities, and Social Sciences (reference) Business Professional Other and undecided Institutional characteristics Carnegie: doctoral institution Carnegie: master’s institution Carnegie: baccalaureate institution Carnegie: other Private City Suburban Town Rural

Seniors

Mean

SD

Min.

Max.

Mean

SD

Min.

Max.

Description

.38

.49

0

1

.42

.49

0

1

.64 .03 .28 .18 .26

.48 .18 .45 .39 .44

0 0 0 0 0

1 1 1 1 1

.65 .13 .26 .17 .28

.48 .33 .44 .37 .45

0 0 0 0 0

1 1 1 1 1

First generation college student is defined as neither parents has a baccalaureate degree from a college. 1 = first generation college student, 0 = all other Gender: 1 = female, 0 = male Enrollment status: 1 = enrolled part-time, 0 = enrolled full-time Ethnicity: 0 = White/Caucasian, 1 = all other Major: 1 = Science, Technology, Engineering, and Mathematics, 0 = all other Major: 1 = Arts, Humanities, and Social Sciences, 0 = all other

.17 .12 .16

.37 .32 .37

0 0 0

1 1 1

.18 .13 .15

.39 .34 .36

0 0 0

1 1 1

Major: 1 = Business, 0 = all other Major: 1 = professional, 0 = all other Major: 1 = Other majors and undecided, 0 = all other

.18

.39

0

1

.18

.39

0

1

Carnegie classification: 1 = doctorate granting universities, 0 = all other

.36

.48

0

1

.36

.48

0

1

Carnegie classification: 1 = master’s colleges and universities, 0 = all other

.4

.5

0

1

.4

.5

0

1

Carnegie classification: 1 = baccalaureate colleges, 0 = all other

.07

.25

0

1

.07

.25

0

1

.69 .6 .13 .16 .11

.47 .5 .34 .37 .32

0 0 0 0 0

1 1 1 1 1

.69 .6 .13 .16 .11

.47 .5 .34 .37 .32

0 0 0 0 0

1 1 1 1 1

Carnegie classification: 1 = special focus institutions, tribal colleges, noneclassified institutions Control: 1 = private, 0 = public Urbanicity: 1 = city, 0 = all other Urbanicity: 1 = suburban, 0 = all other Urbanicity: 1 = town, 0 = all other Urbanicity: 1 = rural, 0 = all other

(0 = White/Caucasian, 1 = minority), first generation college student status (0 = at least one parent has a baccalaureate degree, 1 = neither parent has a baccalaureate degree), and a series of dummy-coded variables for major (with Arts, Humanities, and Social Sciences being the reference category). The outcomes of the random coefficient regression model will be reported jointly with the final model. In the third and final step in the modeling process, we built the between-institution model by allowing the intercept to vary by institution. We then modeled the intercept with institutional characteristics. Included in the level 2 models are 2005 basic Carnegie classifications (doctorate granting universities, master’s colleges and universities, baccalaureate colleges, and others) with the doctorate granting universities serving as the reference category. We also included institution control (public or private) and locale or urbanicity (city, suburban, town, and rural, of which city serves as the reference category). To avoid multicollinearity, we did not include the size of the institution as a control because the size of institution is highly correlated with the Carnegie classification within our sample (r = .71, p < .001). Table 8 illustrates the summary effects of individual and institutional variables on student’s decision to take online courses. It is clear that the factors that affect online course taking for first-year students and seniors are quite different. For first-year students, enrollment in a private institution slightly increases the likelihood (p < .05) of enrollment in online courses while enrollment in a baccalaureate colleges and universities slightly reduces (p < .05) the chance of enrollment in online courses compared with their counterparts enrolled in a doctorate granting institutions. Contrary to their effect on first-year students, institutional variables have no statistically significant effect on senior students’ decision to take online courses. Although individual variables affect both first-year and senior students’ decision to take online courses, they tend to affect seniors more than first-year students. For first-year students, racial and ethnic minorities (p < .001) and part-time students (p < .05) are more likely to enroll in online courses. The same effects can also be found with senior students (both at p < .001). Additionally, seniors who major in the professional fields (e.g. education, nursing, occupational therapy. . . , etc.) are also more likely to enroll in online courses (p < .001). The student’s major has no effect on first-year student’s likelihood of taking online courses except for students in business, who are slightly more likely than students in other majors to enroll in online courses (p < .05). 3.2. Multiple regression models To answer the third research question, which addresses the impact of learning technologies on student engagement and outcomes, Ordinary Least Squares (OLS) multiple regression analysis was used. As can be seen in Tables 9 and 10, the total variance explained by the

1228

Pu-Shih Daniel Chen et al. / Computers & Education 54 (2010) 1222–1232

Table 8 Coefficients from HLM for the ratio of courses taken online by the student. First-Year Students

Seniors

Coefficient

p-value

Coefficient

p-value

Institution-level variables Intercept Carnegie: master’s Carnegie: baccalaureate Carnegie: other Private Locale: suburban Locale: town Locale: rural

.118 .01 .038 .039 .025 .016 .003 .039

.001 .435 .016 .282 .043 .282 .859 .075

.141 .004 .03 .27 .014 .001 .027 .001

.001 .816 .188 .628 .408 .992 .27 .995

Individual-level variables First generation college student Female Part-time Minority Major: STEM Major: business Major: professional Major: Other and undecided

.013 .01 .093 .035 .02 .02 .009 .001

.056 .113 .016 .001 .056 .032 .307 .952

.013 .005 .086 .047 .03 .004 .046 .008

.096 .421 .001 .001 .041 .778 .001 .518

Variance Variance Variance Variance Variance

.0006 69.70% .05407 5.65%

components between institutions between explained within institutions within explained

.00539 57% .06368 5.90%

Table 9 First-year students’ partitioning of variance for the deep learning scales, gains scales, and NSSE Benchmarks in multiple regression models. Variance due to

Studenta and institutionalb characteristics

Delivery of coursesc

Use of learning technologyd

Total variance explained

Deep learning scales Higher order thinking Integrative learning Reflective learning

.046*** .050*** .032***

.005*** .008*** .001***

.116*** .199*** .090***

.167*** .257*** .123***

Gains scales Person and social development Practical competence General education

.070*** .075*** .059***

.007*** .009*** .010***

.129*** .164*** .126***

.206*** .248*** .195***

NSSE Benchmarks Academic challenge Active and collaborative learning Supportive campus environment Student-faculty interaction

.085*** .096*** .076*** .106***

.008*** .004** .013*** .001***

.144*** .185*** .102*** .214***

.237*** .285*** .191*** .321***

a Student characteristics include: gender, enrollment status, parents’ education, grades, SAT scores, transfer status, age, membership in a fraternity/sorority, whether or not a student is a STEM field, race-ethnicity, and US citizenship. b Institutional characteristics include: Carnegie classification and control. c Delivery of courses included: the percentage of courses a student was taking online and the percentage of courses a student was taking face-to-face with Webcomponents. d Use of learning technology included: a single scale combining the seven questions asking students about how often they used certain course-related technology. ** p < .01. *** p < .001.

multiple regression models employed in this study is statistically significant in all cases and quite substantial in many of these models. For first-year students (Table 9), the variance explained by the models ranges from 12.3% to 32.1% while for seniors it ranges from 11.1% to 26.2% (Table 10). Of the variance explained the largest portion by far is students’ use of learning technology. In contrast, the delivery method of the courses in which students are enrolled seems to have a statistically significant but in most cases unsubstantial, impact on the variance explained for the model. In all of these models, the relationship between the NSSE Benchmarks of Effective Education Practices, deep approach of learning, and student self-reported educational gains, and the use of learning technology is positive and relatively strong. Table 11 displays the relative influence of learning technology with other forms of engagement and students learning. Multicollinearity is not a concern for this study as the only moderate correction happens between enrollment status and age (r = .47). All the other independent variables have a Pearson’s r less than .1.

4. Discussion The first research question asked: How often do college students in different types of courses use the Web and Internet technologies for course-related tasks? First, it is important to note that the majority of students in this study had classes that were entirely or partially in the

1229

Pu-Shih Daniel Chen et al. / Computers & Education 54 (2010) 1222–1232 Table 10 Seniors students’ partitioning of variance for the deep learning scales, gains scales, and NSSE Benchmarks in multiple regression models. Variance due to

Studenta and institutionalb characteristics

Delivery of coursesc

Use of learning technologyd

Total variance explained

Deep learning scales Higher order thinking Integrative learning Reflective learning

.143*** .251*** .111***

.032*** .069*** .038***

.005*** .012*** .007***

.106*** .170*** .066***

Gains scales Person and social development Practical competence General education

.091*** .069*** .078***

.004*** .013*** .009***

.119*** .138*** .089***

.214*** .220*** .176***

NSSE benchmarks Academic challenge Active and collaborative learning Supportive campus environment Student-faculty interaction

.045*** .082*** .065*** .074***

.013*** .015*** .008*** .010***

.132*** .165*** .085*** .161***

.190*** .262*** .158*** .245***



p < .01. a Student characteristics include: gender, enrollment status, parents’ education, grades, SAT scores, transfer status, age, membership in a fraternity/sorority, whether or not a student is a STEM field, race-ethnicity, and US citizenship. b Institutional characteristics include: Carnegie classification and control. c Delivery of courses included: the percentage of courses a student was taking online and the percentage of courses a student was taking face-to-face with Webcomponents. d Use of learning technology included: a single scale combining the seven questions asking students about how often they used certain course-related technology. *** p < .001.

Table 11 Net effectsa of use of learning technology on the deep learning scales, gains scales, and NSSE Benchmarks in multiple regression models. Variance due to

First-year students

Seniors

Deep learning scales Higher order thinking Integrative learning Reflective learning

++ ++ ++

++ ++ +

Gains scales Person and social development Practical competence General education

+++ +++ ++

+++ ++ +

NSSE Benchmarks Academic challenge Active and collaborative learning Supportive campus environment Student-faculty interaction

+ ++ + +++

+ ++ + +++

+, p < .001 and unstandardized B > .3; ++, p < .001 and unstandarized B > .4, +++, p < .001 and unstandarized B > .5. a Table reports results from ten multiple regression models (one per row). Student level controls include gender, enrollment status, parents’ education, grades, SAT scores, transfer status, age, membership in a fraternity/sorority, whether or not a student is a STEM field, race-ethnicity, US citizenship, the percentage of courses a student was taking online and the percentage of courses a student was taking face-to-face with Web-components. Institutional controls include Carnegie classification and control.

classroom. Very few were enrolled in all online courses and few were enrolled in hybrid-only or hybrid and online classes. Our finding is consistent with the perception that students who took online courses are more likely to use Web or Internet technologies to enhance their learning and communication with faculty and other students. Our results also indicate that students who took hybrid courses more frequently utilize Web-based library resources in completing assignments than students who took only online or face-to-face courses. Although the cause of this result is unknown, it is possible that not all students who took online courses are aware of the learning resources that are available to them. Instructors must ensure that students who enroll in online courses are provided instruction on how to access the learning resources that are available to them online and offline. Institutions may also want to provide personal assistance in dealing with academic difficulties and technical problems to online students who do not have the benefit of personal contacts with faculty and fellow classmates as in the face-to-face classrooms (LaPadula, 2003). Our second research question asked: Do individual and institutional characteristics affect the likelihood of taking online courses? The results of our analyses indicate that individual and institutional characteristics have small but statistically significant effects on students’ likelihood of taking online courses. We understand that there are many personal and institutional factors that can affect a student’s course taking behavior and we are not trying to imply a casual relationship in our study. Factors like employment, child care, and financial support can and should have a significant impact on a student’s decision of which type of courses he or she would take. Nevertheless, we find that certain types of students including racial and ethnic minorities and part-time students are more likely to take online courses. We also found that senior college students majoring in professional fields and first-year business students more frequently take online courses than students in other fields. In the future, the question that deserves further investigation is whether minority and part-time students take online courses more often because online courses offer better quality of education or because it is more convenient. If the reason is for mere convenience – and our guess is it probably is – then institutions must ensure that online students receive high quality instruction, support services, and other fringe benefits enjoyed by traditional face-to-face students. Things like social and informal interaction with faculty and other students and opportunities to receive personal assistance from faculty and staff are also important for both online and face-to-face

1230

Pu-Shih Daniel Chen et al. / Computers & Education 54 (2010) 1222–1232

students. If online students do not receive the same quality of education and support as their traditional classroom counterparts, another form of unintended educational segregation may develop as increasing numbers of minority, part-time, and working students disproportionately elect to take online courses. In our third research question we asked: does the relative amount of technology employed in a course have a relationship with student engagement, learning approaches, and student self-reported learning outcomes? While one should hesitate to suggest a causal relationship between the use of information technology and learning approaches, educational gains, and other forms of engagement, the results suggest that even after controlling for individual and institutional characteristics, there is a relationship that exists between students who engage in course-related technology and those who engage in other ways, as well as the learning approaches and gains while in college. It would seem that the use of course-related learning technology is another important concept under the umbrella of student engagement. Comparing results from the models for first-year students to those for seniors also suggests that use of technology has a stronger impact earlier in the college experience. Perhaps integrating technology into lower-division courses could be more beneficial in encouraging engagement in other ways of learning in college. The positive correlation between the use of technology and measures of engagement found in this study are not surprising because it replicates previous studies (Hu & Kuh, 2001; Kuh & Hu, 2001; Nelson Laird & Kuh, 2005). This study demonstrates that this positive correlation is persisting even as new technologies are being introduced and students are entering college with increasingly sophisticated uses for and expectations of technology in their lives and on campus. While this study does not explain the precise nature of the relationship between technology and engagement, it does highlight the need for future research exploring the nature of this persisting positive correlation. 4.1. Limitations The most significant limitation of this study is that the results are largely based on responses to an experimental set of questions that are relatively untested for their psychometric properties, including validity and reliability. While the questions have face and content validity, the researchers have not yet performed extensive investigations of the psychometric properties of these questions. Additionally, institutions participating in this study were not randomly selected from the pool of 4-year colleges and universities in the United States – the nature of NSSE allows institutions to self-select into the pool. Although the sample covers a wide range of American higher education institutions in terms of the Carnegie classifications, size, control, and urbanicity, one must be cautious when generalizing the results of this study beyond these students. On a related note, because the limitations of our data, including non-random institutional sample and the nature of the NSSE survey, it is not possible to make conclusions about the direction of causality in this study. For instance, while our findings suggest that students who use online learning technology are more engaged, it is possible that more engaged students tend to use learning technology more. Future studies are needed in order to point out the direction of causality between the use of learning technology and student engagement. Lastly, a large sample size like we had in this study (17,819 first-year and senior students) can be both a blessing and a curse. A large randomly selected student sample improves the external validity of this study, but it also has the potential of making all statistical tests significant. For that reason, we reported effect sizes for all our statistical findings. From our point of view, we believe the benefits of a large sample outweigh the associated disadvantages. 5. Conclusion Overall, the results of this study point to a positive relationship between Web-based learning technology use and student engagement and desirable learning outcomes. Not only do students who utilize the Web and Internet technologies in their learning tend to score higher in the traditional student engagement measures (e.g. level of academic challenge, active and collaborative learning, student-faculty interaction, and supportive campus environment), they also are more likely to make use of deep approaches of learning like higher order thinking, reflective learning, and integrative learning in their study and they reported higher gains in general education, practical competence, and personal and social development. These results are encouraging signs that Internet and Web-based learning technologies continue to have a positive impact on student learning and engagement. New technology also brings new challenges to higher education institutions. As more ethnic minority and part-time students elect to take online courses instead of traditional classroom courses, ensuring the quality of online education and providing good online student support services becomes a mandate for social equity. It is also the responsibility of the institutional administrators and faculty to make certain that all online students received adequate academic and technological support and they are made aware of all the online and offline resources available to them. No one would deny that computers and the Internet technology have offered educational opportunities to many people who would otherwise be excluded from the traditional higher education system. Now the goal should be not just provide the educational opportunities but the highest educational quality for all students.

Appendix A A.1. NSSE 2008 online learning survey items 1. During the current school year, how many courses have you completed in total? (Use a drop down menu for student to select from 0 to 20 or more) 2. During the current school year, about how many of these courses used the Web or Internet as the primary method to deliver course content? (Use a drop down menu for student to select from 0 to 20 or more) 3. During the current school year, about how many of your courses were conducted face-to-face but had a Web component designed to promote interaction among students and instructors? (Use a drop down menu for student to select from 0 to 20 or more) 4. In your experience at your institution during the current school year, about how often have you done each of the following? (Very often, often, sometimes, never) a. Discussed or completed an assignment using a ‘‘synchronous” tool like instant messenger, online chat room, video conference, etc.

Pu-Shih Daniel Chen et al. / Computers & Education 54 (2010) 1222–1232

b. c. d. e. f. g. h. i. j.

1231

Discussed or completed an assignment using an ‘‘asynchronous” tool like e-mail, discussion board, listserv, etc. Asked for help from a tutor or other students outside of required class activities. Participated in discussions about important topics related to your major field or discipline. Participated in course activities that challenged you intellectually. Participated in a study group outside of those required as a class activity. Participated in discussions that enhance your understanding of social responsibility. Used your institution’s Web-based library resources in completing class assignments. Participated in discussions that enhance your understanding of different cultures. Used the Internet to discuss with an instructor topics you would not feel comfortable discussing face-to-face or in a classroom.

References Allen, I. E., & Seaman, J. (2008). Staying the course: Online education in the United States. Needham, MA: Sloan Consortium. . Austin, A. W. (1993). What matters in colleges: Four critical years revisited. San Francisco: Jossey-Bass. Bersin, J. (2004). The blended learning book: Best practices, proven methodologies, and lessons learned. San Francisco, CA: Pfeiffer. Bråten, I., & Streømsø, H. I. (2006). Epistemological beliefs, interest, and gender as predictors of Internet-based learning activities. Computers in Human Behavior, 22(6), 1027–1042. Bernard, R. M., Abrami, P. C., Lou, Y., Borokhovski, E., Wade, A., Wozney, L., et al. (2004). How does distance education compare with classroom instruction? A meta-analysis of the empirical literature. Review of Educational Research, 74, 379–439. Bullen, M. (1998). Participation and critical thinking in online university distance education. Journal of Distance Education, 13(2), 1–32. Carini, R. M., Kuh, G. D., & Klein, S. P. (2006). Student engagement and student learning: Testing the linkages. Research in Higher Education, 47(1), 1–32. . Carnegie Foundation for the Advancement of Teaching (2009). Basic classification. . Chen, P. D., Ted, I., & Davis, L. K. (2007). Engaging African American students: Compare student engagement and student satisfaction at HBCUs and their self-identified PWIs using National Survey of Student Engagement (NSSE) data. In Paper presented at the 32nd annual conference of the Association for the Study of Higher Education, Louisville, KY. Indiana University Bloomington, Center for Postsecondary Research. . Chickering, A. W., & Gamson, Z. F. (1987). Seven principles for good practice in undergraduate education. AAHE Bulletin, 39(7), 3–7. Chickering, A. W., & Ehrmann, S. C. (1996). Implementing the seven principles: Technology as lever. AAHE Bulletin(October), 3–6. Clark, R. E. (2009). Past and future research in online education. In Paper presented at the 2009 annual meeting of the American Education Research Association, San Diego, California. . Driscoll, M. (2002). Blended learning: Let’s get beyond the hype. Learning & training innovations newsline. . Duderstadt, J., Atkins, D., & Houweling, D. (2002). Higher education in the digital age: Technology issues and strategies for American colleges and universities. Westport, CT: Praeger. Ehrmann, S. C. (2004). Beyond computer literacy: Implications of technology for the content of a college education. Liberal Education, 90(4), 6–13. Ellison, N. (2007). Facebook use on campus: A social capital perspective on social network sites. In Program presented at the sixth annual ECAR symposium, Boca Raton, FL. Gladieux, L. E., & Swail, W. S. (1999). The virtual university and educational opportunity: Issues of equity and access for the next generation. Washington, DC: College Board. Graham, C. R. (2006). Blended learning systems: Definition, current trends, and future directions. In C. J. Bonk & C. R. Graham (Eds.), The handbook of blended learning (pp. 3–21). San Francisco, CA: Preiffer. Green, K. (2007). The 2007 campus computing survey. . Hawkins, B. L., & Rudy, J. A. (2008). EDUCAUSE core data service fiscal year 2007 summary report. . Hu, S., & Kuh, G. D. (2001). Computing experience and good practices in undergraduate education: Does the degree of campus ‘‘wiredness” matter? Education Policy Analysis Archives, 9(49). . Indiana University Center for Postsecondary Research (2008a). Institutional report 2008. Indiana University Center for Postsecondary Research. . Indiana University Center for Postsecondary Research. (2008b). Promoting engagement for all students: The imperative to look within 2008 results. Bloomington, IN. . Jenkins, H. (2006). Convergence culture: Where old and new media collide. New York: NYU Press. Kuh, G. D. (2004). The National Survey Of Student Engagement: Conceptual framework and overview of psychometric properties. . Kuh, G. D., & Hu, S. (2001). The relationships between computer and information technology use, student learning, and other college experiences. Journal of College Student Development, 42, 217–232. Kuh, G. D., & Vesper, N. (2001). Do computers enhance or detract from student learning? Research in Higher Education, 42, 87–102. LaNasa, S. M., Cabrera, A. F., & Trangsrud, H. (2009). The construct validity of student engagement: A confirmatory factor analysis approach. Research in Higher Education, 50(4), 315–332. LaPadula, M. (2003). A comprehensive look at online student support services for distance learners. The American Journal of Distance Education, 17(2), 119–128. Means, B., Toyama, Y., Murphy, R., Bakia, M., & Jones, K. (2009). Evaluation of evidence-based practices in online learning: A meta-analysis and review of online learning studies. Washington, DC: US Department of Education. Nelson Laird, T. F., & Kuh, G. D. (2005). Student experiences with information technology and their relationship to other aspects of student engagement. Research in Higher Education, 46(2), 211–233. Nelson Laird, T. F., Shoup, R., & Kuh, G. D. (2005). Measuring deep approaches to learning using the National Survey of Student Engagement. In Paper presented at the annual meeting of the Association for Institutional Research, Chicago, IL. . Ouimet, J. A., Bunnage, J. C., Carini, R. M., Kuh, G. D., & Kennedy, J. (2004). Using focus groups, expert advice, and cognitive interviews to establish the validity of a college student survey. Research in Higher Education, 45(3), 233–250. Pace, C. R. (1980). Measuring the quality of student effort. Current Issues in Higher Education, 2, 10–16. Parsad, B., & Lewis, L. (2008). Distance education at degree-granting Postsecondary Institutions: 2006–2007 (NCES 2009–044). National Center for Education Statistics, Institute of Education Sciences. Washington, DC: US Department of Education. . Pascarella, E. T., & Seifert, T. A. (2008). Validation of the NSSE benchmarks and deep approaches to learning against liberal arts outcomes. Wabash College, Center for Inquiry in the Liberal Arts . Sanders, R. (2006). The ‘‘imponderable bloom”: Reconsidering the role of technology in education. Innovate Journal of Online Education, 2(6). . Sands, P. (2002). Inside outside, upside downside: Strategies for connecting online and face-to-face instruction in hybrid courses. Teaching with Technology Today, 8(6). .

1232

Pu-Shih Daniel Chen et al. / Computers & Education 54 (2010) 1222–1232

Siegel, S., & Castellan, N. J. Jr., (1988). Nonparametric statistics for the behavioral sciences (2nd ed.). Boston: McGraw-Hill. Sitzmann, T., Kraiger, K., Stewart, D., & Wisher, R. (2006). The comparative effectiveness of web-based and classroom instruction: A meta-analysis. Personnel Psychology, 59, 623–664. Tabachnick, B. G., & Fidell, L. S. (2007). Using multivariate statistics (5th ed.). Boston: Pearson. Terrell, S. R., & Dringus, L. (2000). An investigation of the effect of learning style on student success in an online learning environment. Journal of Educational Technology Systems, 28(3), 231–238. Thurmond, V., & Wambach, K. (2004). Understanding interactions in distance education: A review of the literature. International Journal of Instructional Technology & Distance Learning, 1, 9–33. . Ward, J., & LaBranche, G. A. (2003). Blended learning: The convergence of e-learning and meetings. Franchising World, 35(4), 22–23. Ward, M., & Newlands, D. (1998). Use of the Web in undergraduate teaching. Computers and Education, 31(2), 171–184. Zhou, L., & Zhang, D. (2008). Web 2.0 impact on student learning process. In K. McFerrin et al. (Eds.), Proceedings of society for information technology and teacher education international conference (pp. 2880–2882). Chesapeake, VA: AACE.