Advances in Engineering Education: Using Technology To Enhance Learning and Engagement in Engineering
Advances in Engineering Education: Using Technology To Enhance Learning and Engagement in Engineering
Advances in Engineering Education: Using Technology To Enhance Learning and Engagement in Engineering
SPRING 2019
AND
ALEJANDRA J. MAGANA
Purdue University
West Lafayette, IN
ABSTRACT
computer technology for short, influences the way that engineering is taught and learned. The goal
of our analysis it to contribute towards a research, adoption, and policy agenda for propagating
the effective use of computer technology in engineering education and for avoiding pitfalls asso-
ciated with the connectivity that this technology enables. We seek to inform action and generate
conversation amongst instructors, students, researchers, administrators, policy makers, and other
We organize our discussion through six main issues as the focus for systemic change for the ef-
fective integration of technology in engineering education. These issues were identified through a
Delphi study with a group of engineering education experts. Our analysis of the issues then draws
on major policy reports of the use of technology in education and the extant research literature from
engineering education and science education as well as technology studies and science studies.
Specifically, we address the role of technology in learning engineering, including both technologies
developed specifically for learning engineering (learning innovations) and domain-specific computer
issues around instructional design including learning outcomes, assessment, and instructional prac-
tice. We include discussion of professional development that better prepares faculty to effectively
use technology in the classroom. Finally, we outline the broader ways technology interacts with the
work of engineering students and faculty at the systems level - for better and for worse.
Key words: Computer technology, educational innovations, engineering practice, instructional design,
assessment, faculty professional development.
SPRING 2019 1
ADVANCES IN ENGINEERING EDUCATION
People were being drawn out of their familiar worlds into one more free, less personal, in
which associations that once attached to each person, place, and object came undone. It
INTRODUCTION
The above excerpt could be describing the impact of the Internet, personal computers, and mobile
devices on human interactions in the early 21st century. However, it was written about a different
network – the network of trains that fundamentally changed transportation, and western society, in
the early 19th century.1 Similarly, in The Signal and the Noise, Nate Silver (2012) comments about the
invention of the printing press, created 400 years before the railroad infrastructure: “paradoxically,
the result of having so much more shared knowledge was increasing isolation” (p. 3). Engineers
are centrally involved in continued improvement of these technologies punctuated by the creation
of revolutionary technology breakthroughs that change the core ways humans interact with the
world and with one another. With an eye toward liberation, it is ethically incumbent for members of
society to reflect on implications of how technology is used, and how it affects the central practices
The “information age” that has ushered in the 21st century has been built squarely on information,
communications, and computational technology (ICCT), which we call “computer technology,” for
short. In this article, we explore how rapidly changing computer technology has and will substantially
impact the way that engineering is taught and learned. We seek to be proactive and address the
research and policy agenda for propagating the effective use of computer technology in engineering
education and for avoiding pitfalls associated with connectivity that this technology enables. The
goal of this analysis is to inform action and generate conversation amongst instructors, students,
researchers, administrators, policy makers, and other key stakeholders in engineering education.
We consider two broad perspectives to approach this goal. First, engineering practice has always
advanced its own boundaries through changing its technologies (National Research Council, 1985).
As such, over the last several decades, computer technology has fundamentally shifted the ways engi-
neering work is done by supporting discovery, collaboration, and innovation processes (Clough, 2004;
Madhavan and Lindsay, 2014). In tandem, learning technologies promise to provide an unprecedented
opportunity to improve instruction, provide adaptive learning, and foster increased engagement and
1
Tom Standage (1998) used a similar analogy between the internet and the telegraph in The Victorian Internet.
2 SPRING 2019
ADVANCES IN ENGINEERING EDUCATION
broader access to education (Hilton, 2002; Pea et al., 2003; Woolf, 2010). Learning technologies consist
of both tools and resources developed specifically for an educational setting and tools that have been
repurposed and coupled with strategies that make them useful for learning (Johnson, Adams Becker,
Estrada, and Freeman, 2014). However, the engineering education community is facing the challenge
to adopt approaches to computer technology use that are grounded in theory, and educational re-
searchers have called for evidence for the effectiveness of the uses of technology for teaching and
learning (Woolf, 2010). Furthermore, it is unclear what aspects of learning can effectively be sourced
to occur adaptively “in the technology,” or to what degree it is fruitful to view technology primarily
as a tool that augments the person-to-person social processes of learning. In general, like those who
confronted with the opportunities presented by the printing press six centuries ago, educators (includ-
ing engineering educators) have struggled to keep pace with quickly evolving computer technology.
In considering the impact of computer technology, we consider two distinct but related types
of questions:
1. How can educators channel the ever-increasing number of learning technologies in ways that
effectively promote meaningful learning and equitable engagement? How do they identify the
2. How do educators best adapt engineering programs (curricula) to prepare students to contribute
In this white paper, we begin discussion with the first type of question while recognizing that the two
types are interdependent and we cannot answer the first without elaboration of the second. Specifically,
in the context of engineering learning environments, instructors must first identify what it is possible
to accomplish with such technologies; that is their affordances. ‘Affordance’ refers the perceived and
actual properties of an object, as related to its functional properties that define how such things could
potentially be used (Gibson, 1979). Additionally, we consider computer technology as one element in
the larger ecosystem where engineering learning occurs and the issues we present strongly interact with
aspects of the other two articles in this special issue including learning in the classroom and pathways
for an inclusive and diverse engineering community (Finelli & Froyd, 2019; Simmons & Lord, 2019).
Our premises and working scope for this paper include the following:
1. Computer technology is a tool that has potential to productively support teaching and enhance
student learning especially with increasing class sizes and more diverse populations.
3. We choose to emphasize the use of technology for on-campus programs. While many of these
systems can be used for learning at a distance, we are not directly addressing issues of solely
distance education.
SPRING 2019 3
ADVANCES IN ENGINEERING EDUCATION
4. While the use of computer technology for on-campus and distance learning appear interchange-
able at first blush, we believe that where and how technology is situated is foundationally
related to how its uses are conceived and how it affects student learning.
5. We center the discussion on the undergraduate level engineering while acknowledging issues
CLASSIFICATION OF TECHNOLOGY
The uses of technology in education are broad, so it is practical to slice it up into more manage-
able categories. While there are many appropriate ways to approach this categorization, we choose
one that is represented in Figure 1. It shows three ways to classify the use of computer technology
in engineering education: (a) learning innovations specifically developed around instructional de-
sign to foster deep thinking and meaningful learning, (b) computational tools used in engineering
4 SPRING 2019
ADVANCES IN ENGINEERING EDUCATION
practice, and (c) Technology with a capital “T.” Here, the first two categories, learning innovations
and computational tools, sit as overlapping sub-sets within the overencompassing span of Technol-
ogy with a capital “T.” We believe that there is risk for cross-talk in the conversation when different
interlocutors implicitly argue from different lenses. Thus, we make categories explicit and address
each with separate issues. However, we recognize that specific cases can also be considered as
1. Learning innovations. We consider what Fishman and colleagues (2004) call cognitively
o riented technology innovations (COTIs), which we call learning innovations, for short.
Learning innovations form a subset of computer technology that are intentionally devel-
oped for educational uses in classroom settings. They are specifically designed to foster
deep thinking and meaningful learning “rooted in cognitive and constructivist learning
theories” (Fishman et al., p. 45). With these innovations, “technology is employed as a tool
to support teaching and learning, as opposed to the object of learning. These innovations
often use technology to scaffold teaching and learning practices that would be difficult to
achieve otherwise, such as making complex causal modeling accessible to students” (p.
46). In some instances, instruction is delivered entirely though the computer, while in oth-
and learning, such as when audience response systems (e.g. clickers) are used to support
ties, actions, or applied skills where individuals must think and act in the modes of a particular
tice. Important technologies include disciplinary specific design tools like ASPEN in chemical
engineering and SolidWorks in mechanical engineering and more general tools like high-level
scientific computing programming languages (e.g., MATLAB) and software platforms for
3. Technology with a capital “T.” We consider the broad impact of Technology change on
the instructional environment in engineering. In this aspect, there are both positive (e.g.,
increased resources) and negative (e.g., access to solution manuals) ways that Technology
impacts the learning environment. We need to understand and manage the affordances
between Technology as a large force in society and the uses in education. From this lens,
it is useful to identify how Technology fits into the cultures of engineering programs and
the culture of higher education, with the goal of promoting systemic change towards more
effective instruction.
SPRING 2019 5
ADVANCES IN ENGINEERING EDUCATION
ISSUES
Correspondingly, we organize the discussion with six main issues that have been identified through
a Delphi study of engineering education experts as the focus for systemic change for the effective
integration of technology in engineering education (Besterfield-Sacre & Shuman, 2016). Our analy-
sis of the issues then draws on major policy reports of the use of technology in education (Hilton,
2002; Honey & Hilton, 2011; Johnson et al., 2014; Pea et al., 2003; Sharples et al., 2015; Woolf, 2010),
specific discussions of the uses of technology in engineering education (Madhavan & Lindsay, 2014;
Froyd, Wankat, & Smith, 2012; Cheville, 2012), and the general research literature on technology and
learning in engineering, the learning sciences, and other related disciplines (completed Jan. 2017).
3. Alignment of technology affordances with learning outcomes: a case study of virtual laboratories
5. Alignment of technology with instructional practice: faculty beliefs and pedagogical knowledge
As Figure 2 illustrates, for both learning innovations (Issue 1) and use of computational tools in en-
gineering practice (Issue 2), we consider the interacting components of instructional design including
6 SPRING 2019
ADVANCES IN ENGINEERING EDUCATION
learning outcomes (Issue 3), assessment (Issue 4), and instructional practice (Issue 5). These three
components (Issues 3, 4 and 5) are the foundational elements considered when designing learning
experiences and align with the elements of “backwards design” (Wiggins & McTighe, 2005). Finally,
we address how all five issues are situated within the broader considerations of Technology (Issue 6).
For each of these issues, we first present a summary of our argument. We then present a synthesis
of the literature of the current state of practice to support that argument, drawing from the research
reports in the learning sciences, in engineering, and in related fields. We then provide a brief set of
In this section, we argue that educators need to intentionally align the development and use of
learning innovations with learning processes (e.g., sense making, disciplinary practices) and with
student engagement. At the same time, they must seek ways to use technology to scale effective
instructional practices to mitigate ever increasing economic tensions in delivering high quality
We examined a set of exemplar learning innovations that have propagated well in engineering.
These learning innovations were developed by collaborative teams and share the following broad
characteristics:
• Pedagogical Core: They are all grounded in a core pedagogical approach that focuses on en-
hancing the experience of the learner, and the teams include pedagogy experts who centrally
• Emergent Use: They can be used in a diverse set of courses and can be flexibly implemented.
This characteristic builds on a general set of Core Components, which lead to a broad Span
of Participation.
• Community Building: They all contain strong community-building strategies and activities.
• Research-Based: They all have strong research on student learning integrated into the core
project activity that is used to iteratively improve the technology-learning system and also
These four identified characteristics mutually support one another to allow high quality learning
innovations to be developed and to scale. The Pedagogical Core ensures the innovation focuses on
student activity and social interactions that are centered on evidence based practices (e.g., coopera-
tive learning, concept-based active learning). Through Community Building, potential instructors and
students learn about the innovation and are connected to others who use it. Emergent Use allows the
SPRING 2019 7
ADVANCES IN ENGINEERING EDUCATION
innovation to be implemented in a variety of settings, both making it better fit the needs of individual
instructors, and also providing the developers access to variations in learning environments to learn
what works and where there might be opportunities for improvement. Research-Based means that the
learning innovations are grounded in learning theory and that there are continued cycles of design-
based implementation research that lead to iterative improvements and expansion of their scope.
We illustrate these four characteristics with four exemplar learning technologies as shown
in Table 1: CATME, the Concept Warehouse, PhETs, and SCALE-UP. The first two systems were
developed specifically for use in engineering while the second two were initiated in the physics
education research community, but have shown significant propagation to engineering. The values
for the extent of propagation reflect the status as of Jan. 2017.
built upon cooperative learning, a pedagogy developed by Johnson & Johnson (1999) based on
social interdependence theory. Cooperative learning has shown strong positive effect sizes on stu-
dent achievement, interpersonal relations, and psychological health (Johnson et al., 2000; Johnson
& Johnson, 2009). CATME provides a web-based support for cooperative learning that includes
several Core Components to enable instructors to more effectively manage teams. Training modules
and meeting supports help students learn core socio-cognitive concepts of teamwork and teaming
behaviors. One of its tools, the Team-Maker (Layton, et al., 2010), has an algorithm that lets instruc-
tors form teams based on their criteria and information submitted by students. CATME Peer Evalu-
ation (Ohland, et al., 2012) and the associated rater training system allows self- and peer- evaluation
processes that are research-based, building on the team effectiveness research literature to create
a valid and reliable behaviorally anchored rating scale for team-member effectiveness. Through
social media and in-person community building activities, CATME has reached 650,000 students
of 12,000 faculty members at 1,900 institutions since 2005 that includes a Span of Participation
The pedagogical core for the Concept Warehouse is concept-based active learning (Koretsky
et al., 2014). The Concept Warehouse houses a set of tools to lower the activation barrier for in-
structors to implement concept-based active learning in their classes. Its Core Components include
ConcepTests, concept inventories, interactive virtual laboratories, and inquiry-based activities. These
materials support what Chi (2009) calls interactive activities where students are both are cognitively
active (e.g., students responding to conceptual questions) and participate in socially collaborative
discourse where they make connections to the course topics while talking to one another [e.g., by
using peer instruction (Lasry, Mazur, & Watkins, 2008; Mazur, 1997)]. The connectedness of con-
cepts is promoted by having students reason through concept-based questions, link them to more
extensive activities (interactive virtual laboratories, inquiry-based learning activities), reflect on the
8 SPRING 2019
Table 1. Examples of university-developed learning innovations that have evidence of propagation.
Emergent Use
SPRING 2019
Public Pedagogical Span of Community Research Activity
Learning Innovation Release Extent of Propagation Core Core Components Participation Building (Examples)
CATME 2005 12,000 faculty Cooperative Assignment of students Engineering, Linked-in CATME Team assignment
www.catme.org/ 650,000 students Learning – Team into teams; peer business, User group; (Layton et al.,
1,900 institutions skills evaluation; peer rater Workshops at 2010); Peer
USA and International calibration; teaming professional society evaluation (Ohland,
scaffolds meetings et al., 2012;
Loughry et al.,
2007); Teamwork
Skills (Loughry
et al., 2014)
ADVANCES IN ENGINEERING EDUCATION
Concept Warehouse 2012 1,000 faculty Concept-based ConcepTests; concept Core chemical Workshops at Propagation
cw.edudiv.org/ 25,000 students active learning inventories; interactive engineering professional society (Friedrichson et al.,
200 institutions virtual laboratories; courses; some meetings; Webinars; 2016; Koretsky
USA and International inquiry based activities related engineering Community et al. 2014); Role of
disciplines (ME, newsletter written justification
MS) (Koretsky et al.,
2016a, 2016b;
Brooks et al., 2016)
PhETs 2002 ~130 simulations Self-guided Interactive simulations Science (physics, Website enabled Simulation design,
phet.colorado.edu/ 40,000,000 “uses” scientific inquiry (feedback, implicit chemistry, biology, support tools, e.g., scaffolding, and
USA and International scaffolding, multiple earth science) and teacher activity exploration (Adams
representations, mathematics sharing; tips and et al., 2008;
pedagogically useful resources to use Podolefsky et al.,
actions, intuitive PhETs and develop 2010); In-class use
interface) activities (Chamberlain et al.,
2014; Finkelstein et
al., 2005).
SCALE-UP 1997 300 departments Cooperative Classroom architecture STEM courses Workshops; website Student learning
scaleup.ncsu.edu/ 200 institutions Learning – and technology (most in physics, with videos (Beichner et al.,
USA and International Learning space support (e.g., round 30%, engineering 2007; Beichner
tables with 9 students; 7.5%) et al., 2000) and
whiteboards, …) retention (Dori
Using Technology to Enhance Learning and Engagement in Engineering
et al., 2003);
Propagation (Foote
et al., 2014)
9
ADVANCES IN ENGINEERING EDUCATION
activities, and demonstrate understanding through assessment (concept inventories). The Concept
Warehouse has three distinct but complementary functions: (i) a repository of resources with high
quality topic-specific content, (ii) an audience response system and learning management system
to deliver the content, and (iii) learning analytics that provide assessment data of student responses
to instructors and researchers. Although relatively young, its Community Building activities have led
to propagation within chemical engineering and related disciplines (Freidrichson et al., 2016). The
Concept Warehouse has reached 1,000 faculty members at 200 institutions since 2012 that includes
a Span of Participation across chemical engineering and other related engineering disciplines.
Similarly, the two physics initiated technology systems, PhETs and SCALE-UP, share the charac-
teristics of Pedagogical Core, Emergent Use, Community Building, and Research-Based. PhETs are
a set of over 100 simulations that have been developed primarily for the physical and life sciences
and are based on a core pedagogy of scientific inquiry. The simulations use multiple representations
and have built-in, implicit constraints to allow students to learn scientific principles. In engineering,
they are used mostly in core, introductory engineering science courses like material and energy bal-
ances. Like CATME, SCALE-UP is also built on cooperative learning pedagogy. However, it deploys
large classrooms. More detail about how PhETs and SCALE-UP align to the characteristics we have
Recommendations
We recommend that, as much as possible, innovators work together to form collaborative teams
and build technology systems with the four characteristics identified above: Pedagogical Core, Emer-
gent Use, Community Building, and Research-Based. While we acknowledge the role of individual
innovators is important, we recommend funding agencies include this type of broad, collaborative
pedagogically-centered innovation as a key component to their portfolios. To this end, funding agen-
cies could support small symposia or workshops that allow targeted networking of PIs with related
but complementary expertise. These projects have potential to “cross the chasm” into mainstream
To work towards sustainable scalability, we recommend that education policy makers explore stra-
tegic and holistic approaches to technology development, such as considering technology “genera-
tions” to map out stages in potential university - industry partnerships. We could envision the earliest
stages being single investigator for visionary high-risk proof-of-concepts, then collaborative university
projects like the ones identified in Table 1 for broader implementation with pedagogical integrity, and
finally university-industry partnerships or other vehicles (e.g. open source) to commercialize and
bring to scale. Such strategies would lead to technologies that are not only research-based, but where
10 SPRING 2019
ADVANCES IN ENGINEERING EDUCATION
research provides and integral part of the development and propagation. Such an approach would
lead to significant scale through continuous iterative improvement but, importantly, also be likely to
Through these stages, the engineering education community needs to distinguish between
general approaches that are inherently less effective and poor implementation of potentially fruitful
approaches. While the strategy we recommend partially addresses this by having the innovations
used in different ways and in different contexts, research is needed to identify characteristics that
allow the community to distinguish between ineffective tools and poor implementation. The com-
munity also needs to address issues of technology failure and reliability, which make faculty reluctant.
Current research practices discourage communication of this type of information; rather robustness
and reliability studies could accompany a coordinated approach through technology generations.
We encourage programs like the Innovation Corps for Learning (I-Corps L; Chavela Guerra &
Smith, 2016) where academic technology developers form teams and learn entrepreneurial tools
and methods that underlie successful start-up companies. Importantly, like in I-Corps L, innova-
tors should be challenged to step out of their local contexts and identify the value of the learning
We argue that educators need to align expert uses of domain-specific computational tools with
their affordances for connecting to engineering practice and to foundational disciplinary background
knowledge. At the same time, they need to devise proper scaffolding methods or adaptations of the
tools for novice learners to become fluent in the use of professional computational tools.
puter software, tools, and packages that embed mathematics and/or engineering principles (herein
called computational tools). Examples of these technologies include simulation and design tools
like ASPEN in chemical engineering and SPICE in electrical engineering, and more general tools
like numerical computing or analysis tools (e.g., MATLAB, COMSOL). The importance of developing
this type of technical proficiency has been identified by many engineering education stakeholders.
National reports such as the Transforming Undergraduate Education in Engineering report ([ASEE],
2013), recently identified that industry professionals value the ability to use computational tools to
support problem solving and design thinking. For instance, an ability to use the techniques, skills,
and modern engineering tools necessary for engineering practice is a stated ABET (2013) student
outcome and is reflected in the Washington (2011) and Dublin Accords (Patil , & Codner, 2007).
The use of these tools has consistently been identified as relevant to engineering practice over
the last 100 years (e.g., Landau & Rosenberg, 1986; Mann, 1918), and consequentially has become
SPRING 2019 11
ADVANCES IN ENGINEERING EDUCATION
an important part of undergraduate curricula, especially in the senior year. Typically engineering
students use simulation and design tools as an extension of the analytical methods they learn in
engineering science classes to more complicated systems and processes that they will face in engi-
neering practice (Dahm et al., 2002). In this vein, simulations provide a critical computational tool
for practice and just about every engineering program inevitably uses some type of commercial
simulation, e.g., ASPEN, HYSIS, SolidWorks, or Synopsys TCAD, where students are asked to predict
performance of an artifact or process unit (Lewin et al., 2004). For example, the 64 institutions re-
sponding to a survey of how design is taught in chemical engineering all reported the use of some
kind of domain-specific process simulation software (e.g., ASPEN, ChemCAD) (Silverstein et al.,
2013). Similarly, about 75% of the 73 respondents to a survey of mechanical engineering programs
revealed that students were required to take a course to learn a general numerical computing tools
(e.g., MATLAB, MAPLE) and then use that tool in the upper-division (Steele & Hodge, 2001).
In Figure 3, we present a conceptual organizer for the role of computational tools for learning
engineering. Clearly, students need to learn first how to work with the tool and thereby become
reasonably fluent with it (top circle). With fluency, accompanying mental models form about how a
engineering practice.
12 SPRING 2019
ADVANCES IN ENGINEERING EDUCATION
particular type of tool is structured and what the computational sequences entail. This conceptual
understanding allows transfer within tools of the same type. For example, experience with one com-
putational fluid dynamics (or statistics or process design) package can provide a type of knowledge
that makes it easier to use another. We also believe the use of these tools can reinforce foundational
disciplinary principles and concepts, and develop understanding in students of how this knowledge
can be extended and applied to engineering practice (Magana, Falk, Vieira, & Reese, 2016). From
this perspective, the use of computational tools can be seen as a bridge to help connect classroom
learning to practice (dashed line). However, instructional strategies are needed to ensure students
have sufficient fluency with the tool itself so that they are able to make these connections, and to
develop activities and practices that support students in making such connections.
To make expert tools more accessible for novice learners to develop fluency, instructors can
scaffold activity and thereby reduce the cognitive load on students. The cognitive load model
posits that memory resources limit the amount of processing that can occur during problem solv-
ing (Sweller, 1988). When a task exceeds the learner’s cognitive load, the learning benefit becomes
A very common pedagogical approach for developing fluency with computational tools has been
through a guided activity by the instructor (e.g., Khan & Singh, 2015; Toto, Colledge, Frederick, &
Pung, 2014), or by means of a self-paced or online tutorials (e.g., Beg, 2015; Castrellón, Botía, Gómez,
Orozco, & Gil, 2011; El-ZEin, Langrish, & Balaam, 2009; Impelluso, 2009; Uribe, Magana, Bahk, &
Shakouri, 2016). Another scaffolding strategy aimed to reduce cognitive load is the use of worked-
examples. A worked-example is an expert solution to a problem (Atkinson, Derry, Renkl, & Wortham,
2000). By studying worked-examples students can start to solve problems by analogy, until they
reach a stage of fluency in which they are able to solve other problems on their own (Sweller, Ayres,
& Kalyuga, 2011). This scaffolding strategy is especially useful for novice learners who still may not
have background knowledge (i.e. schemata) that would enable them to do problem solving from
the beginning. The use of worked-examples has been identified as a useful strategy in supporting
learning with engineering computational tools (Morrison, Margulieux, & Guzdial, 2015; Vieira, Yan, &
Magana, 2015; Vieira et al., 2019). One way in which students can be prompted to actively explore the
et al., 2000). Although the benefits of explanations in computer programming have been explored
for more than twenty years (e.g., Pirolli & Recker, 1994), specific strategies on how to implement
these as sense making strategies to help engineering students relate programming knowledge with
disciplinary knowledge have just started to emerge (Vieira, Magana, Falk & Garcia, 2017; Vieira, Roy,
SPRING 2019 13
ADVANCES IN ENGINEERING EDUCATION
Magana, Falk, & Reese, 2016; Vieira, Yan, & Magana, 2015). We next address research on how the use
Unfortunately, learning gains resulting from the uses of computational tools in the classroom
have not been thoroughly reported in the literature and the few instances identified are not con-
clusive. Studies have primarily reported students’ level of satisfaction and found that students
commonly believed that computational tools are useful for their learning (e.g., Ayasun & Nwankpa,
2006; Brinson, Belytschko, Moran, & Black, 1997; Castrellón et al., 2011; Hoole, Sivasuthan, Karthik, &
Hoole, 2015; Impelluso, 2009; Khan & Singh, 2015). Other studies have reported learning gains when
students’ solutions to projects, exams or homework assignments are compared to other courses or
with previous offerings of the same course before changes were implemented. For example, a study
with 151 students who performed computational analysis and model validation against experimental
response using Java, Maple or MATLAB reported mixed results. Examination of pre and post-activity
testing revealed that the upper 27% of students showed a significant improvement on most of the
questions following the completion of the module while the lower 27% showed mixed results (Khan
& Singh, 2015). Another study that implemented Computational Fluid Dynamics models and analysis
course identified that introducing the concept of scaling and its application (using computer mod-
els) into undergraduate engineering courses enhanced students’ learning and decision making skills
(Najm, Mohtar, Cherkauer, & French, 2010). Likewise, Alabi and colleagues (2015) identified similar
results when using the Gibbs tool (Cool, García, & Bartol, 2015) to support students’ understanding of
thermodynamics concepts. They concluded that the use of the Gibbs tool might have helped students
develop representational competence. More work is needed to share instructional design strategies
and practices of how students can explicitly connect foundational disciplinary principles to the use
of computational tools, especially the domain specific design tools. Importantly, research is needed
that provides evidence of how these practices lead to students’ developing conceptual understanding.
Engineering Practice
We next address the ways computational tools can be used in educational settings to extend
tools in supporting development of the skills needed in engineering practice have been identified,
including: problem solving (e.g., Delale, Liaw, Jiji, Voiculescu, & Yu, 2011), analysis, calculation and
14 SPRING 2019
ADVANCES IN ENGINEERING EDUCATION
optimization (e.g., Brinson et al., 1997; Castrellón et al., 2011; Hoole et al., 2015; Khan & Singh, 2015),
modeling and simulation (e.g., García–Herreros & Gómez, 2013; Najm et al., 2010; Okamoto, Hsu, &
Bash, 2009); 3D modeling (e.g., Toto et al., 2014), integration of programming skills (e.g., El-ZEin
et al., 2009; Impelluso, 2009; Magana, Falk, & Reese, 2013), connecting or enabling operation of
hardware, equipment, sensors and other cyber-physical systems (Magana & Coutinho, 2017), and
characterization and experimentation skills (e.g., Ayasun & Nwankpa, 2006; Beg, 2015). However,
information of how engineering practices were enacted by students and the effects on their learning
these skills (i.e., design skills, modeling and simulation skills, problem solving skills, and computa-
tion skills, among others) are often incomplete or lacking. Additionally, one might consider doing
engineering work effectively within these computational tools as a skill within itself.
To develop instructional strategies, educators should draw on ethnographic studies that have
identified the ways computational tools are used in engineering practice (Vinck, 2003). For example,
Auregemma and colleagues (2013) investigated the design process of a microfluidic lab-on-a-chip
device. The developers used CAD software, COMSOL, and MATLAB in an iterative process that in-
volved mental models, computational models, and building and testing prototypes. In another study
of an authentic and industrially situated process development design task, Sherrett and colleagues
(2013) found that experts used computational tools in a design process which included informa-
tion gathering, problem formulation, and iterative modeling and experimentation. Results from the
These findings suggest a critical step in the process of designing instruction for the use of com-
putational tools consists of first identifying the ways professionals use them, followed by a clear
definition and proper guidance to enact the practices or competencies the tool affords. Along this
line, Magana and Coutinho (2017) proposed a range of possible affordances of computational tools
for supporting a wide variety of modeling, simulation and experimentation practices. The practices
they proposed were aligned with different curricular levels as informed from experts in industry
and academia. Building on this idea, Magana and colleagues (Magana, Falk, Vieira, & Reese, 2016;
Vieira, Magana, Roy, Falk, & Reese, 2016), explicitly aligned desired professional practices afforded
by the computational tools with disciplinary learning objectives for core undergraduate courses in
While experts fluently apply computational tools to their design work, integrating these practices
into the undergraduate curriculum is challenging. Diefes-Dux and colleagues (2004) investigated the
effectiveness of a multi-level, steady-state food process design tool. They evaluated students uses
of Foods Operations Oriented Design System Block Library (FOODS-LIB) running in the MATLAB-
supported SIMULINK simulation toolbox. The activity was delivered via seven online learning modules
where students used the existing unit operations library to study a single effect evaporator model and
SPRING 2019 15
ADVANCES IN ENGINEERING EDUCATION
construct a new generalized unit operation: compositional split. The implementation took place over
an eleven-week period. While students successfully completed the learning modules, they were unable
to transfer their knowledge and skills to the design of processes. They identified weak MATLAB coding
skills and students’ inability to conceptualize or write out algorithms for their designs as barriers to
productive use of this tool in practice. A similar conclusion was found by García–Herreros and Gómez
(2013) in a study that evaluated students use of process simulators through the modeling and opti-
mization of a crude distillation unit using PRO/II 8.0. They also found a barrier in using computational
tools in practice and concluded that the main problem was the lack of convergence of the model as
a result of poor initial estimates. They attributed it to student shortcuts amidst deadline pressures
and lack of understanding of the tool’s optimization algorithms. These two studies suggest that algo-
rithmic (computational) thinking may be challenging for students and this way of thinking needs to
be progressively developed and supported when using computational tools in engineering practice.
We argue that the use of computational tools for teaching engineering is most effective when
there is alignment between learning outcomes, pedagogical methods and supports, and technol-
ogy uses (Chen et al., 2000; Kadiyala & Crynes, 2000). We believe that the alignment of learning
objectives and affordances of computational tools is not enough. We recommend that instructional
design methods, pedagogical strategies, and scaffolding methods be identified and properly em-
bedded into the instructional environment, so students can fully benefit from the integration of
Instructional design principles, such as the How People Learn (HPL) framework (Bransford, Brown,
& Coocking, 2000) can help educators address issues of the types of integration depicted in Figure 3.
The HPL framework focuses on instructional design that is (a) leaner-centered by first considering
students’ required background knowledge and possible challenges they may encounter when learning
with engineering practices via the computational tool; (c) assessment-centered by providing students
with frequent feedback and opportunities to improve their work; and (d) community-centered by
allowing students to learn from each other. For instance, two studies have reported on scaffolded
instructional design using the HPL framework to plan and integrate engineering computational
tools into engineering courses (Greenberg, Smith & Newman, 2003). These studies show that with
scaffolding students demonstrated significantly better understanding, but they found the demands
extremely challenging compared with other courses. Alternatively, designers have developed scaffolds
based on Cognitive Load Theory (Guzdial, 1994; Sweller, 1994; Vieira, Magana, Roy, Falk, & Reese,
2016). While this approach has elicited positive feedback from instructors and students (Impelluso,
16 SPRING 2019
ADVANCES IN ENGINEERING EDUCATION
2009), more research is needed that identifies how and when students are overloaded and what
Finally, students’ attitudes and beliefs about computational tools can shape how they engage in
learning. One study has found that male students tend to have more positive attitudes than female
students (Hornaes and Royrvik, 2000). In another study (Hutchison, Follman, Sumpter, & Bodner,
2006), researchers found students who listed computing as an important influence on their self-efficacy
beliefs frequently cited their ability to use one or all of the computing tools taught in a course, their
programming abilities, and their ability to use a computer in general. Gender trends emerged in student
responses to factors that affect confidence in success. Specifically, relatively few men saw computing
as negatively affecting their self-efficacy beliefs; while in contrast, nearly one-third of women reported
computing as negatively affecting their self-efficacy beliefs. However, results reported from Magana
and colleagues (2016; Vieira et al., 2018) suggest that frequent exposure to computational tools and
methods may increase students’ self-efficacy beliefs and their value of these tools for their academic
Recommendations
Our recommendations from this issue are addressed to engineering educators and engineering
education researchers. In summary, we posit that students must first develop fluency in using a
domain-specific computational tool, and that associated mental models will develop allowing for
more routinized use and for transfer to other tools of the same type (e.g., between computational
fluid dynamics packages). Fluency can be developed more quickly through pedagogical supports,
embedded scaffolding, or adaptations of the computational tools in order to lower the barrier of
entry and diminish cognitive overload. For example, engineering educators should consider the use
of worked-out examples. However, research is needed to better characterize the mental models that
form in learners while developing fluency, what instructional strategies best develop the models,
and how the models connect to the transfer to other similar tools.
We advocate for two elements to be included in instructional design progressions when integrating
computational tools: (i) have students connect the use of these computational tools to foundational
disciplinary knowledge, and (ii) engage students in uses of the tools that reflects professional practice.
We also posit that increased fluency gives students cognitive bandwidth to make connections to foun-
dational disciplinary knowledge and to engineering practice. But again, research is needed to better
understand these relationships. Indeed, we might imagine that fluent use of computational tools increases
disciplinary knowledge and that such increases feedback to increased capability with the tool itself.
vertical integration of the use of the appropriate set of domain-specific computational tools
SPRING 2019 17
ADVANCES IN ENGINEERING EDUCATION
throughout the undergraduate curriculum (e.g., Hinds & Somerton, 2007; Hinds, Urban-Lurain, Stick-
len, Amey, & Eskil, 2005; Sticklen, Amey, Eskil, Hinds, & Urban-Lurain, 2004; Urban-Lurain, Amey,
Sticklen, Hinds, & Eskil, 2004), so students not only develop technical fluency but also confidence
and self-efficacy and see the value in using these tools. Strategies need to be developed in the vi-
such as the HPL framework, to guide the learning process. For example, educators could consider
a. deliberately connecting the use of the domain-specific computational tool with both foun-
dational disciplinary principles and with realistic engineering practices,
b. taking those connections and then identify a pedagogical approach and a set of scaffolds
c. providing frequent feedback along with opportunities to iteratively improve their work, and
d. properly identify assessment mechanisms that truly evaluate students’ gains of foundational
disciplinary principles, but more importantly how students improve their performance in
enacting engineering practices when engaging with computational tools (see Issue 4 for
We advocate for research that provides concrete evidence of the learning when using these tools
(e.g., Magana, Fennell, Vieira & Falk, 2019), along with transferrable understandings of how computational
tools of engineering practice can be effectively be incorporated into learning activities, particularly in
ways that respond to the rapid changes in the function and capability of these tools. To this end, we also
advocate development of interaction models between industry and academia that allow identification
of the changing ways that practitioners use computational tools in practice and allow translation to
Issue 3: Alignment of Technology Affordances with Learning Outcomes: A Case Analysis of Virtual
Laboratories
With Issue 3, we consider transformative potentials of computer technology for learning. We argue
that the most effective learning innovations are not pallid, clones of traditional learning environments.
Rather, they identify and leverage the affordances of technology to support engaged learning environ-
ments that reconstitute the ways students interact with the content, with each other, and with know-
ing others (e.g., instructors or peer mentors)2. To provide an analogy, we believe too many learning
technologies are developed in a way similar to seeking to provide a credit-card reader for the taxi driver
In analogy, in the Victorian Internet, Tom Standage (1998) describes the initial conceptualizations of the telephone
2
as a “’speaking telegraph’ – an improvement of an existing technology rather than something altogether different”
18 SPRING 2019
ADVANCES IN ENGINEERING EDUCATION
to more readily accept fares rather than reconstituting the environment the way companies like Uber
and Lyft have.3 From this perspective, learning innovations are most effective when they do more than
provide an alternative mode for existing instruction but extend to embody paradigmatic shifts in the
ways learners interact with content, peers and instructors to generate learning. These interactions can
When assessing and evaluating technology, there is an inclination to focus on the “category” of
technology rather than thinking about the opportunities for learning the technology affords, the
ways it can fit into instructional design to take advantage of those affordances, and the potential
obstacles that may need to be addressed in different learning contexts. Without articulating the
thinking processes and social interactions they want students to experience, educators may tend to
revert and use technology to support traditional types of instruction. Alternatively, a more fruitful
approach is to recognize ways that technology enables interactions between the student and other
learners, the instructor, and content that are not otherwise possible. Such an approach more fully
We choose the virtual laboratory (de Jong et al., 2013; Koretsky et al., 2008; Ma & Nickerson,
2006) as an exemplar to illustrate different ways that a single technology category can be integrated
into learning systems towards different instructional purposes and learning goals. In a virtual labo-
ratory (also called a simulation laboratory), computer simulations based on mathematical models
provide values of output variables in response to user-selected input variables. Students work with
of targeted phenomena based on the simulated output. We use the term virtual laboratory to con-
trast the students’ orientation in the learning activity with the more common use of simulation in
the post-secondary engineering classroom, as was discussed with Issue 2. In summary, we refer to
engineering simulations as tools students and practicing engineers use to apply theory to engineer-
ing analysis and design whereas we use virtual laboratory to emphasize facilitated “laboratory-like”
student exploration of the phenomena associated with a specific system, device, or process.
We choose the virtual laboratory as an example of the different ways engineering educators and
technology developers have leveraged the affordances of technology in their instructional designs.
We begin with designs that essentially use the technology in the same way as traditional instruction
and progress to other ways that technology is leveraged in unique and deliberate ways. While the
(p. 197). We argue that in a similar way, educators should conceive how computer technologies can transform the
fundamental social and cognitive interactions that produce learning rather than merely reproducing earlier practices.
3
If we pursue this analogy further, educators might consider how these technologies disrupt existing social
SPRING 2019 19
ADVANCES IN ENGINEERING EDUCATION
case is made in the context of the virtual laboratory, we argue that these fundamental consider-
ations apply to many emerging learning technologies such as electronic textbooks and audience
response systems. In fact, this example was chosen, in part, since it is straight-forward to envision
the design choices we describe in the development of emergent technologies like immersive virtual
reality where students can have the perception of being physically present in the non-physical world
In its most common rendition, the virtual laboratory simulates the same phenomena as a cor-
responding physical laboratory at the university. Virtual laboratories with this instructional design
can be used to replace the analogous physical laboratory where students are asked to do the same
activity, only on a computer screen (Sehati, 2000; Shin et al., 2000; Wiesner and Lan, 2004, Pyatt
and Sims, 2007). In some cases, care is taken to word the tasks identically in the virtual and physical
modes (Finklestein et al., 2005). This use of virtual laboratory technology has benefits: it requires
fewer resources, provides greater flexibility in scheduling, and experiments can be run quickly
providing immediate feedback for students. Learning outcomes between the physical and virtual
modes have been compared via pre-post testing, written and oral exam analysis, and surveys. Find-
ings often show equivalent, and occasionally greater, learning gains in the virtual mode, especially
towards connecting the laboratory activity to foundational disciplinary concepts (Wiesner and Lan,
2004; Campbell, Bourne, and Mosterman, 2002; Powell et al, 2002; Finklestein et al., 2005; Lindsay,
and Good, 2005; Zacharia, 2007; Pryatt and Sims, 2012; Zacharia and Olympiou, 2011). Differences in
learning outcomes between physical and virtual laboratories have been attributed to: shifts in focus
from working with the equipment to collect data to understanding the causal relations between
variables and outputs (Kolloffel and de Jong, 2013); variations in the patterns of collaboration among
students (Corter et al., 2007); and the greater control that each student has over his or her own
learning (Hazel and Baillie, 1998). However, when it is used solely to replace a corresponding physi-
cal laboratory, the enactment of technology sits within a traditional curricular structure. We argue
next that there is greater opportunity for technology to impact student learning when technology
is leveraged to reconceive instructional designs and even more so when it is used to reconceive the
learning system, itself. It is these uses that take greatest advantage of the affordances of technology.
The simplest way to modify an instructional design is to have students use the virtual laboratory
in preparation for the physical laboratory. Here, the virtual laboratory technology allows the ex-
periment to be practiced, stopped and repeated (Abdulwahed and Nagy, 2009; Hodge et al., 2001;
Mosterman, et al., 1994; Rutten at al., 2012; Zacharia, 2007). A number of studies confirm that using
virtual laboratories for preparation enhances the effectiveness of physical laboratories (Akpan &
20 SPRING 2019
ADVANCES IN ENGINEERING EDUCATION
Strayer, 2010; de Jong, Linn, & Zacharia, 2013; Jaakkola et al., 2011; Kolloffel and de Jong, 2013; Pyatt
& Sims, 2012; Toth, et al., 2009; Zacharia, Olympiou, & Papaevripidou, 2008).
Added Representations
Rather than adhering to the sensory limits of a physical laboratory, virtual laboratory technol-
ogy affords added representations that are not possible in the physical world. Indeed, Lindgren
and Schwartz (2009) suggest that making the virtual laboratory as faithful as possible to the real
laboratory might deter from its pedagogical effectiveness. Virtual laboratories have been extended
beyond their physical analogs, usually with the intent of more effectively helping students develop
conceptual understanding. One common enhancement is the inclusion of visual cues or alterna-
tive representations not possible to observe in physical laboratories such as emergent molecular
interactions, the flow of electric current, and vector electric or magnetic fields (Bowen, Reid, and
Koretsky, 2015; Brophy, Magana, & Strachan, 2013; Corter et al., 2007; Dorneich & Jones, 2001;
Finkelstein et al., 2005; Schank & Kozma, 2002; Sengupta and Wilensky, 2009; van Joolingen & de
Jong, 2003; Wieman et al., 2008). Additionally, the instructional design strategies of in preparation
and representation can be combined. In a set of electrical circuits experiments, Zacharia and de
Jong (2014) showed similar learning gains between the physical laboratory alone and the virtual
laboratory as preparation for the physical laboratory in simpler experimental configurations, but
higher gains when the virtual mode is included for more complex configurations. They attribute the
latter result to the technological affordance of visualization of the current flow. These molecular
and field representations allow technology to uncover the invisible and provide a dynamic venue
In another approach, instructional developers have used virtual laboratory technology to situate
learning in disciplinary practice (Koretsky et al., 2015, 2019; Shaffer, 2006). Here technology can
be used to simulate processes that would not otherwise be available at the university due to their
complexity or incompatible length and time scales (e.g., Uribe, Magana, Bahk, & Shakouri, 2016).
This pedagogical strategy supports learning by placing learners in real-world contexts and prompt-
ing them to shift from the role of student to the role of scientist or engineer. Learning activities are
providing the learner an opportunity to apply disciplinary tools and concepts to make meaning of
observed phenomena and design processes and products. This context allows students to engage in
interlocking material, conceptual, epistemic, and social aspects of disciplinary practice in ways that
catalyze student learning (Ford & Forman, 2006; Koretsky et al., 2019; Pickering, 1995, Shaffer, 2006;
SPRING 2019 21
ADVANCES IN ENGINEERING EDUCATION
Windschitl and Calabrese Barton, 2016). In this way, virtual laboratory technology can complement
and extend student experiences in disciplinary practice beyond capstone projects and internships.
As examples presented below illustrate, when technology is deployed in this manner, it does not
directly replicate typical learning systems at the university, but rather allows students to experience
aspects of the professional context that they would not otherwise have access to.
The use of virtual laboratory technology that is situated in contexts of practice is much more
common in the sciences than in engineering. Thus, we first discuss learning systems that place learn-
ers in the role of scientists and center on the practice of inquiry, then we extend the discussion to
Scientific Inquiry. A host of virtual laboratories have been developed in response to educators
and policy calls to engage students in scientific inquiry (de Jong & van Joolingen, 1998; Edelson,
Gordin, & Pea, 1999; Nelson and Ketelhut, 2007). In inquiry-based learning, students engage in
experiments in a manner designed to align, as much as possible, with practicing scientists. Such
ing experiments, revising models, arguing from evidence, and negotiating explanations with peers
(National Research Council, 1996; de Jong & van Joolingen, 1998, Linn et al., 2003; Windschitl and
One approach to using technology in this way is to create immersive worlds for students to en-
gage in similar types of inquiry processes as experienced in practice. For example, multiuser virtual
environments (MUVEs), such as River City (Dede et al., 2004), Quest Atlantis (Barab et al., 2005),
and Habitable Worlds (Horodyskyj et al., 2018) allow students to take on the role of scientist and
observe and measure phenomena. In contrast to most physical laboratories at the university, students
can gather data continuously over days or weeks, an affordance that allows students to iteratively
modify their hypotheses in light of new information and understanding, and also, gives them time
to develop that understanding (Nelson and Ketelhut, 2007). MUVEs afford an iterative element to
Alternatively, the virtual laboratory can provide a resource for dynamic data collection while
student-student and teacher-student interactions in the classroom. For example, Bell and Trundle
(2008) described the virtual laboratory Starry Night Backyard that allow students to observe
moon phases from any location on earth over long time periods without being hindered by ex-
perimental obstacles like weather, physical obstructions, or time of day. This virtual laboratory
is used in conjunction with McDermott’s (1996) instructional design that promotes the cognitive
22 SPRING 2019
ADVANCES IN ENGINEERING EDUCATION
dissonance that can lead to conceptual change. In a variation of the hybrid approach, Raineri
(2001) described an instructional design where students first perform experiments in the physi-
cal laboratory and then use a virtual laboratory to iteratively generate and analyze new data
sets and to perform experiments that are too expensive and time consuming to perform in the
physical laboratory. Raineri’s design allows students to engage in analysis practices and to gain
experience with state-of-the-art techniques in molecular biology that would not be possible in
The description of scientific practice as “inquiry” has been criticized as broad and vague
(Windschitl and Calabrese Barton, 2016). An alternative conception is that modeling is a core
scientific practice. Based on this pedagogical commitment, virtual laboratories have been developed
to guide students to construct scientific models to relate the data from the phenomena to scientific
principles and theories (de Jong and van Joolingen, 2008; Giere, 1999) such as modeling photo-
synthesis in Co-Lab (van Joolingen et al., 2005) or the response of airbags in automobile collisions
Engineering Design. While the practices of science focus on inquiry, engineering work focuses
on the design of products and processes to meet social needs (Dym, Agogino, Eris, Frey, & Leifer,
2005; Simon, 1996); therefore, virtual laboratories focusing on engineering practice inevitably center
on design. The design process is built around iteration, where a design idea is improved based on
the identified shortcomings of previous attempts (Crismond, 2001; Cross, 2006; Dym et al., 2005).
Identified shortcomings allow a feedback cycle where the designers can identify gaps in their
knowledge and understanding, providing impetus for further learning. For engineering professionals,
this practice of knowledge building through iterative design cycles is critical (Vincenti, 1990). The
ability of virtual laboratories to simulate phenomena of a wider range of length scales, time scales,
and complexity than available in the physical laboratory affords iteration in realistic design tasks
(Koretsky et al., 2008). However, virtual laboratories based on design tasks are far less common
Xie and colleagues have developed a set of virtual laboratories in which high-school students
either design a house or an entire city block to maximize energy efficiency (Xie et al., 2014; Purzer
et al., 2015; Xie, 2016). Students learn and apply concepts of radiative heat transfer as they need
to account for how solar radiation varies over a day and over the year. This knowledge is used to
address the open-ended design problem and account for competing constraints. The technology
automatically logs fine-grained student use data and the researchers have developed visual process
analytics as a data mining tool to investigate student activity and learning in these complex and
SPRING 2019 23
ADVANCES IN ENGINEERING EDUCATION
Shaffer and colleagues have developed the virtual laboratory Nephrotex in which first-year,
undergraduate engineering students take on the role of interns at a high-tech bioengineering firm
tasked with designing a kidney dialyzer based on iterative experimentation (Chesler et al., 2013). The
student teams treat the process as a “black box” and are directed towards statistical experimental
design, completing two iterative cycles. In the first cycle they look at one of the input parameters,
and in the second cycle they look at all four input parameters. The computer provides both the
platform for experimentation and interactive correspondence with simulated co-workers and super-
visors. Shaffer and colleagues described the instructional goal of these Epistemic Games as ranging
far beyond the “conceptual understanding” pursued in many learning environments. Rather, they
are used to develop students’ skills, knowledge, identity, values, and epistemology common to the
which students take on the role of process development engineers tasked with determining the
process input parameters (the “recipe”) for a chemical or biological engineering process (Koretsky
et al., 2011). The hybrid design uses virtual laboratory technology to simulate complex industrial
processes but relies on in-person social interactions with others on the engineering development
team (played by other students) and their supervisor (played by the instructor). These projects
are designed for senior-level engineering students and allow for professionally productive social
interactions (Gilbuena et al., 2015). For example, instructor feedback on professional skills helps
nity of practice. Material and conceptual aspects were found to interlock in this learning system
where students developed and used models to make sense of experimental data and move their
Recommendations
In this section, we have argued that technology can enable creation of instructional designs
and learning contexts that can fundamentally shift the type of activity where learning occurs. We
have illustrated this point in detail through the affordances of one “category” of technology, the
virtual laboratory, but such shifts in instructional design apply to other uses of computer technol-
ogy as well. These designs support engaged learning in ways that fundamentally reconstitute the
interactions of students with the content, with each other, and with instructors. We believe that
if educators and developers approach technology genres from this perspective, they can better
utilize the affordances to positively disrupt static and passive learning environments and provide
access to more diverse populations of engineering students. They can also achieve the benefits
at a larger scale.
24 SPRING 2019
ADVANCES IN ENGINEERING EDUCATION
In summary, we recommend that technology developers and engineering educators take a systems
approach guided by findings in the learning and cognitive sciences to conceive what is possible
rather than replicate traditional classroom structures, norms, and interactions. Specifically, they
should explicitly connect the technology affordances to how their proposed designs support learn-
discourse, and process and social skills; understanding the nature of engineering; and motivation
However, engineering educators are only beginning to see enactments of computer technology
used in this way, and the instantiations are often local to the direct sphere of the innovator and have
not propagated broadly. There are several reasons for this limited realization. First, not all attempts
to shift the activities in which learning occurs are going to be effective. Engineering educators
need to understand what works, but just as importantly, what has not worked and why. Second,
innovators and developers need to be grounded with fundamental content knowledge and have
deep understanding of learning, but also stay connected to the possibilities afforded by the wave
of next generation computer technologies. Better understanding is needed of how to create spaces
where content, pedagogy, and technology expertise overlaps. Successful collaborative processes
and environments need to be better characterized and understood. Third, learning and classroom
instruction occur within the culture of higher education. Normative conceptions of learning and
institutional rewards and recognition for teaching can limit the degree to which classes and systems
are ready to take up these innovative uses of technologies. This aspect is addressed in more detail
shifts in the learning environment and that empirically and theoretically study the resulting interac-
tions with student learning, engagement, and equitable access. We also recommend that agencies
support research to better understand characteristics of successful collaborations and ways to shift
We argue that educators need to identify and implement systematic uses of technology that
provide or leverage existing data to inform formative and summative assessment, adaptive instruc-
tion, and research on learning engineering (National Research Council, 2014). Technology-based
assessments have attracted interest in educational contexts because they can enable the design
of learning environments that provide real time feedback, and scalable and personalized support
SPRING 2019 25
ADVANCES IN ENGINEERING EDUCATION
([DOEd], 2010, 2016). At the same time, advancements in cyberinfrastructure, cyberlearning in-
novations and online learning environments have resulted in data that can be used for improving
learning and for educational research purposes (Borgman et al., 2008). Challenges however still
remain, in order to fully realize the potential of uses of data from technology. In addition, concerns
of confidentiality, security and privacy must be addressed (Madhavan & Richey, 2016).
We advocate for the identification of practices, tools, and methods that can provide formative
feedback to faculty (in addition to summative feedback), so they can appropriately adapt instruc-
tion (e.g., just-in-time teaching). This activity includes identifying what types of feedback are most
effective for learning, relating that to information available through technology, and developing
interfaces that provide that information to students and faculty in ways that they can use. Currently,
a number of commercial personalized or adaptive instruction systems tout the ability to provide
students computer-based formative feedback throughout the learning process. Such tools can help
students with spaced repetition, and rapid and adaptive feedback. These platforms can also be useful
for instructors as the feedback is instantaneous and it reduces grading load. Furthermore, advances
combining artificial intelligence, machine learning and learning analytics offer the possibility of real-
izing personalized learning. For example, learning analytics could be combined with student uses
of learning materials, behaviors, and performance already captured with a classroom management
system (e.g., Arnold & Pistilli, 2012). While the prospect of learning that is solely mediated through
a computer is compelling, it is important to establish its limits. Simultaneously, other ways should
be identified for instructors to use computer-based assessment to support student learning (e.g.,
changing pedagogy, in-person feedback, and so forth), and at the same time explore the effective-
ness of non-traditional models of delivery such as inverted classrooms (e.g., Magana, Falk & Reese,
Computer-based assessment methods have had a long tradition in helping educators measure
student learning (e.g., Brown, Race, & Bull, 1999; Mayrath, Clarke-Midura, Robinson, & Schraw, 2012).
According to Recker and colleagues (2016), when compared to paper and pencil assessment meth-
ods, technology-based assessments can offer advantages such as (a) predicting students’ future
learning by creating models that incorporate information such as students’ knowledge, behavior,
motivation, and attitudes; (b) discovering models that characterize the subject matter to be learned
(e.g. math, science, etc.), identify fruitful pedagogical sequences, and suggest how these sequences
might be adapted to students’ needs; (c) studying the effects of varied pedagogical decisions on
student learning; (d) advancing scientific knowledge about learning and learners through building
models of learning processes that incorporate data about students, teachers, understanding of
26 SPRING 2019
ADVANCES IN ENGINEERING EDUCATION
subject matter, pedagogies, and principles from learning sciences; and (e) supporting learning for
all students by adapting learning resources to fit the particular needs identified, including adapta-
In higher education settings, computer-based assessment (Miller, 2009) and intelligent tutors
(Goel, 2016) have been used as a mechanism to provide just-in-time formative feedback and to help
improve students’ performance (Tair & El-Halees, 2012). For instance, by identifying technology-
learner interactions that are more conducive to learning, educators and educational technologists
can support personalization by generating content semi-customized to the learner (Cheville, 2012).
When using analyzed data generated from learners’ engagement with technology, educators can
not only make sense of the outcomes and the impact of using such technology, but can also char-
acterize the student learning process and make data-driven decisions to adapt the environment to
better support student learning (Xie, Zhang, Nourian, Pallant, & Hazzard, 2014). However, it is unclear
to what degree the feedback should be provided by the computer itself, and to what degree the
computer should provide information to the instructor to interpret and interact with the learners.
exams in large enrollment courses (West, Silva, and Herman, 2015) and to create individualized
homework problems for textbooks (Vahid, Edgcomb, & Strawn, 2016). These type of personalized
problems have evolved from simply changing the numbers within the same skeletal problem state-
ment to developing more and more sophisticated methods to apply to other problem features that
make them more like “new” problems from the perspective of the learner. With these problems, like
contrived back-of-the-chapter problems of old, educators need to better understand the degree
that students engage with core disciplinary concepts and the degree they can anticipate problem
Technology-based assessments also offer the opportunity of assessing 21st century competen-
cies such as problem-solving, critical thinking, or design skills ([DOEd], 2010). The open-ended
nature of these processes make them difficult to assess. For example, design skills can neither be
assessed as a product nor as a simple test (Vieira, Goldstein, Purzer, & Magana, 2016); the steps
the students follow to reach a solution is as important as the final solution itself. Specifically for
the case of design skills, by analyzing process data logged through a computer aided design tool,
Xie and collaborators (Xie, Zhang, Nourian, Pallant, & Bailey, 2014; Xie, Zhang, Nourian, Pallant, &
Hazzard, 2014) identified (a) differences in students’ design processes based on gender, (b) different
ways in which students interact with the software, and (c) differences on students’ design processes
after being exposed to an instructional intervention. Likewise, by analyzing logged process data,
design, and Goldstein and colleagues (2015) assessed students’ idea fluency in their design process.
SPRING 2019 27
ADVANCES IN ENGINEERING EDUCATION
However, there is a significant amount of work to be done, particularly on identifying how interac-
tion, process and learning data can result in instruction and feedback tailored for particular learning
Effective practices and guidelines for interpreting and acting on learning and process data are
needed. In order to take advantage of the vast amount of data using information technologies, we
recommend going beyond traditional narrow forms of assessments with the goal of understanding
the learning processes and their nuances (Pellegrino, Chudowsky, & Glaser, 2001) and moving towards
embedded or stealth assessment where the data is provided as part of the core learning activity
(Shute, 2011). To this end, data must be handled and shared properly and mechanisms to guide this
process are highly needed. Such mechanisms should consider not only technological aspects, but
As learners engage with technology, they leave a trail of ways in which they use the technology
(interaction and process data) and the outcomes and the impact of using such technology (learning
data) (Borgman et al., 2008). In conjunction, advancements in natural language processing tech-
niques, machine learning, educational data mining, and learning analytics can result in new ways of
making meaning of these complex process and learning data (e.g., Borgman et al., 2008; Haudek et
al., 2012; Worsley & Blikstein, 2014; Xie et al., 2014). These methods can support the development
Engineering education researchers have used longitudinal datasets to answer research ques-
tions about how students navigate through required engineering curriculum and what courses or
policies present obstacles for graduation (Ohland & Long, 2016). For instance, when used with his-
torical data along with university regulations to identify performance probabilities, graduation and
passing rates for engineering students can be computed (Caro, González, & Mira, 2014). In tandem,
advances in educational data mining and learning analytics can now enable ways to use data for
improving teaching and learning. For example, data mining techniques, such as classification meth-
ods, have been used for prediction purposes on student performance in examinations (Yadav & Pal,
2012). Rawson and Stahovich (2013) captured time-stamped record of students’ solutions to their
homework assignments using a smart pen. They quantified the total amount of ink written and the
time of the day at which the student homework was done, and used those metrics to characterize
homework habits and correlate them with the final course grade. They concluded that by the end of
the third week of the quarter, it was possible to explain a significant amount of the variance in final
course grade by considering homework habits. Machine learning algorithms have also been used
28 SPRING 2019
ADVANCES IN ENGINEERING EDUCATION
to generate models that predict potential dropouts from engineering majors (Pal, 2012). Results
from those predictions could help educators take on-time action in helping lower performing stu-
dents with additional support (Yadav & Pal, 2012). However, while technology tools show promise
in identifying students who are at risk, more work is needed to identify effective ways to interact
Engineering education researchers have recognized the need for better ways to share research
data, but tensions exists regarding legal, regulatory and ethical considerations (Cheville, 2016). Issues
associated with ownership of the data along with and privacy implications have also been identified
(Johri, Yang, Vorvoreanu, & Madhavan, 2016). Other challenges for adoption of sharing practices
relate to technical and cultural barriers (Gilmore, Adolph, Millman, & Gordon, 2016). Because of the
clear value in using and reusing data for discovery purposes, engineering education researchers have
started to devise principles and heuristics to guide the sharing of data (Cheville, 2016). Ongoing
efforts also include devising methods to help researchers move from sharing to enabling partner-
Recommendations
We urge engineering educators and curriculum designers who have successfully taken advantage
of technology-based assessments to share lessons learned and effective practices, tools, and meth-
ods for providing formative feedback. For example, when gaps in student learning are observed,
what are strategies for instructors to respond? What role do advances in machine learning, artificial
intelligence, and learning analytics play in adaptive or personalized feedback? We recommend con-
tinued development of technology and feedback mechanisms to support engineering faculty in (a)
identifying the effects of pedagogical changes, (b) adapting instruction according to students’ prior
knowledge and performance, and (c) advising students based on predictions of future learning and
retention. Critically, exemplary implementations along with lessons learned of such assessments
need to be shared with the wider community to help members evolve in their instructional practice.
To take advantage of the vast amount of data using information technologies for research, we
recommend that engineering educators and engineering education researchers go beyond tradi-
tional narrow forms of outcomes assessments with the goal of understanding the learning process
itself (Pellegrino et al., 2001). For instance, “stealth assessment,” assessment embedded into the
instructional environment, can serve both purposes (Shute, 2011). With these data, inferences can
be made about how to support the learning process as well as how students achieve competency
(Shute & Ventura, 2013). However, care should be taken in over relying on computer-enabled meth-
ods. Learning is complex and requires human interpretation that goes beyond the capability of
technology (Worsley & Blikstein, 2014). Computer-enabled methods should be used as complement
SPRING 2019 29
ADVANCES IN ENGINEERING EDUCATION
to qualitative approaches of data analysis where, for instance, human-scored students’ processes
can be used as an input to technology-based assessment tools (Worsley & Blikstein, 2014).
Finally, we recommend that funding agencies identify and implement monitoring programs to
ensure that research data and other products resulted from federal funds are made available to the
research community following guidelines promised and approved in the proposal (i.e., data manage-
ment plans). It is also recommended to university administrators and investigators to implement and
follow proper procedures for data to be handled and shared properly. Effective practices to guide
data sharing and handling process require deliberate effort and are highly needed.
Issue 5: Alignment of Technology with Instructional Practice: Faculty Beliefs and Pedagogical
Knowledge
We argue that programs, strategies and mechanisms are needed to address faculty pedagogical
beliefs, to help develop faculty technological pedagogical content knowledge, and to connect the
two. Faculty pedagogical attitudes and beliefs are a vital first step toward technology acceptance
essential for instructors to develop sufficient knowledge of pedagogy and technology, and align
Many faculty commonly hold low digital literacy, which limits their ability to effectively integrate
technology into their teaching practice (NMC, 2014). Part of the problem relates to the lack of effec-
tive faculty professional development. However, deeper challenges that exacerbate this issue relate
to: (a) faculty attitudes and pedagogical beliefs that may influence the decision and the ways faculty
integrate technology in the classroom (Ertmer, 2005); (b) faculty development of technological
pedagogical content knowledge (Mishra & Koehler, 2006); and (c) faculty perceptions of usefulness
and ease of incorporating technology (Nolen et al., 2009; Venkatesh & Davis, 2000). A rewards
system that systematically and increasingly values research activities over teaching activities can
limit the effort faculty are willing to devote to improving their teaching practices (Fairweather, 1993).
Faculty attitudes and beliefs are a vital first step toward technology acceptance and eventually
integration (Dusick, 1998; Ertmer, 2005). Faculty attitudes and pedagogical beliefs relate to sup-
positions, opinions, commitments, expectations, and ideologies that they hold about technology
(Ertmer, 2005; Hermans, Tondeur, van Braak, & Valcke, 2008). Faculty first need to identify the value
of technology in helping them achieve instructional goals they perceive to be important (Watson,
2006). Faculty also need to develop their knowledge about technology to a point where they are
confident using it (Faseyitan, Libii, & Hirschbuhl, 1996; Wozney, Venkatesh, & Abrami, 2006) and
30 SPRING 2019
ADVANCES IN ENGINEERING EDUCATION
overcome levels of anxiety when using the technology (Johnson, Wisniewski, Kuhlemeyer, Isaacs,
& Krzykowski, 2012). Importantly, faculty need a clear understanding about how technology can
The way that pedagogical beliefs influence technology integration has been hypothesized as
follows: faculty with more instructor-centered pedagogical beliefs will implement more traditional
or “low level” technology uses, whereas teachers with more student-centered (e.g., constructivist)
pedagogical beliefs will implement more “high-level” technology uses (Ertmer & Ottenbreit-Leftwich,
2010; Roehrig, Kruse, & Kern, 2007). Two studies were identified that provide some insights about
engineering faculty pedagogical beliefs and some of the forms in which they use technology for
teaching and learning. Middleton and colleagues (2015), identified that faculty can hold student-
centered beliefs, teacher-centered beliefs, and non-discriminatory beliefs. In addition, faculty apply
pedagogies that correlated significantly to their attitudes; faculty with student-centered beliefs en-
gaged in more learner centered practices than either teacher-centered faculty or non-discriminating
faculty. As related to engineering faculty uses of technology, it was identified that faculty com-
monly use Internet for simplistic static tasks (e.g., posting of syllabi) as opposed to dynamic more
complicated ones (e.g., online discussion forums) (St. Clair & Baker, 2003). Faculty also reported
that easiest tools to use were more commonly used and at the same time were perceived as most
effective and efficient. On the other hand, the hardest tools to use were used less often and were
perceived as not so effective and efficient (St. Clair & Baker, 2003).
An initial step towards change in pedagogical beliefs that help faculty become comfortable with
using the technology for teaching and learning can be addressed via faculty development programs
(Faseyitan et al., 1996). However, knowing about pedagogical methods, learning theories, or prin-
ciples of good instruction does not mean that faculty will adopt them (Lin, Yu, Wang, & Ho, 2015).
For example, a study that investigated the effect of an online faculty development program to lead
to a change in faculty beliefs and intentions towards more student-centered learning revealed that
after the training, faculty indeed decreased their intention towards knowledge transmission models.
However their practice did not become more student-centered as a result of participation in profes-
sional development (Rienties, Brouwer, & Lygo-Baker, 2013). What has been identified as a predictor
faculty identifying changes in their students’ learning (Levin & Wadmany, 2005), such as the case
presented by Moore and colleagues (Moore et al., 2015) through the use of Model-Eliciting Activities.
outcomes, and perceived usefulness and ease of use, may determine intention to use technology and
subsequently, integration (Taylor & Todd, 1995). Researchers have argued that university faculty do
not choose to use technology in the classroom even though they feel some technologies can improve
SPRING 2019 31
ADVANCES IN ENGINEERING EDUCATION
students’ learning (Ajjan & Hartshorne, 2008). For instance, a predictor for technology integration
for learning is faculty perceived value of technology use (Wozney et al., 2006), along with perceived
usefulness and compatibility (Ajjan & Hartshorne, 2008). That is, faculty may make value judgements
in regards to technology affordances to help them achieve instructional goals they perceive to be
important (Watson, 2006). The more valuable faculty perceive a particular technology is in helping
them achieve instructional goals, the more likely they are to use it (Ertmer & Ottenbreit-Leftwich,
2010). Other predictors for technology integration relate to faculty self-efficacy and confidence in
achieving specific learning outcomes with technology (Faseyitan et al., 1996; Wozney et al., 2006),
as well as the level of anxiety experienced when using the technology (Johnson et al., 2012). For
instance, low personal knowledge about technology may result in stress or anxiety (Kersaint, Hor-
ton, Stohl, & Garofalo, 2003). On the other hand, when faculty believe that the technology is easy
to use and compatible with the way they work, their likelihood of integrating it into the classroom
increases (Ajjan & Hartshorne, 2008). Developers could better align innovative technologies to the
real needs and practices of faculty through processes such as the “customer discovery” interviews
that are central to the I-Corps L program (Guerra & Smith, 2016)
Faculty Knowledge
Studies in higher education level have identified significant correlations between technology
literacy and integration into pedagogical practice (Georgina & Hosford, 2009; Georgina & Olson,
2008). Specifically, low personal knowledge about technology may be a strong barrier to technology
implementation (Kersaint et al., 2003). In addition, educational researchers have identified that in
order to improve student learning with technology, it is essential for instructors to properly orches-
trate the interplay among content, pedagogy, and technology use. Effective technology integration
for meaningful learning requires that educators comprehend the technology tools themselves, along
with the specific affordances of each tool enable for conceptual understanding, problem solving,
design thinking, or other desired outcomes (Angeli & Valanides, 2009). This interplay among faculty
content, pedagogy and technology knowledge has been referred to as Technological Pedagogical
The TPCK framework has emerged as a form to represent educators’ knowledge on how to in-
tegrate technology into their teaching and learning activities (Abbitt, 2011). It was developed as an
extension of Shulman’s (Shulman, 1986) pedagogical content knowledge (PCK). One of the criticisms
of the TPCK framework is that it is complex and difficult to measure (Archambault & Barnett, 2010).
Specifically elements of the TPCK framework are difficult to tease out (Shinas, Yilmaz-Ozden, Mouza,
Karchmer-Klein, & Glutting, 2013). However, clear distinctions have been identified for technology
knowledge (Archambault & Barnett, 2010; Shinas et al., 2013). Similarly, valid and reliable instruments
32 SPRING 2019
ADVANCES IN ENGINEERING EDUCATION
have recently been developed that allow faculty to self-report about their own TPCK (Yurdakul
et al., 2012) and students to indicate their views about their university professors knowledge in
In general, however, understanding of what technology knowledge is needed and the degree
that faculty exhibit such knowledge is sparse. A survey study of 360 engineering faculty from eight
institutions reported on faculty knowledge of specific technologies for teaching and learning (Chen,
Ellis, Lockhart, & Hamoush, 2000). While common technologies such as email, word processing
and the Internet were widely used by engineering faculty, they reported little skill in more specific
class-specific technology uses such as developing multimedia modules, Java applets, or even creat-
ing web pages for a course or holding electronic help-sessions or office hours using conferencing
or collaboration tools. While the landscape has probably shifted somewhat in the years since this
study, more work is needed to identify the types of technological knowledge that faculty have (e.g.,
Magana et al., 2012). Furthermore, valid and reliable ways to measure that knowledge would pro-
vide information to allow support strategies to be developed for more effective use of technology.
Some external factors attributed to faculty not choosing to integrate technology relate to the
lack of organizational support, not enough resources, weak leadership in guiding them through the
adoption process, and lack of training to develop technology knowledge and pedagogical knowledge
(Keengwe, Kidd, & Kyei-Blankson, 2009). Other barriers include the time and effort needed to learn
how to use a technology, the planning to incorporate it into a classroom setting, and concerns about
negative impacts of instructional innovations on faculty teaching evaluations (Berk, 2005; Kersaint
et al., 2003; St. Clair & Baker, 2003). These external factors or barriers can be addressed. Engineering
colleges and departments can support faculty efforts with incentives and time to explore technology
in order to identify effective ways to use it in the classroom, and get rewarded for doing so (Brownell
& Tanner, 2012; Finley & Hartman, 2004; Surry & Land, 2000). Once faculty have used technology for
one specific purpose, along with sufficient technical support and colleague-supported training, they
are more likely to use technology for other purposes (Jacobsen, 2000; Sahin & Thompson, 2007).
For faculty development, we draw on the literature for development of knowledge for complex
pedagogical practices. For example, a qualitative study that measured the effects after a faculty
development program aimed at integrating problem based learning (PBL) among 31 faculty revealed
transformations on their PCK (Major & Palmer, 2006). After conducting interviews with the 31 faculty
and examining their portfolios, Major and Palmer (2006) identified that faculty implementation of
PBL encouraged them to critically examine their teaching and learning processes. Another useful
example comes from a faculty development program in science education that aimed at implementing
SPRING 2019 33
ADVANCES IN ENGINEERING EDUCATION
a cognitive apprenticeship model (Sunal et al., 2001). Part of the workshop dynamics included an
iterative process of exchanging roles from teacher to learner and back again. This exchange of roles
allowed faculty to (a) share beliefs publicly, (b) discuss, reflect and observe alternative approaches for
teaching in order to create cognitive conflict, and (c) reconstruct ideas related to effective learning
and teaching in disciplinary courses. We can draw on this work to develop more effective strategies
for faculty development around complex technological practices in the engineering classroom. Spe-
cific strategies that faculty have identified as useful are interaction of faculty from different colleges,
connections with faculty with similar goals, institutional and administrative support, development
of interpersonal skills, participating in action research, and joining external networks of faculty to
collaborate and disseminate results (Sunal et al., 2001).
Recommendations
Our recommendations for this issue are addressed to university administration and personnel
from units or centers focused on faculty development or instructional innovation programs (e.g.,
centers for instructional excellence). An effective way to shift beliefs and increase faculty knowledge
about technology and how to integrate it effectively into their classrooms starts with participation
on faculty development programs, followed by design of learning materials, and then to actual
integration of the technology in their classrooms (Doering, Veletsianos, Scharber, & Miller, 2009;
Koehler & Mishra, 2005; Koehler, Mishra, & Yahya, 2007). Faculty development programs can not
only help faculty improve their knowledge about how to integrate technologies for teaching and
learning, but can also help them become comfortable with using the technology for such purposes
(Faseyitan et al., 1996). This technology knowledge should go beyond using technology to make
classroom management and delivery more effective. Specifically, faculty need to understand how to
use technology to facilitate and support meaningful learning that can result in students’ connected
knowledge and skills readily available for application to real situations (Lawless & Pellegrino, 2007).
This recommendation for faculty development programs focuses on strategies designed to improve
faculty pedagogical and technological knowledge. We advocate for following the guidelines similar
to those developed by the National Effective Teaching Institute (NETI) (Felder & Brent, 2010; Felder,
Brent, & Prince, 2011), but adapted so technology integration is considered in the proposed practices.
use of technology in engineering. These programs should incorporate the design of technology-
enhanced learning materials, and supports for the actual implementation of the technology in their
classrooms. Long-term and organizational benefit can follow if there is a critical examination and
reflection of individuals’ teaching and learning processes and the creation of communities of practice
34 SPRING 2019
ADVANCES IN ENGINEERING EDUCATION
To this point, we have considered technologies that have been specifically developed for learning
engineering (learning innovations), for domain-specific applications to engineering practice, and for
instructional design including learning outcomes, assessment, and instructional practice. We have
also argued that faculty professional development is a critical component if engineering educators
Our discussion has focused on intentional uses and specific ways that technology can be enacted
broader perspective reveals many ways, both obvious and subtle, that technology interacts with the
work of engineering students and faculty at the university. In this section, we shift the discussion
from developing technologies for specific educational needs towards a broader consideration about
the affordances and tensions with technology towards learning engineering. While it is beyond the
scope of this article to unpack these issues in detail, they are critical to the conversation of how to
As illustrated earlier in Figure 1, the specific technologies that have been developed for engineer-
ing learning and practice are situated within the broader context of the role technology plays in the
lives of humans in the 21st century.4 Mobile, wired devices that provide broad and ready access to
information and to other people have redefined fundamental conceptions of space and time (Castells,
2011; Madhavan & Lindsay, 2014; Standage, 1998). In this section, we address some of the ways that
Technology (with a capital “T”) impacts the engineering learning environment for the better (e.g.,
increased access) and for the worse (e.g., access to solution manuals). Capital “T” Technology is
what Arthur (2009) refers to as the “entire collection of devices and practices available to a culture”
(p. 28). Engineering educators need to understand and negotiate the tensions between Technology
as a large and continually changing force in society and their ability to adapt educational systems
With the goal of promoting systemic change towards more productive learning with equitable
access for all learners, we draw on Arthur’s definition to identify how Technology fits into the cultures
of engineering programs and the culture of higher education. At its core, Technology is a cultural
phenomenon of interlocking material and social practices (Leidner & Kayworth, 2006; Latour &
Woolgar, 2013; Leonardi & Barley, 2008: Pickering, 2010). Engineering faculty and students exist
within the broad culture of higher education, and within specific departmental and institutional cul-
tures, and all of these are changing rapidly. Engineering educators need to create technologies and
4
It is easy to lose sight of how rapidly computer technology is changing, e.g., the widespread use of the personal
computer (began in the 1980s), the internet (1990s), and the smart phone (2000s).
SPRING 2019 35
ADVANCES IN ENGINEERING EDUCATION
enact technology-based practices with the understanding of what culture they want to create. Such
strategies need to account for and shift the values, beliefs, and norms that permeate institutions of
higher education. Faculty, administrators, students, and researchers need to begin by creating the
“time” and “space” to have these conversations. For example, one could argue that a cultural norm
typical in Carnegie Research 1 universities is the need for minimally invasive teaching so that faculty
have more time for research. This norm is reflected in the rewards structures for tenure, promotion,
and merit raises, and also in a growing lower-paid, often marginalized, class of fixed term instructors
to carry the teaching load. If it is important to think about Technology as providing affordances that
enhance student learning, the engineering education community needs to change the way people
think about teaching and the value placed on teaching in the context of their work. In this regard,
there may exist tensions between effectiveness and efficiency. Technology should serve to make
teaching more effective, not less intrusive - to improve teaching, not commoditize it. This cultural
context needs to be considered in developing strategies for faculty development discussed above.
Thus, in considering how Technology fits into the educational system, and extrapolating for how
it will fit into the system into the future, it is necessary to fundamentally address core issues of
Students also bring Technology-mediated cultural behaviors to the engineering classroom. In-
coming students are referred to as digital natives, a term that is often interpreted as digital fluency,
but more appropriately refers to their habits, values and beliefs (Howe & Strauss, 2007; Kirschner
& De Bruyckere, 2017; Oblinger, 2003). In general, this group, the IGens (Twenge, 2017) and before
them the Millennials, is tech-savvy and connected. They are characterized as collaborative (crowd-
sourcing), prone to multi-tasking (which in fact is not effective), and accustomed to instant gratifi-
cation5. As students, they seek a greater work-life balance than their predecessors and seek more
immediate meaning and relevancy to their work. They spend a reported 30% less time studying than
their counterparts 50 years earlier (Cheville, 2012). In addition to less time studying, their activity
may be less focused as their use of computers and other devices for work and fun often coalesce.
They believe that if something is digital, it is everyone’s property – an attitude that can lead to con-
flicts with policies on academic misconduct. However, educators should not lose sight that IGen
characteristics can be productive. They bring to the classroom stronger abilities to use information
technology and to collaborate than predecessors; both these skills are needed in engineering work.
Their outlook that “doing is more important than knowing” can be leveraged in authentic project-
based work around engineering design. Finally, we agree with Bennett, Maton, & Kervin (2008) that
5
Behaviors of instant gratification logically follow the assertion that the internet is compressing or annihilating
36 SPRING 2019
ADVANCES IN ENGINEERING EDUCATION
incoming students’ relationship with technology may be more nuanced and complex than the digital
native characterization implies, and that better understanding is needed to characterize and identify
ways that learning environments can be shifted to better fit this and future generations of learners.
A broad perspective is needed to connect the affordances and tensions of Technology with the
ends for which education should work. Technology has led to tectonic shifts in the amount of infor-
mation available to both students and faculty and the ways they can collaboratively work. However,
the educational processes and systems at the core of many engineering programs remain intact from
the pre-internet era. Engineering education researchers need to better understand tensions between
Technology and the core educational processes and systems through which engineers are formed. We
illustrate this point with three examples. First, faculty and students may have very different perspec-
tives of the role of homework in learning engineering. Faculty could assign homework from a process
orientation, thinking that through engaging in problem solving processes of challenging problems,
students are elicited to make meaning of core concepts and principles. Thus, homework becomes a
central tool in the course design to promote deep learning. On the other hand, students know that
homework problem solutions are often only a click (or a text message) away6. As time constraints and
other stressful factors become salient, they may take a product orientation, prioritizing “getting the
points” for successful completion to genuine engagement. More subtly, students may believe that access
to solutions can be used to “guide” their thinking when stuck, and, even with the best intentions, also
circumvent the meaning making that leads to deep learning in working through challenging problems.
Second, most engineering textbooks are designed for the pre-internet era where the book needed to
serve as both a learning tool and also an encyclopedic resource for future reference. Students have
networked information resources now available rendering textbooks as transient tools only needed
during the term. This shift together with affordances of Technology towards interactivity and adap-
tive feedback discussed above necessitates a need to shift conceptions and manifestations of the
fundamental role of out of class resources in learning engineering (Lee et al., 2013; Edgcomb & Vahid,
2014). Applying the discussion from Issue 3 above, it is better to think about the resources learners
need outside the classroom and use technology to reconstitute those resources to be more effective
rather than simply converting a traditional textbook into an electronic format. From such a perspective,
resources shift from providing static content to interactively developing conceptual understanding
and problem solving skills (Vahid, Edgcomb, & Strawn, 2016). Third, we juxtapose the rich, networked
6
Technology may address this issue by incorporation of “rolling” problems that change in nature for each learner;
however, they may just lead to the next level of a set of students “gaming” homework. An alternative approach is
to develop homework activities within learning environments that prompt students to be interested and to see
SPRING 2019 37
ADVANCES IN ENGINEERING EDUCATION
collaborative learning uses of Technology outside of school and the limited, individualized uses in the
classroom. This contrast suggests that better alignment of the activities and assessments educators
choose in the engineering classroom may lead to higher motivation and more meaningful learning.
Recommendations
Studies are needed that explore how Technology interacts with the cultures of engineering
should draw collaborators with a wide range of expertise, including technology and science studies,
sociology, and anthropology. Rapid changes in Technology are shifting work practices and social
norms. Engineering educators need strategies to keep up with this change. For example, they need
to understand better how students use Technology to get their work done and what aspects of work
practices might misalign with instructors’ conceptions. From such an understanding, educators can
develop strategies to redirect unproductive habits and ideas to leverage others. With the constant
temptation of connectivity, education researchers also need to characterize the relationships between
multi-tasking and deep thinking. Does connectivity form a distraction that subverts complex think-
ing or can it be an affordance to distribute thinking and enhance learning? Ultimately, educators are
responsible for developing learning environments that fit their learners. They must institutionalize
processes that allow them to identify shifts in Technology-based social norms and work practices,
and respond agilely in ways that better match learners’ needs. Importantly, administrators and faculty
need to create space at the university to reward and recognize this challenging work.
CLOSING THOUGHTS
This article seeks to address the questions “What are the potential affordances of computer tech-
nology (ICCT) to make effective tools for teaching and learning engineering?” and “What tools will
scale and propagate?” Often mentioned first is that computers can provide individualized learning
pathways combined with feedback on how the learner is progressing. That is, computers make it
easier to engage in “what if” scenarios. While such scenarios will undoubtedly occupy space in the
educational technology landscape, it is unclear to what extent the social process of learning can be
effectively sourced to a computer tool. If we look at history as a guide, humans are often quick to
project changes to human social systems from the next technology on the horizon, simply by virtue
of the technology itself. For example, a common shared belief was that the telegraph was going
to bring world peace because it would allow humans from different nations to readily interact, and
thereby understand one another (Standage, 1998). But this optimistic prognosis overlooked the hard
38 SPRING 2019
ADVANCES IN ENGINEERING EDUCATION
social work of developing that shared understanding. Similarly, one could take the perspective that
the more general value of computer-based learning innovations is to augment, not replace, the social
work done in learning. From this perspective, computer technology allows more possibilities for the
We have argued that technology-enabled pedagogical systems can fundamentally shift the space
where learning occurs. By that, we do not mean that learning becomes any less human, just that
there are more possibilities for creating meaningful human interactions,7 interactions that can fore-
ground conceptual understanding or foreground disciplinary practices. To realize this use, educators
need to expand their conceptions of the learning environment at the university. There is a tendency
to see technology in terms of the current structures and systems that are in place. The telephone
was first considered as a “talking telegraph” and the automobile first as a “horseless carriage.” In
each of these cases, technology is viewed as making what was once difficult easier. Alternatively,
educators might ask how Technology affords paradigmatic shifts in the fundamental ways they do
the challenging cognitive and social work of helping students learn engineering.
Ultimately, the ways and degree to which educators choose to reconstitute learning environ-
ments in response to Technology change is inextricably linked to their conceptions of learning and
knowing. They might consider knowledge as an entity to be transferred and skills development as
proceeding through orchestrated behavioral responses to external stimuli. Alternatively, they can
view learning as a socially-mediated process where the learner constructs conceptual understand-
ing building on their prior knowledge and experiences through orchestrated activity. Finally, they
might say that learning is a sociocultural process of increasingly greater participation in the valued
educational systems with Technology. One perspective might emphasize personalized and adaptive
learning where the technology provides a responsive diagnostic agent that directly and immedi-
ately responds to the individual learner. Another perspective might prioritize a use of technology
to put students in social situations where they make meaning of activity through interacting with
one another. Failure to make connections to core learning processes can doom instantiations of
technology to the same mistakes as earlier interventions (e.g., see Webel, Krupa, & McManus, 2015).
7
In the mid twentieth century, Claude Shannon (see Shannon, 2001) connected the concept of entropy to informa-
tion. From such a perspective, one could argue that the exponentially increasing amounts of information acces-
sible by computer technology leads to many more system “configurations” in any arena it touches, including the
professional formation of engineers. But, like with entropy, most sets of configurations tend to be disordered. It
takes coordinated input of the right type of work to create “low entropy” configurations that could potentially
SPRING 2019 39
ADVANCES IN ENGINEERING EDUCATION
In some ways, these juxtaposed commitments harken the early debates during the development
of the internet (or ARPAnet) and personal computer (Markoff, 2005). There was one faction that
argued stridently for artificial intelligence (that computers should replace human function) and an-
other who advocated for augmentation (computers should support human thinking and activity).
Clearly, underlying philosophical commitments provide the impetus for the Technology approaches
engineering educators, administrators, and policy makers privilege and where they direct resources.
ACKNOWLEDGMENTS
The work reported here was supported in part by grants from the National Science Foundation,
including DRL 1451263, EEC 1160353, and EEC 1449238. The views represented here are those of
the authors and do not represent the National Science Foundation. The authors are grateful for
the conversations and feedback from Mary Besterfield-Sacre, Alan Cheville, Larry Shuman, Martin
Storksdieck, and Scott Streiner. We are exceedingly grateful to the participants in the Delphi study
and the Workshops at ASEE and the University of Pittsburgh as well as the anonymous reviewers
REFERENCES
Abbitt, J. T. (2011). Measuring technological pedagogical content knowledge in preservice teacher education: A review
of current methods and instruments. Journal of Research on Technology in Education, 43(4), 281–300.
Abdulwahed, M. & Z.K. Nagy. (2009). Applying Kolb’s Experiential Learning Cycle for Laboratory Education. Journal
[ABET]. (2013). 2014-2015 Criteria for Accrediting Engineering Programs. Retrieved from Baltimore, MD: http://www.
abet.org/uploadedFiles/Accreditation/Accreditation_Step_by_Step/Accreditation_Documents/Current/2014_-_2015/
E001%2014-15%20EAC%20Criteria%203-13-14(2).pdf
Adams, R., Radcliffe, D., & Fosmire, M. (2016). Designing for global data sharing, designing for educational transforma-
Adams, W. K., Reid, S., LeMaster, R., McKagan, S. B., Perkins, K. K., Dubson, M., & Wieman, C. E. (2008). A study of
educational simulations part II-interface design. Journal of Interactive Learning Research, 19(4), 551.
Ajjan, H., & Hartshorne, R. (2008). Investigating faculty decisions to adopt Web 2.0 technologies: Theory and empirical
Akpan, J., & Strayer, J. (2010). Which comes first the use of computer simulation of frog dissection or conventional
dissection as academic exercise?. Journal of Computers in Mathematics and Science Teaching, 29(2), 113.
Alabi, O., Magana, A. J., & Garcia, R. E. (2015). Gibbs computational simulation as a teaching tool for students’ under-
40 SPRING 2019
ADVANCES IN ENGINEERING EDUCATION
Angeli, C., & Valanides, N. (2009). Epistemological and methodological issues for the conceptualization, development,
and assessment of ICT–TPCK: Advances in technological pedagogical content knowledge (TPCK). Computers & Education,
52(1), 154–168.
Archambault, L. M., & Barnett, J. H. (2010). Revisiting technological pedagogical content knowledge: Exploring the
Arnold, K. E., & Pistilli, M. D. (2012). Course signals at Purdue: Using learning analytics to increase student success.
Proceedings from the 2nd international conference on learning analytics and knowledge.
Arthur, W. B. (2009). The nature of technology: What it is and how it evolves. Simon and Schuster.
[ASEE]. (2013). Transforming Undergraduate Education in Engineering (TUEE). Retrieved from Arlington, VA:
Atkinson, R. K., Derry, S. J., Renkl, A., & Wortham, D. (2000). Learning from examples: Instructional principles from
Aurigemma, J., Chandrasekharan, S., Nersessian, N. J., & Newstetter, W. (2013). Turning Experiments into Objects: The
Cognitive Processes Involved in the Design of a Lab-on-a-Chip Device. Journal of Engineering Education.
Ayasun, S., & Nwankpa, C. O. (2006). Transformer tests using MATLAB/Simulink and their integration into undergraduate
Barab, S., Thomas, M., Dodge, T., Carteaux, R., & Tuzun, H. (2005). Making learning fun: Quest Atlantis, a game without
Beichner, RJ, Saul, JM, Abbott, DS, Morse, JJ, Deardorff, D, Allain, RJ, & Risley, JS. (2007). The student-centered
activities for large enrollment undergraduate programs (SCALE-UP) project. Research-Based Reform of University
Beichner, R.J, Saul, JM, Allain, RJ, Deardorff, DL, & Abbott, DS (2000). Introduction to SCALE UP: Student-Centered
Activities for Large Enrollment University Physics. In Proceedings of the 2000 American society for engineering education
Beg, A. (2015). A web–based method for building and simulating standard cell circuits—A classroom application.
Bell, R. L., & Trundle, K. C. (2008). The use of a computer simulation to promote scientific conceptions of moon phases.
Bennett, S., Maton, K., & Kervin, L. (2008). The ‘digital natives’ debate: A critical review of the evidence. British journal
of educational technology, 39(5), 775–786.
Berk, R. A. (2005). Survey of 12 strategies to measure teaching effectiveness. International journal of teaching and
Besterfield-Sacre, M. E., & Shuman, L. J. (2016, June). Innovation through Propagation: A Roadmap for Engineering
Education. Paper presented at 2016 ASEE Annual Conference & Exposition, New Orleans, Louisiana.
Borgman, C. L., Abelson, H., Dirks, L., Johnson, R., Koedinger, K. R., Linn, M. C., . . . Salen, K. (2008). Fostering learning
in the networked world: The cyberlearning opportunity and challenge. A 21st century agenda for the National Science
Bowen, A., Reid, D., & Koretsky, M. (2015). Development of Interactive Virtual Laboratories to Help Students Learn
Bransford, J. D., Brown, A. L., & Coocking, R. R. (Eds.). (2000). How people learn: Brain, Mind, Experience, and School.
Brinson, L. C., Belytschko, T., Moran, B., & Black, T. (1997). Design and computational methods in basic mechanics
SPRING 2019 41
ADVANCES IN ENGINEERING EDUCATION
Brophy, S. P., Magana, A. J., & Strachan, A. (2013). Lectures and Simulation Laboratories to Improve Learners’ Conceptual
Brownell, S. E., & Tanner, K. D. (2012). Barriers to faculty pedagogical change: Lack of training, time, incentives, and…
Brooks, B. J., Demaree, D. N., & Koretsky, M. D. (2016). Student Response Times to In-Class Thermodynamics
Concept Questions: A Window into Students’ Thinking Processes. International Journal of Engineering Education,
32(1), 30–38.
Brown, S., Race, P., & Bull, J. (1999). Computer-Assisted Assessment in Higher Education. Staff and Educational
Campbell, J., J. Bourne, P. Mosterman, and A. Brodersen. (2002). The Effectiveness of Learning Simulations for
Castells, M. (2011). The rise of the network society: The information age: Economy, society, and culture (Vol. 1). John
Caro, E., González, C., & Mira, J. M. (2014). Student academic performance stochastic simulator based on the Monte
Castrellón, T., Botía, D., Gómez, R., Orozco, G., & Gil, I. (2011). Using process simulators in the study, design, and
control of distillation columns for undergraduate chemical engineering courses. Computer Applications in Engineering
Chamberlain, J. M., Lancaster, K., Parson, R., & Perkins, K. K. (2014). How guidance affects student engagement with
Chen, J. C., Ellis, M., Lockhart, J., Hamoush, S., Brawner, C. E., & Tront, J. G. (2000). Technology in engineering education:
what do the faculty know and want? Journal of Engineering Education, 89(3), 279–283.
Chesler, N., Arastoopour, G., D’Angelo, C.M., Bagley, E., & Shaffer, D.W. (2013). Design of a professional practice simulator
for educating and motivating first-year engineering students. Advances in Engineering Education 3(3): 1–29.
Cheville, R. A. (2012). Engineering education today: Capturing the afterlife of sisyphus in five snapshots. Proceedings
Cheville, R. A. (2016). Data sharing from a policy perspective. Advances in Engineering Education, 5(2).
Clough, G. W. (2004). The engineer of 2020: Visions of engineering in the new century. National Academy of
Collins, A., Brown, J. S., & Newman, S. E. (1989). Cognitive apprenticeship: Teaching the crafts of reading, writing, and
mathematics. Knowing, learning, and instruction: Essays in honor of Robert Glaser, 18, 32–42.
Cool, T., García, R. E., & Bartol, A. (2015). Gibbs. Retrieved from https://nanohub.org/resources/gibbs
Corter, J. E., Nickerson, J. V., Esche, S. K., Chassapis, C., Im, S., & Ma, J. (2007). Constructing reality: A study of remote,
hands-on, and simulated laboratories. ACM Transactions on Computer-Human Interaction (TOCHI), 14(2), 7.
Crismond, D. (2001). Learning and using science ideas when doing investigate-and-redesign tasks: A study of naive,
novice, and expert designers doing constrained and scaffolded design work. Journal of Research in Science Teaching,
38(7), 791–820.
Dahm, K. D., Hesketh, R. P., & Savelski, M. J. (2002). Is process simulation used effectively in ChE courses? Chemical
42 SPRING 2019
ADVANCES IN ENGINEERING EDUCATION
Dede, C., Nelson, B., Ketelhut, D. J., Clarke, J., & Bowman, C. (2004, June). Design-based research strategies for study-
ing situated learning in a multi-user virtual environment. In Proceedings of the 6th international conference on Learning
de Jong, T., Linn, M. C., & Zacharia, Z. C. (2013). Physical and virtual laboratories in science and engineering education.
de Jong, T., & Van Joolingen, W. R. (1998). Scientific discovery learning with computer simulations of conceptual
de Jong, T., & van Joolingen, W. R. (2008). Model-facilitated learning. Handbook of research on educational
Dori, Y. J., Belcher, J., Bessette, M., Danziger, M., McKinney, A., & Hult, E. (2003). Technology for active learning.
Delale, F., Liaw, B. M., Jiji, L. M., Voiculescu, I., & Yu, H. (2011). Infusion of Emerging Technologies and New Teach-
ing Methods into The Mechanical Engineering Curriculum at The City College of New York. Advances in Engineering
Education, 2(4).
Diefes–Dux, H. A., Samant, C., Johnson, T. E., & O’Connor, D. (2004). Kirkpatrick’s Level 1 Evaluation of the Implementa-
tion of a Computer–Aided Process Design Tool in a Senior–Level Engineering Course. Journal of Engineering Education,
93(4), 321–331.
[DOEd]. (2010). United States. Department of Education. Office of Educational Technology. Transforming American
[DOEd]. (2016). United States. Department of Education. Future Ready Learning: Reimagining the role of technology
in education.
Doering, A., Veletsianos, G., Scharber, C., & Miller, C. (2009). Using the technological, pedagogical, and content
knowledge framework to design online learning environments and professional development. Journal of educational
Dorneich, M.C. & P.M. Jones. (2001). The UIUC Virtual Spectrometer: A Java-Based Collaborative Learning Environment.
Dusick, D. M. (1998). What social cognitive factors influence faculty members’ use of computers for teaching? A
Edelson, D. C., Gordin, D. N., & Pea, R. D. (1999). Addressing the challenges of inquiry-based learning through technology
Edgcomb, A. D., & Vahid, F. (2014, June). Effectiveness of Online Textbooks vs. Interactive Web-Native Content. In Pro-
ceedings of the 2014 American society for engineering education annual conference and exposition. Indianapolis, Indiana.
El-ZEin, A., Langrish, T., & Balaam, N. (2009). Blended teaching and learning of computer programming skills in
Ertmer, P. A. (2005). Teacher pedagogical beliefs: The final frontier in our quest for technology integration? Educational
Ertmer, P. A., & Ottenbreit-Leftwich, A. T. (2010). Teacher technology change: How knowledge, confidence, beliefs,
Fairweather, J. S. (1993). Faculty reward structures: Toward institutional and professional homogenization. Research
SPRING 2019 43
ADVANCES IN ENGINEERING EDUCATION
Faseyitan, S., Libii, J. N., & Hirschbuhl, J. (1996). An in service model for enhancing faculty computer self-efficacy.
Felder, R. M., & Brent, R. (2010). The National Effective Teaching Institute: Assessment of impact and implications for
Felder, R. M., Brent, R., & Prince, M. J. (2011). Engineering instructional development: Programs, best practices, and
Fernández-Balboa, J.-M., & Stiehl, J. (1995). The generic nature of pedagogical content knowledge among college
Fink, L. D., Ambrose, S., & Wheeler, D. (2005). Becoming a professional engineering educator: A new role for a new
Finkelstein, N. D., Adams, W. K., Keller, C. J., Kohl, P. B., Perkins, K. K., Podolefsky, N. S., ... & LeMaster, R. (2005).
When learning about the real world is better done virtually: A study of substituting computer simulations for laboratory
Finley, L., & Hartman, D. (2004). Institutional change and resistance: Teacher preparatory faculty and technology
Finelli, C. J. & Froyd, J. E. (2019). Improving Learning in the Engineering Classroom. Advances in Engineering Education, 7(2).
Fishman, B., Marx, R. W., Blumenfeld, P., Krajcik, J., & Soloway, E. (2004). Creating a framework for research on systemic
Foote, K. T., Neumeyer, X., Henderson, C., Dancy, M. H., & Beichner, R. J. (2014). Diffusion of research-based instruc-
tional strategies: the case of SCALE-UP. International Journal of STEM Education, 1(1), 1.
Ford, M. J., & Forman, E. A. (2006). Redefining disciplinary learning in classroom contexts. Review of research in
Freina, L., & Ott, M. (2015, January). A literature review on immersive virtual reality in education: state of the art and
perspectives. In The International Scientific Conference eLearning and Software for Education (Vol. 1, p. 133). “ Carol I”
Friedrichsen, D. M., Smith, C., & Koretsky, M. D. (2016). Propagation from the start: the spread of a concept-based
Friel, T., Britten, J., Compton, B., Peak, A., Schoch, K., & VanTyle, W. K. (2009). Using pedagogical dialogue as a vehicle
to encourage faculty technology use. Computers & Education, 53(2), 300–307.
Froyd, J. E., Wankat, P. C., & Smith, K. A. (2012). Five major shifts in 100 years of engineering education. Proceedings
García–Herreros, P., & Gómez, J. M. (2013). Modeling and optimization of a crude distillation unit: A case study for
Garrett, K. N. (2014). A quantitative study of higher education faculty self-assessments of technological, pedagogical,
and content knowledge (TPaCK) and technology training. The University of Alabama TUSCALOOSA.
Georgina, D. A., & Hosford, C. C. (2009). Higher education faculty perceptions on technology integration and training.
Georgina, D. A., & Olson, M. R. (2008). Integration of technology in higher education: A review of faculty self-perceptions.
Guerra, R. C. C., & Smith, K. A. (2016, October). I-Corps™ for Learning: Sustaining and scaling STEM education innova-
tions for impact. In Frontiers in Education Conference (FIE), 2016 IEEE (pp. 1–2). IEEE.
44 SPRING 2019
ADVANCES IN ENGINEERING EDUCATION
Gibson, J. (1979). The ecological approach to visual perception: Hillsdale, NJ: Lawrence Erlbaum Associates.
Gilbuena, D. M., Sherrett, B. U., Gummer, E. S., Champagne, A. B., & Koretsky, M. D. (2015). Feedback on Professional
Skills as Enculturation into Communities of Practice. Journal of Engineering Education, 104(1), 7–34.
Gilmore, R. O., Adolph, K. E., Millman, D. S., & Gordon, A. (2016). Transforming education research through open video
Goel, A. (2016). Artificial intelligence course creates AI teaching Assistant. Georgia Institute of Technology News Center. Re-
Goldstein, M. H., Purzer, S‚ ., Vieira, C., Zielinski, M., & Douglas, K. A. (2015). Assessing idea fluency through the student
design process. Paper presented at the Frontiers in Education Conference (FIE), 2015. 32614 2015. IEEE.
Johri, A., Yang, S., Vorvoreanu, M., & Madhavan, K. (2016). Perceptions and practices of data sharing in engineering
Greenberg, J. E., Smith, N. T., & Newman, J. H. (2003). Instructional module in fourier spectral analysis, based on
Guzdial, M. (1994). Software–realized scaffolding to facilitate programming for science learning. Interactive Learning
Haddad, J. A. (1985). Engineering Education and Practice in the United States. Foundations of Our Techno-Economic
Haudek, K. C., Prevost, L. B., Moscarella, R. A., Merrill, J., & Urban-Lurain, M. (2012). What are they thinking? Automated
analysis of student writing about acid–base chemistry in introductory biology. CBE-Life Sciences Education, 11(3), 283–293.
Hazel, E. and C. Baillie. (1998). Improving Teaching and Learning in Laboratories. GoldGuide Series, No. 4 (Canberra:
Heller, D. E., & Rogers, K. R. (2006). Shifting the burden: Public and private financing of higher education in the United
States and implications for Europe. Tertiary Education & Management, 12(2), 91–117.
Hermans, R., Tondeur, J., van Braak, J., & Valcke, M. (2008). The impact of primary school teachers’ educational beliefs
Hilton, M. (Ed.). (2002). Enhancing undergraduate learning with information technology: A workshop summary.
engineering course. Paper presented at the Proceedings of the 2007 American Society for Engineering Education Annual
Hinds, T., Urban-Lurain, M., Sticklen, J., Amey, M., & Eskil, T. (2005). Curricular Integration of Computational Tools: A First
Step. Paper presented at the Proceedings of the 2005 American Society for Engineering Education Annual Conference
Hodge, H., H.S. Hinton, & M. Lightner. (2001). Virtual Circuit Laboratory. Journal of Engineering Education. 90:507–511.
Hoole, S. R. H., Sivasuthan, S., Karthik, V. U., & Hoole, P. R. P. (2015). Flip–teaching engineering optimization, electro-
magnetic product design, and nondestructive evaluation in a semester’s course. Computer Applications in Engineering
Hornaes, H. P., & Royrvik, O. (2000). Aptitude, gender, and computer algebra systems. Journal of Engineering
Horodyskyj, L. B., Mead, C., Belinson, Z., Buxner, S., Semken, S., & Anbar, A. D. (2018). Habitable Worlds: Delivering on
SPRING 2019 45
ADVANCES IN ENGINEERING EDUCATION
Howe, N., & Strauss, W. (2007). Millennials go to college. Great Falls, VA: LifeCourse Associates.
Hutchison, M. A., Follman, D. K., Sumpter, M., & Bodner, G. M. (2006). Factors Influencing the Self-Efficacy Beliefs of
Impelluso, T. J. (2009). Leveraging cognitive load theory, scaffolding, and distance technologies to enhance computer
International Engineering Alliance. (2011). Washington Accord Program Outcomes. Retrieved from Washington, DC:
http://www.ieagreements.org/Rules_and_Procedures.pdf?4504
Jaakkola, T., Nurmi, S., & Veermans, K. (2011). A comparison of students’ conceptual understanding of electric circuits
in simulation only and simulation-laboratory contexts. Journal of Research in Science Teaching, 48, 71–93.
Jacobsen, D. M. (2000). Examining technology adoption patterns by faculty in higher education. Proceedings of ACEC.
Johnson, D. & Johnson, R. (2009). An educational psychology success story: Social interdependence theory and
Johnson, D. W. & Johnson R. T. (1999). Learning Together and Alone: Cooperative, Competitive, and Individualistic
Johnson, D. W., Johnson, R. T., & Stanne, M. B. (2000). Cooperative learning methods: A meta-analysis. Retrieved July,
Johnson, L., Adams Becker, S., Estrada, V., & Freeman, A. (2014). NMC Horizon Report: 2014 Higher Education Edition.
Johnson, T., Wisniewski, M. A., Kuhlemeyer, G., Isaacs, G., & Krzykowski, J. (2012). Technology Adoption in Higher
Education: Overcoming Anxiety through Faculty Bootcamp. Journal of Asynchronous Learning Networks, 16(2), 63–72.
Kadiyala, M., & Crynes, B. L. (2000). A review of literature on effectiveness of use of information technology in
Keengwe, J., Kidd, T., & Kyei-Blankson, L. (2009). Faculty and technology: Implications for faculty training and
Kersaint, G., Horton, B., Stohl, H., & Garofalo, J. (2003). Technology beliefs and practices of mathematics education
Kezar, A. J., & Eckel, P. D. (2002). The effect of institutional culture on change strategies in higher education: Universal
principles or culturally responsive concepts? The Journal of Higher Education, 73(4), 435–460.
Khan, F., & Singh, K. (2015). Curricular Improvements through Computation and Experiment Based Learning Modules.
Kirschner, P. A., & De Bruyckere, P. (2017). The myths of the digital native and the multitasker. Teaching and Teacher
Koehler, M. J., & Mishra, P. (2005). What happens when teachers design educational technology? The development of
technological pedagogical content knowledge. Journal of educational computing research, 32(2), 131–152.
Koehler, M. J., Mishra, P., & Yahya, K. (2007). Tracing the development of teacher knowledge in a design seminar:
Integrating content, pedagogy and technology. Computers & Education, 49(3), 740–762.
Kollöffel, B., & Jong, T. (2013). Conceptual understanding of electrical circuits in secondary vocational engineering educa-
tion: Combining traditional instruction with inquiry learning in a virtual lab. Journal of engineering education, 102(3), 375–393.
Koretsky, M. D., Amatore, D., Barnes, C., & Kimura, S. (2008). Enhancement of student learning in experimental design
Koretsky, M. D., Brooks, B. J., & Higgins, A. Z. (2016a). Written justifications to multiple-choice concept questions
during active learning in class. International Journal of Science Education, 38(11), 1747–1765.
46 SPRING 2019
ADVANCES IN ENGINEERING EDUCATION
Koretsky, M. D., Brooks, B. J., White, R. M., & Bowen, A. S. (2016b). Querying the Questions: Student Responses and
Koretsky, M. D., Falconer, J. L., Brooks, B. J., Gilbuena, D. M., Silverstein, D. L., Smith, C., & Miletic, M. (2014). The AIChE
Concept Warehouse: A Web-Based Tool to Promote Concept-Based Instruction. Advances in Engineering Education, 4(1).
Koretsky, M. D., Kelly, C., & Gummer, E. (2011). Fundamental Research in Engineering Education. Student Learning in
Koretsky, M. D., Nolen S.B., Nefcy, E., Champagne, A.B. (accepted). Affordances of computer and physical laboratory
based design projects for engaging student teams in engineering practice. Cognition & Instruction.
Koretsky, M. D., Nolen, S. B., Tierney, G., & Wetzstein, L. (2015). Productive disciplinary engagement in design tasks.
Landau, R., & Rosenberg, N. (Eds.). (1986). The positive sum strategy: Harnessing technology for economic growth.
Lasry, N., Mazur, E., & Watkins, J. (2008). Peer Instruction: From Harvard to Community Colleges Retrieved from http://
mazur-www.harvard.edu/publications.php?function=display&rowid=613
Latour, B., & Woolgar, S. (2013). Laboratory life: The construction of scientific facts. Princeton University Press.
Lawless, K. A., & Pellegrino, J. W. (2007). Professional development in integrating technology into teaching and learning:
Knowns, unknowns, and ways to pursue better questions and answers. Review of educational research, 77(4), 575–614.
Layton, R. A., Loughry, M. L., Ohland, M. W., & Ricco, G. D. (2010). Design and validation of a web-based system for
assigning members to teams using instructor-specified criteria. Advances in Engineering Education, 2 (1), 1–28.
Lee, C. S., McNeill, N. J., Douglas, E. P., Koro–Ljungberg, M. E., & Therriault, D. J. (2013). Indispensable resource? A phe-
nomenological study of textbook use in engineering problem solving. Journal of Engineering Education, 102(2), 269–288.
Leonardi, P. M., & Barley, S. R. (2008). Materiality and change: Challenges to building better theory about technology
Leidner, D. E., & Kayworth, T. (2006). Review: a review of culture in information systems research: toward a theory of
Levin, T., & Wadmany, R. (2005). Changes in educational beliefs and classroom practices of teachers and students in
rich technology-based classrooms [1]. Technology, Pedagogy and Education, Vol. 14, No. 3, 2005, 14(3), 281–307.
Lewin, D. R., Seider, W. D., & Seader, J. D. (2004). Using Process Simulators in Chemical Engineering [: A Multimedia
Guide for the Core Curriculum. Wiley.
Lin, C. C., Yu, W. W., Wang, J., & Ho, M.-H. (2015). Faculty’s Perceived Integration of Emerging Technologies and Peda-
gogical Knowledge in the Instructional Setting. Procedia-Social and Behavioral Sciences, 176, 854–860.
Lindgren, R., & Schwartz, D. L. (2009). Spatial learning and computer simulations in science. International Journal of
Lindsay, E.D., & Good, M.C. (2005). Effects of Laboratory Access Modes upon Learning Outcomes. IEEE Transactions
on Education. 48:619–631.
Linn, M. C., Clark, D., & Slotta, J. D. (2003). WISE design for knowledge integration. Science education, 87(4), 517–538.
Loughry, M. L., Ohland, M. W., & Moore, D. D. (2007). Development of a theory-based assessment of team member
Loughry, M. L., Ohland, M. L., & Woehr, D. J. (2014). Assessing teamwork skills for assurance of learning using CATME
Ma, J., & Nickerson, J. V. (2006). Hands-on, simulated, and remote laboratories: A comparative literature review. ACM
SPRING 2019 47
ADVANCES IN ENGINEERING EDUCATION
Madhavan, K., & Lindsay, E.D. (2014). Use of information technology in engineering education. In Johri, Aditya, and
Barbara M. Olds, eds. Cambridge handbook of engineering education research. Cambridge University Press.
Madhavan, K., & Richey, M. C. (2016). Problems in Big Data Analytics in Learning. Journal of Engineering Education,
105(1), 6–14.
Mayrath, M. C., Clarke-Midura, J., Robinson, D. H., & Schraw, G. (Eds.). (2012). Technology-based assessments for 21st
century skills: Theoretical and practical implications from modern research. Charlotte, NC: IAP.
Magana, A. J., Brophy, S. P., & Bodner, G. M. (2012a). Instructors’ intended learning outcomes for using computational
Magana, A. J., Brophy, S. P., & Bodner, G. M. (2012b). Student Views of Engineering Professors Technological Pedagogical
Content Knowledge for Integrating Computational Simulation Tools in Nanoscale Science and Engineering. International
Magana, A.J., & Coutinho, G.S. (2017). Modeling and simulation practices for a computational thinking-enabled engi-
Magana, A. J., Falk, L. M., & Reese, J. M. (2013). Introducing Discipline-Based Computing in Undergraduate Engineering
Magana, A. J., Falk, M. L., Vieira, C., & Reese, M. J. (2016). A case study of undergraduate engineering students’ computational
literacy and self-beliefs about computing in the context of authentic practices. Computers in Human Behavior, 61, 427–442.
Magana, A.J., Fennell, H., Vieira, C., & Falk, M.L. (2019). Characterizing the interplay of cognitive and metacognitive
knowledge in computational modeling and Simulation practices. Journal of Engineering Education, 7(2).
Magana, A.J., Ortega-Alvarez, J.D., Lovan, R., Gomez, D., Marulanda, J. & Dyke, S. (2017). Virtual, local and remote
p. 91–105.
Mann, C. R. (1918). A study of engineering education: prepared for the Joint committee on engineering education of
Markoff, J. (2005). What the Dormouse Said: How the Sixties Counterculture Shaped the Personal Computer Industry.
Penguin.
Major, C. H., & Palmer, B. (2006). Reshaping teaching and learning: The transformation of faculty pedagogical content
McDermott, L.C. (1996). Physics by inquiry. New York: John Wiley & Sons, Inc.
McElhaney, K. W., & Linn, M. C. (2011). Investigations of a complex, realistic task: Intentional, unsystematic, and exhaustive
Miller, T. (2009). Formative computer–based assessment in higher education: the effectiveness of feedback in supporting
Middleton, J. A., Krause, S. J., Rae Beeley, K., Judson, E., Ernzen, J., & Chen, Y. (2015, June 14–17). Examining Relation-
ships and Patterns in Pedagogical Beliefs, Attitudes and Classroom Practices for Faculty of Undergraduate Engineering,
Math and Science Foundational Courses. In Proceedings of the 2015 American society for engineering education annual
Mishra, P., & Koehler, M. (2006). Technological pedagogical content knowledge: A framework for teacher knowledge.
Mitcham, C. (1994). Thinking through technology: The path between engineering and philosophy. University of Chicago
Press.
48 SPRING 2019
ADVANCES IN ENGINEERING EDUCATION
Moore, T. J., Guzey, S. S., Roehrig, G. H., Stohlmann, M., Park, M. S., Kim, Y. R., . . . Teo, H. J. (2015). Changes in Faculty Mem-
bers’ Instructional Beliefs while Implementing Model–Eliciting Activities. Journal of Engineering Education, 104(3), 279–302.
Morrison, B. B., Margulieux, L. E., & Guzdial, M. (2015). Subgoals, context, and worked examples in learning computing
problem solving. Paper presented at the Proceedings of the eleventh annual International Conference on International
Mosterman, P.J., M.A.M. Dorlandt, J.O. Campbell, C. Burow, R. Bouw, A.J. Brodersen, & J. Bourne. (1994). Virtual Engi-
Najm, M. R. A., Mohtar, R. H., Cherkauer, K. A., & French, B. F. (2010). Effect of integrating hydrologic scaling concepts
on students learning and decision making experiences. Advances in Engineering Education, 2(1).
National Research Council (Ed.). (1985). Engineering Education and Practice in the United States. Foundations of Our
National Research Council (Ed.). (1996). National science education standards. Washington, DC: National Academy
Press.
National Research Council (Ed.). (2011). Learning Science through Computer Games and Simulations. Washington,
National Research Council (Ed.). (2014). Developing Assessments for the Next Generation Science Standards.
Nelson, B. C., & Ketelhut, D. J. (2007). Scientific inquiry in educational multi-user virtual environments. Educational
Nolen, S. B., Ward, C. J., Horn, I. S., Childers, S., Campbell, S. S., & Mahna, K. (2009). Motivation Development in Novice
Ohland, M. W., & Long, R. A. (2016). The multiple-institution database for investigating engineering longitudinal
development: An experiential case study of data sharing and reuse. Advances in Engineering Education, 5(2).
Ohland, M. W., Loughry, M. L., Woehr, D. J., Finelli, C. J., Bullard, L. G., Felder, R. M., Layton, R. A., Pomeranz, H. R., &
Schmucker, D. G. (2012). The comprehensive assessment of team member effectiveness: Development of a behavior-
ally anchored rating scale for self and peer evaluation. Academy of Management Learning & Education, 11 (4), 609-630.
Okamoto, N., Hsu, T.-R., & Bash, C. E. (2009). A thermal management of electronics course and laboratory for under-
graduates. Advances in Engineering Education, 1(3).
Pal, S. (2012). Mining educational data to reduce dropout rates of engineering students. International Journal of
Paas, F., Renkl, A., & Sweller, J. (2004). Cognitive load theory: Instructional implications of the interaction between
Patil, A., & Codner, G. (2007). Accreditation of engineering education: review, observations and proposal for global
Pea, R., Wulf, W. A., Elliott, S. W., & Darling, M. A. (Eds.). (2003). Planning for two transformations in education and
Pellegrino, J. W., Chudowsky, N., & Glaser, R. (2001). Knowing what students know: Washington, DC: National Academy
Press.
Pickering, A. (2010). The mangle of practice: Time, agency, and science. University of Chicago Press.
Pirolli, P., & Recker, M. (1994). Learning strategies and transfer in the domain of programming. Cognition and instruction,
12(3), 235–275.
SPRING 2019 49
ADVANCES IN ENGINEERING EDUCATION
Podolefsky, N. S., Perkins, K. K., & Adams, W. K. (2010). Factors promoting engaged exploration with computer
Powell, R. M., Anderson, H., Van der Spiegel, J., & Pope, D. P. (2002). Using web–based technology in laboratory
Pyatt, K. & R. Sims. (2007). Learner performance and attitudes in traditional versus simulated laboratory experiences.
In ICT: Providing choices for learners and learning. Proceedings ascilite Singapore 2007.
Pyatt, K., & Sims, R. (2012). Virtual and physical experimentation in inquiry-based science labs: Attitudes, performance
Purzer, S‚ ., Goldstein, M. H., Adams, R. S., Xie, C., & Nourian, S. (2015). An exploratory study of informed engineering
design behaviors associated with scientific explanations. International Journal of STEM Education, 2(1), 1–12.
Rawson, K., & Stahovich, T. (2013). Predicting course performance from homework habits. In Proceedings of the 2013
American society for engineering education annual conference and exposition. Atlanta, GA.
Raineri, D. (2001). Virtual laboratories enhance traditional undergraduate biology laboratories. Biochemistry and
Recker, M., Krumm, A., Feng, M., Grover, S., & Koedinger, K. R. (2016). Educational data mining and learning analytics.
learning-analytics/
Rupp, A. A., Gushta, M., Mislevy, R. J., & Shaffer, D. W. (2010). Evidence-centered design of epistemic games: Measure-
ment principles for complex learning environments. The Journal of Technology, Learning and Assessment, 8(4).
Rutten, N., van Joolingen, W. R., & van der Veen, J. T. (2012). The learning effects of computer simulations in science
Schank, P., & Kozma, R. (2002). Learning chemistry through the use of a representation-based knowledge building
Sehati, S. (2000). Re-engineering the practical laboratory session. International Journal of Electrical Engineering
Sengupta, P., & Wilensky, U. (2009). Learning electricity with NIELS: Thinking with electrons and thinking in levels.
Shaffer, D. W. (2006). Epistemic frames for epistemic games. Computers & education, 46(3), 223–234.
Shannon, C. E. (2001). A mathematical theory of communication. ACM SIGMOBILE Mobile Computing and
Sharples, M., Adams, A., Alozie, N., Ferguson, R., FitzGerald, E., Gaved, M., McAndrew, P., Means, B., Remold, J., Rienties,
B., Roschelle, J., Vogt, K., Whitelock, D. & Yarnall, L. (2015). Innovating Pedagogy 2015: Open University Innovation Report
Sherrett, B. U., Nefcy, E. J., Gummer, E. S., & Koretsky, M. D. (2013). An Expert Solution to Assess an Industrially Situated,
Shin, Dongil, En Sup Yoon, Sang Jin Park, and Euy Soo Lee. (2000) Web-based interactive virtual laboratory
system for unit operations and process systems engineering education. Computers & Chemical Engineering. 24,
1381-1385.
Shute, V. J. (2011). Stealth assessment in computer-based games to support learning. Computer games and instruction,
55(2), 503–524.
Shute, V. J., & Ventura, M. (2013). Stealth assessment: Measuring and supporting learning in video games: MIT Press.
Silver, N. (2012). The signal and the noise: Why so many predictions fail-but some don’t. Penguin.
50 SPRING 2019
ADVANCES IN ENGINEERING EDUCATION
Silverstein, D. L., Bullard, L. G., Seider, W. D., & Vigeant, M. A. (2013, June). How We Teach: Capstone Design.
In P roceedings of the 2013 American society for engineering education annual conference and exposition. Atlanta,
Georgia.
Simmons, D. R, & Lord, S. A. (2019). Removing Invisible Barriers and Changing Mindsets to Improve and Diversify
Simon, H. A. (1996). The sciences of the artificial. Cambridge, MA: MIT Press.
Solnit, R. (2003). River of shadows. Eadweard Muybridge and the Technological West. New York: Viking.
Standage, T. (1998). The Victorian Internet: The remarkable story of the telegraph and the nineteenth century’s online
Sticklen, J., Amey, M., Eskil, T., Hinds, T., & Urban-Lurain, M. (2004). Application of object-centered scaffolding to
introductory MATLAB. Paper presented at the Proceedings of the 2004 American Society for Engineering Education
Sweller, J. (1988). Cognitive load during problem solving: effects on learning. Cognitive Science, 12(2), 257–285.
Sweller, J. (1994). Cognitive load theory, learning difficulty, and instructional design. Learning and instruction, 4(4),
295–312.
Sweller, J., Ayres, P., & Kalyuga, S. (2011). Cognitive load theory. In J. M. Spector & S. P. Lajoie (Eds.), Explora-
tions in the Learning Sciences, Instructional Systems and Performance Technologies (Vol. 10, pp. 978-971). New
York: Springer.
Steele, W., & Hodge, B. (2001, June). Computational Paradigms in Undergraduate Mechanical Engineering Education.
In Proceedings of the 2001 American society for engineering education annual conference and exposition. Albuquerque,
New Mexico
Tair, M. M. A., & El-Halees, A. M. (2012). Mining educational data to improve students’ performance: a case study.
Toth, E. E., Morrow, B. L., & Ludvico, L. R. (2009). Designing blended inquiry learning in a laboratory context: A study
of incorporating hands-on and virtual laboratories. Innovative Higher Education, 33(5), 333–344.
Toto, R., Colledge, T., Frederick, D., & Pung, W. H. (2014). Instructional Strategies to Promote Student Strategic Thinking
Twenge, J. M. (2017). IGen: Why Today’s Super-connected Kids are Growing Up Less Rebellious, More Tolerant, Less
Happy–and Completely Unprepared for Adulthood--and what that Means for the Rest of Us. Simon and Schuster.
Urban-Lurain, M., Amey, M., Sticklen, J., Hinds, T., & T., E. (2004). Curricular Integration of Computational Tools by
Evolutionary Steps. Paper presented at the Proceedings of the 2004 American Society for Engineering Education Annual
Uribe, M. R., Magana, A. J., Bahk, J. H., & Shakouri, A. (2016). Computational simulations as virtual laboratories for online
engineering education: A case study in the field of thermoelectricity. Computer Applications in Engineering Education,
24(3), 428–442.
Vahid, F., Edgcomb, A., & Strawn, S. (2016). U.S. Patent Application No. 15/010,964.
van Joolingen, W.R. and T. de Jong. (2003). SimQuest, authoring educational simulations. In: T. Murray, S. Blessing, S.
Ainsworth: Authoring Tools for Advanced Technology Learning Environments: Toward cost-effective adaptive, interactive,
van Joolingen, W. R., de Jong, T., Lazonder, A. W., Savelsbergh, E. R., & Manlove, S. (2005). Co-Lab: research and
development of an online learning environment for collaborative scientific discovery learning. Computers in human
SPRING 2019 51
ADVANCES IN ENGINEERING EDUCATION
Venkatesh, V., & Davis, F. D. (2000). A theoretical extension of the technology acceptance model: Four longitudinal
Vieira, C., Goldstein, M. H., Purzer, S‚ ., & Magana, A. J. (2016). Using learning analytics to characterize student
experimentation strategies in the context of engineering design. Journal of Learning Analytics, 3(3).
Vieira, C., Magana, A.J., Falk, M.L. & Garcia, R.E. (2017). Writing in-code comments to self-explain in computational
science and engineering education. ACM Transactions on Computing Education (TOCE). 17(4). DOI:10.1145/3058751
Vieira, C., Magana, A.J., Garcia, R.E., Jana, A., & Krafcik, M. (2018). Integrating computational science tools into a
Thermodynamics course. Journal of Science Education and Technology (JOST), 27(4), 322–333.
Vieira, C., Magana, A. J., Roy, A., Falk, L. M., & Reese, J. M. (2016). Exploring undergraduate students’ computational
literacy in the context of problem solving. Computers in Education Journal, 7(1), 100–112.
Vieira C., Magana, A.J., Roy, A., & Falk M.L. (2019). Student explanations in the context of computational s cience
Vieira, C., Roy, A., Magana, A. J., Falk, L. M., & Reese, J. M. (2016). In-code comments as a self-explanation strategy
for computational science education. In Proceedings of the 2016 American society for engineering education annual
Vieira, C., Yan, J., & Magana, A. J. (2015). Exploring design characteristics of worked examples to support program-
ming and algorithm design. Journal of Computational Science Education., 6(1), 2–15.
Vincenti, W. G. (1990). What engineers know and how they know it: Analytical studies from aeronautical history.
Vinck, D. (2003). Everyday engineering: An ethnography of design and innovation. Cambridge, Massachusetts: MIT
Press.
West, M., & Silva , M., & Herman, G. L. (2015, June). Randomized Exams for Large STEM Courses Spread via Communities
of Practice. In Proceedings of the 2015 American society for engineering education annual conference and exposition.
Seattle, Washington.
Webel, C., Krupa, E., & McManus, J. (2015). Benny goes to college: Is the “Math Emporium” reinventing Individually
Wieman, C. E., Adams, W. K., & Perkins, K. K. (2008). PhET: Simulations that enhance learning. Science, 322(5902), 682–683.
Wiesner, T.F. and W. Lan. (2004). Comparison of Student Learning in Physical and Simulated Unit Operations Experiments.
Journal of Engineering Education. 93:195–204.
Wiggins, G., & McTighe, J. (2005). Understanding by design (2nd ed.). Alexandria, VA.: The Association for Supervision
Windschitl, M. & Calabrese Barton, A. (2016) Rigor and Equity By Design: Seeking a Core of Practices for the Science
Woolf, Beverly Park. A roadmap for education technology. Research report. <hal-00588291> (2010)
Worsley, M., & Blikstein, P. (2014). Analyzing Engineering Design through the Lens of Computation. Journal of Learn-
Xie, C., Zhang, Z., Nourian, S., Pallant, A., & Bailey, S. (2014). On the instructional sensitivity of CAD logs. International
Xie, C., Zhang, Z., Nourian, S., Pallant, A., & Hazzard, E. (2014). A time series analysis method for assessing engineering
design processes using a CAD tool. International Journal of Engineering Education, 30(1), 218–230.
Xie, C Visualizing Engineering Design Processes of High School Students Using a CAD System, AERA 2016, Washington
DC, USA
52 SPRING 2019
ADVANCES IN ENGINEERING EDUCATION
Yadav, S. K., & Pal, S. (2012). Data mining: A prediction for performance improvement of engineering students using
Zacharia, Z.C. (2007). Comparing and combining real and virtual experimentation: An effort to enhance students’
conceptual understanding of electric circuits. Journal of Computer Assisted Learning. 23, 120–132.
Zacharia, Z. C., & de Jong, T. (2014). The Effects on Students’ Conceptual Understanding of Electric Circuits of Introducing
Virtual Manipulatives Within a Physical Manipulatives-Oriented Curriculum. Cognition and instruction, 32(2), 101–158.
Zacharia, Z. C., & Olympiou, G. (2011). Physical versus virtual manipulative experimentation in physics learning. Learning
Zacharia, Z. C., Olympiou, G., & Papaevripidou, M. (2008). Effects of experimenting with physical and virtual manipu-
latives on students’ conceptual understanding in heat and temperature. Journal of Research in Science Teaching, 45(9),
1021–1035.
AUTHORS
State University. He received his B.S. and M.S. degrees from UC San
on what prevents students from being able to integrate and extend the knowledge developed in
specific courses in the core curriculum to the more complex, authentic problems and projects they
face as professionals. Dr. Koretsky is one of the founding members of the Center for Lifelong STEM
Education Research at OSU.
means of expert technological and computing tools such as cyber-physical systems, visualizations
SPRING 2019 53