Digital Leadership, Mohan 2021
Digital Leadership, Mohan 2021
1. Introduction
The pace of change in technology, the rising expectations of customers, and the
number of digital companies entering the market to address gaps in customer
expectations have all begun to influence the nature of the leadership needed in the
digital age. Large service organizations, such as those in health care, education,
and government, have started to recognize the impact of the digital age dynamic
on their customers (e.g., patients, students, and citizens). They have started to
explore opportunities to innovate to meet changing customer expectations. In
this chapter, our focus is on the expectations of students in the digital age and
the impact of these expectations on the strategies that higher educational institu-
tions need to employ by leveraging information and communication technologies
(ICT). Note that higher educational institutions also have other stakeholders,
such as teachers, employers, government, and alumni. ICT can be leveraged to
address the needs of these stakeholders as well, but the focus in this chapter is on
students as “consumers.”
Students’ expectations of technology use in educational settings are influenced
by their use of the same technologies when they purchase other products and
services in the commercial marketplace. They prefer to use digital tools, such as
smart phones and apps, to connect with other students, and they use social media
to communicate and share information. As newer technologies improve access to
information using search engines and provide services on demand using wireless
access to connected devices, the challenge for educational institutions is how best
to leverage student interest in technology to help support their reading, participat-
ing, learning, and working together as teams. With the volatility in student demand
for such on-demand services and the availability of options (e.g., online programs,
remote access to learning tools, learning management systems to support team
work, etc.), higher educational institutions will continue to have the following
challenge: How can they best fulfill their academic goals in the digital age?
2002), some of these activities are considered knowledge processes that will help
a learner remember, understand, apply, analyze, evaluate, and create as shown in
Fig. 1. Also, the cognitive processes in the early phases are to learn the knowledge
of chosen disciplines using many traditional techniques of recollection and appli-
cation of a number of simulated cases in class discussions.
David A. Kolb (Kolb, 1984; Smith, 2010), as part of Experiential Learning
theory, discusses four different steps to learning: concrete experience, abstract
conceptualization, reflective observation, and active experimentation (see
Fig. 2). All four of these steps are considered essential for learning. Learners
should have an opportunity to actively experiment (put their knowledge to work)
and get concrete experiences from case studies and projects, either simulated or
in the real world. Repeating such experiences should lead to reflective observa-
tion on what worked or did not work and why, and this will help leaners develop
a conceptualization of processes/practices that are critical to consider in future
experiments. Both academic institutions and business mentors can play a role in
supporting a learner’s reflective observation and abstraction, with the expectation
that learners will develop these skills over time and with experience.
Bloom’s taxonomy views the iterative process of Kolb as helping a learner
acquire four types of knowledge: factual, conceptual, procedural, and meta-
cognitive (as shown in Fig. 3). If a learner is “doing” various activities to acquire
and retain knowledge, then the knowledge acquired should enable a learner to
move from the factual to the meta-cognitive state. This will help learners deal
with VUCA environment and adapt their decision processes. The challenge for
academic institutions is how tightly academic preparation and experiential learn-
ing (i.e., learning by doing) should be combined to support deeper learning and
adaption. Such a preparation has to support a learner’s capability to
use analytics engines to learn and adapt the teaching pedagogy or administrative
mechanisms to support learning both inside and outside a classroom setting. The
rest of the section looks at digital service opportunities within each quadrant.
The next section (Section 4) focuses on one platform used to accumulate knowl-
edge in support of such adaptation.
While the early use of internships and projects with nearby companies has
been the practice to support second-order learning, students now have the abil-
ity to use virtual communication tools to connect with students far and wide to
engage in real world or self-developed projects and gain wider exposure to issues
and challenges. In addition, learning systems allow students from multiple insti-
tutions to work online on projects as teams and experience some of the cultural
and geographical challenges and explore ethical issues when political and social
considerations are brought into the picture.
Independent of the way technology is being used to support learning, the
goal of educational institutions is still to focus on the knowledge acquisition and
retention of the learner, moderated by the learner’s maturity level (e.g., early stage
undergraduate to postgraduate student or students with experience). These are
the learning activities along the x-axis of Fig. 4. However, there has been a signifi-
cant debate, intensifying recently, about learners’ capability to meet today’s chal-
lenges. Competitive and global markets and changing customer expectations are
challenging a learner’s capability to not only acquire relevant knowledge but also
adapt it to varying situations to be effective. Environmental uncertainties make it
difficult to decide which factors influence a decision and what knowledge about
these factors is needed to reduce this uncertainty. This calls for the use multiple
external stakeholders (employees, mentors, alumni, etc.) to provide experiential
learning opportunities for students to see or simulate the real world (activities
on the y-axis of Fig. 4). The goal in this section is to leverage technologies that
students, as customers in today’s marketplace, use to provide both basic as well as
other skills – critical thinking and experiential learning.
1
See Underhill and McDonald (2010).
Digital Leadership in Education 81
schemes provided by De Wever, Schellens, Valcke, and Van Keer (2006), there is
a wide variety of work related to content analysis of forum posts. The majority
explored the effectiveness of online forums as platforms for innovating the edu-
cational practices of teachers (Chávez, Montaño, & Barrera, 2016) facilitating
the teaching process (Brace-Govan, 2003), and providing indicators of student
learning that can assist in student assessment (Dennen, 2008; McKenzie & Mur-
phy, 2000; Premagowrie, Kalai, & Ho, 2014). The information provided teachers
and forum administrators with an indication of how to use these forums to sup-
port critical thinking skills and assess individual learning, as well as help manage
online courses and design learning strategies (Guzdial & Turns, 2000).
According to Hoogeveen, Wang, Baldwin, and Verspoor (2018), discussion forum
research can be divided into two main groups: community question-answering
(CQA) archives and discussion forums. Both groups promote community interaction
and information sharing. CQA archives are intended to assist people with problem-
solving and question answering (QA). As soon as someone posts a good answer to a
new question, the interaction is considered to be finished. Discussion forums, on the
other hand, are designed as a platform for discussion. The distinction between CQA
archives and discussion forums is not very clear, as some discussion forums also focus
on answering questions (for instance, Linux Questions – http://www.linuxquestions.
org/. Accessed on October 20, 2019), and specific CQA archives contain questions
that are indeed conversations (for instance, Yahoo! Answers – https://answers.yahoo.
com/. Accessed on October 20, 2019).
In summary, while forums capture knowledge that can be used to improve
student learning and tailor teaching to support this learning, the challenge is one
of having the tools to analyze this knowledge content quickly and effectively.
AI approaches based on NLP tools can offer value in this regard. The goal of
these tools is to make assimilation of the content and course tracking easier for
students and help teachers manage the learning platforms, especially in online
courses – MOOCs and those that call for collaborative learning (Abri, Jamoussi,
Kraiem, & Khanjari, 2017) – which use online forums. The rest of this section will
briefly discuss the role of an analytics engines in support of learning adaptation.
Our research focuses on the use of AI approaches based on NLP to manage ODL
learning platforms, and specifically online forums, by teachers and administra-
tors. Online forums have a high impact on knowledge transferred among students
(Baxter & Haycock, 2014; Mora, Ferrández, Gil, & Peral, 2017), but no effec-
tive solutions have been proposed to facilitate the automatic monitoring of these
online forums to turn them into a good tool for interaction and communication,
especially when excessive messaging occurs in a disorderly and unstructured man-
ner. The issues being addressed include finding appropriate answers in the forums
to questions (question and answer searches in CQA) and determining when to
create a new thread for discussion in the forum.
Some of the challenges the tool must consider include (a) the existence of irrel-
evant material; (b) data that contains a lot of noise making it difficult to retrieve
relevant material; (c) informal use of language; and (d) spelling and grammati-
cal errors. Also, there is a high computational cost involved in processing large
amount of data, and there are inefficiencies in small and specific data sets. Fur-
thermore, one must consider extracting the Key Performance Indicators (KPIs)
of educational institutions, such as participation, motivation, dropout rates,
and performance of students. In addition, it is necessary to identify similarities
between questions so that similar questions can be grouped and used to detect
common problems in students’ learning. This helps teachers and administrators
determine how well students are developing critical thinking skills and if learning
is actually taking place.
Using a set of learning analytics techniques based on NLP, a set of automatic
analysis tools are being designed to facilitate the monitoring of online learning
communities. The NLP techniques considered focused on lexical, syntactic, and
semantic analysis, ellipsis and anaphora resolution, information retrieval (IR),
QA, and clustering, as well as data management procedures used to infer the
KPIs.
The proposed general architecture presents two phases, offline and online. The
offline phase of processing the forums is carried out before the online execution.
The forums are pre-processed using the NLP techniques (lexical, syntactic and
semantic analysis; and ellipsis and anaphora resolution). After that, the process
of indexing is performed – similar to IR or QA systems – in which the infor-
mation in the collected forums is organized for easier access in the information
search process. Given the highly dynamic nature of online forums, the indexation
techniques must be robust, especially for incremental and update processes.
With regard to lexical-morphological analysis, it consists of carrying out the
Part-of-Speech (PoS) tagging of the text. The output is a set of pairs <word,
PoS tag>, where the PoS tag identifies the grammatical category (noun, verb,
adjective, adverb, conjunction, etc.) and the morphological information (singu-
lar, plural, masculine, feminine, etc.) of the word. FreeLing PoS tagger (Padró &
Stanilovsky, 2012) or TreeTagger (Schmid, 1994) tools could be used for this task.
The syntactic analysis stage consists of performing the parsing of the text.
Full or partial (shallow) parsing of the text can be carried out. In the proposed
architecture, the text is partially parsed to extract only noun phrases (NP), prep-
ositional phrases (PP), and verbal phrases. These phrases represent the “main
Digital Leadership in Education 85
concepts” involved in the text. The text chunks not included in these phrases are
skipped in the parsing. By contrast, NPs can have nested structures such as PPs,
appositions, or relative clauses; therefore, these phrases are fully parsed. Moreo-
ver, coordinated NPs and PPs are parsed. The objective is to extract the rep-
resentative concepts of each sentence and find out their syntactic function. We
have used the partial parser presented in the work of Ferrández, Palomar, and
Moreno (1999). Among other parsers, different parsing services offered by FreeL-
ing library (such as chart-based shallow parsing or statistical dependency pars-
ing) may be carried out in this stage.
The objective of the semantic analysis stage is to enrich the text with semantic
information (in addition to the lexical and syntactic information). It is obtained
from additional semantic resources, such as WordNet (http://wordnetweb.prince-
ton.edu/perl/webwn. Accessed on January 4, 2020). Each noun, verb, adjective,
and adverb is labeled with its synset (number that identifies a set of synonyms
that represent the same semantic concept) and its type. For instance, the first
sense of plant (plant #1) has the synset number 03963198, belongs to the type
noun.artifact and expresses “buildings for carrying on industrial labor.” In this
way, semantic comparisons between different concepts can be done using the syn-
onymy or hyponym/hypernym chains. Furthermore, Word Sense Disambiguation
techniques must be applied to identify the meaning of words in context: the cor-
rect word synset according to the surrounding words (Navigli, 2009).
The final stage of the offline indexation process is linguistic phenomena resolu-
tion (such as ellipsis or anaphora resolution) to resolve the referential ambiguity
of the text. For this purpose, the anaphoric expressions and elided elements are
resolved and replaced by the entities to which they refer. With linguistic phenom-
ena resolution, the comprehension and coherence of the text are improved when
the omitted elements are replaced by the entities/concepts to which they refer.
The output of the offline phase is the meaning data structure, with lexical, syn-
tactic, and semantic information that unambiguously (including the resolution
of linguistic phenomena) represents the meaning of the text. It will be used for
communication between the different modules of the architecture.
The online process refers to the interaction with the user. The system offers
three main functionalities: navigation, information search, and KPI extraction.
The navigating function consists in being able to browse automatically the struc-
ture of the forum generated by the application. Clustering techniques could be
used (Mora et al., 2017). With regard to information search for questions and
answers, the traditional format of a general-purpose search engine (QA system)
is used, in which the user makes a request for information, and the system pre-
sents the most relevant results. These results may include questions similar to user
request (for a broader generation of the request) or a most likely answer to the
specific user request or the generalized request.
Finally, the functionality of KPI inference will report different KPIs extracted
from the forums, which is especially useful for teachers and administrators of
the ODL platform. The results will be presented through dashboards. Indicators
that express participation, motivation, achievement, and the student drop-out
rate are very useful for managing online courses and designing learning strategies.
86 Mohan Tanniru and Jesús Peral
The method presented in the work of Peral, Maté, and Marco (2017) for the
identification of relevant KPIs, which has been evaluated in an educational con-
text, could be used for this module. The authors presented in their case study
identified indicators obtained in a generic way, which may be applicable to other
online courses: (1) increment in the number of students; (2) dropout ratio; (3)
student recovery ratio; (4) percent of active students; (5) percent of students who
fail the course; (6) percent of students passing the exams without seeing the cor-
responding lessons; and (7) percent of students taking the course in a continuous
or sequential pattern.
5. Conclusions
The increasing use of ICT in higher education institutions (HEI) calls for digital
leadership, whereby new innovative solutions to support student learning in the
digital age are explored using advanced technologies and evaluated to support
value creation. Knowledge extracted from student–teacher interactions is then
used to analyze gaps using other technologies, such as AI and NLP, to improve
learning outcomes by adapting various pedagogical activities. Just as other
organizations have learned, digital leadership in higher educational institutions
has to develop digital services, that is, service innovations that leverage advanced
technologies, to engage students in their learning process and analyze their learn-
ing to improve student performance.
The agility with which organizations need to adapt to changing customer
expectations while at the same time maintaining stability to create and sustain
value in today’s fast changing environment creates complexity. Complex adap-
tive systems have learned to use multiple leadership processes (administrative,
enabling, and adaptive) to move between stability and instability or operate in a
state of bounded instability (Uhl-Bien, Marion, & McKelvey, 2007). While most
service provider ecosystems (i.e., business and service organizations) use external
partners and customers to learn about value in context in the customer ecosystem
to create and sustain value, the HEI have a unique opportunity. The customers
of these HEI (students) spend significant amount of time in the provider (HEI)
ecosystem, thus allowing the HEI to use multiple technology platforms (e.g.,
online discussion platforms) to engage students in creating value and learn about
value-in-context to revise the HEI processes. As discussed in the chapter, analytics
engines and AI/NLP technologies can be used to support learning by the HEIs as
they continue to adapt their pedagogical innovations to support student learning
in the digital age.
Some areas for future research include tailoring the type of analysis needed to
address the type of learning students need to engage in. The knowledge frame-
work proposed identifies four different types of learning: first- and second-order
learning and academic and experiential learning to support both reflective learn-
ing and capability for abstraction. Depending on where students are on their edu-
cational ladder – undergraduate to graduate/postgraduate, full-time or part-time
working professional – digital services designed and explored can be tailored to
address their learning needs in each of these four quadrants. Also, by learning
Digital Leadership in Education 87
about the speed with which the VUCA environment influences the disciplines
(e.g., science and engineering fields vis-á-vis business and social sciences), the
technologies used to design digital services that engage students may vary both in
the frequency of engagement and the richness of interactions to seek knowledge.
In conclusion, HEIs have an opportunity to exhibit digital leadership to inno-
vate, explore, and evaluate new methods to engage students in their learning, not
only to teach the basic knowledge of the academic discipline but also allow them
to contextualize this knowledge as they prepare to enter the VUCA world to solve
problems and professionally grow in their chosen disciplines.
References
Abri, A. A., Jamoussi, Y., Kraiem, N., & Khanjari. Z. A. (2017). Comprehensive classification
of collaboration approaches in E-learning. Telematics and Informatics, 34(6), 878–893.
Armstrong, J. S. (2010, April 16). Natural learning in higher education. Available at SSRN:
https://ssrn.com/abstract=1928831 or http://dx.doi.org/10.2139/ssrn.1928831
Astin, A. W., Vogelgesang, L. J., Ikeda, E. K., & Yee, J. A. (2000). How service learning
affects students. Los Angeles, CA: Higher Education Research Institute, UCLA.
Baxter, J. A., & Haycock, J. (2014). Roles and student identities in online large course
forums: Implications for practice. The International Review of Research in Open and
Distributed Learning, 15(1), 20–40.
Brace-Govan, J. (2003). A method to track discussion forum activity: The Moderators’
assessment matrix. The Internet and Higher Education, 6(4), 303–325.
Cavallaro, F., & Tan, K. (2006). Computer-mediated peer-to-peer mentoring. AACE
Journal, 14(2), 129–138.
Chávez J., Montaño, R., & Barrera, R. (2016). Structure and content of messages in
an online environment: An approach from participation. Computers in Human
Behavior, 54, 560–568.
De Azevedo, V. L. L., & Borges, M. (2015). More collaboration, more collective intelligence.
International Journal of Knowledge Society Research (IJKSR), 6(3), 1–18.
De Laat M., Lally, V., Lipponen, L., & Simon, R. J. (2007). Investigating patterns of
interaction in networked learning and computer-supported collaborative learning:
A role for social network analysis. Journal of Computer Supported Collaborative
Learning, 2(1), 87–103.
De Wever, B., Schellens, T., Valcke, M., & Van Keer, H. (2006). Content analysis schemes to
analyze transcripts of online asynchronous discussion groups: A review. Computers
& Education, 46(1), 6–28.
Dennen, V. P. (2008). Looking for evidence in learning: Assessment and analysis methods
for online discourse. Computers in Human Behavior, 24(2), 205–219.
Dioso-Henson, L. (2012). The effect of reciprocal peer tutoring and non-reciprocal peer
tutoring on the performance of students in college physics. Research in Education,
87(1), 34–49.
Eyler, J., & Giles, D. E., Jr. (1999). Where’s the learning in service-learning. Jossey-Bass
Higher and Adult Education Series. San Francisco, CA: Jossey-Bass, Inc.
Ferrández, A., Palomar, M., & Moreno, L. (1999). An empirical approach to Spanish
anaphora resolution. Machine Translation, 14(3/4), 191–216.
Glass, N., & Walter, R. (2000). An experience of peer mentoring with student nurses:
Enhancement of personal and professional growth. Journal of Nursing Education,
39(4), 155–160.
88 Mohan Tanniru and Jesús Peral
Padró, L., & Stanilovsky, E. (2012). FreeLing 3.0: Towards wider multilinguality. In
Proceedings of the language resources and evaluation conference (pp. 2473–2479).
Istanbul.
Peral, J., Ferrández, A., Mora, H., Gil, D., & Kauffmann, E. (2019). A review of the
analytics techniques for an efficient management of online forums: An architecture
proposal. IEEE Access, 7, 12220–12240.
Peral, J., Maté, A., & Marco, M. (2017). Application of data mining techniques to identify
relevant key performance indicators. Computer Standards & Interfaces, 54, 76–85.
Premagowrie, S., Kalai, R. V., & Ho, R. C. (2014). Online forum: A platform that affects
students’ learning. American International Journal of Social Science, 3(7), 107–116.
Ramesh, A., Goldwasser, D., Huang, B., Daume, H., & Getoor, L. (2014). Understanding
MOOC discussion forums using seeded LDA. In Proceedings of the ACL workshop
on innovative use of NLP for building educational applications (pp. 28–33). Baltimore,
Maryland, USA.
Rodger, S., & Tremlbay, P. F. (2003). The effects of a peer mentoring program on academic
success among first year university students. The Canadian Journal of Higher
Education, 33(3), 1–18.
Romero, C., López, M. I., Luna, J. M., & Ventura, S. (2013). Predicting students’ final
performance from participation in on-line discussion forums. Computers &
Education, 68, 458–472.
Sanchez, R. J., Bauer, T. N., & Paronto, M. E. (2006). Peer-mentoring freshmen:
Implications for satisfaction, commitment and retention to graduation. Academy of
Management Learning & Education, 5(1), 25–37.
Schmid, H. (1994). Probabilistic part-of-speech tagging using decision trees. In Proceedings
of the conference on new methods in language processing, (pp. 154–164). Manchester.
Shea, P., & Bidjerano, T. (2009). Community of inquiry as a theoretical framework to foster
“epistemic engagement” and “cognitive presence” in online education. Computers &
Education, 52(3), 543–553.
Smith, M. K. (2010). David A. Kolb on experiential learning. The Encyclopedia of Informal
Education. Retrieved from http://infed.org/mobi/david-a-kolb-on-experiential-
learning/. Accessed on November 7, 2019.
Suh, H. J., & Lee, S. W. (2006). Collaborative learning agent for promoting group
interaction. ETRI Journal, 28(4), 461–474.
Swan, K., Shea, P., Fredericksen, E., Pickett, A., Pelz, W., & Maher, G. (2000). Building
knowledge building communities: Consistency, contact and communication in the
virtual classroom. Journal of Educational Computing Research, 23(4), 359–383.
Tanniru, M., & Agarwal, R. (2002). Applied technology in business (ATiB) program: an
industry-academia partnership to support student learning. E-Service, 1(2), 5–23.
Tanniru, M, Eveslage, B., Ellis, D., & Turaka, N. (2013). Peer-to-peer mentoring with a
global reach. In Proceedings of the 6th annual mentoring conference: Impact and
effectiveness of developmental relationships: Local to global mementorship (pp. 1–12).
Albuquerque, NM.
Tanniru. M., Khuntia, J., & Weiner, J. (2018). Hospital leadership in support of digital
transformation. Pacific Asia Journal of Association of Information Systems, 10(3),
1–24.
Uhl-Bien, M., Marion, R., & McKelvey, B. (2007). Complexity leadership theory: Shifting
leadership from the industrial age to the knowledge era. Leadership Quarterly, 18(4),
298–318.
Underhill, J., & McDonald, J. (2010). Collaborative tutor development: Enabling a
transformative paradigm in a South African university. Mentoring & Tutoring:
Partnership in Learning, 18(2), 91–106.
UNESCO. (2002). Open and distance learning: Trends, policy and strategy considerations.
Paris: UNESCO.
90 Mohan Tanniru and Jesús Peral
Vargo, S. L., & Lusch, R. F. (2008). Why “service”? Journal of the Academy of Marketing
Science, 36(1), 25–38.
Vygotsky, L. S. (1986). Thought and language. Cambridge, MA: MIT Press.
Wells, G. (2007). Semiotic mediation, dialogue and the construction of knowledge. Human
Development, 50(5), 244–274.
Wink, J., & Putney, L. A. (2002). A vision of Vygotsky. Boston, MA: Allyn & Bacon,.
Yang, D., Sinha, T., Adamson, D., & Rosé, C. P. (2013). Turn on, tune in, drop out:
Anticipating student dropouts in massive open online courses. In Proceedings of the
NIPS data-driven education workshop (Vol. 11, p. 14). Lake Tahoe, Nevada.
Digital Leadership in Education 91