You are currently viewing a new version of our website. To view the old version click .
Sensors
  • Communication
  • Open Access

9 August 2011

Smart Learning Services Based on Smart Cloud Computing

,
and
Department of Multimedia Science, Sookmyung Women’s University, Chungpa-Dong 2-Ga, Yongsan-Gu 140-742, Seoul, Korea
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue Selected Papers from FGIT 2010

Abstract

Context-aware technologies can make e-learning services smarter and more efficient since context-aware services are based on the user’s behavior. To add those technologies into existing e-learning services, a service architecture model is needed to transform the existing e-learning environment, which is situation-aware, into the environment that understands context as well. The context-awareness in e-learning may include the awareness of user profile and terminal context. In this paper, we propose a new notion of service that provides context-awareness to smart learning content in a cloud computing environment. We suggest the elastic four smarts (E4S)—smart pull, smart prospect, smart content, and smart push—concept to the cloud services so smart learning services are possible. The E4S focuses on meeting the users’ needs by collecting and analyzing users’ behavior, prospecting future services, building corresponding contents, and delivering the contents through cloud computing environment. Users’ behavior can be collected through mobile devices such as smart phones that have built-in sensors. As results, the proposed smart e-learning model in cloud computing environment provides personalized and customized learning services to its users.

1. Introduction

Thanks to the advances in information technologies and high-speed networking, the e-learning environment can offer a new paradigm of learning to learners. Traditionally e-learning offers teaching and learning by wired computers and in a lecture-style classroom setup only. Even though learners were able to browse and download information anytime and anywhere through the existing e-learning platform, they were limited to wired classroom setups.

Smart learning (s-learning) is an important and new paradigm of learning today. The concept of s-learning plays an important role in the creation of an efficient learning environment that offers personalized contents and easy adaptation to current education model. It also provides learners with a convenient communication environment and rich resources. However, the existing-learning infrastructure is still not complete. For example, it does not allocate necessary computing resources for s-learning system dynamically [1,2]. Currently, the majority of s-learning systems have difficulty in interfacing and sharing data with other systems, i.e., it falls short of systematic arrangement, digestion and absorption of the learning contents in other systems. This may lead to duplication in creating teaching resources and low utilization of existing resources. To resolve this problem, it is recommended to use cloud computing to support resource management.

The cloud computing environment provides the necessary foundation for the integration of platform and technology. It integrates teaching and research resources distributed over various locations by utilizing existing conditions as much as possible to meet the demands of the teaching and research activities. The cloud computing environment with respect to s-learning offers new ideas and solutions in achieving interoperability among heterogeneous resources and systems. The cloud services mean that the Internet can be used as huge workspace, repository, platform, and infrastructure. Learners can access to the Internet from anywhere at anytime, using widely spread mobile devices but the existing cloud computing technologies are only passively responsive to users’ needs. This situation necessitates proactive cloud services rather than passive services. Since learners typically carry mobile devices of some kind at their hands, the volume of information and services processed through the devices continues to increase.

In this paper, we propose a smart cloud computing (SCC) model for smart learning contents through the E4S—smart pull, smart prospect, smart content and smart push—based on the user behavior acquired by the sensors in the users’ mobile devices. As results, SCC can provide customized contents to each user. Hence users are able to receive customized contents or services they want, without explicit searches. In the subsequent chapters, we briefly describe existing cloud computing services and introduce the proposed context-aware cloud computing environment. Then, we explain the smart cloud service and formally define SCC (Smart Cloud Computing) in detail.

3. Smart Cloud Computing

The Smart Cloud Computing (SCC for short) based on elastic computing for 4S model has the capability to provide a smart learning environment. It encourages learning system standardization and provides a means for managing it. A traditional e-learning system can display single content on a single device or multiple contents on one device. The SCC can deliver s-learning to the users so they can use multiple devices to render multi learning contents. The multi learning contents can be played in different devices separately to form a “virtual class”. For this, the SCC uses context-aware sensing. Sensing through the location and IP address of each device, it can orchestrate all devices. The architecture of the model is shown in the Figure 2.

Figure 2. Smart Cloud Computing architecture.

Figure 2 shows how the SCC provides smart learning to the user. It is using Elastic 4S based on information obtained from the user. The information of the user includes the information about the user and the device, received by context-aware sensors. Context-aware monitoring monitors user request(s) and the kind of device(s) that the user is currently using. By using the information collected by the sensors, SCC can provide user-aware service based on Elastic 4S. Elastic 4S is performed through an intelligent learning engine that consists of the rules based on the four services—Smart Pull, Smart Prospect, Smart Content and Smart Push. These four smart services provide high quality services according to the definition of E4S that is described as follows:Definition 1: The SCC model for an s-learning service is defined as a set of four tuples.

{ E 4 S i } = { ( Spul l i , Spro s i , Sco n i , Spus h i ) } , 1 i N
where:
  • Spulli: Smart pull—analyze the extractable content from the sensing information.

  • Sprosi: Smart prospect—description of the content for target devices and delivery time.

  • Sconi: Smart content—connection establishment between server and target devices.

  • Spushi: Smart push—synchronized delivery of contents to target devices.

As shown in the definition, the E4S pulls the sensing data and analyzes the extractable contents. The context-aware module is functioning as an information filter that extracts only the intended information from the sensing data. There can be multiple contexts in sensing data depending on the services available in the learning management system. They are individually synchronized and pushed by the smart learning service. The sensing data analysis process identifies the different backgrounds of each learner and accommodates each learner’s needs individually. Learning contents are customized based on the background and learning needs. It means that different learning content may have unique technical and functional characteristics and may use different communication channels. Each customized learning contents may differ in modality (i.e., text-based, audio, video, etc.), capability (i.e., bandwidth), and timing (i.e., types of synchronization).

4. The SCC

4.1. Context-Aware Module

To provide the smart learning service to each individual user, the context-aware module must automatically deduce the actual situation from the user’s behavior. The context-aware module as shown in the Table 1 presents the context model based on a hybrid situation that consists of the user situation and the physical situation. The context model includes static factors and dynamic factors that describe the hybrid situation. The context model deals with the context objects and the relations among them. Since the context-aware module considers the characteristics of each user individually, such as learners’ knowledge interests, needs, expertise, and experiences, it can provide highly customized and relevant learning services to each user.

Table 1. Context model based on Hybrid Situation.

The user situation contains the detailed information about users. The user preferences in the user situation specify user actions and required services. A user action indicates some preference in user’s requests for learning services. The required service should help users acquire knowledge in the area of interest, share experience, and collaborate with each other in learning. Each user’s personal information such as personal context is secured by some security setting such as user’s schedule and location.

The physical situation includes each terminal’s MAC address, capability, software interface status, and types of software applications. The terminal capability describes the process speed, memory, screen size, resolution, and interface types. The terminal application type describes software applications installed in the terminal. The application type is based on quality of service (QoS) parameters, such as response time, delay, jitters and bandwidth. It can be categorized into four types, namely (1) conversational service, such as VoIP; (2) real-time (RT) service such as Internet Protocol Television (IPTV) and mobile TV; (3) non-real-time (NRT) services such as email or ftp; (4) interactive services such as web browsing. The context model has user situation information such as user’s requests and the devices they are using. Using this information, the SCC can provide user-aware smart learning service based on E4S. The E4S handles the pulling of sensing information, the analysis of context from the pulled information, the generation of smart content, and the push of smart learning service to individual terminals with different contexts.

4.2. The Elastic 4S (E4S) System

In this section, we discuss the E4S—Smart Pull, Smart Prospect, Smart Content, and Smart Push—in detail.

4.2.1. Smart Pull

The Smart Pull is the process of extracting learning content in the fusion learning DB through a variety of sensors from the user’s device including user situation and technical situation information based on the context model. The context model is used to filter information from the sensing information. The filtered information is analyzed to determine user’s behavior patterns.

The Smart Pull identifies a right action service in fusion learning DB based on the user action in context model. The fusion learning DB consists of various multimedia learning materials such as video, text, PPT and image scattered in different learning DB. For example, if a user action in the context model requests the topic of “multimedia”, the Smart Pull extracts a related action of “multimedia” in fusion learning DB. The fusion contents have some information for teaching with one action scenario (ActionNo) as shown in the Figure 3.

Figure 3. ActionNo in Fusion learning DB.

Figure 4 shows a Smart Pull method for extracting an action scenario (ActionNo) based on a user request action in the context model. For this, the Smart Pull checks the user situation and technical situation information to obtain a correct set of values of the requested learning. The set consists of the user’s learning request action, learning interests, needs, and the application type currently in use. The Smart Pull causes a matching search of learning object in the Fusion learning DB. The matching search uses two conditions to find the right ActionNo in the DB. First, the ActionNo object is the mapping to user action object. Second, the ActionNo content type is the mapping to the running application type in the technical situation. It may occur when a user uses multiple devices. If an ActionNo satisfies both conditions, the ActionNo is sent to the next step for synchronization.

Figure 4. Matching search of Smart Pull.

4.2.2. Smart Prospect

The Smart Prospect is mainly responsible for describing the contents in ActionNo—time, memory, resolution and supported application types. The description is needed for fusion content delivery, because ActionNo specifies the fusion learning content in the Fusion learning DB. For the delivery of fusion content to a user’s device, the SCC is required for harmony adaptation. In the harmony adaptation, the most important part is synchronization. The synchronization part controls the time for synchronization among fusion contents in the same ActionNo. To access the information such as memory, resolution, or application type of the contents in ActionNo, the Smart Prospect uses a Semantic Description using of UVA (Universal Video Adaptation) model that has been developed by Yoon [13]. The UVA model uses the video content description in MPEG-7 standard and MPEG-21 multimedia framework.

The Semantic Description based on the UVA model includes the effort to build a new architecture that supports content with formal semantics. The semantic description provides the accurate and meaningful information for the fusion content. The semantic description uses XML, ontology and Resource Description Framework (RDF) that help define fusion content clearly and precisely. It also represents systematic information about the contents. The role of the ontology is to formally describe the shared meaning of vocabulary used. The ontology describes the basic fusion learning contents of some domain where learning takes place (e.g., history of science). It includes the relations between these concepts and some basic properties. Based on the ontology, all learning content in the ActionNo are associated each other. For example, the description of the video content used in semantic description can be related to the scenes of video as shown in Figure 5.

Figure 5. Video analysis based Semantic Description.

Figure 5 shows a result of a detailed description of video content in ActionNo based on the UVA semantic description model. The Semantic description provides the detail information about video content for the Smart Prospect. The detailed description consists of time, memory, resolution and supported application types for each content. The time information has two types—<All time> and <time offset>. The <All time> represents playback time of the entire video content. The <time> represents the playback time of a shot in the video content. Based on the two types of time information, the next step—Smart Content—performs a synchronization process that utilizes three elements—<type>, <memory>, and <resolution>. The <type> indicates supporting types for some application, such as video, image, PPT or text. The <memory> and <resolution> specifies memory and resolution information of video content, respectively. The process of synchronization between device and content is accomplished by using the <type, memory, resolution> information.

4.2.3. Smart Content

The Smart Content generates the fusion content for the user’s device using the harmony adaptation. The harmony adaptation has two steps—Fusion Content Adaptation and Device Synchronization process. The Fusion Content Adaptation presents the synchronization among the fusion contents in ActionNo. The Device Synchronization performs the process of synchronization between devices. For the synchronization of the fusion contents, the Fusion Content Adaptation (FCA) uses the contents that are indicated by Semantic description from Smart Prospect. The adapted contents include <All time {start, duration, delay, end}> and <time {start, duration, delay, end}> information. For the time synchronization, the FCA uses Interpreter Playout Schedule (IPS) to schedule the order of playout. The IPS uses the duration time of the adapted contents with consideration of the synchronization among connected devices. The simulation results for sample IPS are illustrated in Figure 6.

Figure 6. Synchronization of fusion contents.

According to the IPS, every media stream can be modeled as an agent, or alternatively, any number of the same media type can present a scenario sequence that controls different media streams at different times.

For the process of synchronization between multiple devices, the Device Synchronization creates a channel for each device based on the technical situation in the context model. Each channel includes the unique information for the associated device. The major role of channel management is the synchronization of the channels to communicate with each other when multiple devices are playing. It means that the fusion contents can be played on multi devices at the same time, so it is required to have the coordination among them. The Device Synchronization uses SyncML (Synchronization Markup Language) to set the synchronization between devices. The SyncML is an international standard language for matching data between different devices and applications at any network company (network Ltd.). Through the synchronization between devices, the adapted contents can be used in Smart Push step.

4.2.4. Smart Push

The last step of E4S is the Smart Push process. The Smart Push delivers the smart learning service for user after the fusion contents are adapted and device and content are synchronized. As for the content delivery, the situation analyzer will be used. The situation analysis links the contents and devices so learners can use the learning contents at their device. The situation analyzer uses the physical situation information to find the related details of the contents and devices. The related details for the contents are time, memory, resolution and application types for the contents. The related details for device are process speed, memory, screen size, resolution and supported interface types for the terminal. The details of terminal indicate the information about available resources in the current device. If the details of a device and contents are of the same or compatible type, a link can be established and the contents are delivered to the device.

After a device and contents are synchronized by the situation analyzer, the Smart Push delivers the smart learning service to its users. The Smart Push delivers a complete set of smart contents to the user using terminal’s AP (Access Point) and MAC-Address in the context model of physical situation. The access point (AP) recognizes the physical location area of mobile devices that enter to the location range of AP. The AP assigns an IP-address to a mobile device based on its MAC-Address. The mobile device is connected to the Internet via AP, so we can get the physical location information when a mobile device gets its IP-address. The AP usually connects to a router (via a wired network), and can relay data between the wireless devices (such as computers or printers) and wired devices on the network. The Table 2 shows the context model of physical situation.

Table 2. Context model of physical situation.

Table 2 shows the information of the access point (AP) based on the user’s location. According to the user’s location, the AP handles multiple devices that are within the range of the AP. An access point connected directly to a wired LAN provides a connection point for wireless users. If more than one access point is connected to the LAN, users can roam from one area of a facility to another without losing their connection to the network. As users move out of range of one AP, they automatically connect to the network (associate) through another AP. The roaming process is seamless and transparent to the user. The Figure 7 shows an access point in an all-wireless network.

Figure 7. AP as central unit in All-Wireless Network.

The AP’s generic information describes the status of the network connection, such as connected or disconnected. As shown in the Table 2, we can know a position of the user from the location area of AP-1 and AP-2. It is because APs configures a MAC (Medium Access Control)-Address and IP-address within its location range. For example of Table 2, the three user devices—MAC addresses are 00:11:93:0d:a7:f5, 00:11:93:8d:f1:50 and 00:11:93:25:t4:bb—are located in physical range of AP-1. However, the AP 2 recognized one device with one MAC-Address. This means that mobile device (MAC-Address is 00:1b:93:e8:c4:1f) did not connected since the mobile device located out-of-range for AP 2. To conserve and effectively manage the resource of the devices, the AP performs the AP Controller. The AP Controller manages the MAC-Addresses for transmitting the learning contents to the linked devices according to the context model.

5. Implementation

As a part of the proof of the proposed concept, we have implemented the fusion media generator. The Figure 8 shows screen shots for fusion content generation and synchronization based on the fusion content adaptation step for an s-learning service. When a list of synchronized fusion-media is uploaded, they can be played in the preview.

Figure 8. Synchronization of Fusion contents.

The left side of the Figure 8 shows a screen shot of the audio and video synchronizer. The audio and video should have the same duration schedule—start, duration, delay and end time. After synchronization of the video and audio contents is completed, other fusion media (PPT, image and text) may be synchronized according to the duration schedule. The right side of the figure shows the action of synchronized learning contents based on the IPS. The graph in the figure shows the progress of the synchronization. The upper left part shows a synchronization of the video and audio. The part below shows the synchronization with other fusion media. The synchronized fusion media can be played in the preview. In the bottom side of the figure shows the start and duration time of the schedule for the fusion media that are displayed on the screen. Currently the figure shows four fusion media—video, audio, PPT and text. Accordingly, the time schedule of the fusion media is displayed in four lines.

To transfer the fusion media to the multiple devices, the SCC uses the Smart Push process. This process identifies the type of contents being transmitted, determines the order of media, and synchronizes media streams from different sources. The Smart Push may be applied more efficiently in many other multimedia applications, such as video conferencing.

In the implementation, we have delivered the synchronized fusion media to two different devices that have different MAC-Addresses—00:11:93:0d:a7:f5 and 00:11:93:8d:f1:50. The AP controller transmits the fusion media to the devices, as shown in Figure 9.

Figure 9. Fusion media being transferred to multiple devices.

6. Conclusions

Modern learning services typically deal with multi-media resources such as graphics, video, images, text etc., since such resources provide an efficient learning environment that helps learners understand the topic of interest better. The awareness of user behavior in the learning process can be very helpful in providing the right contents at the right time. The learning services that include the concept of such awareness and the capability of handling multi-media resources efficiently can be termed smart learning systems. In this paper, we have introduced the use of context-awareness for user behavior and a way to deliver the corresponding contents to the users.

The concept of the context model in context-awareness was introduced, which includes the static and dynamic descriptions of the user and physical situation. The context model deals with the context objects and the relations among them. The results of the context-awareness allow learning efficiency and outcomes for smart learning, such as learners’ knowledge interests, needs, expertise, and experiences. Using the context model, the smart cloud model (SCC) can provide the necessary contents to users precisely. In order to collect user’s behavior, sensors in users’ devices were used. Based on the sensing information, the cloud computing environment can forecast and prepare the contents by analyzing the collected information. Such a process enables smart learning services to provide the contents to the users at an appropriate time.

In order to deliver such customized contents to the users at right time, the SCC followed Elastic 4S—Smart-Pull, Smart-Push, Smart-Prospect and Smart-Contents. All the services are based on the collected data through the sensors in user’s device. We have utilized the E4S model and analyzed the sensed information within the category of context. The context-aware model handles the fusion media adaptation, synchronization, and transmission for a smart learning service. We have considered various requirements that for the users, the networks, and the cloud. As a future work, the protocols for smart cloud computing and domain specific ontology will be investigated.

Acknowledgments

This Research was supported by the Sookmyung Women’s University Research Grants 2010.

References

  1. Uden, L; Damiani, E. The future of E-learning: E-learning ecosystem. Proceedings of the First IEEE International Conference on Digital Ecosystems and Technologies, Cairns, Australia, 21–23 February 2007; pp. 113–117.
  2. Chang, V; Guetl, C. E-learning ecosystem (ELES)—A holistic approach for the development of more effective learning environment for small-and-medium sized enterprises. Proceedings of the First IEEE International Conference on Digital Ecosystems and Technologies, Cairns, Australia, 21–23 February 2007; pp. 420–425.
  3. Yoon, Y; Kim, S. Mobile OS; Korea Institute of Information Scientists and Engineers: Seoul, Korea, 2010; Volume 28, pp. 79–87. [Google Scholar]
  4. Yoon, Y; Kim, S. Mobile Cloud Computing Technology; National IT Industry Promotion Agency (NIPA): Seoul, Korea, 2010; Volume 1439, pp. 28–39. [Google Scholar]
  5. Perez, S. Why Cloud Computing is the Future of Mobile, Available online: http://www.readwriteweb.com/archives/why_cloud_computing_is_the_future_of_mobile.php/ (accessed on 19 July 2011).
  6. Schilit, B; Theimer, M. Disseminating active map information to mobile hosts. IEEE Netw 1994, 8, 22–32. [Google Scholar]
  7. Schmidt, A; Aidoo, KA; Takaluoma, A; Tuomela, U; Laerhoven, KV; de Velde, WV. Advanced interaction in context. Proceedings of the 1st International Symposium on Handheld and Ubiquitous Computing Table of Contents, Karlsruhe, Germany, 27–29 September 1999; 1707, pp. 89–101.
  8. Casquero, O; Portillo, J; Ramón Ovelar, R; Romo, J; Benito, M. iGoogle and gadgets as a platform for integrating institutional and external services. MUPPLE 2008, 8, 37–41. [Google Scholar]
  9. Facebook Platform, Available online: http://developers.facebook.com/ (accessed on 19 July 2011).
  10. Sedayao, J. Implementing and operating an internet scale distributed application using service oriented architecture principles and cloud computing infrastructure. Proceedings of the 10th International Conference on Information Integration and Web-Based Applications & Services, iiWAS2008, Linz, Austria, 24–26 November 2008; pp. 417–421.
  11. Zheng, YL; Li, LY; Ogata, H; Yano, Y. Toward awareness services in CSCL: Classifications and contextual providing mechanism. Proceedings of ICCE 2004, Melbourne, Australia, 30 November–30 December 2004; pp. 803–810.
  12. Chen, C-M; Li, Y-L; Chen, M-C. Personalized context-aware ubiquitous learning system for supporting effectively English vocabulary learning. Proceedings of the Seventh IEEE International Conference on Advanced Learning Technologies, Niigata, Japan, 18–20 July 2007.
  13. Yoon, Y; Kim, S; Lee, J. Universal video adaptation for content delivery in ubiquitous computing. Proceedings of the 8th Asia-Pacific Conference on Computer Human Interface (APCHI-LNCS), Seoul, Korea, 6–9 July 2008; 5068, pp. 193–202.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.