Search results for: BIM service based cloud computing
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 30236

Search results for: BIM service based cloud computing

29486 Using Lean Six-Sigma in the Improvement of Service Quality at Aviation Industry: Case Study at the Departure Area in KKIA

Authors: Tareq Al Muhareb, Jasper Graham-Jones

Abstract:

The service quality is a significant element in aviation industry especially in the international airports. Through this paper, the researchers built a model based on Lean six sigma methodologies and applied it in the departure area at KKIA (King Khalid International Airport) in order to assess it. This model characterized with many special features that can become over the cultural differences in aviation industry since it is considered the most critical circumstance in this field. Applying the model of this study is depending on following the DMAIC procedure systemized in lean thinking aspects. This model of Lean-six-sigma as a managerial procedure is mostly focused on the change management culture that requires high level of planning, organizing, modifying, and controlling in order to benefit from strengths as well as revoke weaknesses.

Keywords: lean-six-sigma, service quality, aviation industry, KKIA (King Khalid International Airport), SERVQUAL

Procedia PDF Downloads 405
29485 Continuous Functions Modeling with Artificial Neural Network: An Improvement Technique to Feed the Input-Output Mapping

Authors: A. Belayadi, A. Mougari, L. Ait-Gougam, F. Mekideche-Chafa

Abstract:

The artificial neural network is one of the interesting techniques that have been advantageously used to deal with modeling problems. In this study, the computing with artificial neural network (CANN) is proposed. The model is applied to modulate the information processing of one-dimensional task. We aim to integrate a new method which is based on a new coding approach of generating the input-output mapping. The latter is based on increasing the neuron unit in the last layer. Accordingly, to show the efficiency of the approach under study, a comparison is made between the proposed method of generating the input-output set and the conventional method. The results illustrated that the increasing of the neuron units, in the last layer, allows to find the optimal network’s parameters that fit with the mapping data. Moreover, it permits to decrease the training time, during the computation process, which avoids the use of computers with high memory usage.

Keywords: neural network computing, continuous functions generating the input-output mapping, decreasing the training time, machines with big memories

Procedia PDF Downloads 263
29484 Investigation of Delivery of Triple Play Service in GE-PON Fiber to the Home Network

Authors: Anurag Sharma, Dinesh Kumar, Rahul Malhotra, Manoj Kumar

Abstract:

Fiber based access networks can deliver performance that can support the increasing demands for high speed connections. One of the new technologies that have emerged in recent years is Passive Optical Networks. This paper is targeted to show the simultaneous delivery of triple play service (data, voice and video). The comparative investigation and suitability of various data rates is presented. It is demonstrated that as we increase the data rate, number of users to be accommodated decreases due to increase in bit error rate.

Keywords: BER, PON, TDMPON, GPON, CWDM, OLT, ONT

Procedia PDF Downloads 711
29483 Guests’ Satisfaction and Intention to Revisit Smart Hotels: Qualitative Interviews Approach

Authors: Raymond Chi Fai Si Tou, Jacey Ja Young Choe, Amy Siu Ian So

Abstract:

Smart hotels can be defined as the hotel which has an intelligent system, through digitalization and networking which achieve hotel management and service information. In addition, smart hotels include high-end designs that integrate information and communication technology with hotel management fulfilling the guests’ needs and improving the quality, efficiency and satisfaction of hotel management. The purpose of this study is to identify appropriate factors that may influence guests’ satisfaction and intention to revisit Smart Hotels based on service quality measurement of lodging quality index and extended UTAUT theory. Unified Theory of Acceptance and Use of Technology (UTAUT) is adopted as a framework to explain technology acceptance and use. Since smart hotels are technology-based infrastructure hotels, UTATU theory could be as the theoretical background to examine the guests’ acceptance and use after staying in smart hotels. The UTAUT identifies four key drivers of the adoption of information systems: performance expectancy, effort expectancy, social influence, and facilitating conditions. The extended UTAUT modifies the definitions of the seven constructs for consideration; the four previously cited constructs of the UTAUT model together with three new additional constructs, which including hedonic motivation, price value and habit. Thus, the seven constructs from the extended UTAUT theory could be adopted to understand their intention to revisit smart hotels. The service quality model will also be adopted and integrated into the framework to understand the guests’ intention of smart hotels. There are rare studies to examine the service quality on guests’ satisfaction and intention to revisit in smart hotels. In this study, Lodging Quality Index (LQI) will be adopted to measure the service quality in smart hotels. Using integrated UTAUT theory and service quality model because technological applications and services require using more than one model to understand the complicated situation for customers’ acceptance of new technology. Moreover, an integrated model could provide more perspective insights to explain the relationships of the constructs that could not be obtained from only one model. For this research, ten in-depth interviews are planned to recruit this study. In order to confirm the applicability of the proposed framework and gain an overview of the guest experience of smart hotels from the hospitality industry, in-depth interviews with the hotel guests and industry practitioners will be accomplished. In terms of the theoretical contribution, it predicts that the integrated models from the UTAUT theory and the service quality will provide new insights to understand factors that influence the guests’ satisfaction and intention to revisit smart hotels. After this study identifies influential factors, smart hotel practitioners could understand which factors may significantly influence smart hotel guests’ satisfaction and intention to revisit. In addition, smart hotel practitioners could also provide outstanding guests experience by improving their service quality based on the identified dimensions from the service quality measurement. Thus, it will be beneficial to the sustainability of the smart hotels business.

Keywords: intention to revisit, guest satisfaction, qualitative interviews, smart hotels

Procedia PDF Downloads 188
29482 International Service Learning 3.0: Using Technology to Improve Outcomes and Sustainability

Authors: Anthony Vandarakis

Abstract:

Today’s International Service Learning practices require an update: modern technologies, fresh educational frameworks, and a new operating system to accountably prosper. This paper describes a model of International Service Learning (ISL), which combines current technological hardware, electronic platforms, and asynchronous communications that are grounded in inclusive pedagogy. This model builds on the work around collaborative field trip learning, extending the reach to international partnerships across continents. Mobile technology, 21st century skills and summit-basecamp modeling intersect to support novel forms of learning that tread lightly on fragile natural ecosystems, affirm local reciprocal partnership in projects, and protect traveling participants from common yet avoidable cultural pitfalls.

Keywords: International Service Learning, ISL, field experiences, mobile technology, out there in here, summit basecamp pedagogy

Procedia PDF Downloads 160
29481 Investigation of Effects of Geomagnetic Storms Produced by Different Solar Sources on the Total Electron Content (TEC)

Authors: P. K. Purohit, Azad A. Mansoori, Parvaiz A. Khan, Purushottam Bhawre, Sharad C. Tripathi, A. M. Aslam, Malik A. Waheed, Shivangi Bhardwaj, A. K. Gwal

Abstract:

The geomagnetic storm represents the most outstanding example of solar wind-magnetospheric interaction, which causes global disturbances in the geomagnetic field as well as the trigger ionospheric disturbances. We study the behaviour of ionospheric Total Electron Content (TEC) during the geomagnetic storms. For the present investigation we have selected 47 intense geomagnetic storms (Dst ≤ -100nT) that were observed during the solar cycle 23 i.e. during 1998-2006. We then categorized these storms into four categories depending upon their solar sources like Magnetic Cloud (MC), Co-rotating Interaction Region (CIR), SH+ICME and SH+MC. We then studied the behaviour of ionospheric TEC at a mid latitude station Usuda (36.13N, 138.36E), Japan during these storm events produced by four different solar sources. During our study we found that the smooth variations in TEC are replaced by rapid fluctuations and the value of TEC is strongly enhanced during the time of these storms belonging to all the four categories. However, the greatest enhancements in TEC are produced during those geomagnetic storms which are either caused by sheath driven magnetic cloud (SH+MC) or sheath driven ICME (SH+ICME). We also derived the correlation between the TEC enhancements produced during storms of each category with the minimum Dst. We found the strongest correlation exists for the SH+ICME category followed by SH+MC, MC and finally CIR. Since the most intense storms were either caused by SH+ICME or SH+MC while the least intense storms were caused by CIR, consequently the correlation was the strongest with SH+ICME and SH+MC and least with CIR.

Keywords: GPS, TEC, geomagnetic storm, sheath driven magnetic cloud

Procedia PDF Downloads 525
29480 Heat Transfer and Trajectory Models for a Cloud of Spray over a Marine Vessel

Authors: S. R. Dehghani, G. F. Naterer, Y. S. Muzychka

Abstract:

Wave-impact sea spray creates many droplets which form a spray cloud traveling over marine objects same as marine vessels and offshore structures. In cold climates such as Arctic reigns, sea spray icing, which is ice accretion on cold substrates, is strongly dependent on the wave-impact sea spray. The rate of cooling of droplets affects the process of icing that can yield to dry or wet ice accretion. Trajectories of droplets determine the potential places for ice accretion. Combining two models of trajectories and heat transfer for droplets can predict the risk of ice accretion reasonably. The majority of the cooling of droplets is because of droplet evaporations. In this study, a combined model using trajectory and heat transfer evaluate the situation of a cloud of spray from the generation to impingement. The model uses some known geometry and initial information from the previous case studies. The 3D model is solved numerically using a standard numerical scheme. Droplets are generated in various size ranges from 7 mm to 0.07 mm which is a suggested range for sea spray icing. The initial temperature of droplets is considered to be the sea water temperature. Wind velocities are assumed same as that of the field observations. Evaluations are conducted using some important heading angles and wind velocities. The characteristic of size-velocity dependence is used to establish a relation between initial sizes and velocities of droplets. Time intervals are chosen properly to maintain a stable and fast numerical solution. A statistical process is conducted to evaluate the probability of expected occurrences. The medium size droplets can reach the highest heights. Very small and very large droplets are limited to lower heights. Results show that higher initial velocities create the most expanded cloud of spray. Wind velocities affect the extent of the spray cloud. The rate of droplet cooling at the start of spray formation is higher than the rest of the process. This is because of higher relative velocities and also higher temperature differences. The amount of water delivery and overall temperature for some sample surfaces over a marine vessel are calculated. Comparing results and some field observations show that the model works accurately. This model is suggested as a primary model for ice accretion on marine vessels.

Keywords: evaporation, sea spray, marine icing, numerical solution, trajectory

Procedia PDF Downloads 207
29479 An Analysis of New Service Interchange Designs

Authors: Joseph E. Hummer

Abstract:

An efficient freeway system will be essential to the development of Africa, and interchanges are a key to that efficiency. Around the world, many interchanges between freeways and surface streets, called service interchanges, are of the diamond configuration, and interchanges using roundabouts or loop ramps are also popular. However, many diamond interchanges have serious operational problems, interchanges with roundabouts fail at high demand levels, and loops use lots of expensive land. Newer service interchange designs provide other options. The most popular new interchange design in the US at the moment is the double crossover diamond (DCD), also known as the diverging diamond. The DCD has enormous potential, but also has several significant limitations. The objectives of this paper are to review new service interchange options and to highlight some of the main features of those alternatives. The paper tests four conventional and seven unconventional designs using seven measures related to efficiency, cost, and safety. The results show that there is no superior design in all measures investigated. The DCD is better than most designs tested on most measures examined. However, the DCD was only superior to all other designs for bridge width. The DCD performed relatively poorly for capacity and for serving pedestrians. Based on the results, African freeway designers are encouraged to investigate the full range of alternatives that could work at the spot of interest. Diamonds and DCDs have their niches, but some of the other designs investigated could be optimum at some spots.

Keywords: interchange, diamond, diverging diamond, capacity, safety, cost

Procedia PDF Downloads 232
29478 3D Classification Optimization of Low-Density Airborne Light Detection and Ranging Point Cloud by Parameters Selection

Authors: Baha Eddine Aissou, Aichouche Belhadj Aissa

Abstract:

Light detection and ranging (LiDAR) is an active remote sensing technology used for several applications. Airborne LiDAR is becoming an important technology for the acquisition of a highly accurate dense point cloud. A classification of airborne laser scanning (ALS) point cloud is a very important task that still remains a real challenge for many scientists. Support vector machine (SVM) is one of the most used statistical learning algorithms based on kernels. SVM is a non-parametric method, and it is recommended to be used in cases where the data distribution cannot be well modeled by a standard parametric probability density function. Using a kernel, it performs a robust non-linear classification of samples. Often, the data are rarely linearly separable. SVMs are able to map the data into a higher-dimensional space to become linearly separable, which allows performing all the computations in the original space. This is one of the main reasons that SVMs are well suited for high-dimensional classification problems. Only a few training samples, called support vectors, are required. SVM has also shown its potential to cope with uncertainty in data caused by noise and fluctuation, and it is computationally efficient as compared to several other methods. Such properties are particularly suited for remote sensing classification problems and explain their recent adoption. In this poster, the SVM classification of ALS LiDAR data is proposed. Firstly, connected component analysis is applied for clustering the point cloud. Secondly, the resulting clusters are incorporated in the SVM classifier. Radial basic function (RFB) kernel is used due to the few numbers of parameters (C and γ) that needs to be chosen, which decreases the computation time. In order to optimize the classification rates, the parameters selection is explored. It consists to find the parameters (C and γ) leading to the best overall accuracy using grid search and 5-fold cross-validation. The exploited LiDAR point cloud is provided by the German Society for Photogrammetry, Remote Sensing, and Geoinformation. The ALS data used is characterized by a low density (4-6 points/m²) and is covering an urban area located in residential parts of the city Vaihingen in southern Germany. The class ground and three other classes belonging to roof superstructures are considered, i.e., a total of 4 classes. The training and test sets are selected randomly several times. The obtained results demonstrated that a parameters selection can orient the selection in a restricted interval of (C and γ) that can be further explored but does not systematically lead to the optimal rates. The SVM classifier with hyper-parameters is compared with the most used classifiers in literature for LiDAR data, random forest, AdaBoost, and decision tree. The comparison showed the superiority of the SVM classifier using parameters selection for LiDAR data compared to other classifiers.

Keywords: classification, airborne LiDAR, parameters selection, support vector machine

Procedia PDF Downloads 134
29477 Performance Analysis of Elliptic Curve Cryptography Using Onion Routing to Enhance the Privacy and Anonymity in Grid Computing

Authors: H. Parveen Begam, M. A. Maluk Mohamed

Abstract:

Grid computing is an environment that allows sharing and coordinated use of diverse resources in dynamic, heterogeneous and distributed environment using Virtual Organization (VO). Security is a critical issue due to the open nature of the wireless channels in the grid computing which requires three fundamental services: authentication, authorization, and encryption. The privacy and anonymity are considered as an important factor while communicating over publicly spanned network like web. To ensure a high level of security we explored an extension of onion routing, which has been used with dynamic token exchange along with protection of privacy and anonymity of individual identity. To improve the performance of encrypting the layers, the elliptic curve cryptography is used. Compared to traditional cryptosystems like RSA (Rivest-Shamir-Adelman), ECC (Elliptic Curve Cryptosystem) offers equivalent security with smaller key sizes which result in faster computations, lower power consumption, as well as memory and bandwidth savings. This paper presents the estimation of the performance improvements of onion routing using ECC as well as the comparison graph between performance level of RSA and ECC.

Keywords: grid computing, privacy, anonymity, onion routing, ECC, RSA

Procedia PDF Downloads 380
29476 Efficacy of Deep Learning for Below-Canopy Reconstruction of Satellite and Aerial Sensing Point Clouds through Fractal Tree Symmetry

Authors: Dhanuj M. Gandikota

Abstract:

Sensor-derived three-dimensional (3D) point clouds of trees are invaluable in remote sensing analysis for the accurate measurement of key structural metrics, bio-inventory values, spatial planning/visualization, and ecological modeling. Machine learning (ML) holds the potential in addressing the restrictive tradeoffs in cost, spatial coverage, resolution, and information gain that exist in current point cloud sensing methods. Terrestrial laser scanning (TLS) remains the highest fidelity source of both canopy and below-canopy structural features, but usage is limited in both coverage and cost, requiring manual deployment to map out large, forested areas. While aerial laser scanning (ALS) remains a reliable avenue of LIDAR active remote sensing, ALS is also cost-restrictive in deployment methods. Space-borne photogrammetry from high-resolution satellite constellations is an avenue of passive remote sensing with promising viability in research for the accurate construction of vegetation 3-D point clouds. It provides both the lowest comparative cost and the largest spatial coverage across remote sensing methods. However, both space-borne photogrammetry and ALS demonstrate technical limitations in the capture of valuable below-canopy point cloud data. Looking to minimize these tradeoffs, we explored a class of powerful ML algorithms called Deep Learning (DL) that show promise in recent research on 3-D point cloud reconstruction and interpolation. Our research details the efficacy of applying these DL techniques to reconstruct accurate below-canopy point clouds from space-borne and aerial remote sensing through learned patterns of tree species fractal symmetry properties and the supplementation of locally sourced bio-inventory metrics. From our dataset, consisting of tree point clouds obtained from TLS, we deconstructed the point clouds of each tree into those that would be obtained through ALS and satellite photogrammetry of varying resolutions. We fed this ALS/satellite point cloud dataset, along with the simulated local bio-inventory metrics, into the DL point cloud reconstruction architectures to generate the full 3-D tree point clouds (the truth values are denoted by the full TLS tree point clouds containing the below-canopy information). Point cloud reconstruction accuracy was validated both through the measurement of error from the original TLS point clouds as well as the error of extraction of key structural metrics, such as crown base height, diameter above root crown, and leaf/wood volume. The results of this research additionally demonstrate the supplemental performance gain of using minimum locally sourced bio-inventory metric information as an input in ML systems to reach specified accuracy thresholds of tree point cloud reconstruction. This research provides insight into methods for the rapid, cost-effective, and accurate construction of below-canopy tree 3-D point clouds, as well as the supported potential of ML and DL to learn complex, unmodeled patterns of fractal tree growth symmetry.

Keywords: deep learning, machine learning, satellite, photogrammetry, aerial laser scanning, terrestrial laser scanning, point cloud, fractal symmetry

Procedia PDF Downloads 81
29475 Behavioral Pattern of 2G Mobile Internet Subscribers: A Study on an Operator of Bangladesh

Authors: Azfar Adib

Abstract:

Like many other countries of the world, mobile internet has been playing a key role in the growth of internet subscriber base in Bangladesh. This study has attempted to identify particular behavioral or usage patterns of 2G mobile internet subscribers who were using the service of the topmost internet service provider (as well as the top mobile operator) of Bangladesh prior to the launching of 3G services (when 2G was fully dominant). It contains some comprehensive analysis carried on different info regarding 2G mobile internet subscribers, obtained from the operator’s own network insights.This is accompanied by the results of a survey conducted among 40 high-frequency users of this service.

Keywords: mobile internet, Symbian, Android, iPhone

Procedia PDF Downloads 419
29474 The Interrelationship between Formal and Informal Institutions and Its Impacts on the Autonomy of Public Service Delivery Units: The Case of Vietnam

Authors: Minh Thi Hai Vo

Abstract:

This article draws on in-depth interviews with state employees at public hospitals and universities in its institutional analysis of the autonomy practices of public service delivery units in Vietnam. Unlike many empirical and theoretical studies that view formal and informal institutions as complements or substitutes, this article finds no evidence of complementary or substitutive relationships. Instead, the article finds that formal institutions accommodate informal ones and that informal institutions tend to compete and interfere, with the existing and ineffective formal institutions. The result of such conflicting relationship is that the actual autonomy of public service delivery units is, in most cases, perceived to be greater than the formal autonomy they are given. In the condition of poor regulation, the informal autonomy may result in unethical practices including rent-seeking and corruption. The implication of the study finding is policy-makers need to redesign and reorganize the autonomisation of public service delivery units to make informal institutions support and reinforce formal ones in a complementary manner.

Keywords: autonomy, formal institutions, informal institutions, public service delivery units, Vietnam

Procedia PDF Downloads 184
29473 Event Monitoring Based On Web Services for Heterogeneous Event Sources

Authors: Arne Koschel

Abstract:

This article discusses event monitoring options for heterogeneous event sources as they are given in nowadays heterogeneous distributed information systems. It follows the central assumption, that a fully generic event monitoring solution cannot provide complete support for event monitoring; instead, event source specific semantics such as certain event types or support for certain event monitoring techniques have to be taken into account. Following from this, the core result of the work presented here is the extension of a configurable event monitoring (Web) service for a variety of event sources. A service approach allows us to trade genericity for the exploitation of source specific characteristics. It thus delivers results for the areas of SOA, Web services, CEP and EDA.

Keywords: event monitoring, ECA, CEP, SOA, web services

Procedia PDF Downloads 720
29472 Applying Quadrant Analysis in Identifying Business-to-Business Customer-Driven Improvement Opportunities in Third Party Logistics Industry

Authors: Luay Jum'a

Abstract:

Many challenges are facing third-party logistics (3PL) providers in the domestic and global markets which create a volatile decision making environment. All these challenges such as managing changes in consumer behaviour, demanding expectations from customers and time compressions have turned into complex problems for 3PL providers. Since the movement towards increased outsourcing outpaces movement towards insourcing, the need to achieve a competitive advantage over competitors in 3PL market increases. This trend continues to grow over the years and as a result, areas of strengths and improvements are highlighted through the analysis of the LSQ factors that lead to B2B customers’ satisfaction which become a priority for 3PL companies. Consequently, 3PL companies are increasingly focusing on the most important issues from the perspective of their customers and relying more on this value of information in making their managerial decisions. Therefore, this study is concerned with providing guidance for improving logistics service quality (LSQ) levels in the context of 3PL industry in Jordan. The study focused on the most important factors in LSQ and used a managerial tool that guides 3PL companies in making LSQ improvements based on a quadrant analysis of two main dimensions: LSQ declared importance and LSQ inferred importance. Although, a considerable amount of research has been conducted to investigate the relationship between logistics service quality (LSQ) and customer satisfaction, there remains a lack of developing managerial tools to aid in the process of LSQ improvement decision-making. Moreover, the main advantage for the companies to use 3PL service providers as a trend is due to the realised percentage of cost reduction on the total cost of logistics operations and the incremental improvement in customer service. In this regard, having a managerial tool that help 3PL service providers in managing the LSQ factors portfolio effectively and efficiently would be a great investment for service providers. One way of suggesting LSQ improvement actions for 3PL service providers is via the adoption of analysis tools that perform attribute categorisation such as Importance–Performance matrix. In mind of the above, it can be stated that the use of quadrant analysis will provide a valuable opportunity for 3PL service providers to identify improvement opportunities as customer service attributes or factors importance are identified in two different techniques that complete each other. Moreover, the data were collected through conducting a survey and 293 questionnaires were returned from business-to-business (B2B) customers of 3PL companies in Jordan. The results showed that the LSQ factors vary in their importance and 3PL companies should focus on some LSQ factors more than other factors. Moreover, ordering procedures, timeliness/responsiveness LSQ factors considered being crucial in 3PL businesses and therefore they need to have more focus and development by 3PL service providers in the Jordanian market.

Keywords: logistics service quality, managerial decisions, quadrant analysis, third party logistics service provider

Procedia PDF Downloads 117
29471 Improving Fault Tolerance and Load Balancing in Heterogeneous Grid Computing Using Fractal Transform

Authors: Saad M. Darwish, Adel A. El-Zoghabi, Moustafa F. Ashry

Abstract:

The popularity of the Internet and the availability of powerful computers and high-speed networks as low-cost commodity components are changing the way we use computers today. These technical opportunities have led to the possibility of using geographically distributed and multi-owner resources to solve large-scale problems in science, engineering, and commerce. Recent research on these topics has led to the emergence of a new paradigm known as Grid computing. To achieve the promising potentials of tremendous distributed resources, effective and efficient load balancing algorithms are fundamentally important. Unfortunately, load balancing algorithms in traditional parallel and distributed systems, which usually run on homogeneous and dedicated resources, cannot work well in the new circumstances. In this paper, the concept of a fast fractal transform in heterogeneous grid computing based on R-tree and the domain-range entropy is proposed to improve fault tolerance and load balancing algorithm by improve connectivity, communication delay, network bandwidth, resource availability, and resource unpredictability. A novel two-dimension figure of merit is suggested to describe the network effects on load balance and fault tolerance estimation. Fault tolerance is enhanced by adaptively decrease replication time and message cost while load balance is enhanced by adaptively decrease mean job response time. Experimental results show that the proposed method yields superior performance over other methods.

Keywords: Grid computing, load balancing, fault tolerance, R-tree, heterogeneous systems

Procedia PDF Downloads 466
29470 SIP Flooding Attacks Detection and Prevention Using Shannon, Renyi and Tsallis Entropy

Authors: Neda Seyyedi, Reza Berangi

Abstract:

Voice over IP (VOIP) network, also known as Internet telephony, is growing increasingly having occupied a large part of the communications market. With the growth of each technology, the related security issues become of particular importance. Taking advantage of this technology in different environments with numerous features put at our disposal, there arises an increasing need to address the security threats. Being IP-based and playing a signaling role in VOIP networks, Session Initiation Protocol (SIP) lets the invaders use weaknesses of the protocol to disable VOIP service. One of the most important threats is denial of service attack, a branch of which in this article we have discussed as flooding attacks. These attacks make server resources wasted and deprive it from delivering service to authorized users. Distributed denial of service attacks and attacks with a low rate can mislead many attack detection mechanisms. In this paper, we introduce a mechanism which not only detects distributed denial of service attacks and low rate attacks, but can also identify the attackers accurately. We detect and prevent flooding attacks in SIP protocol using Shannon (FDP-S), Renyi (FDP-R) and Tsallis (FDP-T) entropy. We conducted an experiment to compare the percentage of detection and rate of false alarm messages using any of the Shannon, Renyi and Tsallis entropy as a measure of disorder. Implementation results show that, according to the parametric nature of the Renyi and Tsallis entropy, by changing the parameters, different detection percentages and false alarm rates will be gained with the possibility to adjust the sensitivity of the detection mechanism.

Keywords: VOIP networks, flooding attacks, entropy, computer networks

Procedia PDF Downloads 383
29469 Resource Allocation Scheme For IEEE802.16 Networks

Authors: Elmabruk Laias

Abstract:

IEEE Standard 802.16 provides QoS (Quality of Service) for the applications such as Voice over IP, video streaming and high bandwidth file transfer. With the ability of broadband wireless access of an IEEE 802.16 system, a WiMAX TDD frame contains one downlink subframe and one uplink subframe. The capacity allocated to each subframe is a system parameter that should be determined based on the expected traffic conditions. a proper resource allocation scheme for packet transmissions is imperatively needed. In this paper, we present a new resource allocation scheme, called additional bandwidth yielding (ABY), to improve transmission efficiency of an IEEE 802.16-based network. Our proposed scheme can be adopted along with the existing scheduling algorithms and the multi-priority scheme without any change. The experimental results show that by using our ABY, the packet queuing delay could be significantly improved, especially for the service flows of higher-priority classes.

Keywords: IEEE 802.16, WiMAX, OFDMA, resource allocation, uplink-downlink mapping

Procedia PDF Downloads 455
29468 Context-Aware Alert Method in Hajj Pilgrim Location-Based Tracking System

Authors: Syarif Hidayat

Abstract:

As millions of people with different backgrounds perform hajj every year in Saudi Arabia, it brings out several problems. Missing people is among many crucial problems need to be encountered. Some people might have had insufficient knowledge of using tracking system equipment. Other might become a victim of an accident, lose consciousness, or even died, prohibiting them to perform certain activity. For those reasons, people could not send proper SOS message. The major contribution of this paper is the application of the diverse alert method in pilgrims tracking system. It offers a simple yet robust solution to send SOS message by pilgrims during Hajj. Knowledge of context aware computing is assumed herein. This study presents four methods that could be utilized by pilgrims to send SOS. The first method is simple mobile application contains only a button. The second method is based on behavior analysis based off GPS location movement anomaly. The third method is by introducing pressing pattern to smartwatch physical button as a panic button. The fourth method is by identifying certain accelerometer pattern recognition as a sign of emergency situations. Presented method in this paper would be an important part of pilgrims tracking system. The discussion provided here includes easy to use design whilst maintaining tracking accuracy, privacy, and security of its users.

Keywords: context aware computing, emergency alert system, GPS, hajj pilgrim tracking, location-based services

Procedia PDF Downloads 199
29467 Automatic Aggregation and Embedding of Microservices for Optimized Deployments

Authors: Pablo Chico De Guzman, Cesar Sanchez

Abstract:

Microservices are a software development methodology in which applications are built by composing a set of independently deploy-able, small, modular services. Each service runs a unique process and it gets instantiated and deployed in one or more machines (we assume that different microservices are deployed into different machines). Microservices are becoming the de facto standard for developing distributed cloud applications due to their reduced release cycles. In principle, the responsibility of a microservice can be as simple as implementing a single function, which can lead to the following issues: - Resource fragmentation due to the virtual machine boundary. - Poor communication performance between microservices. Two composition techniques can be used to optimize resource fragmentation and communication performance: aggregation and embedding of microservices. Aggregation allows the deployment of a set of microservices on the same machine using a proxy server. Aggregation helps to reduce resource fragmentation, and is particularly useful when the aggregated services have a similar scalability behavior. Embedding deals with communication performance by deploying on the same virtual machine those microservices that require a communication channel (localhost bandwidth is reported to be about 40 times faster than cloud vendor local networks and it offers better reliability). Embedding can also reduce dependencies on load balancer services since the communication takes place on a single virtual machine. For example, assume that microservice A has two instances, a1 and a2, and it communicates with microservice B, which also has two instances, b1 and b2. One embedding can deploy a1 and b1 on machine m1, and a2 and b2 are deployed on a different machine m2. This deployment configuration allows each pair (a1-b1), (a2-b2) to communicate using the localhost interface without the need of a load balancer between microservices A and B. Aggregation and embedding techniques are complex since different microservices might have incompatible runtime dependencies which forbid them from being installed on the same machine. There is also a security concern since the attack surface between microservices can be larger. Luckily, container technology allows to run several processes on the same machine in an isolated manner, solving the incompatibility of running dependencies and the previous security concern, thus greatly simplifying aggregation/embedding implementations by just deploying a microservice container on the same machine as the aggregated/embedded microservice container. Therefore, a wide variety of deployment configurations can be described by combining aggregation and embedding to create an efficient and robust microservice architecture. This paper presents a formal method that receives a declarative definition of a microservice architecture and proposes different optimized deployment configurations by aggregating/embedding microservices. The first prototype is based on i2kit, a deployment tool also submitted to ICWS 2018. The proposed prototype optimizes the following parameters: network/system performance, resource usage, resource costs and failure tolerance.

Keywords: aggregation, deployment, embedding, resource allocation

Procedia PDF Downloads 180
29466 Adoption and Diffusion of E-Government Services in India: The Impact of User Demographics and Service Quality

Authors: Sayantan Khanra, Rojers P. Joseph

Abstract:

This study attempts to analyze the impact of demography and service quality on the adoption and diffusion of e-Government services in the context of India. The objective of this paper is to study the users' perception about e-Government services and investigate the key variables that are most salient to the Indian populace. At the completion of this study, a research model that would help to understand the relationship involving the demographic variables and service quality dimensions, and the willingness to adopt e-Government services is expected to be developed. Dedicated authorities, particularly those in developing economies, may use that model or its augmented versions to design and update e-Government services and promote their use among citizens. After all, enhanced public participation is required to improve efficiency, engagement and transparency in the implementation of the aforementioned services.

Keywords: adoption and diffusion of e-government services, demographic variables, hierarchical regression analysis, service quality dimensions

Procedia PDF Downloads 269
29465 R Data Science for Technology Management

Authors: Sunghae Jun

Abstract:

Technology management (TM) is important issue in a company improving the competitiveness. Among many activities of TM, technology analysis (TA) is important factor, because most decisions for management of technology are decided by the results of TA. TA is to analyze the developed results of target technology using statistics or Delphi. TA based on Delphi is depended on the experts’ domain knowledge, in comparison, TA by statistics and machine learning algorithms use objective data such as patent or paper instead of the experts’ knowledge. Many quantitative TA methods based on statistics and machine learning have been studied, and these have been used for technology forecasting, technological innovation, and management of technology. They applied diverse computing tools and many analytical methods case by case. It is not easy to select the suitable software and statistical method for given TA work. So, in this paper, we propose a methodology for quantitative TA using statistical computing software called R and data science to construct a general framework of TA. From the result of case study, we also show how our methodology is applied to real field. This research contributes to R&D planning and technology valuation in TM areas.

Keywords: technology management, R system, R data science, statistics, machine learning

Procedia PDF Downloads 440
29464 Hypothesis about the Origin of the Lighting

Authors: Igor Kuzminov

Abstract:

Till now, the nature of lightning is not established. A hypothesis of the origin of lightning is proposed. The lightning charge is formed by electromagnetic induction. The role of the conductor is performed by the air mass of the cloud. This conductor moves in the Earth's magnetic field. The upper and lower edges of the cloud are the plates of the capacitor. Lightning is a special case of electromagnetic processes in an atmosphere. The category of lightning occurs in the process of accumulation of a charge. The process of accumulation goes constantly, but the charge is not fixed. Naturally, the hypothesis demands the carrying out of additional experiments and official acknowledgement. As the proof of a hypothesis can serve that the maximal lighting activity in an equatorial zone where cosφ it is close to 1. An experiment conducted privately showed that there is a potential difference in the atmosphere at different levels. The probability of applied value development of power installation is great.

Keywords: electromagnetic induction, Earth's magnetic field, plates of the capacitors, charge accumulation

Procedia PDF Downloads 72
29463 Analysis and Performance of European Geostationary Navigation Overlay Service System in North of Algeria for GPS Single Point Positioning

Authors: Tabti Lahouaria, Kahlouche Salem, Benadda Belkacem, Beldjilali Bilal

Abstract:

The European Geostationary Navigation Overlay Service (EGNOS) provides an augmentation signal to GPS (Global Positioning System) single point positioning. Presently EGNOS provides data correction and integrity information using the GPS L1 (1575.42 MHz) frequency band. The main objective of this system is to provide a better real-time positioning precision than using GPS only. They are expected to be used with single-frequency code observations. EGNOS offers navigation performance for an open service (OS), in terms of precision and availability this performance gradually degrades as moving away from the service area. For accurate system performance, the service will become less and less available as the user moves away from the EGNOS service. The improvement in position solution is investigated using the two collocated dual frequency GPS, where no EGNOS Ranging and Integrity Monitoring Station (RIMS) exists. One of the pseudo-range was kept as GPS stand-alone and the other was corrected by EGNOS to estimate the planimetric and altimetric precision for different dates. It is found that precision in position improved significantly in the second due to EGNOS correction. The performance of EGNOS system in the north of Algeria is also investigated in terms of integrity. The results show that the horizontal protection level (HPL) value is below 18.25 meters (95%) and the vertical protection level (VPL) is below 42.22 meters (95 %). These results represent good integrity information transmitted by EGNOS for APV I service. This service is thus compliant with the aviation requirements for Approaches with Vertical Guidance (APV-I), which is characterised by 40 m HAL (horizontal alarm limit) and 50 m VAL (vertical alarm limit).

Keywords: EGNOS, GPS, positioning, integrity, protection level

Procedia PDF Downloads 210
29462 Convectory Policing-Reconciling Historic and Contemporary Models of Police Service Delivery

Authors: Mark Jackson

Abstract:

Description: This paper is based on an theoretical analysis of the efficacy of the dominant model of policing in western jurisdictions. Those results are then compared with a similar analysis of a traditional reactive model. It is found that neither model provides for optimal delivery of services. Instead optimal service can be achieved by a synchronous hybrid model, termed the Convectory Policing approach. Methodology and Findings: For over three decades problem oriented policing (PO) has been the dominant model for western police agencies. Initially based on the work of Goldstein during the 1970s the problem oriented framework has spawned endless variants and approaches, most of which embrace a problem solving rather than a reactive approach to policing. This has included the Area Policing Concept (APC) applied in many smaller jurisdictions in the USA, the Scaled Response Policing Model (SRPM) currently under trial in Western Australia and the Proactive Pre-Response Approach (PPRA) which has also seen some success. All of these, in some way or another, are largely based on a model that eschews a traditional reactive model of policing. Convectory Policing (CP) is an alternative model which challenges the underpinning assumptions which have seen proliferation of the PO approach in the last three decades and commences by questioning the economics on which PO is based. It is argued that in essence, the PO relies on an unstated, and often unrecognised assumption that resources will be available to meet demand for policing services, while at the same time maintaining the capacity to deploy staff to develop solutions to the problems which were ultimately manifested in those same calls for service. The CP model relies on the observations from a numerous western jurisdictions to challenge the validity of that underpinning assumption, particularly in fiscally tight environment. In deploying staff to pursue and develop solutions to underpinning problems, there is clearly an opportunity cost. Those same staff cannot be allocated to alternative duties while engaged in a problem solution role. At the same time, resources in use responding to calls for service are unavailable, while committed to that role, to pursue solutions to the problems giving rise to those same calls for service. The two approaches, reactive and PO are therefore dichotomous. One cannot be optimised while the other is being pursued. Convectory Policing is a pragmatic response to the schism between the competing traditional and contemporary models. If it is not possible to serve either model with any real rigour, it becomes necessary to taper an approach to deliver specific outcomes against which success or otherwise might be measured. CP proposes that a structured roster-driven approach to calls for service, combined with the application of what is termed a resource-effect response capacity has the potential to resolve the inherent conflict between traditional and models of policing and the expectations of the community in terms of community policing based problem solving models.

Keywords: policing, reactive, proactive, models, efficacy

Procedia PDF Downloads 465
29461 Determinants of Successful Accounting Information System Outsourcing for the Egyptian Small and Medium Enterprises: An Empirical Study

Authors: Maram Elkady

Abstract:

Purpose: The purpose behind this study is to determine the impact of some factors on achieving successful accounting information systems (AIS) outsourcing in Egypt, taking into account two factors: the selection of an effective accounting service provider and the quality relationships between the client firm and the accounting service provider. The researcher measured outsourcing success through the perceived benefits, including (strategic, technological, and economic benefits). Design/Methodology/Approach: A survey was carried out by means of questionnaires answered by 152 small and medium Egyptian firms outsourcing their accounting activities. The researcher targeted the personnel in the client firms who were in direct contact with the accounting outsourcer. The hypotheses were tested through multiple regression analysis using SPSS 24 and AMOS 22. Findings: Building a quality relationship with the provider is found to have more impact than the effective selection of the AIS provider on the success of the AIS outsourcing process. Originality/Value: The researcher found that some proxies of each success determinant can be more influential than others based on type of benefits perceived from AIS outsourcing (strategic, technological, and economic).

Keywords: accounting information system, AIS, outsourcing, successful outsourcing, AIS service provider selection, relationship with the accounting service provider

Procedia PDF Downloads 138
29460 Determining the Collaboration and Challenges of Public Employment Service with Stakeholders, Employers and Job Seekers: In Case of Amhara National Regional State, Ethiopia

Authors: Redie Bezabih Hailu

Abstract:

Unemployment is a problem of nations that needs a continuous research. This study aimed to determine the collaborations and challenges of public employment service (PES) with special emphasis of stakeholders, employers and job seekers. The researcher used pragmatic philosophy, exploratory design and inductive approach to collect data from the respondents using interview and focused group discussion techniques. PES provides job market information, vocational counseling, and training. As PES is not fully furnished with man power, budget, modern technologies, it is providing less adequate services to the employers and job seekers. Matching job seekers with job vacancies is the major challenge for the center and using paper-based data management system too. There is also a number of job seekers in spite of very limited number of vacancies that the service provision is poor due to the fact that there is low level of vacancies and high level of job seekers. The center has collaboration with AFE, AYA, BoTVED, BoWCY, and CETU. The major challenges with this collaborations was the absence of operational guidelines to evaluate effectiveness and performance, lottery method of selecting candidates for vacancies and nepotism or favoritism were challenges for job seekers. On the other hand, (COVID-19) pandemic, inability to get skilled labor, absence of standardized payment, expectation of job seekers and less educational quality and mass graduation were another challenges for employment services. The study recommended quality education and training, operational guideline for collaboration, technology based labor market information system and suggested further studies on quality of PES.

Keywords: public employment service, collaborations, stakeholders, employers, job seekers

Procedia PDF Downloads 19
29459 Implementing a Hospitalist Co-Management Service in Orthopaedic Surgery

Authors: Diane Ghanem, Whitney Kagabo, Rebecca Engels, Uma Srikumaran, Babar Shafiq

Abstract:

Hospitalist co-management of orthopaedic surgery patients is a growing trend across the country. It was created as a collaborative effort to provide overarching care to patients with the goal of improving their postoperative care and decreasing in-hospital medical complications. The aim of this project is to provide a guide for implementing and optimizing a hospitalist co-management service in orthopaedic surgery. Key leaders from the hospitalist team, orthopaedic team and quality, safety and service team were identified. Multiple meetings were convened to discuss the comanagement service and determine the necessary building blocks behind an efficient and well-designed co-management framework. After meticulous deliberation, a consensus was reached on the final service agreement and a written guide was drafted. Fundamental features of the service include the identification of service stakeholders and leaders, frequent consensus meetings, a well-defined framework, with goals, program metrics and unified commands, and a regular satisfaction assessment to update and improve the program. Identified pearls for co-managing orthopaedic surgery patients are standardization, timing, adequate patient selection, and two-way feedback between hospitalists and orthopaedic surgeons to optimize the protocols. Developing a service agreement is a constant work in progress, with meetings, discussions, revisions, and multiple piloting attempts before implementation. It is a partnership created to provide hospitals with a streamlined admission process where at-risk patients are identified early, and patient care is optimized regardless of the number or nature of medical comorbidities. A wellestablished hospitalist co-management service can increase patient care quality and safety, as well as health care value.

Keywords: co-management, hospitalist co-management, implementation, orthopaedic surgery, quality improvement

Procedia PDF Downloads 63
29458 Non-Singular Gravitational Collapse of a Dust Cloud in Einstein-Cartan Theory

Authors: Amir Hadi Ziaie, Mostafa Hashemi, Shahram Jalalzadeh

Abstract:

It is now known that the end state of the collapse process of a dense star under its own gravity is the formation of a spacetime singularity. This is the spacetime event where the energy density and spacetime curvature diverge, and the classical general relativity breaks down. As we know, a realistic star is composed of fermions so that their spin effects could alter the final fate of the collapse scenario. The underlying theory within which the inclusion of spin effects can be worked out is the Einstein-Cartan theory. In this theory, the spacetime torsion which is defined as a geometrical quantity, is related to an intrinsic angular momentum of fermions (spin). In this work, we study the collapse process of a homogeneous spin fluid in such a framework and show that taking into account the spin effects of the collapsing cloud could prevent the formation of spacetime singularity.

Keywords: gravitational collapse, einstein-cartan theory, spacetime singularity, black hole physics

Procedia PDF Downloads 381
29457 Financial Service of Financial Institution for SME in Thailand

Authors: Charawee Butbumrung

Abstract:

This research aim to study the financial service of the Thailand financial Institution, second is to identify "best practices" offered by four financial institutions, namely, Kasikornthai Bank, Bangkok Bank, Siam Commercial Bank, and Thanachart Bank. In-depth interviews with managers of financial institution and borrowers reveal best practices from each financial institution. Close monitoring of and a close relationship with borrowers appear to be important for early detection of any problem. Another aspect that may be important is building up loyalty and developing reliability among members. A close and informal relationship with borrowers may also help in monitoring and early detection of problems that may arise in non-repayment of loans. Other factors that may be considered important to the success of a financial service scheme are cooperation and coordination among various agencies that provide additional support to borrowers. Indirectly, these support systems contribute to the success of a SME in Thailand.

Keywords: best practices, financial service, financial institution, SME in Thailand

Procedia PDF Downloads 272