Search results for: cloud monitoring
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 3563

Search results for: cloud monitoring

3293 Flexible 3D Virtual Desktop Using Handles for Cloud Environments

Authors: J. K. Lee, S. L. Lee

Abstract:

Due to the improvement in performance of computer hardware and the development of operating systems, a multi-tasking for several programs has become one of the basic functions to computer users. It is natural for computer users to want more functional, convenient, and visual GUI functions (Graphic User Interface). In this paper, a 3D virtual desktop system was proposed to meet users’ requirements for cloud environments such as a virtual desktop function in the Windows environment. The proposed system uses the handles of the windows to hide or restore several windows. It connects the list of task spaces using the circular double linked list to manage the handles. Each handle list is registered in the corresponding task space being executed. The 3D virtual desktop is efficient and flexible in handling the numbers of task spaces and can help users to work under more comfortable environments. Acknowledgment: This research was supported by the Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Korea government (MSIP) (NRF-2015R1D1A1A01057680).

Keywords: virtual desktop, GUI, cloud, virtualization

Procedia PDF Downloads 182
3292 Big Data Analytics and Data Security in the Cloud via Fully Homomorphic Encyption Scheme

Authors: Victor Onomza Waziri, John K. Alhassan, Idris Ismaila, Noel Dogonyara

Abstract:

This paper describes the problem of building secure computational services for encrypted information in the Cloud. Computing without decrypting the encrypted data; therefore, it meets the yearning of computational encryption algorithmic aspiration model that could enhance the security of big data for privacy or confidentiality, availability and integrity of the data and user’s security. The cryptographic model applied for the computational process of the encrypted data is the Fully Homomorphic Encryption Scheme. We contribute a theoretical presentations in a high-level computational processes that are based on number theory that is derivable from abstract algebra which can easily be integrated and leveraged in the Cloud computing interface with detail theoretic mathematical concepts to the fully homomorphic encryption models. This contribution enhances the full implementation of big data analytics based on cryptographic security algorithm.

Keywords: big data analytics, security, privacy, bootstrapping, Fully Homomorphic Encryption Scheme

Procedia PDF Downloads 442
3291 Flow Conservation Framework for Monitoring Software Defined Networks

Authors: Jesús Antonio Puente Fernández, Luis Javier Garcia Villalba

Abstract:

New trends on streaming videos such as series or films require a high demand of network resources. This fact results in a huge problem within traditional IP networks due to the rigidity of its architecture. In this way, Software Defined Networks (SDN) is a new concept of network architecture that intends to be more flexible and it simplifies the management in networks with respect to the existing ones. These aspects are possible due to the separation of control plane (controller) and data plane (switches). Taking the advantage of this separated control, it is easy to deploy a monitoring tool independent of device vendors since the existing ones are dependent on the installation of specialized and expensive hardware. In this paper, we propose a framework that optimizes the traffic monitoring in SDN networks that decreases the number of monitoring queries to improve the network traffic and also reduces the overload. The performed experiments (with and without the optimization) using a video streaming delivery between two hosts demonstrate the feasibility of our monitoring proposal.

Keywords: optimization, monitoring, software defined networking, statistics, query

Procedia PDF Downloads 297
3290 3D Point Cloud Model Color Adjustment by Combining Terrestrial Laser Scanner and Close Range Photogrammetry Datasets

Authors: M. Pepe, S. Ackermann, L. Fregonese, C. Achille

Abstract:

3D models obtained with advanced survey techniques such as close-range photogrammetry and laser scanner are nowadays particularly appreciated in Cultural Heritage and Archaeology fields. In order to produce high quality models representing archaeological evidences and anthropological artifacts, the appearance of the model (i.e. color) beyond the geometric accuracy, is not a negligible aspect. The integration of the close-range photogrammetry survey techniques with the laser scanner is still a topic of study and research. By combining point cloud data sets of the same object generated with both technologies, or with the same technology but registered in different moment and/or natural light condition, could construct a final point cloud with accentuated color dissimilarities. In this paper, a methodology to uniform the different data sets, to improve the chromatic quality and to highlight further details by balancing the point color will be presented.

Keywords: color models, cultural heritage, laser scanner, photogrammetry

Procedia PDF Downloads 254
3289 Vibration-Based Monitoring of Tensioning Stay Cables of an Extradosed Bridge

Authors: Chun-Chung Chen, Bo-Han Lee, Yu-Chi Sung

Abstract:

Monitoring the status of tensioning force of stay cables is a significant issue for the assessment of structural safety of extradosed bridges. Moreover, it is known that there is a high correlation between the existing tension force and the vibration frequencies of cables. This paper presents the characteristic of frequencies of stay cables of a field extradosed bridge by using vibration-based monitoring methods. The vibration frequencies of each stay cables were measured in stages from the beginning to the completion of bridge construction. The result shows that the vibration frequency variation trend of different lengths of cables at each measured stage is different. The observed feature can help the application of the bridge long-term monitoring system and contribute to the assessment of bridge safety.

Keywords: vibration-based method, extradosed bridges, bridge health monitoring, bridge stay cables

Procedia PDF Downloads 128
3288 Global Indicators of Successful Remote Monitoring Adoption Applying Diffusion of Innovation Theory

Authors: Danika Tynes

Abstract:

Innovations in technology have implications for sustainable development in health and wellness. Remote monitoring is one innovation for which the evidence-base has grown to support its viability as a quality healthcare delivery adjunct. This research reviews global data on telehealth adoption, in particular, remote monitoring, and the conditions under which its success becomes more likely. System-level indicators were selected to represent four constructs of DoI theory (relative advantage, compatibility, complexity, and observability) and assessed against 5 types of Telehealth (Teleradiology, Teledermatology, Telepathology, Telepsychology, and Remote Monitoring) using ordinal logistic regression. Analyses include data from 84 countries, as extracted from the World Health Organization, World Bank, ICT (Information Communications Technology) Index, and HDI (Human Development Index) datasets. Analyses supported relative advantage and compatibility as the strongest influencers of remote monitoring adoption. Findings from this research may help focus on the allocation of resources, as a sustainability concern, through consideration of systems-level factors that may influence the success of remote monitoring adoption.

Keywords: remote monitoring, diffusion of innovation, telehealth, digital health

Procedia PDF Downloads 101
3287 Monitoring of Hydrological Parameters in the Alexandra Jukskei Catchment in South Africa

Authors: Vhuhwavho Gadisi, Rebecca Alowo, German Nkhonjera

Abstract:

It has been noted that technical programming for handling groundwater resources is not accessible. The lack of these systems hinders groundwater management processes necessary for decision-making through monitoring and evaluation regarding the Jukskei River of the Crocodile River (West) Basin in Johannesburg, South Africa. Several challenges have been identified in South Africa's Jukskei Catchment concerning groundwater management. Some of those challenges will include the following: Gaps in data records; there is a need for training and equipping of monitoring staff; formal accreditation of monitoring capacities and equipment; there is no access to regulation terms (e.g., meters). Taking into consideration necessities and human requirements as per typical densities in various regions of South Africa, there is a need to construct several groundwater level monitoring stations in a particular segment; the available raw data on groundwater level should be converted into consumable products for example, short reports on delicate areas (e.g., Dolomite compartments, wetlands, aquifers, and sole source) and considering the increasing civil unrest there has been vandalism and theft of groundwater monitoring infrastructure. GIS was employed at the catchment level to plot the relationship between those identified groundwater parameters in the catchment area and the identified borehole. GIS-based maps were designed for groundwater monitoring to be pretested on one borehole in the Jukskei catchment. This data will be used to establish changes in the borehole compared to changes in the catchment area according to identified parameters.

Keywords: GIS, monitoring, Jukskei, catchment

Procedia PDF Downloads 64
3286 Indirect Environmental Benefits from Cloud Computing Information and Communications Technology Integration in Rural Agricultural Communities

Authors: Jeana Cadby, Kae Miyazawa

Abstract:

With rapidly expanding worldwide adoption of mobile technologies, Information and Communication Technology (ITC) is a major energy user and a contributor to global carbon emissions, due to infrastructure and operational energy consumption. The agricultural sector is also significantly responsible for contributing to global carbon emissions. However, ICT cloud computing using mobile technology can directly reduce environmental impacts in the agricultural sector through applications and mobile connectivity, such as precision fertilizer and pesticide applications, or access to weather data, for example. While direct impacts are easily calculated, indirect environmental impacts from ICT cloud computing usage have not been thoroughly investigated. For example, while women may be more poorly equipped for adaptation to environmentally sustainable agricultural practices due to resource constraints, this research concludes that indirect environmental benefits can be achieved by improving rural access to mobile technology for women. Women in advanced roles and secure land tenure are more likely to invest in long-term agricultural conservation strategies, which protect against environmental degradation. This study examines how ICT using mobile technology advances the role of women in rural agricultural systems and indirectly reduces environmental impacts from agricultural production, through literature examination from secondary sources. Increasing access for women to ICT mobile technology provides indirect environmental and social benefits in the rural agricultural sector.

Keywords: cloud computing, environmental benefits, mobile technology, women

Procedia PDF Downloads 140
3285 A Novel Meta-Heuristic Algorithm Based on Cloud Theory for Redundancy Allocation Problem under Realistic Condition

Authors: H. Mousavi, M. Sharifi, H. Pourvaziri

Abstract:

Redundancy Allocation Problem (RAP) is a well-known mathematical problem for modeling series-parallel systems. It is a combinatorial optimization problem which focuses on determining an optimal assignment of components in a system design. In this paper, to be more practical, we have considered the problem of redundancy allocation of series system with interval valued reliability of components. Therefore, during the search process, the reliabilities of the components are considered as a stochastic variable with a lower and upper bounds. In order to optimize the problem, we proposed a simulated annealing based on cloud theory (CBSAA). Also, the Monte Carlo simulation (MCS) is embedded to the CBSAA to handle the random variable components’ reliability. This novel approach has been investigated by numerical examples and the experimental results have shown that the CBSAA combining MCS is an efficient tool to solve the RAP of systems with interval-valued component reliabilities.

Keywords: redundancy allocation problem, simulated annealing, cloud theory, monte carlo simulation

Procedia PDF Downloads 389
3284 Bearing Condition Monitoring with Acoustic Emission Techniques

Authors: Faisal AlShammari, Abdulmajid Addali

Abstract:

Monitoring the conditions of rotating machinery as bearing is important in order to improve its stability of works. Acoustic emission (AE) and vibration analysis are some of the most accomplished techniques used for this purpose. Acoustic emission has the ability to detect the initial phase of component degradation. Moreover, it has been observed that the success of vibration analysis does not take place below 100 rpm rotational speed. This because the energy generated below 100 rpm rotational speed is not detectable using conventional vibration. From this pint, this paper has presented a focused review of using acoustic emission techniques for monitoring bearings condition.

Keywords: condition monitoring, stress wave analysis, low-speed bearings, bearing defect diagnosis

Procedia PDF Downloads 287
3283 Intrusion Detection Based on Graph Oriented Big Data Analytics

Authors: Ahlem Abid, Farah Jemili

Abstract:

Intrusion detection has been the subject of numerous studies in industry and academia, but cyber security analysts always want greater precision and global threat analysis to secure their systems in cyberspace. To improve intrusion detection system, the visualisation of the security events in form of graphs and diagrams is important to improve the accuracy of alerts. In this paper, we propose an approach of an IDS based on cloud computing, big data technique and using a machine learning graph algorithm which can detect in real time different attacks as early as possible. We use the MAWILab intrusion detection dataset . We choose Microsoft Azure as a unified cloud environment to load our dataset on. We implement the k2 algorithm which is a graphical machine learning algorithm to classify attacks. Our system showed a good performance due to the graphical machine learning algorithm and spark structured streaming engine.

Keywords: Apache Spark Streaming, Graph, Intrusion detection, k2 algorithm, Machine Learning, MAWILab, Microsoft Azure Cloud

Procedia PDF Downloads 114
3282 Design of Quality Assessment System for On-Orbit 3D Printing Based on 3D Reconstruction Technology

Authors: Jianning Tang, Trevor Hocksun Kwan, Xiaofeng Wu

Abstract:

With the increasing demand for space use in multiple sectors (navigation, telecommunication, imagery, etc.), the deployment and maintenance demand of satellites are growing. Considering the high launching cost and the restrictions on weight and size of the payload when using launch vehicle, the technique of on-orbit manufacturing has obtained more attention because of its significant potential to support future space missions. 3D printing is the most promising manufacturing technology that could be applied in space. However, due to the lack of autonomous quality assessment, the operation of conventional 3D printers still relies on human presence to supervise the printing process. This paper is proposed to develop an automatic 3D reconstruction system aiming at detecting failures on the 3D printed objects through application of point cloud technology. Based on the data obtained from the point cloud, the 3D printer could locate the failure and repair the failure. The system will increase automation and provide 3D printing with more feasibilities for space use without human interference.

Keywords: 3D printing, quality assessment, point cloud, on-orbit manufacturing

Procedia PDF Downloads 91
3281 An Empirical Investigation of the Challenges of Secure Edge Computing Adoption in Organizations

Authors: Hailye Tekleselassie

Abstract:

Edge computing is a spread computing outline that transports initiative applications closer to data sources such as IoT devices or local edge servers, and possible happenstances would skull the action of new technologies. However, this investigation was attained to investigation the consciousness of technology and communications organization workers and computer users who support the service cloud. Surveys were used to achieve these objectives. Surveys were intended to attain these aims, and it is the functional using survey. Enquiries about confidence are also a key question. Problems like data privacy, integrity, and availability are the factors affecting the company’s acceptance of the service cloud.

Keywords: IoT, data, security, edge computing

Procedia PDF Downloads 59
3280 Risk Assessment of Natural Gas Pipelines in Coal Mined Gobs Based on Bow-Tie Model and Cloud Inference

Authors: Xiaobin Liang, Wei Liang, Laibin Zhang, Xiaoyan Guo

Abstract:

Pipelines pass through coal mined gobs inevitably in the mining area, the stability of which has great influence on the safety of pipelines. After extensive literature study and field research, it was found that there are a few risk assessment methods for coal mined gob pipelines, and there is a lack of data on the gob sites. Therefore, the fuzzy comprehensive evaluation method is widely used based on expert opinions. However, the subjective opinions or lack of experience of individual experts may lead to inaccurate evaluation results. Hence the accuracy of the results needs to be further improved. This paper presents a comprehensive approach to achieve this purpose by combining bow-tie model and cloud inference. The specific evaluation process is as follows: First, a bow-tie model composed of a fault tree and an event tree is established to graphically illustrate the probability and consequence indicators of pipeline failure. Second, the interval estimation method can be scored in the form of intervals to improve the accuracy of the results, and the censored mean algorithm is used to remove the maximum and minimum values of the score to improve the stability of the results. The golden section method is used to determine the weight of the indicators and reduce the subjectivity of index weights. Third, the failure probability and failure consequence scores of the pipeline are converted into three numerical features by using cloud inference. The cloud inference can better describe the ambiguity and volatility of the results which can better describe the volatility of the risk level. Finally, the cloud drop graphs of failure probability and failure consequences can be expressed, which intuitively and accurately illustrate the ambiguity and randomness of the results. A case study of a coal mine gob pipeline carrying natural gas has been investigated to validate the utility of the proposed method. The evaluation results of this case show that the probability of failure of the pipeline is very low, the consequences of failure are more serious, which is consistent with the reality.

Keywords: bow-tie model, natural gas pipeline, coal mine gob, cloud inference

Procedia PDF Downloads 221
3279 A Resource Optimization Strategy for CPU (Central Processing Unit) Intensive Applications

Authors: Junjie Peng, Jinbao Chen, Shuai Kong, Danxu Liu

Abstract:

On the basis of traditional resource allocation strategies, the usage of resources on physical servers in cloud data center is great uncertain. It will cause waste of resources if the assignment of tasks is not enough. On the contrary, it will cause overload if the assignment of tasks is too much. This is especially obvious when the applications are the same type because of its resource preferences. Considering CPU intensive application is one of the most common types of application in the cloud, we studied the optimization strategy for CPU intensive applications on the same server. We used resource preferences to analyze the case that multiple CPU intensive applications run simultaneously, and put forward a model which can predict the execution time for CPU intensive applications which run simultaneously. Based on the prediction model, we proposed the method to select the appropriate number of applications for a machine. Experiments show that the model can predict the execution time accurately for CPU intensive applications. To improve the execution efficiency of applications, we propose a scheduling model based on priority for CPU intensive applications. Extensive experiments verify the validity of the scheduling model.

Keywords: cloud computing, CPU intensive applications, resource optimization, strategy

Procedia PDF Downloads 256
3278 Internet of Things, Edge and Cloud Computing in Rock Mechanical Investigation for Underground Surveys

Authors: Esmael Makarian, Ayub Elyasi, Fatemeh Saberi, Olusegun Stanley Tomomewo

Abstract:

Rock mechanical investigation is one of the most crucial activities in underground operations, especially in surveys related to hydrocarbon exploration and production, geothermal reservoirs, energy storage, mining, and geotechnics. There is a wide range of traditional methods for driving, collecting, and analyzing rock mechanics data. However, these approaches may not be suitable or work perfectly in some situations, such as fractured zones. Cutting-edge technologies have been provided to solve and optimize the mentioned issues. Internet of Things (IoT), Edge, and Cloud Computing technologies (ECt & CCt, respectively) are among the most widely used and new artificial intelligence methods employed for geomechanical studies. IoT devices act as sensors and cameras for real-time monitoring and mechanical-geological data collection of rocks, such as temperature, movement, pressure, or stress levels. Structural integrity, especially for cap rocks within hydrocarbon systems, and rock mass behavior assessment, to further activities such as enhanced oil recovery (EOR) and underground gas storage (UGS), or to improve safety risk management (SRM) and potential hazards identification (P.H.I), are other benefits from IoT technologies. EC techniques can process, aggregate, and analyze data immediately collected by IoT on a real-time scale, providing detailed insights into the behavior of rocks in various situations (e.g., stress, temperature, and pressure), establishing patterns quickly, and detecting trends. Therefore, this state-of-the-art and useful technology can adopt autonomous systems in rock mechanical surveys, such as drilling and production (in hydrocarbon wells) or excavation (in mining and geotechnics industries). Besides, ECt allows all rock-related operations to be controlled remotely and enables operators to apply changes or make adjustments. It must be mentioned that this feature is very important in environmental goals. More often than not, rock mechanical studies consist of different data, such as laboratory tests, field operations, and indirect information like seismic or well-logging data. CCt provides a useful platform for storing and managing a great deal of volume and different information, which can be very useful in fractured zones. Additionally, CCt supplies powerful tools for predicting, modeling, and simulating rock mechanical information, especially in fractured zones within vast areas. Also, it is a suitable source for sharing extensive information on rock mechanics, such as the direction and size of fractures in a large oil field or mine. The comprehensive review findings demonstrate that digital transformation through integrated IoT, Edge, and Cloud solutions is revolutionizing traditional rock mechanical investigation. These advanced technologies have empowered real-time monitoring, predictive analysis, and data-driven decision-making, culminating in noteworthy enhancements in safety, efficiency, and sustainability. Therefore, by employing IoT, CCt, and ECt, underground operations have experienced a significant boost, allowing for timely and informed actions using real-time data insights. The successful implementation of IoT, CCt, and ECt has led to optimized and safer operations, optimized processes, and environmentally conscious approaches in underground geological endeavors.

Keywords: rock mechanical studies, internet of things, edge computing, cloud computing, underground surveys, geological operations

Procedia PDF Downloads 26
3277 Critical Evaluation of Groundwater Monitoring Networks for Machine Learning Applications

Authors: Pedro Martinez-Santos, Víctor Gómez-Escalonilla, Silvia Díaz-Alcaide, Esperanza Montero, Miguel Martín-Loeches

Abstract:

Groundwater monitoring networks are critical in evaluating the vulnerability of groundwater resources to depletion and contamination, both in space and time. Groundwater monitoring networks typically grow over decades, often in organic fashion, with relatively little overall planning. The groundwater monitoring networks in the Madrid area, Spain, were reviewed for the purpose of identifying gaps and opportunities for improvement. Spatial analysis reveals the presence of various monitoring networks belonging to different institutions, with several hundred observation wells in an area of approximately 4000 km2. This represents several thousand individual data entries, some going back to the early 1970s. Major issues included overlap between the networks, unknown screen depth/vertical distribution for many observation boreholes, uneven time series, uneven monitored species, and potentially suboptimal locations. Results also reveal there is sufficient information to carry out a spatial and temporal analysis of groundwater vulnerability based on machine learning applications. These can contribute to improve the overall planning of monitoring networks’ expansion into the future.

Keywords: groundwater monitoring, observation networks, machine learning, madrid

Procedia PDF Downloads 50
3276 Classification of Manufacturing Data for Efficient Processing on an Edge-Cloud Network

Authors: Onyedikachi Ulelu, Andrew P. Longstaff, Simon Fletcher, Simon Parkinson

Abstract:

The widespread interest in 'Industry 4.0' or 'digital manufacturing' has led to significant research requiring the acquisition of data from sensors, instruments, and machine signals. In-depth research then identifies methods of analysis of the massive amounts of data generated before and during manufacture to solve a particular problem. The ultimate goal is for industrial Internet of Things (IIoT) data to be processed automatically to assist with either visualisation or autonomous system decision-making. However, the collection and processing of data in an industrial environment come with a cost. Little research has been undertaken on how to specify optimally what data to capture, transmit, process, and store at various levels of an edge-cloud network. The first step in this specification is to categorise IIoT data for efficient and effective use. This paper proposes the required attributes and classification to take manufacturing digital data from various sources to determine the most suitable location for data processing on the edge-cloud network. The proposed classification framework will minimise overhead in terms of network bandwidth/cost and processing time of machine tool data via efficient decision making on which dataset should be processed at the ‘edge’ and what to send to a remote server (cloud). A fast-and-frugal heuristic method is implemented for this decision-making. The framework is tested using case studies from industrial machine tools for machine productivity and maintenance.

Keywords: data classification, decision making, edge computing, industrial IoT, industry 4.0

Procedia PDF Downloads 149
3275 Focusing of Technology Monitoring Activities Using Indicators

Authors: Günther Schuh, Christina König, Toni Drescher

Abstract:

One of the key factors for the competitiveness and market success of technology-driven companies is the timely provision of information about emerging technologies, changes in existing technologies, as well as relevant related changes in the market's structures and participants. Therefore, many companies conduct technology intelligence (TI) activities to ensure an early identification of appropriate technologies and other (weak) signals. One base activity of TI is technology monitoring, which is defined as the systematic tracking of developments within a specified topic of interest as well as related trends over a long period of time. Due to the very large number of dynamically changing parameters within the technological and the market environment of a company as well as their possible interdependencies, it is necessary to focus technology monitoring on specific indicators or other criteria, which are able to point out technological developments and market changes. In addition to the execution of a literature review on existing approaches, which mainly propose patent-based indicators, it is examined in this paper whether indicator systems from other branches such as risk management or economic research could be transferred to technology monitoring in order to enable an efficient and focused technology monitoring for companies.

Keywords: technology forecasting, technology indicator, technology intelligence, technology management, technology monitoring

Procedia PDF Downloads 445
3274 Modeling Metrics for Monitoring Software Project Performance Based on the GQM Model

Authors: Mariayee Doraisamy, Suhaimi bin Ibrahim, Mohd Naz’ri Mahrin

Abstract:

There are several methods to monitor software projects and the objective for monitoring is to ensure that the software projects are developed and delivered successfully. A performance measurement is a method that is closely associated with monitoring and it can be scrutinized by looking at two important attributes which are efficiency and effectiveness both of which are factors that are important for the success of a software project. Consequently, a successful steering is achieved by monitoring and controlling a software project via the performance measurement criteria and metrics. Hence, this paper is aimed at identifying the performance measurement criteria and the metrics for monitoring the performance of a software project by using the Goal Question Metrics (GQM) approach. The GQM approach is utilized to ensure that the identified metrics are reliable and useful. These identified metrics are useful guidelines for project managers to monitor the performance of their software projects.

Keywords: component, software project performance, goal question metrics, performance measurement criteria, metrics

Procedia PDF Downloads 320
3273 RFID Laptop Monitoring and Management System

Authors: Francis E. Idachaba, Sarah Uyimeh Tommy

Abstract:

This paper describes the design of an RFID laptop monitoring and management system. Laptops embedded with RFID chips are monitored and tracked to provide a monitoring system for the purpose of tracking as well as monitoring movement of the laptops in and out of a building. The proposed system is implemented with both hardware and software components. The hardware architecture consists of RFID passive tag, RFID module (reader), and a server hosting the application and database. The RFID readers are distributed at major exits of a building or premises. The tags are programmed with owner laptop details are concealed in the laptops. The software architecture consists of application software that has the APIs (Applications Programming Interface) necessary to interface the RFID system with the PC, to achieve automated laptop monitoring system. A friendly graphic user interface (GUI) and a database that saves all readings and owners details. The system is capable of reducing laptop theft especially in students’ hostels as laptops can be monitored as they are taken either in or out of the building.

Keywords: asset tracking, GUI, laptop monitoring, radio frequency identification, passive tags

Procedia PDF Downloads 360
3272 Day/Night Detector for Vehicle Tracking in Traffic Monitoring Systems

Authors: M. Taha, Hala H. Zayed, T. Nazmy, M. Khalifa

Abstract:

Recently, traffic monitoring has attracted the attention of computer vision researchers. Many algorithms have been developed to detect and track moving vehicles. In fact, vehicle tracking in daytime and in nighttime cannot be approached with the same techniques, due to the extreme different illumination conditions. Consequently, traffic-monitoring systems are in need of having a component to differentiate between daytime and nighttime scenes. In this paper, a HSV-based day/night detector is proposed for traffic monitoring scenes. The detector employs the hue-histogram and the value-histogram on the top half of the image frame. Experimental results show that the extraction of the brightness features along with the color features within the top region of the image is effective for classifying traffic scenes. In addition, the detector achieves high precision and recall rates along with it is feasible for real time applications.

Keywords: day/night detector, daytime/nighttime classification, image classification, vehicle tracking, traffic monitoring

Procedia PDF Downloads 530
3271 Optical Signal-To-Noise Ratio Monitoring Based on Delay Tap Sampling Using Artificial Neural Network

Authors: Feng Wang, Shencheng Ni, Shuying Han, Shanhong You

Abstract:

With the development of optical communication, optical performance monitoring (OPM) has received more and more attentions. Since optical signal-to-noise ratio (OSNR) is directly related to bit error rate (BER), it is one of the important parameters in optical networks. Recently, artificial neural network (ANN) has been greatly developed. ANN has strong learning and generalization ability. In this paper, a method of OSNR monitoring based on delay-tap sampling (DTS) and ANN has been proposed. DTS technique is used to extract the eigenvalues of the signal. Then, the eigenvalues are input into the ANN to realize the OSNR monitoring. The experiments of 10 Gb/s non-return-to-zero (NRZ) on–off keying (OOK), 20 Gb/s pulse amplitude modulation (PAM4) and 20 Gb/s return-to-zero (RZ) differential phase-shift keying (DPSK) systems are demonstrated for the OSNR monitoring based on the proposed method. The experimental results show that the range of OSNR monitoring is from 15 to 30 dB and the root-mean-square errors (RMSEs) for 10 Gb/s NRZ-OOK, 20 Gb/s PAM4 and 20 Gb/s RZ-DPSK systems are 0.36 dB, 0.45 dB and 0.48 dB respectively. The impact of chromatic dispersion (CD) on the accuracy of OSNR monitoring is also investigated in the three experimental systems mentioned above.

Keywords: artificial neural network (ANN), chromatic dispersion (CD), delay-tap sampling (DTS), optical signal-to-noise ratio (OSNR)

Procedia PDF Downloads 82
3270 Providing Reliability, Availability and Scalability Support for Quick Assist Technology Cryptography on the Cloud

Authors: Songwu Shen, Garrett Drysdale, Veerendranath Mannepalli, Qihua Dai, Yuan Wang, Yuli Chen, David Qian, Utkarsh Kakaiya

Abstract:

Hardware accelerator has been a promising solution to reduce the cost of cloud data centers. This paper investigates the QoS enhancement of the acceleration of an important datacenter workload: the webserver (or proxy) that faces high computational consumption originated from secure sockets layer (SSL) or transport layer security (TLS) procession in the cloud environment. Our study reveals that for the accelerator maintenance cases—need to upgrade driver/firmware or hardware reset due to hardware hang; we still can provide cryptography services by switching to software during maintenance phase and then switching back to accelerator after maintenance. The switching is seamless to server application such as Nginx that runs inside a VM on top of the server. To achieve this high availability goal, we propose a comprehensive fallback solution based on Intel® QuickAssist Technology (QAT). This approach introduces an architecture that involves the collaboration between physical function (PF) and virtual function (VF), and collaboration among VF, OpenSSL, and web application Nginx. The evaluation shows that our solution could provide high reliability, availability, and scalability (RAS) of hardware cryptography service in a 7x24x365 manner in the cloud environment.

Keywords: accelerator, cryptography service, RAS, secure sockets layer/transport layer security, SSL/TLS, virtualization fallback architecture

Procedia PDF Downloads 119
3269 Scouring Rate Pattern/Monitoring at Coastal and Offshore Structures

Authors: Ahmad Saifullah Mazlan, Hossein Basser, Shatirah Akib

Abstract:

Scouring pattern evaluation and measuring its depth around coastal and offshore structures is very essential issue to assure the safety of the structures as well as providing needed design parameters. Scouring is known as one of the important phenomena which threatens the safety of infrastructures. Several countermeasures have been developed to control scouring by protecting the structures against water flow attack directly or indirectly by changing the water flow pattern. Recently, monitoring methods for estimating water flow pattern and scour depth are studied to track the safety of structures. Since most of studies regarding scouring is related to monitoring scouring around piers in rivers therefore it is necessary to develop researches investigating scouring around piers in coastal and offshore areas. This paper describes a review of monitoring methods may be used for detecting scour depth around piers in coastal and offshore structures.

Keywords: scour, monitoring, pier, coastal, offshore

Procedia PDF Downloads 609
3268 Use Cloud-Based Watson Deep Learning Platform to Train Models Faster and More Accurate

Authors: Susan Diamond

Abstract:

Machine Learning workloads have traditionally been run in high-performance computing (HPC) environments, where users log in to dedicated machines and utilize the attached GPUs to run training jobs on huge datasets. Training of large neural network models is very resource intensive, and even after exploiting parallelism and accelerators such as GPUs, a single training job can still take days. Consequently, the cost of hardware is a barrier to entry. Even when upfront cost is not a concern, the lead time to set up such an HPC environment takes months from acquiring hardware to set up the hardware with the right set of firmware, software installed and configured. Furthermore, scalability is hard to achieve in a rigid traditional lab environment. Therefore, it is slow to react to the dynamic change in the artificial intelligent industry. Watson Deep Learning as a service, a cloud-based deep learning platform that mitigates the long lead time and high upfront investment in hardware. It enables robust and scalable sharing of resources among the teams in an organization. It is designed for on-demand cloud environments. Providing a similar user experience in a multi-tenant cloud environment comes with its own unique challenges regarding fault tolerance, performance, and security. Watson Deep Learning as a service tackles these challenges and present a deep learning stack for the cloud environments in a secure, scalable and fault-tolerant manner. It supports a wide range of deep-learning frameworks such as Tensorflow, PyTorch, Caffe, Torch, Theano, and MXNet etc. These frameworks reduce the effort and skillset required to design, train, and use deep learning models. Deep Learning as a service is used at IBM by AI researchers in areas including machine translation, computer vision, and healthcare. 

Keywords: deep learning, machine learning, cognitive computing, model training

Procedia PDF Downloads 181
3267 The Need for Multi-Edge Strategies and Solutions

Authors: Hugh Taylor

Abstract:

Industry analysts project that edge computing will be generating tens of billions in revenue in coming years. It’s not clear, however, if this will actually happen, and who, if anyone, will make it happen. Edge computing is seen as a critical success factor in industries ranging from telecom, enterprise IT and co-location. However, will any of these industries actually step up to make edge computing into a viable technology business? This paper looks at why the edge seems to be in a chasm, on the edge of realization, so to speak, but failing to coalesce into a coherent technology category like the cloud—and how the segment’s divergent industry players can come together to build a viable business at the edge.

Keywords: edge computing, multi-edge strategies, edge data centers, edge cloud

Procedia PDF Downloads 74
3266 Operating System Based Virtualization Models in Cloud Computing

Authors: Dev Ras Pandey, Bharat Mishra, S. K. Tripathi

Abstract:

Cloud computing is ready to transform the structure of businesses and learning through supplying the real-time applications and provide an immediate help for small to medium sized businesses. The ability to run a hypervisor inside a virtual machine is important feature of virtualization and it is called nested virtualization. In today’s growing field of information technology, many of the virtualization models are available, that provide a convenient approach to implement, but decision for a single model selection is difficult. This paper explains the applications of operating system based virtualization in cloud computing with an appropriate/suitable model with their different specifications and user’s requirements. In the present paper, most popular models are selected, and the selection was based on container and hypervisor based virtualization. Selected models were compared with a wide range of user’s requirements as number of CPUs, memory size, nested virtualization supports, live migration and commercial supports, etc. and we identified a most suitable model of virtualization.

Keywords: virtualization, OS based virtualization, container based virtualization, hypervisor based virtualization

Procedia PDF Downloads 277
3265 The Impact of Vertical Velocity Parameter Conditions and Its Relationship with Weather Parameters in the Hail Event

Authors: Nadine Ayasha

Abstract:

Hail happened in Sukabumi (August 23, 2020), Sekadau (August 22, 2020), and Bogor (September 23, 2020), where this extreme weather phenomenon occurred in the dry season. This study uses the ERA5 reanalysis model data, it aims to examine the vertical velocity impact on the hail occurrence in the dry season, as well as its relation to other weather parameters such as relative humidity, streamline, and wind velocity. Moreover, HCAI product satellite data is used as supporting data for the convective cloud development analysis. Based on the results of graphs, contours, and Hovmoller vertical cut from ERA5 modeling, the vertical velocity values in the 925 Mb-300 Mb layer in Sukabumi, Sekadau, and Bogor before the hail event ranged between -1.2-(-0.2), -1.5-(-0.2), -1-0 Pa/s. A negative value indicates that there is an upward motion from the air mass that trigger the convective cloud growth, which produces hail. It is evidenced by the presence of Cumulonimbus cloud on HCAI product when the hail falls. Therefore, the vertical velocity has significant effect on the hail event. In addition, the relative humidity in the 850-700 Mb layer is quite wet, which ranges from 80-90%. Meanwhile, the streamline and wind velocity in the three regions show the convergence with slowing wind velocity ranging from 2-4 knots. These results show that the upward motion of the vertical velocity is enough to form the wet atmospheric humidity and form a convergence for the growth of the convective cloud, which produce hail in the dry season.

Keywords: hail, extreme weather, vertical velocity, relative humidity, streamline

Procedia PDF Downloads 130
3264 Wireless Sensor Networks for Water Quality Monitoring: Prototype Design

Authors: Cesar Eduardo Hernández Curiel, Victor Hugo Benítez Baltazar, Jesús Horacio Pacheco Ramírez

Abstract:

This paper is devoted to present the advances in the design of a prototype that is able to supervise the complex behavior of water quality parameters such as pH and temperature, via a real-time monitoring system. The current water quality tests that are performed in government water quality institutions in Mexico are carried out in problematic locations and they require taking manual samples. The water samples are then taken to the institution laboratory for examination. In order to automate this process, a water quality monitoring system based on wireless sensor networks is proposed. The system consists of a sensor node which contains one pH sensor, one temperature sensor, a microcontroller, and a ZigBee radio, and a base station composed by a ZigBee radio and a PC. The progress in this investigation shows the development of a water quality monitoring system. Due to recent events that affected water quality in Mexico, the main motivation of this study is to address water quality monitoring systems, so in the near future, a more robust, affordable, and reliable system can be deployed.

Keywords: pH measurement, water quality monitoring, wireless sensor networks, ZigBee

Procedia PDF Downloads 367