Search results for: unreliable server
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 363

Search results for: unreliable server

123 Performance Comparison of Thread-Based and Event-Based Web Servers

Authors: Aikaterini Kentroti, Theodore H. Kaskalis

Abstract:

Today, web servers are expected to serve thousands of client requests concurrently within stringent response time limits. In this paper, we evaluate experimentally and compare the performance as well as the resource utilization of popular web servers, which differ in their approach to handle concurrency. More specifically, Central Processing Unit (CPU)- and I/O intensive tests were conducted against the thread-based Apache and Go as well as the event-based Nginx and Node.js under increasing concurrent load. The tests involved concurrent users requesting a term of the Fibonacci sequence (the 10th, 20th, 30th) and the content of a table from the database. The results show that Go achieved the best performance in all benchmark tests. For example, Go reached two times higher throughput than Node.js and five times higher than Apache and Nginx in the 20th Fibonacci term test. In addition, Go had the smallest memory footprint and demonstrated the most efficient resource utilization, in terms of CPU usage. Instead, Node.js had by far the largest memory footprint, consuming up to 90% more memory than Nginx and Apache. Regarding the performance of Apache and Nginx, our findings indicate that Hypertext Preprocessor (PHP) becomes a bottleneck when the servers are requested to respond by performing CPU-intensive tasks under increasing concurrent load.

Keywords: apache, Go, Nginx, node.js, web server benchmarking

Procedia PDF Downloads 63
122 Supply Chain of Energy Resources and Its Alternatives Due to the Arab Spring: The Case of Egyptian Natural Gas Flow to Jordan

Authors: Moh’d Anwer Al-Shboul

Abstract:

The year 2011 was a challenging year for Jordanian economy, which felt a variety of effects from the Arab Spring which took place in neighboring countries. Since February, 5th 2012, the Arab Gas Supply Pipeline, which carries natural gas from Egypt through the Sinai Peninsula and to Jordan and Israel, has been attacked more than 39 times. Jordan imported about 80 percent of its necessity of natural gas (about 250 million cubic feet of natural gas per day) from Egypt to generate particularly electricity, with the reminder of being produced locally. Jordan has utilized multiple alternatives to address the interruption of available natural gas supply from Egypt. The Jordanian distributed power plants now rely on the use of heavy fuel oil and diesel for electricity generation, in this case, it costs Jordan about four times than natural gas. The substitution of Egyptian natural gas supplies by fuel oil and diesel, coupled with the 32 percent rise in global fuel prices, has increased Jordan’s energy import bill by over 50 percent in 2011, reaching more than 16 percent of the 2011 GDP. The increase in the cost of electricity generation pushed the Jordanian economy to borrow from multiple internal and external resource channels, thus increasing the public debt. The Jordanian government’s short-term solution to the reduced natural gas supply from Egypt was alternatively purchasing the necessary quantities from some Gulf countries such as Qatar and/or Saudi Arabia, which can be imported with two possible methods. The first method is to rent a ship equipped with a liquefied natural gas (LNG) terminal, which is currently operating. The second method requires equipping the Aqaba port with an LNG terminal, which also currently is operating. In the long-term, a viable solution to depending on importing expensive and often unreliable natural gas supplies from surrounding countries is to depend more heavily on renewable supply energy, including solar, wind, and water energy.

Keywords: energy supply resources, Arab spring, liquefied natural gas, pipeline, Jordan

Procedia PDF Downloads 110
121 Comprehensive Review of Ultralightweight Security Protocols

Authors: Prashansa Singh, Manjot Kaur, Rohit Bajaj

Abstract:

The proliferation of wireless sensor networks and Internet of Things (IoT) devices in the quickly changing digital landscape has highlighted the urgent need for strong security solutions that can handle these systems’ limited resources. A key solution to this problem is the emergence of ultralightweight security protocols, which provide strong security features while respecting the strict computational, energy, and memory constraints imposed on these kinds of devices. This in-depth analysis explores the field of ultralightweight security protocols, offering a thorough examination of their evolution, salient features, and the particular security issues they resolve. We carefully examine and contrast different protocols, pointing out their advantages and disadvantages as well as the compromises between resource limitations and security resilience. We also study these protocols’ application domains, including the Internet of Things, RFID systems, and wireless sensor networks, to name a few. In addition, the review highlights recent developments and advancements in the field, pointing out new trends and possible avenues for future research. This paper aims to be a useful resource for researchers, practitioners, and developers, guiding the design and implementation of safe, effective, and scalable systems in the Internet of Things era by providing a comprehensive overview of ultralightweight security protocols.

Keywords: wireless sensor network, machine-to-machine, MQTT broker, server, ultralightweight, TCP/IP

Procedia PDF Downloads 23
120 Comparative Advantage of Mobile Agent Application in Procuring Software Products on the Internet

Authors: Michael K. Adu, Boniface K. Alese, Olumide S. Ogunnusi

Abstract:

This paper brings to fore the inherent advantages in application of mobile agents to procure software products rather than downloading software content on the Internet. It proposes a system whereby the products come on compact disk with mobile agent as deliverable. The client/user purchases a software product, but must connect to the remote server of the software developer before installation. The user provides an activation code that activates mobile agent which is part of the software product on compact disk. The validity of the activation code is checked on connection at the developer’s end to ascertain authenticity and prevent piracy. The system is implemented by downloading two different software products as compare with installing same products on compact disk with mobile agent’s application. Downloading software contents from developer’s database as in the traditional method requires a continuously open connection between the client and the developer’s end, a fixed network is not economically or technically feasible. Mobile agent after being dispatched into the network becomes independent of the creating process and can operate asynchronously and autonomously. It can reconnect later after completing its task and return for result delivery. Response Time and Network Load are very minimal with application of Mobile agent.

Keywords: software products, software developer, internet, activation code, mobile agent

Procedia PDF Downloads 267
119 Statistical Analysis with Prediction Models of User Satisfaction in Software Project Factors

Authors: Katawut Kaewbanjong

Abstract:

We analyzed a volume of data and found significant user satisfaction in software project factors. A statistical significance analysis (logistic regression) and collinearity analysis determined the significance factors from a group of 71 pre-defined factors from 191 software projects in ISBSG Release 12. The eight prediction models used for testing the prediction potential of these factors were Neural network, k-NN, Naïve Bayes, Random forest, Decision tree, Gradient boosted tree, linear regression and logistic regression prediction model. Fifteen pre-defined factors were truly significant in predicting user satisfaction, and they provided 82.71% prediction accuracy when used with a neural network prediction model. These factors were client-server, personnel changes, total defects delivered, project inactive time, industry sector, application type, development type, how methodology was acquired, development techniques, decision making process, intended market, size estimate approach, size estimate method, cost recording method, and effort estimate method. These findings may benefit software development managers considerably.

Keywords: prediction model, statistical analysis, software project, user satisfaction factor

Procedia PDF Downloads 89
118 Computer Assisted Learning Module (CALM) for Consumer Electronics Servicing

Authors: Edicio M. Faller

Abstract:

The use of technology in the delivery of teaching and learning is vital nowadays especially in education. Computer Assisted Learning Module (CALM) software is the use of computer in the delivery of instruction with a tailored fit program intended for a specific lesson or a set of topics. The CALM software developed in this study is intended to supplement the traditional teaching methods in technical-vocational (TECH-VOC) instruction specifically the Consumer Electronics Servicing course. There are three specific objectives of this study. First is to create a learning enhancement and review materials on the selected lessons. Second, is to computerize the end-of-chapter quizzes. Third, is to generate a computerized mock exam and summative assessment. In order to obtain the objectives of the study the researcher adopted the Agile Model where the development of the study undergoes iterative and incremental process of the Software Development Life Cycle. The study conducted an acceptance testing using a survey questionnaire to evaluate the CALM software. The results showed that CALM software was generally interpreted as very satisfactory. To further improve the CALM software it is recommended that the program be updated, enhanced and lastly, be converted from stand-alone to a client/server architecture.

Keywords: computer assisted learning module, software development life cycle, computerized mock exam, consumer electronics servicing

Procedia PDF Downloads 368
117 Approaches to Estimating the Radiation and Socio-Economic Consequences of the Fukushima Daiichi Nuclear Power Plant Accident Using the Data Available in the Public Domain

Authors: Dmitry Aron

Abstract:

Major radiation accidents carry not only the potential risks of negative consequences for public health due to exposure but also because of large-scale emergency measures were taken by authorities to protect the population, which can lead to unreasonable social and economic damage. It is technically difficult, as a rule, to assess the possible costs and damages from decisions on evacuation or resettlement of residents in the shortest possible time, since it requires specially prepared information systems containing relevant information on demographic, economic parameters and incoming data on radiation conditions. Foreign observers also face the difficulties in assessing the consequences of an accident in a foreign territory, since they usually do not have official and detailed statistical data on the territory of foreign state beforehand. Also, they can suppose the application of unofficial data from open Internet sources is an unreliable and overly labor-consuming procedure. This paper describes an approach to prompt creation of relational database that contains detailed actual data on economics, demographics and radiation situation at the Fukushima Prefecture during the Fukushima Daiichi NPP accident, received by the author from open Internet sources. This database was developed and used to assess the number of evacuated population, radiation doses, expected financial losses and other parameters of the affected areas. The costs for the areas with temporarily evacuated and long-term resettled population were investigated, and the radiological and economic effectiveness of the measures taken to protect the population was estimated. Some of the results are presented in the article. The study showed that such a tool for analyzing the consequences of radiation accidents can be prepared in a short space of time for the entire territory of Japan, and it can serve for the modeling of social and economic consequences for hypothetical accidents for any nuclear power plant in its territory.

Keywords: Fukushima, radiation accident, emergency measures, database

Procedia PDF Downloads 159
116 CanVis: Towards a Web Platform for Cancer Progression Tree Analysis

Authors: Michael Aupetit, Mahmoud Al-ismail, Khaled Mohamed

Abstract:

Cancer is a major public health problem all over the world. Breast cancer has the highest incidence rate over all cancers for women in Qatar making its study a top priority of the country. Human cancer is a dynamic disease that develops over an extended period through the accumulation of a series of genetic alterations. A Darwinian process drives the tumor cells toward higher malignancy growing the branches of a progression tree in the space of genes expression. Although it is not possible to track these genetic alterations dynamically for one patient, it is possible to reconstruct the progression tree from the aggregation of thousands of tumor cells’ genetic profiles from thousands of different patients at different stages of the disease. Analyzing the progression tree is a way to detect pivotal molecular events that drive the malignant evolution and to provide a guide for the development of cancer diagnostics, prognostics and targeted therapeutics. In this work we present the development of a Visual Analytic web platform CanVis enabling users to upload gene-expression data and analyze their progression tree. The server computes the progression tree based on state-of-the-art techniques and allows an interactive visual exploration of this tree and the gene-expression data along its branching structure helping to discover potential driver genes.

Keywords: breast cancer, progression tree, visual analytics, web platform

Procedia PDF Downloads 384
115 Health Economics in the Cost-Benefit Analysis of Transport Schemes

Authors: Henry Kelly, Helena Shaw

Abstract:

This paper will seek how innovative methods from Health Economics and, to a lesser extent, wellbeing analysis can be applied in the Cost-Benefit Analysis (CBA) of transport infrastructure and policy interventions. The context for this will focus on the framework articulated by the UK Treasury (finance department) and the English Department for Transport. Both have well-established methods for undertaking CBA, but there is increased policy interest, particularly at a regional level of exploring broader strategic goals beyond those traditionally associated with transport user benefits, productivity gains, and labour market access. Links to different CBA approaches internationally, such as New Zealand, France, and Wales will be referenced. By exploring a complementary method of accessing the impacts of policies through the quantification of health impacts is a fruitful line to explore. In a previous piece of work, 14 impact pathways were identified, mapping the relationship between transport and health. These are wide-ranging, from improved employment prospects, the stress of unreliable journey times, and air quality to isolation and loneliness. Importantly, we will consider these different measures of health from an intersectional point of view to ensure that the basis that remains in the health industry does not get translated across to this work. The objective is to explore how a CBA based on these pathways may, through quantifying forecast impacts in terms of Quality-Adjusted Life Years may, produce different findings than a standard approach. Of particular interest is how a health-based approach may have different distributional impacts on socio-economic groups and may favour distinct types of interventions. Consideration will be given to the degree this approach may double-count impacts or if it is possible to identify additional benefits to the established CBA approach. The investigation will explore a range of schemes, from a high-speed rail link, highway improvements, rural mobility hubs, and coach services to cycle lanes. The conclusions should aid the progression of methods concerning the assessment of publicly funded infrastructure projects.

Keywords: cost-benefit analysis, health, QALYs transport

Procedia PDF Downloads 57
114 Study on Resource Allocation of Cloud Operating System Based on Multi-Tenant Data Resource Sharing Technology

Authors: Lin Yunuo, Seow Xing Quan, Burra Venkata Durga Kumar

Abstract:

In this modern era, the cloud operating system is the world trend applied in various industries such as business, healthy, etc. In order to deal with the large capacity of requirements in cloud computing, research come up with multi-tenant cloud computing to maximize the benefits of server providers and clients. However, there are still issues in multi-tenant cloud computing especially regarding resource allocation. Issues such as inefficient resource utilization, large latency, lack of scalability and elasticity and poor data isolation had caused inefficient resource allocation in multi-tenant cloud computing. Without a doubt, these issues prevent multitenancy reaches its best condition. In fact, there are multiple studies conducted to determine the optimal resource allocation to solve these problems these days. This article will briefly introduce the cloud operating system, Multi-tenant cloud computing and resource allocation in cloud computing. It then discusses resource allocation in multi-tenant cloud computing and the current challenges it faces. According to the issue ‘ineffective resource utilization’, we will discuss an efficient dynamic scheduling technique for multitenancy, namely Multi-tenant Dynamic Resource Scheduling Model (MTDRSM). Moreover, there also have some recommendations to improve the shortcoming of this model in this paper’s final section.

Keywords: cloud computing, cloud operation system, multitenancy, resource allocation, utilization of cloud resources

Procedia PDF Downloads 59
113 Using Surface Entropy Reduction to Improve the Crystallization Properties of a Recombinant Antibody Fragment RNA Crystallization Chaperone

Authors: Christina Roman, Deepak Koirala, Joseph A. Piccirilli

Abstract:

Phage displaying synthetic Fab libraries have been used to obtain Fabs that bind to specific RNA targets with high affinity and specificity. These Fabs have been demonstrated to facilitate RNA crystallization. However, the antibody framework used in the construction of these phage display libraries contains numerous bulky, flexible, and charged residues, which facilitate solubility and hinder aggregation. These residues can interfere with crystallization due to the entropic cost associated with burying them within crystal contacts. To systematically reduce the surface entropy of the Fabs and improve their crystallization properties, a protein engineering strategy termed surface entropy reduction (SER) is being applied to the Fab framework. In this approach, high entropy residues are mutated to smaller ones such as alanine or serine. Focusing initially on Fab BL3-6, which binds an RNA AAACA pentaloop with 20nM affinity, the SER P server (http://services.mbi.ucla.edu/SER/) was used and analysis was performed on existing RNA-Fab BL3-6 co-crystal structures. From this analysis twelve surface entropy reduced mutants were designed. These SER mutants were expressed and are now being measured for their crystallization and diffraction performance with various RNA targets. So far, one mutant has generated 3.02 angstrom diffraction with the yjdF riboswitch RNA. Ultimately, the most productive mutations will be combined into a new Fab framework to be used in a optimized phage displayed Fab library.

Keywords: antibody fragment, crystallography, RNA, surface entropy reduction

Procedia PDF Downloads 153
112 Lovely, Lyrical, Lilting: Kubrick’s Translation of Lolita’s Voice

Authors: Taylor La Carriere

Abstract:

“What I had madly possessed was not she, but my own creation, another, fanciful Lolita perhaps, more real than Lolita; overlapping, encasing he and having no will, no consciousness indeed, no life of her own,” Vladimir Nabokov writes in his seminal work, Lolita. Throughout Nabokov’s novel, the eponymous character is rendered nonexistent through unreliable narrator Humbert Humbert’s impenetrable narrative, infused with lyrical rationalization. Instead, Lolita is “safely solipsised,” as Humbert muses, solidifying the potential for the erasure of Lolita’s agency and identity. In this literary work, Lolita’s voice is reduced to a nearly invisible presence, only seen through the eyes of her captor. However, in Stanley Kubrick’s film adaptation of Lolita (1962), the “nymphet,” as Nabokov coins, reemerges with a voice of her own, fueled by a lyric impulse, that displaces Humbert’s first-person narration. The lyric, as defined by Catherine Ing, is the voice of the invisible; it is also characterized by performance, the concentrated utterance of individual emotion, and the appearance of spontaneity. The novel’s lyricism is largely in the service of Humbert’s “seductive” voice, while the film reorients it more to Lolita’s subjectivity. Through a close analysis of Kubrick’s cinematic techniques, this paper examines the emergence and translation of Lolita’s voice in contrast with Humbert’s attempts to silence her in Nabokov’s Lolita, hypothesizing that Kubrick translates Lolita’s presence into a visual and aural voice with lyrical attributes, exemplified through the establishment of an altered power dynamic, Sue Lyon’s transformative performance as the titular character, Nelson Riddle and Bob Harris’ musical score, and the omission of Humbert’s first-person point-of-view. In doing so, the film reclaims Lolita’s agency by taking instances of Lolita’s voice in the novel as depicted in the last half of the work and expanding upon them in a way only cinematic depictions could allow. The results of this study suggest that Lolita’s voice in Kubrick’s adaptation functions without disrupting the lyricism present in Nabokov’s source text, materializing through the actions, expressions, and performance of Sue Lyon in the film. This voice, fueled by a lyric impulse of its own, refutes the silence bestowed upon the titular character and enables its ultimate reclamation upon the silver screen.

Keywords: cinema, adaptation, Lolita, lyric voice

Procedia PDF Downloads 165
111 On the Design of a Secure Two-Party Authentication Scheme for Internet of Things Using Cancelable Biometrics and Physically Unclonable Functions

Authors: Behnam Zahednejad, Saeed Kosari

Abstract:

Widespread deployment of Internet of Things (IoT) has raised security and privacy issues in this environment. Designing a secure two-factor authentication scheme between the user and server is still a challenging task. In this paper, we focus on Cancelable Biometric (CB) as an authentication factor in IoT. We show that previous CB-based scheme fail to provide real two-factor security, Perfect Forward Secrecy (PFS) and suffer database attacks and traceability of the user. Then we propose our improved scheme based on CB and Physically Unclonable Functions (PUF), which can provide real two-factor security, PFS, user’s unlinkability, and resistance to database attack. In addition, Key Compromise Impersonation (KCI) resilience is achieved in our scheme. We also prove the security of our proposed scheme formally using both Real-Or-Random (RoR) model and the ProVerif analysis tool. For the usability of our scheme, we conducted a performance analysis and showed that our scheme has the least communication cost compared to the previous CB-based scheme. The computational cost of our scheme is also acceptable for the IoT environment.

Keywords: IoT, two-factor security, cancelable biometric, key compromise impersonation resilience, perfect forward secrecy, database attack, real-or-random model, ProVerif

Procedia PDF Downloads 71
110 Synthetic Method of Contextual Knowledge Extraction

Authors: Olga Kononova, Sergey Lyapin

Abstract:

Global information society requirements are transparency and reliability of data, as well as ability to manage information resources independently; particularly to search, to analyze, to evaluate information, thereby obtaining new expertise. Moreover, it is satisfying the society information needs that increases the efficiency of the enterprise management and public administration. The study of structurally organized thematic and semantic contexts of different types, automatically extracted from unstructured data, is one of the important tasks for the application of information technologies in education, science, culture, governance and business. The objectives of this study are the contextual knowledge typologization, selection or creation of effective tools for extracting and analyzing contextual knowledge. Explication of various kinds and forms of the contextual knowledge involves the development and use full-text search information systems. For the implementation purposes, the authors use an e-library 'Humanitariana' services such as the contextual search, different types of queries (paragraph-oriented query, frequency-ranked query), automatic extraction of knowledge from the scientific texts. The multifunctional e-library «Humanitariana» is realized in the Internet-architecture in WWS-configuration (Web-browser / Web-server / SQL-server). Advantage of use 'Humanitariana' is in the possibility of combining the resources of several organizations. Scholars and research groups may work in a local network mode and in distributed IT environments with ability to appeal to resources of any participating organizations servers. Paper discusses some specific cases of the contextual knowledge explication with the use of the e-library services and focuses on possibilities of new types of the contextual knowledge. Experimental research base are science texts about 'e-government' and 'computer games'. An analysis of the subject-themed texts trends allowed to propose the content analysis methodology, that combines a full-text search with automatic construction of 'terminogramma' and expert analysis of the selected contexts. 'Terminogramma' is made out as a table that contains a column with a frequency-ranked list of words (nouns), as well as columns with an indication of the absolute frequency (number) and the relative frequency of occurrence of the word (in %% ppm). The analysis of 'e-government' materials showed, that the state takes a dominant position in the processes of the electronic interaction between the authorities and society in modern Russia. The media credited the main role in these processes to the government, which provided public services through specialized portals. Factor analysis revealed two factors statistically describing the used terms: human interaction (the user) and the state (government, processes organizer); interaction management (public officer, processes performer) and technology (infrastructure). Isolation of these factors will lead to changes in the model of electronic interaction between government and society. In this study, the dominant social problems and the prevalence of different categories of subjects of computer gaming in science papers from 2005 to 2015 were identified. Therefore, there is an evident identification of several types of contextual knowledge: micro context; macro context; dynamic context; thematic collection of queries (interactive contextual knowledge expanding a composition of e-library information resources); multimodal context (functional integration of iconographic and full-text resources through hybrid quasi-semantic algorithm of search). Further studies can be pursued both in terms of expanding the resource base on which they are held, and in terms of the development of appropriate tools.

Keywords: contextual knowledge, contextual search, e-library services, frequency-ranked query, paragraph-oriented query, technologies of the contextual knowledge extraction

Procedia PDF Downloads 325
109 Mining the Proteome of Fusobacterium nucleatum for Potential Therapeutics Discovery

Authors: Abdul Musaweer Habib, Habibul Hasan Mazumder, Saiful Islam, Sohel Sikder, Omar Faruk Sikder

Abstract:

The plethora of genome sequence information of bacteria in recent times has ushered in many novel strategies for antibacterial drug discovery and facilitated medical science to take up the challenge of the increasing resistance of pathogenic bacteria to current antibiotics. In this study, we adopted subtractive genomics approach to analyze the whole genome sequence of the Fusobacterium nucleatum, a human oral pathogen having association with colorectal cancer. Our study divulged 1499 proteins of Fusobacterium nucleatum, which has no homolog in human genome. These proteins were subjected to screening further by using the Database of Essential Genes (DEG) that resulted in the identification of 32 vitally important proteins for the bacterium. Subsequent analysis of the identified pivotal proteins, using the KEGG Automated Annotation Server (KAAS) resulted in sorting 3 key enzymes of F. nucleatum that may be good candidates as potential drug targets, since they are unique for the bacterium and absent in humans. In addition, we have demonstrated the 3-D structure of these three proteins. Finally, determination of ligand binding sites of the key proteins as well as screening for functional inhibitors that best fitted with the ligands sites were conducted to discover effective novel therapeutic compounds against Fusobacterium nucleatum.

Keywords: colorectal cancer, drug target, Fusobacterium nucleatum, homology modeling, ligands

Procedia PDF Downloads 357
108 Design and Realization of Computer Network Security Perception Control System

Authors: El Miloudi Djelloul

Abstract:

Based on analysis on applications by perception control technology in computer network security status and security protection measures, from the angles of network physical environment and network software system environmental security, this paper provides network security system perception control solution using Internet of Things (IOT), telecom and other perception technologies. Security Perception Control System is in the computer network environment, utilizing Radio Frequency Identification (RFID) of IOT and telecom integration technology to carry out integration design for systems. In the network physical security environment, RFID temperature, humidity, gas and perception technologies are used to do surveillance on environmental data, dynamic perception technology is used for network system security environment, user-defined security parameters, security log are used for quick data analysis, extends control on I/O interface, by development of API and AT command, Computer Network Security Perception Control based on Internet and GSM/GPRS is achieved, which enables users to carry out interactive perception and control for network security environment by WEB, E-MAIL as well as PDA, mobile phone short message and Internet. In the system testing, through middle ware server, security information data perception in real time with deviation of 3-5% was achieved; it proves the feasibility of Computer Network Security Perception Control System.

Keywords: computer network, perception control system security strategy, Radio Frequency Identification (RFID)

Procedia PDF Downloads 412
107 Comparison of Authentication Methods in Internet of Things Technology

Authors: Hafizah Che Hasan, Fateen Nazwa Yusof, Maslina Daud

Abstract:

Internet of Things (IoT) is a powerful industry system, which end-devices are interconnected and automated, allowing the devices to analyze data and execute actions based on the analysis. The IoT technology leverages the technology of Radio-Frequency Identification (RFID) and Wireless Sensor Network (WSN), including mobile and sensor. These technologies contribute to the evolution of IoT. However, due to more devices are connected each other in the Internet, and data from various sources exchanged between things, confidentiality of the data becomes a major concern. This paper focuses on one of the major challenges in IoT; authentication, in order to preserve data integrity and confidentiality are in place. A few solutions are reviewed based on papers from the last few years. One of the proposed solutions is securing the communication between IoT devices and cloud servers with Elliptic Curve Cryptograhpy (ECC) based mutual authentication protocol. This solution focuses on Hyper Text Transfer Protocol (HTTP) cookies as security parameter.  Next proposed solution is using keyed-hash scheme protocol to enable IoT devices to authenticate each other without the presence of a central control server. Another proposed solution uses Physical Unclonable Function (PUF) based mutual authentication protocol. It emphasizes on tamper resistant and resource-efficient technology, which equals a 3-way handshake security protocol.

Keywords: Internet of Things (IoT), authentication, PUF ECC, keyed-hash scheme protocol

Procedia PDF Downloads 230
106 Conceptual Model of a Residential Waste Collection System Using ARENA Software

Authors: Bruce G. Wilson

Abstract:

The collection of municipal solid waste at the curbside is a complex operation that is repeated daily under varying circumstances around the world. There have been several attempts to develop Monte Carlo simulation models of the waste collection process dating back almost 50 years. Despite this long history, the use of simulation modeling as a planning or optimization tool for waste collection is still extremely limited in practice. Historically, simulation modeling of waste collection systems has been hampered by the limitations of computer hardware and software and by the availability of representative input data. This paper outlines the development of a Monte Carlo simulation model that overcomes many of the limitations contained in previous models. The model uses a general purpose simulation software program that is easily capable of modeling an entire waste collection network. The model treats the stops on a waste collection route as a queue of work to be processed by a collection vehicle (or server). Input data can be collected from a variety of sources including municipal geographic information systems, global positioning system recorders on collection vehicles, and weigh scales at transfer stations or treatment facilities. The result is a flexible model that is sufficiently robust that it can model the collection activities in a large municipality, while providing the flexibility to adapt to changing conditions on the collection route.

Keywords: modeling, queues, residential waste collection, Monte Carlo simulation

Procedia PDF Downloads 375
105 To Ensure Maximum Voter Privacy in E-Voting Using Blockchain, Convolutional Neural Network, and Quantum Key Distribution

Authors: Bhaumik Tyagi, Mandeep Kaur, Kanika Singla

Abstract:

The advancement of blockchain has facilitated scholars to remodel e-voting systems for future generations. Server-side attacks like SQL injection attacks and DOS attacks are the most common attacks nowadays, where malicious codes are injected into the system through user input fields by illicit users, which leads to data leakage in the worst scenarios. Besides, quantum attacks are also there which manipulate the transactional data. In order to deal with all the above-mentioned attacks, integration of blockchain, convolutional neural network (CNN), and Quantum Key Distribution is done in this very research. The utilization of blockchain technology in e-voting applications is not a novel concept. But privacy and security issues are still there in a public and private blockchains. To solve this, the use of a hybrid blockchain is done in this research. This research proposed cryptographic signatures and blockchain algorithms to validate the origin and integrity of the votes. The convolutional neural network (CNN), a normalized version of the multilayer perceptron, is also applied in the system to analyze visual descriptions upon registration in a direction to enhance the privacy of voters and the e-voting system. Quantum Key Distribution is being implemented in order to secure a blockchain-based e-voting system from quantum attacks using quantum algorithms. Implementation of e-voting blockchain D-app and providing a proposed solution for the privacy of voters in e-voting using Blockchain, CNN, and Quantum Key Distribution is done.

Keywords: hybrid blockchain, secure e-voting system, convolutional neural networks, quantum key distribution, one-time pad

Procedia PDF Downloads 54
104 Routing Medical Images with Tabu Search and Simulated Annealing: A Study on Quality of Service

Authors: Mejía M. Paula, Ramírez L. Leonardo, Puerta A. Gabriel

Abstract:

In telemedicine, the image repository service is important to increase the accuracy of diagnostic support of medical personnel. This study makes comparison between two routing algorithms regarding the quality of service (QoS), to be able to analyze the optimal performance at the time of loading and/or downloading of medical images. This study focused on comparing the performance of Tabu Search with other heuristic and metaheuristic algorithms that improve QoS in telemedicine services in Colombia. For this, Tabu Search and Simulated Annealing heuristic algorithms are chosen for their high usability in this type of applications; the QoS is measured taking into account the following metrics: Delay, Throughput, Jitter and Latency. In addition, routing tests were carried out on ten images in digital image and communication in medicine (DICOM) format of 40 MB. These tests were carried out for ten minutes with different traffic conditions, reaching a total of 25 tests, from a server of Universidad Militar Nueva Granada (UMNG) in Bogotá-Colombia to a remote user in Universidad de Santiago de Chile (USACH) - Chile. The results show that Tabu search presents a better QoS performance compared to Simulated Annealing, managing to optimize the routing of medical images, a basic requirement to offer diagnostic images services in telemedicine.

Keywords: medical image, QoS, simulated annealing, Tabu search, telemedicine

Procedia PDF Downloads 189
103 Prioritized Processor-Sharing with a Maximum Permissible Sojourn Time

Authors: Yoshiaki Shikata

Abstract:

A prioritized processor-sharing (PS) system with a maximum permissible sojourn time (MPST) is proposed. In this PS system, a higher-priority request is allocated a larger service ratio than a lower-priority request. Moreover, each request receiving service is guaranteed the maximum permissible sojourn time determined by each priority class, regardless of its service time. Arriving requests that cannot receive service due to this guarantee are rejected. We further propose a guarantee method for implementing such a system, and discuss performance evaluation procedures for the resulting system. Practical performance measures, such as the relationships between the loss probability or mean sojourn time of each class request and the maximum permissible sojourn time are evaluated via simulation. At the arrival of each class request, its acceptance or rejection is judged using extended sojourn times of all requests receiving service in the server. As the MPST increases, the mean sojourn time increases almost linearly. However, the logarithm of the loss probability decreases almost linearly. Moreover with an MPST, the difference in the mean sojourn time for different MPSTs increases with the traffic rate. Conversely, the difference in the loss probability for different MPSTs decreases as the traffic rate increases.

Keywords: prioritized processor sharing, priority ratio, permissible sojourn time, loss probability, mean sojourn time, simulation

Procedia PDF Downloads 164
102 A Bi-Objective Model to Optimize the Total Time and Idle Probability for Facility Location Problem Behaving as M/M/1/K Queues

Authors: Amirhossein Chambari

Abstract:

This article proposes a bi-objective model for the facility location problem subject to congestion (overcrowding). Motivated by implementations to locate servers in internet mirror sites, communication networks, one-server-systems, so on. This model consider for situations in which immobile (or fixed) service facilities are congested (or queued) by stochastic demand to behave as M/M/1/K queues. We consider for this problem two simultaneous perspectives; (1) Customers (desire to limit times of accessing and waiting for service) and (2) Service provider (desire to limit average facility idle-time). A bi-objective model is setup for facility location problem with two objective functions; (1) Minimizing sum of expected total traveling and waiting time (customers) and (2) Minimizing the average facility idle-time percentage (service provider). The proposed model belongs to the class of mixed-integer nonlinear programming models and the class of NP-hard problems. In addition, to solve the model, controlled elitist non-dominated sorting genetic algorithms (Controlled NSGA-II) and controlled elitist non-dominated ranking genetic algorithms (NRGA-I) are proposed. Furthermore, the two proposed metaheuristics algorithms are evaluated by establishing standard multiobjective metrics. Finally, the results are analyzed and some conclusions are given.

Keywords: bi-objective, facility location, queueing, controlled NSGA-II, NRGA-I

Procedia PDF Downloads 551
101 Reconfigurable Device for 3D Visualization of Three Dimensional Surfaces

Authors: Robson da C. Santos, Carlos Henrique de A. S. P. Coutinho, Lucas Moreira Dias, Gerson Gomes Cunha

Abstract:

The article refers to the development of an augmented reality 3D display, through the control of servo motors and projection of image with aid of video projector on the model. Augmented Reality is a branch that explores multiple approaches to increase real-world view by viewing additional information along with the real scene. The article presents the broad use of electrical, electronic, mechanical and industrial automation for geospatial visualizations, applications in mathematical models with the visualization of functions and 3D surface graphics and volumetric rendering that are currently seen in 2D layers. Application as a 3D display for representation and visualization of Digital Terrain Model (DTM) and Digital Surface Models (DSM), where it can be applied in the identification of canyons in the marine area of the Campos Basin, Rio de Janeiro, Brazil. The same can execute visualization of regions subject to landslides, as in Serra do Mar - Agra dos Reis and Serranas cities both in the State of Rio de Janeiro. From the foregoing, loss of human life and leakage of oil from pipelines buried in these regions may be anticipated in advance. The physical design consists of a table consisting of a 9 x 16 matrix of servo motors, totalizing 144 servos, a mesh is used on the servo motors for visualization of the models projected by a retro projector. Each model for by an image pre-processing, is sent to a server to be converted and viewed from a software developed in C # Programming Language.

Keywords: visualization, 3D models, servo motors, C# programming language

Procedia PDF Downloads 306
100 DEEPMOTILE: Motility Analysis of Human Spermatozoa Using Deep Learning in Sri Lankan Population

Authors: Chamika Chiran Perera, Dananjaya Perera, Chirath Dasanayake, Banuka Athuraliya

Abstract:

Male infertility is a major problem in the world, and it is a neglected and sensitive health issue in Sri Lanka. It can be determined by analyzing human semen samples. Sperm motility is one of many factors that can evaluate male’s fertility potential. In Sri Lanka, this analysis is performed manually. Manual methods are time consuming and depend on the person, but they are reliable and it can depend on the expert. Machine learning and deep learning technologies are currently being investigated to automate the spermatozoa motility analysis, and these methods are unreliable. These automatic methods tend to produce false positive results and false detection. Current automatic methods support different techniques, and some of them are very expensive. Due to the geographical variance in spermatozoa characteristics, current automatic methods are not reliable for motility analysis in Sri Lanka. The suggested system, DeepMotile, is to explore a method to analyze motility of human spermatozoa automatically and present it to the andrology laboratories to overcome current issues. DeepMotile is a novel deep learning method for analyzing spermatozoa motility parameters in the Sri Lankan population. To implement the current approach, Sri Lanka patient data were collected anonymously as a dataset, and glass slides were used as a low-cost technique to analyze semen samples. Current problem was identified as microscopic object detection and tackling the problem. YOLOv5 was customized and used as the object detector, and it achieved 94 % mAP (mean average precision), 86% Precision, and 90% Recall with the gathered dataset. StrongSORT was used as the object tracker, and it was validated with andrology experts due to the unavailability of annotated ground truth data. Furthermore, this research has identified many potential ways for further investigation, and andrology experts can use this system to analyze motility parameters with realistic accuracy.

Keywords: computer vision, deep learning, convolutional neural networks, multi-target tracking, microscopic object detection and tracking, male infertility detection, motility analysis of human spermatozoa

Procedia PDF Downloads 69
99 TACTICAL: Ram Image Retrieval in Linux Using Protected Mode Architecture’s Paging Technique

Authors: Sedat Aktas, Egemen Ulusoy, Remzi Yildirim

Abstract:

This article explains how to get a ram image from a computer with a Linux operating system and what steps should be followed while getting it. What we mean by taking a ram image is the process of dumping the physical memory instantly and writing it to a file. This process can be likened to taking a picture of everything in the computer’s memory at that moment. This process is very important for tools that analyze ram images. Volatility can be given as an example because before these tools can analyze ram, images must be taken. These tools are used extensively in the forensic world. Forensic, on the other hand, is a set of processes for digitally examining the information on any computer or server on behalf of official authorities. In this article, the protected mode architecture in the Linux operating system is examined, and the way to save the image sample of the kernel driver and system memory to disk is followed. Tables and access methods to be used in the operating system are examined based on the basic architecture of the operating system, and the most appropriate methods and application methods are transferred to the article. Since there is no article directly related to this study on Linux in the literature, it is aimed to contribute to the literature with this study on obtaining ram images. LIME can be mentioned as a similar tool, but there is no explanation about the memory dumping method of this tool. Considering the frequency of use of these tools, the contribution of the study in the field of forensic medicine has been the main motivation of the study due to the intense studies on ram image in the field of forensics.

Keywords: linux, paging, addressing, ram-image, memory dumping, kernel modules, forensic

Procedia PDF Downloads 77
98 The Implementation of Sexual and Reproductive Health Education Policy in Schools in Asia and Africa: A Scoping Review

Authors: Rhea Khosla, Victoria Tzortziou-Brown

Abstract:

Introduction: Adolescent SRH has been neglected since the start of the millennium. Adolescents comprise 16% of the global population, with the largest proportion living in Asia (650 million). By late adolescence, individuals in these regions are likely to become sexually active, and thus they must understand their SRH rights. Many lack knowledge of SRH, using unreliable sources for such information. Sex education is necessary to standardize and inform sexual knowledge, which empowers adolescents to make informed SRH decisions. School is an appropriate environment for this, however, SRH education requires effective policy to enforce. Nonetheless, this issue remains of low political priority in Asia and Africa. Current literature on sex education policy in schools in these regions is scarce and tends to have broad aims. Thus, a scoping review was necessary. Methods: Literature searches were conducted in February 2023 using six databases, including grey literature databases (PubMed, Scopus, Embase, Web of Science, Google Scholar, Global Index Medicus), returning a total of 1537 unique articles. After screening titles, abstracts and full text, 17 articles remained. References of included articles were additionally searched, producing a further 7 articles, which then underwent thematic analysis Results: Most countries in Africa and Asia did not have studies on this topic. Studies derived data from interviews with key stakeholders and quantitative methods quantified questionnaire responses. Barriers were: policy/curriculum issues, societal opinions, teaching discomfort, and lack of educator training. Limitations were insufficient timing, inconsistent implementation, insufficient hours dedicated to teaching, education received late into schooling, and discrepancies between teachers, schools, and students about whether policies were being implemented. Discussion: Based on the existing limited evidence, a cultural shift to reduce stigma seems necessary, alongside teacher and student involvement in policy formulation with effective implementation monitoring and educator training.

Keywords: adolescent, Africa, Asia, education, sexual and reproductive health, policy

Procedia PDF Downloads 25
97 IoT Based Agriculture Monitoring Framework for Sustainable Rice Production

Authors: Armanul Hoque Shaon, Md Baizid Mahmud, Askander Nobi, Md. Raju Ahmed, Md. Jiabul Hoque

Abstract:

In the Internet of Things (IoT), devices are linked to the internet through a wireless network, allowing them to collect and transmit data without the need for a human operator. Agriculture relies heavily on wireless sensors, which are a vital component of the Internet of Things (IoT). This kind of wireless sensor network monitors physical or environmental variables like temperatures, sound, vibration, pressure, or motion without relying on a central location or sink and collaboratively passes its data across the network to be analyzed. As the primary source of plant nutrients, the soil is critical to the agricultural industry's continued growth. We're excited about the prospect of developing an Internet of Things (IoT) solution. To arrange the network, the sink node collects groundwater levels and sends them to the Gateway, which centralizes the data and forwards it to the sensor nodes. The sink node gathers soil moisture data, transmits the mean to the Gateways, and then forwards it to the website for dissemination. The web server is in charge of storing and presenting the moisture in the soil data to the web application's users. Soil characteristics may be collected using a networked method that we developed to improve rice production. Paddy land is running out as the population of our nation grows. The success of this project will be dependent on the appropriate use of the existing land base.

Keywords: IoT based agriculture monitoring, intelligent irrigation, communicating network, rice production

Procedia PDF Downloads 124
96 SIM (Subscriber Identity Module) Banking

Authors: Okanta Andrew, Richmond Kweku Frempong

Abstract:

As mobile networks are upgraded with technologies like WAP, GPRS and UMTS to deliver next-generation multimedia services, so are the banks and other financial institutions also getting ready to unleash the financial products on the mobile platform to meet growing demand for mobile based application services. Hence, the onset of Unstructured Supplementary Services (USSD) Banking which would make banking services available at anywhere, anytime through a string of interactive SMS sessions between a mobile device and an application server of a service provider. The aim of this studies was to find out whether the public will accept the sim banking service when it is implemented. Our target group includes: Working class. E. g. Businessmen/women, office workers, fishermen, market women, teachers etc. Nonworking class. E. g. Students (Tertiary, Senior High School), housewives. etc. The survey was in the form of a questionnaire and a verbal interview (video) which was to investigate their idea about the current banking system and the yet to be introduced sim banking concept. Meanwhile, some challenges accompanied the progression of data gathering because some populace showed reluctance in freeing their information. One other suggestion was that government should put measures against foremost challenges obstructing sim banking in Ghana counter to computers hackers. Government and individual have a key role to undertake to give suitable support to facelift the sim banking industry in the country. It was also suggested that Government put strong regulations on the use of sim banking products and services to streamline all the activities and also create awareness of the need for sim banking and emphasize its relevance in the aspect of national GDP.

Keywords: banking, mobile banking, SIM banking, mobile banking in Ghana

Procedia PDF Downloads 455
95 Classification of Manufacturing Data for Efficient Processing on an Edge-Cloud Network

Authors: Onyedikachi Ulelu, Andrew P. Longstaff, Simon Fletcher, Simon Parkinson

Abstract:

The widespread interest in 'Industry 4.0' or 'digital manufacturing' has led to significant research requiring the acquisition of data from sensors, instruments, and machine signals. In-depth research then identifies methods of analysis of the massive amounts of data generated before and during manufacture to solve a particular problem. The ultimate goal is for industrial Internet of Things (IIoT) data to be processed automatically to assist with either visualisation or autonomous system decision-making. However, the collection and processing of data in an industrial environment come with a cost. Little research has been undertaken on how to specify optimally what data to capture, transmit, process, and store at various levels of an edge-cloud network. The first step in this specification is to categorise IIoT data for efficient and effective use. This paper proposes the required attributes and classification to take manufacturing digital data from various sources to determine the most suitable location for data processing on the edge-cloud network. The proposed classification framework will minimise overhead in terms of network bandwidth/cost and processing time of machine tool data via efficient decision making on which dataset should be processed at the ‘edge’ and what to send to a remote server (cloud). A fast-and-frugal heuristic method is implemented for this decision-making. The framework is tested using case studies from industrial machine tools for machine productivity and maintenance.

Keywords: data classification, decision making, edge computing, industrial IoT, industry 4.0

Procedia PDF Downloads 149
94 An Efficient Traceability Mechanism in the Audited Cloud Data Storage

Authors: Ramya P, Lino Abraham Varghese, S. Bose

Abstract:

By cloud storage services, the data can be stored in the cloud, and can be shared across multiple users. Due to the unexpected hardware/software failures and human errors, which make the data stored in the cloud be lost or corrupted easily it affected the integrity of data in cloud. Some mechanisms have been designed to allow both data owners and public verifiers to efficiently audit cloud data integrity without retrieving the entire data from the cloud server. But public auditing on the integrity of shared data with the existing mechanisms will unavoidably reveal confidential information such as identity of the person, to public verifiers. Here a privacy-preserving mechanism is proposed to support public auditing on shared data stored in the cloud. It uses group signatures to compute verification metadata needed to audit the correctness of shared data. The identity of the signer on each block in shared data is kept confidential from public verifiers, who are easily verifying shared data integrity without retrieving the entire file. But on demand, the signer of the each block is reveal to the owner alone. Group private key is generated once by the owner in the static group, where as in the dynamic group, the group private key is change when the users revoke from the group. When the users leave from the group the already signed blocks are resigned by cloud service provider instead of owner is efficiently handled by efficient proxy re-signature scheme.

Keywords: data integrity, dynamic group, group signature, public auditing

Procedia PDF Downloads 364