Search results for: throughput
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 339

Search results for: throughput

129 Application of Systems Engineering Tools and Methods to Improve Healthcare Delivery Inside the Emergency Department of a Mid-Size Hospital

Authors: Mohamed Elshal, Hazim El-Mounayri, Omar El-Mounayri

Abstract:

Emergency department (ED) is considered as a complex system of interacting entities: patients, human resources, software and hardware systems, interfaces, and other systems. This paper represents a research for implementing a detailed Systems Engineering (SE) approach in a mid-size hospital in central Indiana. This methodology will be applied by “The Initiative for Product Lifecycle Innovation (IPLI)” institution at Indiana University to study and solve the crowding problem with the aim of increasing throughput of patients and enhance their treatment experience; therefore, the nature of crowding problem needs to be investigated with all other problems that leads to it. The presented SE methods are workflow analysis and systems modeling where SE tools such as Microsoft Visio are used to construct a group of system-level diagrams that demonstrate: patient’s workflow, documentation and communication flow, data systems, human resources workflow and requirements, leadership involved, and integration between ER different systems. Finally, the ultimate goal will be managing the process through implementation of an executable model using commercialized software tools, which will identify bottlenecks, improve documentation flow, and help make the process faster.

Keywords: systems modeling, ED operation, workflow modeling, systems analysis

Procedia PDF Downloads 152
128 Performance Evaluation of a Spouted Bed Bioreactor (SBBR) for the Biodegradation of 2, 4 Dichlorophenol

Authors: Taghreed Al-Khalid, Muftah El-Naas

Abstract:

As an economical and environmentally friendly technology, biological treatment has been shown to be one of the most promising approaches for the removal of numerous types of organic water pollutants such as Chlorophenols, which are hazardous pollutants commonly encountered in wastewater generated by the petroleum and petrochemical industries. This study aimed at evaluating the performance of a spouted bed bioreactor (SBBR) for aerobic biodegradation of 2, 4 dichlorophenol (DCP) by a commercial strain of Pseudomonas putida immobilized in polyvinyl alcohol (PVA) gel particles. The SBBR is characterized by systematic intense mixing, resulting in improvement of the biodegradation rates through reducing the mass transfer limitations. The reactor was evaluated in both batch and continuous mode in order to evaluate its hydrodynamics in terms of stability and response to shock loads. The SBBR was able to maintain a stable operation and recovered quickly to its normal operating mode once the shock load had been removed. In comparison to a packed bed reactor bioreactor, the SBBR proved to be more efficient and more stable, achieving a removal percentage and throughput of 80% and 1414 g/m3day, respectively. In addition, the biodegradation of chlorophenols was mathematically modeled using a dynamic modeling approach in order to assess reaction and mass transfer limitations. The results confirmed the effectiveness of the use of the PVA immobilization technique for the biodegradation of phenols.

Keywords: biodegradation, 2, 4 dichlorophenol, immobilization, polyvinyl alcohol (PVA) gel

Procedia PDF Downloads 149
127 Development and Power Characterization of an IoT Network for Agricultural Imaging Applications

Authors: Jacob Wahl, Jane Zhang

Abstract:

This paper describes the development and characterization of a prototype IoT network for use with agricultural imaging and monitoring applications. The sensor and gateway nodes are designed using the ESP32 SoC with integrated Bluetooth Low Energy 4.2 and Wi-Fi. A development board, the Arducam IoTai ESP32, is used for prototyping, testing, and power measurements. Google’s Firebase is used as the cloud storage site for image data collected by the sensor. The sensor node captures images using the OV2640 2MP camera module and transmits the image data to the gateway via Bluetooth Low Energy. The gateway then uploads the collected images to Firebase via a known nearby Wi-Fi network connection. This image data can then be processed and analyzed by computer vision and machine learning pipelines to assess crop growth or other needs. The sensor node achieves a wireless transmission data throughput of 220kbps while consuming 150mA of current; the sensor sleeps at 162µA. The sensor node device lifetime is estimated to be 682 days on a 6600mAh LiPo battery while acquiring five images per day based on the development board power measurements. This network can be utilized by any application that requires high data rates, low power consumption, short-range communication, and large amounts of data to be transmitted at low-frequency intervals.

Keywords: Bluetooth low energy, ESP32, firebase cloud, IoT, smart farming

Procedia PDF Downloads 110
126 Hybrid Weighted Multiple Attribute Decision Making Handover Method for Heterogeneous Networks

Authors: Mohanad Alhabo, Li Zhang, Naveed Nawaz

Abstract:

Small cell deployment in 5G networks is a promising technology to enhance capacity and coverage. However, unplanned deployment may cause high interference levels and high number of unnecessary handovers, which in turn will result in an increase in the signalling overhead. To guarantee service continuity, minimize unnecessary handovers, and reduce signalling overhead in heterogeneous networks, it is essential to properly model the handover decision problem. In this paper, we model the handover decision according to Multiple Attribute Decision Making (MADM) method, specifically Technique for Order Preference by Similarity to an Ideal Solution (TOPSIS). In this paper, we propose a hybrid TOPSIS method to control the handover in heterogeneous network. The proposed method adopts a hybrid weighting, which is a combination of entropy and standard deviation. A hybrid weighting control parameter is introduced to balance the impact of the standard deviation and entropy weighting on the network selection process and the overall performance. Our proposed method shows better performance, in terms of the number of frequent handovers and the mean user throughput, compared to the existing methods.

Keywords: handover, HetNets, interference, MADM, small cells, TOPSIS, weight

Procedia PDF Downloads 113
125 Routing Medical Images with Tabu Search and Simulated Annealing: A Study on Quality of Service

Authors: Mejía M. Paula, Ramírez L. Leonardo, Puerta A. Gabriel

Abstract:

In telemedicine, the image repository service is important to increase the accuracy of diagnostic support of medical personnel. This study makes comparison between two routing algorithms regarding the quality of service (QoS), to be able to analyze the optimal performance at the time of loading and/or downloading of medical images. This study focused on comparing the performance of Tabu Search with other heuristic and metaheuristic algorithms that improve QoS in telemedicine services in Colombia. For this, Tabu Search and Simulated Annealing heuristic algorithms are chosen for their high usability in this type of applications; the QoS is measured taking into account the following metrics: Delay, Throughput, Jitter and Latency. In addition, routing tests were carried out on ten images in digital image and communication in medicine (DICOM) format of 40 MB. These tests were carried out for ten minutes with different traffic conditions, reaching a total of 25 tests, from a server of Universidad Militar Nueva Granada (UMNG) in Bogotá-Colombia to a remote user in Universidad de Santiago de Chile (USACH) - Chile. The results show that Tabu search presents a better QoS performance compared to Simulated Annealing, managing to optimize the routing of medical images, a basic requirement to offer diagnostic images services in telemedicine.

Keywords: medical image, QoS, simulated annealing, Tabu search, telemedicine

Procedia PDF Downloads 191
124 A Hybrid Traffic Model for Smoothing Traffic Near Merges

Authors: Shiri Elisheva Decktor, Sharon Hornstein

Abstract:

Highway merges and unmarked junctions are key components in any urban road network, which can act as bottlenecks and create traffic disruption. Inefficient highway merges may trigger traffic instabilities such as stop-and-go waves, pose safety conditions and lead to longer journey times. These phenomena occur spontaneously if the average vehicle density exceeds a certain critical value. This study focuses on modeling the traffic using a microscopic traffic flow model. A hybrid traffic model, which combines human-driven and controlled vehicles is assumed. The controlled vehicles obey different driving policies when approaching the merge, or in the vicinity of other vehicles. We developed a co-simulation model in SUMO (Simulation of Urban Mobility), in which the human-driven cars are modeled using the IDM model, and the controlled cars are modeled using a dedicated controller. The scenario chosen for this study is a closed track with one merge and one exit, which could be later implemented using a scaled infrastructure on our lab setup. This will enable us to benchmark the results of this study obtained in simulation, to comparable results in similar conditions in the lab. The metrics chosen for the comparison of the performance of our algorithm on the overall traffic conditions include the average speed, wait time near the merge, and throughput after the merge, measured under different travel demand conditions (low, medium, and heavy traffic).

Keywords: highway merges, traffic modeling, SUMO, driving policy

Procedia PDF Downloads 76
123 The Coalescence Process of Droplet Pairs in Different Junctions

Authors: Xiang Wang, Yan Pang, Zhaomiao Liu

Abstract:

Droplet-based microfluidics have been studied extensively with the development of the Micro-Electro-Mechanical System (MEMS) which bears the advantages of high throughput, high efficiency, low cost and low polydispersity. Droplets, worked as versatile carriers, could provide isolated chambers as the internal dispersed phase is protected from the outside continuous phase. Droplets are used to add reagents to start or end bio-chemical reactions, to generate concentration gradients, to realize hydrate crystallization or protein analyses, while droplets coalescence acts as an important control technology. In this paper, deionized water is used as the dispersed phase, and several kinds of oil are used as the continuous phase to investigate the influence of the viscosity ratio of the two phases on the coalescence process. The microchannels are fabricated by coating a polydimethylsiloxane (PDMS) layer onto another PDMS flat plate after corona treatment. All newly made microchannels are rinsed with the continuous oil phase for hours before experiments to ensure the swelling fully developed. High-speed microscope system is used to document the serial videos with a maximum speed of 2000 frames per second. The critical capillary numbers (Ca*) of droplet pairs in various junctions are studied and compared. Ca* varies with different junctions or different liquids within the range of 0.002 to 0.01. However, droplets without extra control would have the problem of synchronism which reduces the coalescence efficiency.

Keywords: coalescence, concentration, critical capillary number, droplet pair, split

Procedia PDF Downloads 212
122 Ultra-Fast pH-Gradient Ion Exchange Chromatography for the Separation of Monoclonal Antibody Charge Variants

Authors: Robert van Ling, Alexander Schwahn, Shanhua Lin, Ken Cook, Frank Steiner, Rowan Moore, Mauro de Pra

Abstract:

Purpose: Demonstration of fast high resolution charge variant analysis for monoclonal antibody (mAb) therapeutics within 5 minutes. Methods: Three commercially available mAbs were used for all experiments. The charge variants of therapeutic mAbs (Bevacizumab, Cetuximab, Infliximab, and Trastuzumab) are analyzed on a strong cation exchange column with a linear pH gradient separation method. The linear gradient from pH 5.6 to pH 10.2 is generated over time by running a linear pump gradient from 100% Thermo Scientific™ CX-1 pH Gradient Buffer A (pH 5.6) to 100% CX-1 pH Gradient Buffer B (pH 10.2), using the Thermo Scientific™ Vanquish™ UHPLC system. Results: The pH gradient method is generally applicable to monoclonal antibody charge variant analysis. In conjunction with state-of-the-art column and UHPLC technology, ultra fast high-resolution separations are consistently achieved in under 5 minutes for all mAbs analyzed. Conclusion: The linear pH gradient method is a platform method for mAb charge variant analysis. The linear pH gradient method can be easily optimized to improve separations and shorten cycle times. Ultra-fast charge variant separation is facilitated with UHPLC that complements, and in some instances outperforms CE approaches in terms of both resolution and throughput.

Keywords: charge variants, ion exchange chromatography, monoclonal antibody, UHPLC

Procedia PDF Downloads 405
121 Dynamic Cellular Remanufacturing System (DCRS) Design

Authors: Tariq Aljuneidi, Akif Asil Bulgak

Abstract:

Remanufacturing may be defined as the process of bringing used products to “like-new” functional state with warranty to match, and it is one of the most popular product end-of-life scenarios. An efficient remanufacturing network lead to an efficient design of sustainable manufacturing enterprise. In remanufacturing network, products are collected from the customer zone, disassembled and remanufactured at a suitable remanufacturing facility. In this respect, another issue to consider is how the returned product to be remanufactured, in other words, what is the best layout for such facility. In order to achieve a sustainable manufacturing system, Cellular Manufacturing System (CMS) designs are highly recommended, CMSs combine high throughput rates of line layouts with the flexibility offered by functional layouts (job shop). Introducing the CMS while designing a remanufacturing network will benefit the utilization of such a network. This paper presents and analyzes a comprehensive mathematical model for the design of Dynamic Cellular Remanufacturing Systems (DCRSs). In this paper, the proposed model is the first one to date that consider CMS and remanufacturing system simultaneously. The proposed DCRS model considers several manufacturing attributes such as multi-period production planning, dynamic system reconfiguration, duplicate machines, machine capacity, available time for workers, worker assignments, and machine procurement, where the demand is totally satisfied from a returned product. A numerical example is presented to illustrate the proposed model.

Keywords: cellular manufacturing system, remanufacturing, mathematical programming, sustainability

Procedia PDF Downloads 352
120 Asynchronous Low Duty Cycle Media Access Control Protocol for Body Area Wireless Sensor Networks

Authors: Yasin Ghasemi-Zadeh, Yousef Kavian

Abstract:

Wireless body area networks (WBANs) technology has achieved lots of popularity over the last decade with a wide range of medical applications. This paper presents an asynchronous media access control (MAC) protocol based on B-MAC protocol by giving an application for medical issues. In WBAN applications, there are some serious problems such as energy, latency, link reliability (quality of wireless link) and throughput which are mainly due to size of sensor networks and human body specifications. To overcome these problems and improving link reliability, we concentrated on MAC layer that supports mobility models for medical applications. In the presented protocol, preamble frames are divided into some sub-frames considering the threshold level. Actually, the main reason for creating shorter preambles is the link reliability where due to some reasons such as water, the body signals are affected on some frequency bands and causes fading and shadowing on signals, therefore by increasing the link reliability, these effects are reduced. In case of mobility model, we use MoBAN model and modify that for some more areas. The presented asynchronous MAC protocol is modeled by OMNeT++ simulator. The results demonstrate increasing the link reliability comparing to B-MAC protocol where the packet reception ratio (PRR) is 92% also covers more mobility areas than MoBAN protocol.

Keywords: wireless body area networks (WBANs), MAC protocol, link reliability, mobility, biomedical

Procedia PDF Downloads 343
119 Of an 80 Gbps Passive Optical Network Using Time and Wavelength Division Multiplexing

Authors: Malik Muhammad Arslan, Muneeb Ullah, Dai Shihan, Faizan Khan, Xiaodong Yang

Abstract:

Internet Service Providers are driving endless demands for higher bandwidth and data throughput as new services and applications require higher bandwidth. Users want immediate and accurate data delivery. This article focuses on converting old conventional networks into passive optical networks based on time division and wavelength division multiplexing. The main focus of this research is to use a hybrid of time-division multiplexing and wavelength-division multiplexing to improve network efficiency and performance. In this paper, we design an 80 Gbps Passive Optical Network (PON), which meets the need of the Next Generation PON Stage 2 (NGPON2) proposed in this paper. The hybrid of the Time and Wavelength division multiplexing (TWDM) is said to be the best solution for the implementation of NGPON2, according to Full-Service Access Network (FSAN). To co-exist with or replace the current PON technologies, many wavelengths of the TWDM can be implemented simultaneously. By utilizing 8 pairs of wavelengths that are multiplexed and then transmitted over optical fiber for 40 Kms and on the receiving side, they are distributed among 256 users, which shows that the solution is reliable for implementation with an acceptable data rate. From the results, it can be concluded that the overall performance, Quality Factor, and bandwidth of the network are increased, and the Bit Error rate is minimized by the integration of this approach.

Keywords: bit error rate, fiber to the home, passive optical network, time and wavelength division multiplexing

Procedia PDF Downloads 44
118 Dynamic Risk Identification Using Fuzzy Failure Mode Effect Analysis in Fabric Process Industries: A Research Article as Management Perspective

Authors: A. Sivakumar, S. S. Darun Prakash, P. Navaneethakrishnan

Abstract:

In and around Erode District, it is estimated that more than 1250 chemical and allied textile processing fabric industries are affected, partially closed and shut off for various reasons such as poor management, poor supplier performance, lack of planning for productivity, fluctuation of output, poor investment, waste analysis, labor problems, capital/labor ratio, accumulation of stocks, poor maintenance of resources, deficiencies in the quality of fabric, low capacity utilization, age of plant and equipment, high investment and input but low throughput, poor research and development, lack of energy, workers’ fear of loss of jobs, work force mix and work ethic. The main objective of this work is to analyze the existing conditions in textile fabric sector, validate the break even of Total Productivity (TP), analyze, design and implement fuzzy sets and mathematical programming for improvement of productivity and quality dimensions in the fabric processing industry. It needs to be compatible with the reality of textile and fabric processing industries. The highly risk events from productivity and quality dimension were found by fuzzy systems and results are wrapped up among the textile fabric processing industry.

Keywords: break even point, fuzzy crisp data, fuzzy sets, productivity, productivity cycle, total productive maintenance

Procedia PDF Downloads 305
117 Simulating Elevated Rapid Transit System for Performance Analysis

Authors: Ran Etgar, Yuval Cohen, Erel Avineri

Abstract:

One of the major challenges of transportation in medium sized inner-cities (such as Tel-Aviv) is the last-mile solution. Personal rapid transit (PRT) seems like an applicable candidate for this, as it combines the benefits of personal (car) travel with the operational benefits of transit. However, the investment required for large area PRT grid is significant and there is a need to economically justify such investment by correctly evaluating the grid capacity. PRT main elements are small automated vehicles (sometimes referred to as podcars) operating on a network of specially built guideways. The research is looking at a specific concept of elevated PRT system. Literature review has revealed the drawbacks PRT modelling and simulation approaches, mainly due to the lack of consideration of technical and operational features of the system (such as headways, acceleration, safety issues); the detailed design of infrastructure (guideways, stations, and docks); the stochastic and sessional characteristics of demand; and safety regulations – all of them have a strong effect on the system performance. A highly detailed model of the system, developed in this research, is applying a discrete event simulation combined with an agent-based approach, to represent the system elements and the podecars movement logic. Applying a case study approach, the simulation model is used to study the capacity of the system, the expected throughput of the system, the utilization, and the level of service (journey time, waiting time, etc.).

Keywords: capacity, productivity measurement, PRT, simulation, transportation

Procedia PDF Downloads 125
116 Artificial Intelligent-Based Approaches for Task ‎Offloading, ‎Resource ‎Allocation and Service ‎Placement of ‎Internet of Things ‎Applications: State of the Art

Authors: Fatima Z. Cherhabil, Mammar Sedrati, Sonia-Sabrina Bendib‎

Abstract:

In order to support the continued growth, critical latency of ‎IoT ‎applications, and ‎various obstacles of traditional data centers, ‎mobile edge ‎computing (MEC) has ‎emerged as a promising solution that extends cloud data-processing and decision-making to edge devices. ‎By adopting a MEC structure, IoT applications could be executed ‎locally, on ‎an edge server, different fog nodes, or distant cloud ‎data centers. However, we are ‎often ‎faced with wanting to optimize conflicting criteria such as ‎minimizing energy ‎consumption of limited local capabilities (in terms of CPU, RAM, storage, bandwidth) of mobile edge ‎devices and trying to ‎keep ‎high performance (reducing ‎response time, increasing throughput and service availability) ‎at the same ‎time‎. Achieving one goal may affect the other, making task offloading (TO), ‎resource allocation (RA), and service placement (SP) complex ‎processes. ‎It is a nontrivial multi-objective optimization ‎problem ‎to study the trade-off between conflicting criteria. ‎The paper provides a survey on different TO, SP, and RA recent multi-‎objective optimization (MOO) approaches used in edge computing environments, particularly artificial intelligent (AI) ones, to satisfy various objectives, constraints, and dynamic conditions related to IoT applications‎.

Keywords: mobile edge computing, multi-objective optimization, artificial ‎intelligence ‎approaches, task offloading, resource allocation, ‎ service placement

Procedia PDF Downloads 84
115 FTIR Spectroscopy for in vitro Screening in Microbial Biotechnology

Authors: V. Shapaval, N. K. Afseth, D. Tzimorotas, A. Kohler

Abstract:

Globally there is a dramatic increase in the demand for food, energy, materials and clean water since natural resources are limited. As a result, industries are looking for ways to reduce rest materials and to improve resource efficiency. Microorganisms have a high potential to be used as bio factories for the production of primary and secondary metabolites that represent high-value bio-products (enzymes, polyunsaturated fatty acids, bio-plastics, glucans, etc.). In order to find good microbial producers, to design suitable substrates from food rest materials and to optimize fermentation conditions, rapid analytical techniques for quantifying target bio products in microbial cells are needed. In the EU project FUST (R4SME, Fp7), we have developed a fully automated high-throughput FUST system based on micro-cultivation and FTIR spectroscopy that facilitates the screening of microorganisms, substrates and fermentation conditions for the optimization of the production of different high-value metabolites (single cell oils, bio plastics). The automated system allows the preparation of 100 samples per hour. Currently, The FUST system is in use for screening of filamentous fungi in order to find oleaginous strains with the ability to produce polyunsaturated fatty acids, and the optimization of cheap substrates, derived from food rest materials, and the optimization of fermentation conditions for the high yield of single cell oil.

Keywords: FTIR spectroscopy, FUST system, screening, biotechnology

Procedia PDF Downloads 418
114 Population Structure Analysis of Pakistani Indigenous Cattle Population by Using High Density SNP Array

Authors: Hamid Mustafa, Huson J. Heather, Kim Eiusoo, McClure Matt, Khalid Javed, Talat Nasser Pasha, Afzal Ali1, Adeela Ajmal, Tad Sonstegard

Abstract:

Genetic differences associated with speciation, breed formation or local adaptation can help to preserve and effective utilization of animals in selection programs. Analyses of population structure and breed diversity have provided insight into the origin and evolution of cattle. In this study, we used a high-density panel of SNP markers to examine population structure and diversity among ten Pakistani indigenous cattle breeds. In total, 25 individuals from three cattle populations, including Achi (n=08), Bhagnari (n=04) and Cholistani (n=13) were genotyped for 777, 962 single nucleotide polymorphism (SNP) markers. Population structure was examined using the linkage model in the program STRUCTURE. After characterizing SNP polymorphism in the different populations, we performed a detailed analysis of genetic structure at both the individual and population levels. The whole-genome SNP panel identified several levels of population substructure in the set of examined cattle breeds. We further searched for spatial patterns of genetic diversity among these breeds under the recently developed spatial principal component analysis framework. Overall, such high throughput genotyping data confirmed a clear partitioning of the cattle genetic diversity into distinct breeds. The resulting complex historical origins associated with both natural and artificial selection have led to the differentiation of numerous different cattle breeds displaying a broad phenotypic variety over a short period of time.

Keywords: Pakistan, cattle, genetic diversity, population structure

Procedia PDF Downloads 586
113 Virtual Prototyping of Ventilated Corrugated Fibreboard Carton of Fresh Fruit for Improved Containerized Transportation

Authors: Alemayehu Ambaw, Matia Mukama, Umezuruike Linus Opara

Abstract:

This study introduces a comprehensive method for designing ventilated corrugated fiberboard carton for fresh fruit packaging utilising virtual prototyping. The technique efficiently assesses and analyses the mechanical and thermal capabilities of fresh fruit packing boxes prior to making production investments. Comprehensive structural, aerodynamic, and thermodynamic data from designs were collected and evaluated in comparison to real-world packaging needs. Physical prototypes of potential designs were created and evaluated afterward. The virtual prototype is created with computer-aided graphics, computational structural dynamics, and computational fluid dynamics technologies. The virtual prototyping quickly generated data on carton compression strength, airflow resistance, produce cooling rate, spatiotemporal temperature, and product quality map in the cold chain within a few hours. Six distinct designs were analysed. All the various carton designs showed similar effectiveness in preserving the quality of the goods. The innovative packaging box design is more compact, resulting in a higher freight density of 1720 kg more fruit per reefer compared to the commercial counterpart. The precooling process was improved, resulting in a 17% increase in throughput and a 30% reduction in power usage.

Keywords: postharvest, container logistics, space/volume usage, computational method, packaging technology

Procedia PDF Downloads 26
112 Integrated Microsystem for Multiplexed Genosensor Detection of Biowarfare Agents

Authors: Samuel B. Dulay, Sandra Julich, Herbert Tomaso, Ciara K. O'Sullivan

Abstract:

An early, rapid and definite detection for the presence of biowarfare agents, pathogens, viruses and toxins is required in different situations which include civil rescue and security units, homeland security, military operations, public transportation securities such as airports, metro and railway stations due to its harmful effect on the human population. In this work, an electrochemical genosensor array that allows simultaneous detection of different biowarfare agents within an integrated microsystem that provides an easy handling of the technology which combines a microfluidics setup with a multiplexing genosensor array has been developed and optimised for the following targets: Bacillus anthracis, Brucella abortis and melitensis, Bacteriophage lambda, Francisella tularensis, Burkholderia mallei and pseudomallei, Coxiella burnetii, Yersinia pestis, and Bacillus thuringiensis. The electrode array was modified via co-immobilisation of a 1:100 (mol/mol) mixture of a thiolated probe and an oligoethyleneglycol-terminated monopodal thiol. PCR products from these relevant biowarfare agents were detected reproducibly through a sandwich assay format with the target hybridised between a surface immobilised probe into the electrode and a horseradish peroxidase-labelled secondary reporter probe, which provided an enzyme based electrochemical signal. The potential of the designed microsystem for multiplexed genosensor detection and cross-reactivity studies over potential interfering DNA sequences has demonstrated high selectivity using the developed platform producing high-throughput.

Keywords: biowarfare agents, genosensors, multipled detection, microsystem

Procedia PDF Downloads 234
111 Network Analysis of Genes Involved in the Biosynthesis of Medicinally Important Naphthodianthrone Derivatives of Hypericum perforatum

Authors: Nafiseh Noormohammadi, Ahmad Sobhani Najafabadi

Abstract:

Hypericins (hypericin and pseudohypericin) are natural napthodianthrone derivatives produced by Hypericum perforatum (St. John’s Wort), which have many medicinal properties such as antitumor, antineoplastic, antiviral, and antidepressant activities. Production and accumulation of hypericin in the plant are influenced by both genetic and environmental conditions. Despite the existence of different high-throughput data on the plant, genetic dimensions of hypericin biosynthesis have not yet been completely understood. In this research, 21 high-quality RNA-seq data on different parts of the plant were integrated into metabolic data to reconstruct a coexpression network. Results showed that a cluster of 30 transcripts was correlated with total hypericin. The identified transcripts were divided into three main groups based on their functions, including hypericin biosynthesis genes, transporters, detoxification genes, and transcription factors (TFs). In the biosynthetic group, different isoforms of polyketide synthase (PKSs) and phenolic oxidative coupling proteins (POCPs) were identified. Phylogenetic analysis of protein sequences integrated into gene expression analysis showed that some of the POCPs seem to be very important in the biosynthetic pathway of hypericin. In the TFs group, six TFs were correlated with total hypericin. qPCR analysis of these six TFs confirmed that three of them were highly correlated. The identified genes in this research are a rich resource for further studies on the molecular breeding of H. perforatum in order to obtain varieties with high hypericin production.

Keywords: hypericin, St. John’s Wort, data mining, transcription factors, secondary metabolites

Procedia PDF Downloads 54
110 Transcriptomic Analyses of Kappaphycus alvarezii under Different Wavelengths of Light

Authors: Vun Yee Thien, Kenneth Francis Rodrigues, Clemente Michael Vui Ling Wong, Wilson Thau Lym Yong

Abstract:

Transcriptomes associated with the process of photosynthesis have offered insights into the mechanism of gene regulation in terrestrial plants; however, limited information is available as far as macroalgae are concerned. This investigation aims to decipher the underlying mechanisms associated with photosynthesis in the red alga, Kappaphycus alvarezii, by performing a differential expression analysis on a de novo assembled transcriptomes. Comparative analysis of gene expression was designed to examine the alteration of light qualities and its effect on physiological mechanisms in the red alga. High-throughput paired-end RNA-sequencing was applied to profile the transcriptome of K. alvarezii irradiated with different wavelengths of light (blue 492-455 nm, green 577-492 nm and red 780-622 nm) as compared to the full light spectrum, resulted in more than 60 million reads individually and assembled using Trinity and SOAPdenovo-Trans. The transcripts were annotated in the NCBI non-redundant (nr) protein, SwissProt, KEGG and COG databases with a cutoff E-value of 1e-5 and nearly 30% of transcripts were assigned to functional annotation by Blast searches. Differential expression analysis was performed using edgeR. The DEGs were designated to six categories: BL (blue light) regulated, GL (green light) regulated, RL (red light) regulated, BL or GL regulated, BL or RL regulated, GL or RL regulated, and either BL, GL or RL regulated. These DEGs were mapped to terms in KEGG database and compared with the whole transcriptome background to search for genes that regulated by light quality. The outcomes of this study will enhance our understanding of molecular mechanisms underlying light-induced responses in red algae.

Keywords: de novo transcriptome sequencing, differential gene expression, Kappaphycus alvareziired, red alga

Procedia PDF Downloads 488
109 Health Care using Queuing Theory

Authors: S. Vadivukkarasi, K. Karthi, M. Karthick, C. Dinesh, S. Santhosh, A. Yogaraj

Abstract:

The appointment system was designed to minimize patient’s idle time overlooking patients waiting time in hospitals. This is no longer valid in today’s consumer oriented society. Long waiting times for treatment in the outpatient department followed by short consultations has long been a complaint. Nowadays, customers use waiting time as a decisive factor in choosing a service provider. Queuing theory constitutes a very powerful tool because queuing models require relatively little data and are simple and fast to use. Because of this simplicity and speed, modelers can be used to quickly evaluate and compare various alternatives for providing service. The application of queuing models in the analysis of health care systems is increasingly accepted by health care decision makers. Timely access to care is a key component of high-quality health care. However, patient delays are prevalent throughout health care systems, resulting in dissatisfaction and adverse clinical consequences for patients as well as potentially higher costs and wasted capacity for providers. Arguably, the most critical delays for health care are the ones associated with health care emergencies. The allocation of resources can be divided into three general areas: bed management, staff management, and room facility management. Effective and efficient patient flow is indicated by high patient throughput, low patient waiting times, a short length of stay at the hospital and overtime, while simultaneously maintaining adequate staff utilization rates and low patient’s idle times.

Keywords: appointment system, patient scheduling, bed management, queueing calculation, system analysis

Procedia PDF Downloads 275
108 CMPD: Cancer Mutant Proteome Database

Authors: Po-Jung Huang, Chi-Ching Lee, Bertrand Chin-Ming Tan, Yuan-Ming Yeh, Julie Lichieh Chu, Tin-Wen Chen, Cheng-Yang Lee, Ruei-Chi Gan, Hsuan Liu, Petrus Tang

Abstract:

Whole-exome sequencing focuses on the protein coding regions of disease/cancer associated genes based on a priori knowledge is the most cost-effective method to study the association between genetic alterations and disease. Recent advances in high throughput sequencing technologies and proteomic techniques has provided an opportunity to integrate genomics and proteomics, allowing readily detectable mutated peptides corresponding to mutated genes. Since sequence database search is the most widely used method for protein identification using Mass spectrometry (MS)-based proteomics technology, a mutant proteome database is required to better approximate the real protein pool to improve disease-associated mutated protein identification. Large-scale whole exome/genome sequencing studies were launched by National Cancer Institute (NCI), Broad Institute, and The Cancer Genome Atlas (TCGA), which provide not only a comprehensive report on the analysis of coding variants in diverse samples cell lines but a invaluable resource for extensive research community. No existing database is available for the collection of mutant protein sequences related to the identified variants in these studies. CMPD is designed to address this issue, serving as a bridge between genomic data and proteomic studies and focusing on protein sequence-altering variations originated from both germline and cancer-associated somatic variations.

Keywords: TCGA, cancer, mutant, proteome

Procedia PDF Downloads 560
107 Managing Data from One Hundred Thousand Internet of Things Devices Globally for Mining Insights

Authors: Julian Wise

Abstract:

Newcrest Mining is one of the world’s top five gold and rare earth mining organizations by production, reserves and market capitalization in the world. This paper elaborates on the data acquisition processes employed by Newcrest in collaboration with Fortune 500 listed organization, Insight Enterprises, to standardize machine learning solutions which process data from over a hundred thousand distributed Internet of Things (IoT) devices located at mine sites globally. Through the utilization of software architecture cloud technologies and edge computing, the technological developments enable for standardized processes of machine learning applications to influence the strategic optimization of mineral processing. Target objectives of the machine learning optimizations include time savings on mineral processing, production efficiencies, risk identification, and increased production throughput. The data acquired and utilized for predictive modelling is processed through edge computing by resources collectively stored within a data lake. Being involved in the digital transformation has necessitated the standardization software architecture to manage the machine learning models submitted by vendors, to ensure effective automation and continuous improvements to the mineral process models. Operating at scale, the system processes hundreds of gigabytes of data per day from distributed mine sites across the globe, for the purposes of increased improved worker safety, and production efficiency through big data applications.

Keywords: mineral technology, big data, machine learning operations, data lake

Procedia PDF Downloads 84
106 Enzymatic Repair Prior To DNA Barcoding, Aspirations, and Restraints

Authors: Maxime Merheb, Rachel Matar

Abstract:

Retrieving ancient DNA sequences which in return permit the entire genome sequencing from fossils have extraordinarily improved in recent years, thanks to sequencing technology and other methodological advances. In any case, the quest to search for ancient DNA is still obstructed by the damage inflicted on DNA which accumulates after the death of a living organism. We can characterize this damage into three main categories: (i) Physical abnormalities such as strand breaks which lead to the presence of short DNA fragments. (ii) Modified bases (mainly cytosine deamination) which cause errors in the sequence due to an incorporation of a false nucleotide during DNA amplification. (iii) DNA modifications referred to as blocking lesions, will halt the PCR extension which in return will also affect the amplification and sequencing process. We can clearly see that the issues arising from breakage and coding errors were significantly decreased in recent years. Fast sequencing of short DNA fragments was empowered by platforms for high-throughput sequencing, most of the coding errors were uncovered to be the consequences of cytosine deamination which can be easily removed from the DNA using enzymatic treatment. The methodology to repair DNA sequences is still in development, it can be basically explained by the process of reintroducing cytosine rather than uracil. This technique is thus restricted to amplified DNA molecules. To eliminate any type of damage (particularly those that block PCR) is a process still pending the complete repair methodologies; DNA detection right after extraction is highly needed. Before using any resources into extensive, unreasonable and uncertain repair techniques, it is vital to distinguish between two possible hypotheses; (i) DNA is none existent to be amplified to begin with therefore completely un-repairable, (ii) the DNA is refractory to PCR and it is worth to be repaired and amplified. Hence, it is extremely important to develop a non-enzymatic technique to detect the most degraded DNA.

Keywords: ancient DNA, DNA barcodong, enzymatic repair, PCR

Procedia PDF Downloads 378
105 Cooperative Cross Layer Topology for Concurrent Transmission Scheduling Scheme in Broadband Wireless Networks

Authors: Gunasekaran Raja, Ramkumar Jayaraman

Abstract:

In this paper, we consider CCL-N (Cooperative Cross Layer Network) topology based on the cross layer (both centralized and distributed) environment to form network communities. Various performance metrics related to the IEEE 802.16 networks are discussed to design CCL-N Topology. In CCL-N topology, nodes are classified as master nodes (Master Base Station [MBS]) and serving nodes (Relay Station [RS]). Nodes communities are organized based on the networking terminologies. Based on CCL-N Topology, various simulation analyses for both transparent and non-transparent relays are tabulated and throughput efficiency is calculated. Weighted load balancing problem plays a challenging role in IEEE 802.16 network. CoTS (Concurrent Transmission Scheduling) Scheme is formulated in terms of three aspects – transmission mechanism based on identical communities, different communities and identical node communities. CoTS scheme helps in identifying the weighted load balancing problem. Based on the analytical results, modularity value is inversely proportional to that of the error value. The modularity value plays a key role in solving the CoTS problem based on hop count. The transmission mechanism for identical node community has no impact since modularity value is same for all the network groups. In this paper three aspects of communities based on the modularity value which helps in solving the problem of weighted load balancing and CoTS are discussed.

Keywords: cross layer network topology, concurrent scheduling, modularity value, network communities and weighted load balancing

Procedia PDF Downloads 229
104 Fabrication of Uniform Nanofibers Using Gas Dynamic Virtual Nozzle Based Microfluidic Liquid Jet System

Authors: R. Vasireddi, J. Kruse, M. Vakili, M. Trebbin

Abstract:

Here we present a gas dynamic virtual nozzle (GDVN) based microfluidic jetting devices for spinning of nano/microfibers. The device is fabricated by soft lithography techniques and is based on the principle of a GDVN for precise three-dimensional gas focusing of the spinning solution. The nozzle device is used to produce micro/nanofibers of a perfluorinated terpolymer (THV), which were collected on an aluminum substrate for scanning electron microscopy (SEM) analysis. The influences of air pressure, polymer concentration, flow rate and nozzle geometry on the fiber properties were investigated. It was revealed that surface properties are controlled by air pressure and polymer concentration while the diameter and shape of the fibers are influenced mostly by the concentration of the polymer solution and pressure. Alterations of the nozzle geometry had a negligible effect on the fiber properties, however, the jetting stability was affected. Round and flat fibers with differing surface properties from craters, grooves to smooth surfaces could be fabricated by controlling the above-mentioned parameters. Furthermore, the formation of surface roughness was attributed to the fast evaporation rate and velocity (mis)match between the polymer solution jet and the surrounding air stream. The diameter of the fibers could be tuned from ~250 nm to ~15 µm. Because of the simplicity of the setup, the precise control of the fiber properties, access to biocompatible nanofiber fabrication and the easy scale-up of parallel channels for high throughput, this method offers significant benefits compared to existing solution-based fiber production methods.

Keywords: gas dynamic virtual nozzle (GDVN) principle, microfluidic device, spinning, uniform nanofibers

Procedia PDF Downloads 126
103 Performance Analysis of Search Medical Imaging Service on Cloud Storage Using Decision Trees

Authors: González A. Julio, Ramírez L. Leonardo, Puerta A. Gabriel

Abstract:

Telemedicine services use a large amount of data, most of which are diagnostic images in Digital Imaging and Communications in Medicine (DICOM) and Health Level Seven (HL7) formats. Metadata is generated from each related image to support their identification. This study presents the use of decision trees for the optimization of information search processes for diagnostic images, hosted on the cloud server. To analyze the performance in the server, the following quality of service (QoS) metrics are evaluated: delay, bandwidth, jitter, latency and throughput in five test scenarios for a total of 26 experiments during the loading and downloading of DICOM images, hosted by the telemedicine group server of the Universidad Militar Nueva Granada, Bogotá, Colombia. By applying decision trees as a data mining technique and comparing it with the sequential search, it was possible to evaluate the search times of diagnostic images in the server. The results show that by using the metadata in decision trees, the search times are substantially improved, the computational resources are optimized and the request management of the telemedicine image service is improved. Based on the experiments carried out, search efficiency increased by 45% in relation to the sequential search, given that, when downloading a diagnostic image, false positives are avoided in management and acquisition processes of said information. It is concluded that, for the diagnostic images services in telemedicine, the technique of decision trees guarantees the accessibility and robustness in the acquisition and manipulation of medical images, in improvement of the diagnoses and medical procedures in patients.

Keywords: cloud storage, decision trees, diagnostic image, search, telemedicine

Procedia PDF Downloads 180
102 Performance Evaluation of Routing Protocol in Cognitive Radio with Multi Technological Environment

Authors: M. Yosra, A. Mohamed, T. Sami

Abstract:

Over the past few years, mobile communication technologies have seen significant evolution. This fact promoted the implementation of many systems in a multi-technological setting. From one system to another, the Quality of Service (QoS) provided to mobile consumers gets better. The growing number of normalized standards extends the available services for each consumer, moreover, most of the available radio frequencies have already been allocated, such as 3G, Wifi, Wimax, and LTE. A study by the Federal Communications Commission (FCC) found that certain frequency bands are partially occupied in particular locations and times. So, the idea of Cognitive Radio (CR) is to share the spectrum between a primary user (PU) and a secondary user (SU). The main objective of this spectrum management is to achieve a maximum rate of exploitation of the radio spectrum. In general, the CR can greatly improve the quality of service (QoS) and improve the reliability of the link. The problem will reside in the possibility of proposing a technique to improve the reliability of the wireless link by using the CR with some routing protocols. However, users declared that the links were unreliable and that it was an incompatibility with QoS. In our case, we choose the QoS parameter "bandwidth" to perform a supervised classification. In this paper, we propose a comparative study between some routing protocols, taking into account the variation of different technologies on the existing spectral bandwidth like 3G, WIFI, WIMAX, and LTE. Due to the simulation results, we observe that LTE has significantly higher availability bandwidth compared with other technologies. The performance of the OLSR protocol is better than other on-demand routing protocols (DSR, AODV and DSDV), in LTE technology because of the proper receiving of packets, less packet drop and the throughput. Numerous simulations of routing protocols have been made using simulators such as NS3.

Keywords: cognitive radio, multi technology, network simulator (NS3), routing protocol

Procedia PDF Downloads 37
101 Method Development for the Determination of Gamma-Aminobutyric Acid in Rice Products by Lc-Ms-Ms

Authors: Cher Rong Matthew Kong, Edmund Tian, Seng Poon Ong, Chee Sian Gan

Abstract:

Gamma-aminobutyric acid (GABA) is a non-protein amino acid that is a functional constituent of certain rice varieties. When consumed, it decreases blood pressure and reduces the risk of hypertension-related diseases. This has led to more research dedicated towards the development of functional food products (e.g. germinated brown rice) with enhanced GABA content, and the development of these functional food products has led to increased demand for instrument-based methods that can efficiently and effectively determine GABA content. Current analytical methods require analyte derivatisation, and have significant disadvantages such as being labour intensive and time-consuming, and being subject to analyte loss due to the increased complexity of the sample preparation process. To address this, an LC-MS-MS method for the determination of GABA in rice products has been developed and validated. This developed method involves a relatively simple sample preparation process before analysis using HILIC LC-MS-MS. This method eliminates the need for derivatisation, thereby significantly reducing the labour and time associated with such an analysis. Using LC-MS-MS also allows for better differentiation of GABA from any potential co-eluting compounds in the sample matrix. Results obtained from the developed method demonstrated high linearity, accuracy, and precision for the determination of GABA (1ng/L to 8ng/L) in a variety of brown rice products. The method can significantly simplify sample preparation steps, improve the accuracy of quantitation, and increase the throughput of analyses, thereby providing a quick but effective alternative to established instrumental analysis methods for GABA in rice.

Keywords: functional food, gamma-aminobutyric acid, germinated brown rice, method development

Procedia PDF Downloads 232
100 Amino Acid Responses of Wheat Cultivars under Glasshouse Drought Accurately Predict Yield-Based Drought Tolerance in the Field

Authors: Arun K. Yadav, Adam J. Carroll, Gonzalo M. Estavillo, Greg J. Rebetzke, Barry J. Pogson

Abstract:

Water limits crop productivity, so selecting for minimal yield-gap in drier environments is critical to mitigate against climate change and land-use pressures. To date, no markers measured in glasshouses have been reported to predict field-based drought tolerance. In the field, the best measure of drought tolerance is yield-gap; but this requires multisite trials that are an order of magnitude more resource intensive and can be impacted by weather variation. We investigated the responses of relative water content (RWC), stomatal conductance (gs), chlorophyll content and metabolites in flag leaves of commercial wheat (Triticum aestivum L.) cultivars to three drought treatments in the glasshouse and field environments. We observed strong genetic associations between glasshouse-based RWC, metabolites and Yield gap-based Drought Tolerance (YDT): the ratio of yield in water-limited versus well-watered conditions across 24 field environments spanning sites and seasons. Critically, RWC response to glasshouse drought was strongly associated with both YDT (r2 = 0.85, p < 8E-6) and RWC under field drought (r2 = 0.77, p < 0.05). Multiple regression analyses revealed that 98% of genetic YDT variance was explained by drought responses of four metabolites: serine, asparagine, methionine and lysine (R2 = 0.98; p < 0.01). Fitted coefficients suggested that, for given levels of serine and asparagine, stronger methionine and lysine accumulation was associated with higher YDT. Collectively, our results demonstrate that high-throughput, targeted metabolic phenotyping of glasshouse-grown plants may be an effective tool for the selection of wheat cultivars with high YDT in the field.

Keywords: drought stress, grain yield, metabolomics, stomatal conductance, wheat

Procedia PDF Downloads 233