Search results for: decision uncertainty distance (DUD)
744 Tuberculosis Modelling Using Bio-PEPA Approach
Authors: Dalila Hamami, Baghdad Atmani
Abstract:
Modelling is a widely used tool to facilitate the evaluation of disease management. The interest of epidemiological models lies in their ability to explore hypothetical scenarios and provide decision makers with evidence to anticipate the consequences of disease incursion and impact of intervention strategies.
All models are, by nature, simplification of more complex systems. Models that involve diseases can be classified into different categories depending on how they treat the variability, time, space, and structure of the population. Approaches may be different from simple deterministic mathematical models, to complex stochastic simulations spatially explicit.
Thus, epidemiological modelling is now a necessity for epidemiological investigations, surveillance, testing hypotheses and generating follow-up activities necessary to perform complete and appropriate analysis.
The state of the art presented in the following, allows us to position itself to the most appropriate approaches in the epidemiological study.
Keywords: Bio-PEPA, Cellular automata, Epidemiological modelling, multi agent system, ordinary differential equations, PEPA, Process Algebra, Tuberculosis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2158743 Playing Games with Genetic Algorithms: Application on Price-QoS Competition in Telecommunications Market
Authors: M’hamed Outanoute, Mohamed Baslam, Belaid Bouikhalene
Abstract:
The customers use the best compromise criterion between price and quality of service (QoS) to select or change their Service Provider (SP). The SPs share the same market and are competing to attract more customers to gain more profit. Due to the divergence of SPs interests, we believe that this situation is a non-cooperative game of price and QoS. The game converges to an equilibrium position known Nash Equilibrium (NE). In this work, we formulate a game theoretic framework for the dynamical behaviors of SPs. We use Genetic Algorithms (GAs) to find the price and QoS strategies that maximize the profit for each SP and illustrate the corresponding strategy in NE. In order to quantify how this NE point is performant, we perform a detailed analysis of the price of anarchy induced by the NE solution. Finally, we provide an extensive numerical study to point out the importance of considering price and QoS as a joint decision parameter.
Keywords: Pricing, QoS, Market share game, Genetic algorithms, Nash equilibrium, Learning, Price of anarchy.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1806742 Historical Landscape Affects Present Tree Density in Paddy Field
Authors: Ha T. Pham, Shuichi Miyagawa
Abstract:
Ongoing landscape transformation is one of the major causes behind disappearance of traditional landscapes, and lead to species and resource loss. Tree in paddy fields in the northeast of Thailand is one of those traditional landscapes. Using three different historical time layers, we acknowledged the severe deforestation and rapid urbanization happened in the region. Despite the general thinking of decline in tree density as consequences, the heterogeneous trend of changes in total tree density in three studied landscapes denied the hypothesis that number of trees in paddy field depend on the length of land use practice. On the other hand, due to selection of planting new trees on levees, existence of trees in paddy field now relies on their values for human use. Besides, changes in land use and landscape structure had a significant impact on decision of which tree density level is considered as suitable for the landscape.
Keywords: Aerial photographs, land use change, traditional landscape, tree in paddy fields.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1864741 Trust Building Mechanisms for Electronic Business Networks and Their Relation to eSkills
Authors: Radoslav Delina, Michal Tkáč
Abstract:
Globalization, supported by information and communication technologies, changes the rules of competitiveness and increases the significance of information, knowledge and network cooperation. In line with this trend, the need for efficient trust-building tools has emerged. The absence of trust building mechanisms and strategies was identified within several studies. Through trust development, participation on e-business network and usage of network services will increase and provide to SMEs new economic benefits. This work is focused on effective trust building strategies development for electronic business network platforms. Based on trust building mechanism identification, the questionnairebased analysis of its significance and minimum level of requirements was conducted. In the paper, we are confirming the trust dependency on e-Skills which play crucial role in higher level of trust into the more sophisticated and complex trust building ICT solutions.Keywords: Correlation analysis, decision trees, e-marketplace, trust building
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1927740 Grain Size Characteristics and Sediments Distribution in the Eastern Part of Lekki Lagoon
Authors: Mayowa Philips Ibitola, Abe Oluwaseun Banji, Olorunfemi Akinade-Solomon
Abstract:
A total of 20 bottom sediment samples were collected from the Lekki Lagoon during the wet and dry season. The study was carried out to determine the textural characteristics, sediment distribution pattern and energy of transportation within the lagoon system. The sediment grain sizes and depth profiling was analyzed using dry sieving method and MATLAB algorithm for processing. The granulometric reveals fine grained sand both for the wet and dry season with an average mean value of 2.03 ϕ and -2.88 ϕ, respectively. Sediments were moderately sorted with an average inclusive standard deviation of 0.77 ϕ and -0.82 ϕ. Skewness varied from strongly coarse and near symmetrical 0.34- ϕ and 0.09 ϕ. The kurtosis average value was 0.87 ϕ and -1.4 ϕ (platykurtic and leptokurtic). Entirely, the bathymetry shows an average depth of 4.0 m. The deepest and shallowest area has a depth of 11.2 m and 0.5 m, respectively. High concentration of fine sand was observed at deep areas compared to the shallow areas during wet and dry season. Statistical parameter results show that the overall sediments are sorted, and deposited under low energy condition over a long distance. However, sediment distribution and sediment transport pattern of Lekki Lagoon is controlled by a low energy current and the down slope configuration of the bathymetry enhances the sorting and the deposition rate in the Lekki Lagoon.Keywords: Lekki Lagoon, marine sediment, bathymetry, grain size distribution.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1056739 Fuzzy C-Means Clustering Algorithm for Voltage Stability in Large Power Systems
Authors: Mohamad R. Khaldi, Christine S. Khoury, Guy M. Naim
Abstract:
The steady-state operation of maintaining voltage stability is done by switching various controllers scattered all over the power network. When a contingency occurs, whether forced or unforced, the dispatcher is to alleviate the problem in a minimum time, cost, and effort. Persistent problem may lead to blackout. The dispatcher is to have the appropriate switching of controllers in terms of type, location, and size to remove the contingency and maintain voltage stability. Wrong switching may worsen the problem and that may lead to blackout. This work proposed and used a Fuzzy CMeans Clustering (FCMC) to assist the dispatcher in the decision making. The FCMC is used in the static voltage stability to map instantaneously a contingency to a set of controllers where the types, locations, and amount of switching are induced.Keywords: Fuzzy logic, Power system control, Reactive power control, Voltage control
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1885738 Application of Spreadsheet and Queuing Network Model to Capacity Optimization in Product Development
Authors: Muhammad Marsudi, Dzuraidah Abdul Wahab, Che Hassan Che Haron
Abstract:
Modeling of a manufacturing system enables one to identify the effects of key design parameters on the system performance and as a result to make correct decision. This paper proposes a manufacturing system modeling approach using a spreadsheet model based on queuing network theory, in which a static capacity planning model and stochastic queuing model are integrated. The model was used to improve the existing system utilization in relation to product design. The model incorporates few parameters such as utilization, cycle time, throughput, and batch size. The study also showed that the validity of developed model is good enough to apply and the maximum value of relative error is 10%, far below the limit value 32%. Therefore, the model developed in this study is a valuable alternative model in evaluating a manufacturing systemKeywords: Manufacturing system, product design, spreadsheet model, utilization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1920737 Scheduling a Flexible Flow Shops Problem using DEA
Authors: Fatemeh Dadkhah, Hossein Ali Akbarpour
Abstract:
This paper considers a scheduling problem in flexible flow shops environment with the aim of minimizing two important criteria including makespan and cumulative tardiness of jobs. Since the proposed problem is known as an Np-hard problem in literature, we have to develop a meta-heuristic to solve it. We considered general structure of Genetic Algorithm (GA) and developed a new version of that based on Data Envelopment Analysis (DEA). Two objective functions assumed as two different inputs for each Decision Making Unit (DMU). In this paper we focused on efficiency score of DMUs and efficient frontier concept in DEA technique. After introducing the method we defined two different scenarios with considering two types of mutation operator. Also we provided an experimental design with some computational results to show the performance of algorithm. The results show that the algorithm implements in a reasonable time.Keywords: Data envelopment analysis, Efficiency, Flexible flow shops, Genetic algorithm
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1814736 An Efficient and Optimized Multi Constrained Path Computation for Real Time Interactive Applications in Packet Switched Networks
Authors: P.S. Prakash, S. Selvan
Abstract:
Quality of Service (QoS) Routing aims to find path between source and destination satisfying the QoS requirements which efficiently using the network resources and underlying routing algorithm and to fmd low-cost paths that satisfy given QoS constraints. One of the key issues in providing end-to-end QoS guarantees in packet networks is determining feasible path that satisfies a number of QoS constraints. We present a Optimized Multi- Constrained Routing (OMCR) algorithm for the computation of constrained paths for QoS routing in computer networks. OMCR applies distance vector to construct a shortest path for each destination with reference to a given optimization metric, from which a set of feasible paths are derived at each node. OMCR is able to fmd feasible paths as well as optimize the utilization of network resources. OMCR operates with the hop-by-hop, connectionless routing model in IP Internet and does not create any loops while fmding the feasible paths. Nodes running OMCR not necessarily maintaining global view of network state such as topology, resource information and routing updates are sent only to neighboring nodes whereas its counterpart link-state routing method depend on complete network state for constrained path computation and that incurs excessive communication overhead.
Keywords: QoS Routing, Optimization, feasible path, multiple constraints.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1125735 Classification Based on Deep Neural Cellular Automata Model
Authors: Yasser F. Hassan
Abstract:
Deep learning structure is a branch of machine learning science and greet achievement in research and applications. Cellular neural networks are regarded as array of nonlinear analog processors called cells connected in a way allowing parallel computations. The paper discusses how to use deep learning structure for representing neural cellular automata model. The proposed learning technique in cellular automata model will be examined from structure of deep learning. A deep automata neural cellular system modifies each neuron based on the behavior of the individual and its decision as a result of multi-level deep structure learning. The paper will present the architecture of the model and the results of simulation of approach are given. Results from the implementation enrich deep neural cellular automata system and shed a light on concept formulation of the model and the learning in it.Keywords: Cellular automata, neural cellular automata, deep learning, classification.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 866734 An Overview of Nano-Particles Effect on Mechanical Properties of Composites
Authors: Olatunde I. Sekunowo, Stephen I. Durowaye, Ganiyu I. Lawal
Abstract:
Composites depending on the nature of their constituents and mode of production are regarded as one of the advanced materials that drive today’s technology. This paper attempts a short review of the subject matter with a general aim of pushing to the next level the frontier of knowledge as it impacts the technology of nano-particles manufacturing. The objectives entail an effort to; aggregate recent research efforts in this field, analyse research findings and observations, streamline research efforts and support industry in taking decision on areas of fund deployment. It is envisaged that this work will serve as a quick hand-on compendium material for researchers in this field and a guide to relevant government departments wishing to fund a research whose outcomes have the potential of improving the nation’s GDP.
Keywords: Advanced materials, Composites, Mechanical properties, Nano-particles.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4520733 2.5D Face Recognition Using Gabor Discrete Cosine Transform
Authors: Ali Cheraghian, Farshid Hajati, Soheila Gheisari, Yongsheng Gao
Abstract:
In this paper, we present a novel 2.5D face recognition method based on Gabor Discrete Cosine Transform (GDCT). In the proposed method, the Gabor filter is applied to extract feature vectors from the texture and the depth information. Then, Discrete Cosine Transform (DCT) is used for dimensionality and redundancy reduction to improve computational efficiency. The system is combined texture and depth information in the decision level, which presents higher performance compared to methods, which use texture and depth information, separately. The proposed algorithm is examined on publically available Bosphorus database including models with pose variation. The experimental results show that the proposed method has a higher performance compared to the benchmark.Keywords: Gabor filter, discrete cosine transform, 2.5D face recognition, pose.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1754732 Assessment of Pollution Reduction
Authors: Katarzyna Strzała-Osuch
Abstract:
Environmental investments, including ecological projects, relating to the protection of atmosphere are today a need. However, investing in the environment should be based on rational management rules. This comes across a problem of selecting a method to assess substances reduced during projects. Therefore, a method allowing for the assessment of decision rationality has to be found. The purpose of this article is to present and systematise pollution reduction assessment methods and illustrate theoretical analyses with empirical data. Empirical results confirm theoretical considerations, which proved that the only method for judging pollution reduction, free of apparent disadvantages, is the Eco 99-ratio method. To make decisions on environmental projects, financing institutions should take into account a rationality rule. Therefore the Eco 99-ratio method could be applied to make decisions relating to environmental investments in the area of air protection.Keywords: Assessment of pollution reduction, costs of environmental protection, efficiency of environmental investments.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1327731 The System for Root Canal Length Measurement Based on Multifrequency Impedance Method
Authors: Zheng Zhang, Xin Chen, Guoqing Ding
Abstract:
Electronic apex locators (EAL) has been widely used clinically for measuring root canal working length with high accuracy, which is crucial for successful endodontic treatment. In order to maintain high accuracy in different measurement environments, this study presented a system for root canal length measurement based on multifrequency impedance method. This measuring system can generate a sweep current with frequencies from 100 Hz to 1 MHz through a direct digital synthesizer. Multiple impedance ratios with different combinations of frequencies were obtained and transmitted by an analog-to-digital converter and several of them with representatives will be selected after data process. The system analyzed the functional relationship between these impedance ratios and the distance between the file and the apex with statistics by measuring plenty of teeth. The position of the apical foramen can be determined by the statistical model using these impedance ratios. The experimental results revealed that the accuracy of the system based on multifrequency impedance ratios method to determine the position of the apical foramen was higher than the dual-frequency impedance ratio method. Besides that, for more complex measurement environments, the performance of the system was more stable.Keywords: Root canal length, apex locator, multifrequency impedance, sweep frequency.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 744730 Parametric Analysis in the Electronic Sensor Frequency Adjustment Process
Authors: Rungchat Chompu-Inwai, Akararit Charoenkasemsuk
Abstract:
The use of electronic sensors in the electronics industry has become increasingly popular over the past few years, and it has become a high competition product. The frequency adjustment process is regarded as one of the most important process in the electronic sensor manufacturing process. Due to inaccuracies in the frequency adjustment process, up to 80% waste can be caused due to rework processes; therefore, this study aims to provide a preliminary understanding of the role of parameters used in the frequency adjustment process, and also make suggestions in order to further improve performance. Four parameters are considered in this study: air pressure, dispensing time, vacuum force, and the distance between the needle tip and the product. A full factorial design for experiment 2k was considered to determine those parameters that significantly affect the accuracy of the frequency adjustment process, where a deviation in the frequency after adjustment and the target frequency is expected to be 0 kHz. The experiment was conducted on two levels, using two replications and with five center-points added. In total, 37 experiments were carried out. The results reveal that air pressure and dispensing time significantly affect the frequency adjustment process. The mathematical relationship between these two parameters was formulated, and the optimal parameters for air pressure and dispensing time were found to be 0.45 MPa and 458 ms, respectively. The optimal parameters were examined by carrying out a confirmation experiment in which an average deviation of 0.082 kHz was achieved.Keywords: Design of Experiment, Electronic Sensor, Frequency Adjustment, Parametric Analysis
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1397729 An Intelligent Optimization Model for Multi-objective Order Allocation Planning
Authors: W. K. Wong, Z. X. Guo, P.Y. Mok
Abstract:
This paper presents a multi-objective order allocation planning problem with the consideration of various real-world production features. A novel hybrid intelligent optimization model, integrating a multi-objective memetic optimization process, a Monte Carlo simulation technique and a heuristic pruning technique, is proposed to handle this problem. Experiments based on industrial data are conducted to validate the proposed model. Results show that (1) the proposed model can effectively solve the investigated problem by providing effective production decision-making solutions, which outperformsan NSGA-II-based optimization process and an industrial method.Keywords: Multi-objective order allocation planning, Pareto optimization, Memetic algorithm, Mento Carlo simulation
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1639728 Detecting Major Misconceptions about Employment in ICT: A Study of the Myths about ICT Work among Females
Authors: Eneli Kindsiko, Kulno Türk
Abstract:
The purpose of the current article is to reveal misconceptions about ICT occupations that keep females away from the field. The study focuses on the three phases in one’s career life cycle: pre-university, university and workplace with the aim of investigating how to attract more females into an ICT-related career. By studying nearly 300 secondary school graduates, 102 university students and 18 female ICT specialists, the study revealed six myths that influence the decision-making process of young girls in pursuing an ICT-related education and career. Furthermore, discriminating conception of ICT as a primarily man’s world is developed before the university period. Stereotypical barriers should be brought out to the public debate, so that a remarkable proportion of possible employees (women) would not stay away from the tech-related fields. Countries could make a remarkable leap in efficiency, when turning their attention to the gender-related issues in the labour market structure.Keywords: ICT, women, education, stereotypes, computers.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1640727 Determination and Assessment of Ground Motion and Spectral Parameters for Iran
Authors: G. Ghodrati Amiri, M. Khorasani, Razavian Ameri, M.Mohamadi Dehcheshmeh, S.Fathi
Abstract:
Many studies have been conducted for derivation of attenuation relationships worldwide, however few relationships have been developed to use for the seismic region of Iranian plateau and only few of these studies have been conducted for derivation of attenuation relationships for parameters such as uniform duration. Uniform duration is the total time during which the acceleration is larger than a given threshold value (default is 5% of PGA). In this study, the database was same as that used previously by Ghodrati Amiri et al. (2007) with same correction methods for earthquake records in Iran. However in this study, records from earthquakes with MS< 4.0 were excluded from this database, each record has individually filtered afterward, and therefore the dataset has been expanded. These new set of attenuation relationships for Iran are derived based on tectonic conditions with soil classification into rock and soil. Earthquake parameters were chosen to be hypocentral distance and magnitude in order to make it easier to use the relationships for seismic hazard analysis. Tehran is the capital city of Iran wit ha large number of important structures. In this study, a probabilistic approach has been utilized for seismic hazard assessment of this city. The resulting uniform duration against return period diagrams are suggested to be used in any projects in the area.Keywords: Attenuation Relationships, Iran, Probabilistic Seismic Hazard Analysis, Tehran, Uniform Duration
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1690726 Position Based Routing Protocol with More Reliability in Mobile Ad Hoc Network
Authors: Mahboobeh Abdoos, Karim Faez, Masoud Sabaei
Abstract:
Position based routing protocols are the kinds of routing protocols, which they use of nodes location information, instead of links information to routing. In position based routing protocols, it supposed that the packet source node has position information of itself and it's neighbors and packet destination node. Greedy is a very important position based routing protocol. In one of it's kinds, named MFR (Most Forward Within Radius), source node or packet forwarder node, sends packet to one of it's neighbors with most forward progress towards destination node (closest neighbor to destination). Using distance deciding metric in Greedy to forward packet to a neighbor node, is not suitable for all conditions. If closest neighbor to destination node, has high speed, in comparison with source node or intermediate packet forwarder node speed or has very low remained battery power, then packet loss probability is increased. Proposed strategy uses combination of metrics distancevelocity similarity-power, to deciding about giving the packet to which neighbor. Simulation results show that the proposed strategy has lower lost packets average than Greedy, so it has more reliability.Keywords: Mobile Ad Hoc Network, Position Based, Reliability, Routing.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1763725 Assuming the Decision of Having One (More) Child: The New Dimensions of the Post Communist Romanian Family
Authors: Raluca-Ioana Horea-Şerban, Marinela Istrate
Abstract:
The first part of the paper analyzes the dynamics of the total fertility rate both at the national and regional level, pointing out the regional disparities in the distribution of this indicator. At the same time, we also focus on the collapse of the number of live births, on the changes in the fertility rate by birth rank, as well as on the failure of acquiring the desired number of children. The second part of the study centres upon a survey applied to urban families with 3 and more than 3 offspring. The preliminary analysis highlights the fact that an increased fertility (more than 3rd rank) is triggered by the parents’ above the average material condition and superior education. The current situation of Romania, which is still passing through a period of relatively rapid demographic changes, marked by numerous convulsions, requires a new approach, in compliance with the recent interpretations appropriate to a new post-transitional demographic regime.Keywords: Family size intention, fertility rate, regional disparities, third birth rank.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1719724 Multistage Condition Monitoring System of Aircraft Gas Turbine Engine
Authors: A. M. Pashayev, D. D. Askerov, C. Ardil, R. A. Sadiqov, P. S. Abdullayev
Abstract:
Researches show that probability-statistical methods application, especially at the early stage of the aviation Gas Turbine Engine (GTE) technical condition diagnosing, when the flight information has property of the fuzzy, limitation and uncertainty is unfounded. Hence the efficiency of application of new technology Soft Computing at these diagnosing stages with the using of the Fuzzy Logic and Neural Networks methods is considered. According to the purpose of this problem training with high accuracy of fuzzy multiple linear and non-linear models (fuzzy regression equations) which received on the statistical fuzzy data basis is made. For GTE technical condition more adequate model making dynamics of skewness and kurtosis coefficients- changes are analysed. Researches of skewness and kurtosis coefficients values- changes show that, distributions of GTE work parameters have fuzzy character. Hence consideration of fuzzy skewness and kurtosis coefficients is expedient. Investigation of the basic characteristics changes- dynamics of GTE work parameters allows drawing conclusion on necessity of the Fuzzy Statistical Analysis at preliminary identification of the engines' technical condition. Researches of correlation coefficients values- changes shows also on their fuzzy character. Therefore for models choice the application of the Fuzzy Correlation Analysis results is offered. At the information sufficiency is offered to use recurrent algorithm of aviation GTE technical condition identification (Hard Computing technology is used) on measurements of input and output parameters of the multiple linear and non-linear generalised models at presence of noise measured (the new recursive Least Squares Method (LSM)). The developed GTE condition monitoring system provides stageby- stage estimation of engine technical conditions. As application of the given technique the estimation of the new operating aviation engine technical condition was made.Keywords: aviation gas turbine engine, technical condition, fuzzy logic, neural networks, fuzzy statistics
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1570723 Improvement of the Q-System Using the Rock Engineering System: A Case Study of Water Conveyor Tunnel of Azad Dam
Authors: S. Golmohammadi, M. Noorian Bidgoli
Abstract:
Because the status and mechanical parameters of discontinuities in the rock mass are included in the calculations, various methods of rock engineering classification are often used as a starting point for the design of different types of structures. The Q-system is one of the most frequently used methods for stability analysis and determination of support systems of underground structures in rock, including tunnel. In this method, six main parameters of the rock mass, namely, the Rock Quality Designation (RQD), joint set number (Jn), joint roughness number (Jr), joint alteration number (Ja), joint water parameter (Jw) and Stress Reduction Factor (SRF) are required. In this regard, in order to achieve a reasonable and optimal design, identifying the effective parameters for the stability of the mentioned structures is one of the most important goals and the most necessary actions in rock engineering. Therefore, it is necessary to study the relationships between the parameters of a system and how they interact with each other and, ultimately, the whole system. In this research, it has been attempted to determine the most effective parameters (key parameters) from the six parameters of rock mass in the Q-system using the Rock Engineering System (RES) method to improve the relationships between the parameters in the calculation of the Q value. The RES system is, in fact, a method by which one can determine the degree of cause and effect of a system's parameters by making an interaction matrix. In this research, the geomechanical data collected from the water conveyor tunnel of Azad Dam were used to make the interaction matrix of the Q-system. For this purpose, instead of using the conventional methods that are always accompanied by defects such as uncertainty, the Q-system interaction matrix is coded using a technique that is actually a statistical analysis of the data and determining the correlation coefficient between them. So, the effect of each parameter on the system is evaluated with greater certainty. The results of this study show that the formed interaction matrix provides a reasonable estimate of the effective parameters in the Q-system. Among the six parameters of the Q-system, the SRF and Jr parameters have the maximum and minimum impact on the system, respectively, and also the RQD and Jw parameters have the maximum and minimum impact on the system, respectively. Therefore, by developing this method, we can obtain a more accurate relation to the rock mass classification by weighting the required parameters in the Q-system.
Keywords: Q-system, Rock Engineering System, statistical analysis, rock mass, tunnel.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 295722 Analysis of Translational Ship Oscillations in a Realistic Environment
Authors: Chen Zhang, Bernhard Schwarz-Röhr, Alexander Härting
Abstract:
To acquire accurate ship motions at the center of gravity, a single low-cost inertial sensor is utilized and applied on board to measure ship oscillating motions. As observations, the three axes accelerations and three axes rotational rates provided by the sensor are used. The mathematical model of processing the observation data includes determination of the distance vector between the sensor and the center of gravity in x, y, and z directions. After setting up the transfer matrix from sensor’s own coordinate system to the ship’s body frame, an extended Kalman filter is applied to deal with nonlinearities between the ship motion in the body frame and the observation information in the sensor’s frame. As a side effect, the method eliminates sensor noise and other unwanted errors. Results are not only roll and pitch, but also linear motions, in particular heave and surge at the center of gravity. For testing, we resort to measurements recorded on a small vessel in a well-defined sea state. With response amplitude operators computed numerically by a commercial software (Seaway), motion characteristics are estimated. These agree well with the measurements after processing with the suggested method.
Keywords: Extended Kalman filter, nonlinear estimation, sea trial, ship motion estimation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1054721 A Systems Approach to Gene Ranking from DNA Microarray Data of Cervical Cancer
Authors: Frank Emmert Streib, Matthias Dehmer, Jing Liu, Max Mühlhauser
Abstract:
In this paper we present a method for gene ranking from DNA microarray data. More precisely, we calculate the correlation networks, which are unweighted and undirected graphs, from microarray data of cervical cancer whereas each network represents a tissue of a certain tumor stage and each node in the network represents a gene. From these networks we extract one tree for each gene by a local decomposition of the correlation network. The interpretation of a tree is that it represents the n-nearest neighbor genes on the n-th level of a tree, measured by the Dijkstra distance, and, hence, gives the local embedding of a gene within the correlation network. For the obtained trees we measure the pairwise similarity between trees rooted by the same gene from normal to cancerous tissues. This evaluates the modification of the tree topology due to progression of the tumor. Finally, we rank the obtained similarity values from all tissue comparisons and select the top ranked genes. For these genes the local neighborhood in the correlation networks changes most between normal and cancerous tissues. As a result we find that the top ranked genes are candidates suspected to be involved in tumor growth and, hence, indicates that our method captures essential information from the underlying DNA microarray data of cervical cancer.Keywords: Graph similarity, DNA microarray data, cancer.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1757720 Remaining Useful Life Prediction Using Elliptical Basis Function Network and Markov Chain
Authors: Yi Yu, Lin Ma, Yong Sun, Yuantong Gu
Abstract:
This paper presents a novel method for remaining useful life prediction using the Elliptical Basis Function (EBF) network and a Markov chain. The EBF structure is trained by a modified Expectation-Maximization (EM) algorithm in order to take into account the missing covariate set. No explicit extrapolation is needed for internal covariates while a Markov chain is constructed to represent the evolution of external covariates in the study. The estimated external and the unknown internal covariates constitute an incomplete covariate set which are then used and analyzed by the EBF network to provide survival information of the asset. It is shown in the case study that the method slightly underestimates the remaining useful life of an asset which is a desirable result for early maintenance decision and resource planning.Keywords: Elliptical Basis Function Network, Markov Chain, Missing Covariates, Remaining Useful Life
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1662719 Spatial Distribution and Risk Assessment of As, Hg, Co and Cr in Kaveh Industrial City, using Geostatistic and GIS
Authors: Abbas Hani
Abstract:
The concentrations of As, Hg, Co, Cr and Cd were tested for each soil sample, and their spatial patterns were analyzed by the semivariogram approach of geostatistics and geographical information system technology. Multivariate statistic approaches (principal component analysis and cluster analysis) were used to identify heavy metal sources and their spatial pattern. Principal component analysis coupled with correlation between heavy metals showed that primary inputs of As, Hg and Cd were due to anthropogenic while, Co, and Cr were associated with pedogenic factors. Ordinary kriging was carried out to map the spatial patters of heavy metals. The high pollution sources evaluated was related with usage of urban and industrial wastewater. The results of this study helpful for risk assessment of environmental pollution for decision making for industrial adjustment and remedy soil pollution.Keywords: Geographic Information system, Geostatistics, Kaveh, Multivariate Statistical Analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1981718 Ensembling Classifiers – An Application toImage Data Classification from Cherenkov Telescope Experiment
Authors: Praveen Boinee, Alessandro De Angelis, Gian Luca Foresti
Abstract:
Ensemble learning algorithms such as AdaBoost and Bagging have been in active research and shown improvements in classification results for several benchmarking data sets with mainly decision trees as their base classifiers. In this paper we experiment to apply these Meta learning techniques with classifiers such as random forests, neural networks and support vector machines. The data sets are from MAGIC, a Cherenkov telescope experiment. The task is to classify gamma signals from overwhelmingly hadron and muon signals representing a rare class classification problem. We compare the individual classifiers with their ensemble counterparts and discuss the results. WEKA a wonderful tool for machine learning has been used for making the experiments.Keywords: Ensembles, WEKA, Neural networks [NN], SupportVector Machines [SVM], Random Forests [RF].
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1765717 Design Optimization of Ferrocement-Laminated Plate Using Genetic Algorithm
Authors: M. Rokonuzzaman, Z. Gürdal
Abstract:
This paper describes the design optimization of ferrocement-laminated plate made up of reinforcing steel wire mesh(es) and cement mortar. For the improvement of the designing process, the plate is modeled as a multi-layer medium, dividing the ferrocement plate into layers of mortar and ferrocement. The mortar layers are assumed to be isotropic in nature and the ferrocement layers are assumed to be orthotropic. The ferrocement layers are little stiffer, but much more costlier, than the mortar layers due the presence of steel wire mesh. The optimization is performed for minimum weight design of the laminate using a genetic algorithm. The optimum designs are discussed for different plate configurations and loadings, and it is compared with the worst designs obtained at the final generation. The paper provides a procedure for the designers in decision-making process.
Keywords: Buckling, Ferrocement-Laminated Plate, Genetic Algorithm, Plate Theory.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2186716 e-Collaborative Learning Circles
Authors: C. Ardil
Abstract:
In this paper, we introduce an e-collaborative learning circles methodology which utilizes the information and communication technologies (ICTs) in e-educational processes. In e-collaborative learning circles methodology, the teachers and students announce their research projects on various mailing lists and discussion boards using available ICTs. The teachers & moderators and students who are already members of the e-forums, discuss the project proposals in their classrooms sent out by the potential global partner schools and return the requested feed back to the proposing school(s) about their level of the participation and contribution in the research. In general, an e-collaborative learning circle project is implemented with a small and diverse group (usually 8-10 participants) from around the world. The students meet regularly over a period of weeks/months through the ICTs during the ecollaborative learning process. When the project is completed, a project product (e-book / DVD) is prepared and sent to the circle members. In this research, when taking into account the interests and motivation of the participating students with the facilitating role of the teacher(s), the students in each circle do research to obtain new data and information, thus enabling them to have the opportunity to meet both different cultures and international understandings across the globe. However, while the participants communicate along with the members in the circle they also practice and develop their communication language skills. Finally, teachers and students find the possibility to develop their skills in using the ICTs as well.
Keywords: Distance Education, Online Learning, Web BasedLearning, Learning Circles, e-Collaborative Learning Circles
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1692715 Bayesian Network Based Intelligent Pediatric System
Authors: Jagmohan Mago, Parvinder S. Sandhu, Neeru Chawla
Abstract:
In this paper, a Bayesian Network (BN) based system is presented for providing clinical decision support to healthcare practitioners in rural or remote areas of India for young infants or children up to the age of 5 years. The government is unable to appoint child specialists in rural areas because of inadequate number of available pediatricians. It leads to a high Infant Mortality Rate (IMR). In such a scenario, Intelligent Pediatric System provides a realistic solution. The prototype of an intelligent system has been developed that involves a knowledge component called an Intelligent Pediatric Assistant (IPA); and User Agents (UA) along with their Graphical User Interfaces (GUI). The GUI of UA provides the interface to the healthcare practitioner for submitting sign-symptoms and displaying the expert opinion as suggested by IPA. Depending upon the observations, the IPA decides the diagnosis and the treatment plan. The UA and IPA form client-server architecture for knowledge sharing.Keywords: Network, Based Intelligent, Pediatric System
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2217