Search results for: usual stochastic order
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 14194

Search results for: usual stochastic order

13624 Role-Governed Categorization and Category Learning as a Result from Structural Alignment: The RoleMap Model

Authors: Yolina A. Petrova, Georgi I. Petkov

Abstract:

The paper presents a symbolic model for category learning and categorization (called RoleMap). Unlike the other models which implement learning in a separate working mode, role-governed category learning and categorization emerge in RoleMap while it does its usual reasoning. The model is based on several basic mechanisms known as reflecting the sub-processes of analogy-making. It steps on the assumption that in their everyday life people constantly compare what they experience and what they know. Various commonalities between the incoming information (current experience) and the stored one (long-term memory) emerge from those comparisons. Some of those commonalities are considered to be highly important, and they are transformed into concepts for further use. This process denotes the category learning. When there is missing knowledge in the incoming information (i.e. the perceived object is still not recognized), the model makes anticipations about what is missing, based on the similar episodes from its long-term memory. Various such anticipations may emerge for different reasons. However, with time only one of them wins and is transformed into a category member. This process denotes the act of categorization.

Keywords: analogy-making, categorization, category learning, cognitive modeling, role-governed categories

Procedia PDF Downloads 127
13623 Semi-Supervised Learning for Spanish Speech Recognition Using Deep Neural Networks

Authors: B. R. Campomanes-Alvarez, P. Quiros, B. Fernandez

Abstract:

Automatic Speech Recognition (ASR) is a machine-based process of decoding and transcribing oral speech. A typical ASR system receives acoustic input from a speaker or an audio file, analyzes it using algorithms, and produces an output in the form of a text. Some speech recognition systems use Hidden Markov Models (HMMs) to deal with the temporal variability of speech and Gaussian Mixture Models (GMMs) to determine how well each state of each HMM fits a short window of frames of coefficients that represents the acoustic input. Another way to evaluate the fit is to use a feed-forward neural network that takes several frames of coefficients as input and produces posterior probabilities over HMM states as output. Deep neural networks (DNNs) that have many hidden layers and are trained using new methods have been shown to outperform GMMs on a variety of speech recognition systems. Acoustic models for state-of-the-art ASR systems are usually training on massive amounts of data. However, audio files with their corresponding transcriptions can be difficult to obtain, especially in the Spanish language. Hence, in the case of these low-resource scenarios, building an ASR model is considered as a complex task due to the lack of labeled data, resulting in an under-trained system. Semi-supervised learning approaches arise as necessary tasks given the high cost of transcribing audio data. The main goal of this proposal is to develop a procedure based on acoustic semi-supervised learning for Spanish ASR systems by using DNNs. This semi-supervised learning approach consists of: (a) Training a seed ASR model with a DNN using a set of audios and their respective transcriptions. A DNN with a one-hidden-layer network was initialized; increasing the number of hidden layers in training, to a five. A refinement, which consisted of the weight matrix plus bias term and a Stochastic Gradient Descent (SGD) training were also performed. The objective function was the cross-entropy criterion. (b) Decoding/testing a set of unlabeled data with the obtained seed model. (c) Selecting a suitable subset of the validated data to retrain the seed model, thereby improving its performance on the target test set. To choose the most precise transcriptions, three confidence scores or metrics, regarding the lattice concept (based on the graph cost, the acoustic cost and a combination of both), was performed as selection technique. The performance of the ASR system will be calculated by means of the Word Error Rate (WER). The test dataset was renewed in order to extract the new transcriptions added to the training dataset. Some experiments were carried out in order to select the best ASR results. A comparison between a GMM-based model without retraining and the DNN proposed system was also made under the same conditions. Results showed that the semi-supervised ASR-model based on DNNs outperformed the GMM-model, in terms of WER, in all tested cases. The best result obtained an improvement of 6% relative WER. Hence, these promising results suggest that the proposed technique could be suitable for building ASR models in low-resource environments.

Keywords: automatic speech recognition, deep neural networks, machine learning, semi-supervised learning

Procedia PDF Downloads 324
13622 Measuring Financial Asset Return and Volatility Spillovers, with Application to Sovereign Bond, Equity, Foreign Exchange and Commodity Markets

Authors: Petra Palic, Maruska Vizek

Abstract:

We provide an in-depth analysis of interdependence of asset returns and volatilities in developed and developing countries. The analysis is split into three parts. In the first part, we use multivariate GARCH model in order to provide stylized facts on cross-market volatility spillovers. In the second part, we use a generalized vector autoregressive methodology developed by Diebold and Yilmaz (2009) in order to estimate separate measures of return spillovers and volatility spillovers among sovereign bond, equity, foreign exchange and commodity markets. In particular, our analysis is focused on cross-market return, and volatility spillovers in 19 developed and developing countries. In order to estimate named spillovers, we use daily data from 2008 to 2017. In the third part of the analysis, we use a generalized vector autoregressive framework in order to estimate total and directional volatility spillovers. We use the same daily data span for one developed and one developing country in order to characterize daily volatility spillovers across stock, bond, foreign exchange and commodities markets.

Keywords: cross-market spillovers, sovereign bond markets, equity markets, value at risk (VAR)

Procedia PDF Downloads 244
13621 Design of Effective Decoupling Point in Build-To-Order Systems: Focusing on Trade-Off Relation between Order-To-Delivery Lead Time and Work in Progress

Authors: Zhiyong Li, Hiroshi Katayama

Abstract:

Since 1990s, e-commerce and internet business have been grown gradually over the word and customers tend to express their demand attributes in terms of specification requirement on parts, component, product structure etc. This paper deals with designing effective decoupling points for build to order systems under e-commerce environment, which can be realized through tradeoff relation analysis between two major criteria, customer order lead time and value of work in progress. These KPIs are critical for successful BTO business, namely time-based service effectiveness on coping with customer requirements for the first issue and cost effective ness with risk aversive operations for the second issue. Approach of this paper consists of investigation of successful business standing for BTO scheme, manufacturing model development of this scheme, quantitative evaluation of proposed models by calculation of two KPI values under various decoupling point distributions and discussion of the results brought by pattern of decoupling point distribution, where some cases provide the pareto optimum performances. To extract the relevant trade-off relation between considered KPIs among 2-dimensional resultant performance, useful logic developed by former research work, i.e. Katayama and Fonseca, is applied. Obtained characteristics are evaluated as effective information for managing BTO manufacturing businesses.

Keywords: build-to-order (BTO), decoupling point, e-commerce, order-to-delivery lead time (ODLT), work in progress (WIP)

Procedia PDF Downloads 310
13620 Integrated Modeling Approach for Energy Planning and Climate Change Mitigation Assessment in the State of Florida

Authors: K. Thakkar, C. Ghenai

Abstract:

An integrated modeling approach was used in this study to (1) track energy consumption, production, and resource extraction, (2) track greenhouse gases emissions and (3) analyze emissions for local and regional air pollutions. The model was used in this study for short and long term energy and GHG emissions reduction analysis for the state of Florida. The integrated modeling methodology will help to evaluate the alternative energy scenarios and examine emissions-reduction strategies. The mitigation scenarios have been designed to describe the future energy strategies. They consist of various demand and supply side scenarios. One of the GHG mitigation scenarios is crafted by taking into account the available renewable resources potential for power generation in the state of Florida to compare and analyze the GHG reduction measure against ‘Business As Usual’ and ‘Florida State Policy’ scenario. Two more ‘integrated’ scenarios, (‘Electrification’ and ‘Efficiency and Lifestyle’) are crafted through combination of various mitigation scenarios to assess the cumulative impact of the reduction measures such as technological changes and energy efficiency and conservation.

Keywords: energy planning, climate change mitigation assessment, integrated modeling approach, energy alternatives, and GHG emission reductions

Procedia PDF Downloads 430
13619 Social Media Mining with R. Twitter Analyses

Authors: Diana Codat

Abstract:

Tweets' analysis is part of text mining. Each document is a written text. It's possible to apply the usual text search techniques, in particular by switching to the bag-of-words representation. But the tweets induce peculiarities. Some may enrich the analysis. Thus, their length is calibrated (at least as far as public messages are concerned), special characters make it possible to identify authors (@) and themes (#), the tweet and retweet mechanisms make it possible to follow the diffusion of the information. Conversely, other characteristics may disrupt the analyzes. Because space is limited, authors often use abbreviations, emoticons to express feelings, and they do not pay much attention to spelling. All this creates noise that can complicate the task. The tweets carry a lot of potentially interesting information. Their exploitation is one of the main axes of the analysis of the social networks. We show how to access Twitter-related messages. We will initiate a study of the properties of the tweets, and we will follow up on the exploitation of the content of the messages. We will work under R with the package 'twitteR'. The study of tweets is a strong focus of analysis of social networks because Twitter has become an important vector of communication. This example shows that it is easy to initiate an analysis from data extracted directly online. The data preparation phase is of great importance.

Keywords: data mining, language R, social networks, Twitter

Procedia PDF Downloads 161
13618 Study of Effect of Gear Tooth Accuracy on Transmission Mount Vibration

Authors: Kalyan Deepak Kolla, Ketan Paua, Rajkumar Bhagate

Abstract:

Transmission dynamics occupy major role in customer perception of the product in both senses of touch and quality of sound. The quantity and quality of sound perceived is more concerned with the whine noise of the gears engaged. Whine noise is tonal in nature and tonal noises cause fatigue and irritation to customers, which in turn affect the quality of the product. Transmission error is the usual suspect for whine noise, which can be caused due to misalignments, tolerances, manufacturing variabilities. In-cabin noise is also more sensitive to the gear design. As the details of the gear tooth design and manufacturing are in microns, anything out of the tolerance zone, either in design or manufacturing, will cause a whine noise. This will also cause high variation in stress and deformation due to change in the load and leads to the fatigue failure of the gears. Hence gear design and development take priority in the transmission development process. This paper aims to study such variability by considering five pairs of helical spur gears and their effect on the transmission error, contact pattern and vibration level on the transmission.

Keywords: gears, whine noise, manufacturing variability, mount vibration variability

Procedia PDF Downloads 135
13617 A Study on Computational Fluid Dynamics (CFD)-Based Design Optimization Techniques Using Multi-Objective Evolutionary Algorithms (MOEA)

Authors: Ahmed E. Hodaib, Mohamed A. Hashem

Abstract:

In engineering applications, a design has to be as fully perfect as possible in some defined case. The designer has to overcome many challenges in order to reach the optimal solution to a specific problem. This process is called optimization. Generally, there is always a function called “objective function” that is required to be maximized or minimized by choosing input parameters called “degrees of freedom” within an allowed domain called “search space” and computing the values of the objective function for these input values. It becomes more complex when we have more than one objective for our design. As an example for Multi-Objective Optimization Problem (MOP): A structural design that aims to minimize weight and maximize strength. In such case, the Pareto Optimal Frontier (POF) is used, which is a curve plotting two objective functions for the best cases. At this point, a designer should make a decision to choose the point on the curve. Engineers use algorithms or iterative methods for optimization. In this paper, we will discuss the Evolutionary Algorithms (EA) which are widely used with Multi-objective Optimization Problems due to their robustness, simplicity, suitability to be coupled and to be parallelized. Evolutionary algorithms are developed to guarantee the convergence to an optimal solution. An EA uses mechanisms inspired by Darwinian evolution principles. Technically, they belong to the family of trial and error problem solvers and can be considered global optimization methods with a stochastic optimization character. The optimization is initialized by picking random solutions from the search space and then the solution progresses towards the optimal point by using operators such as Selection, Combination, Cross-over and/or Mutation. These operators are applied to the old solutions “parents” so that new sets of design variables called “children” appear. The process is repeated until the optimal solution to the problem is reached. Reliable and robust computational fluid dynamics solvers are nowadays commonly utilized in the design and analyses of various engineering systems, such as aircraft, turbo-machinery, and auto-motives. Coupling of Computational Fluid Dynamics “CFD” and Multi-Objective Evolutionary Algorithms “MOEA” has become substantial in aerospace engineering applications, such as in aerodynamic shape optimization and advanced turbo-machinery design.

Keywords: mathematical optimization, multi-objective evolutionary algorithms "MOEA", computational fluid dynamics "CFD", aerodynamic shape optimization

Procedia PDF Downloads 243
13616 An Observation Approach of Reading Order for Single Column and Two Column Layout Template

Authors: In-Tsang Lin, Chiching Wei

Abstract:

Reading order is an important task in many digitization scenarios involving the preservation of the logical structure of a document. From the paper survey, it finds that the state-of-the-art algorithm could not fulfill to get the accurate reading order in the portable document format (PDF) files with rich formats, diverse layout arrangement. In recent years, most of the studies on the analysis of reading order have targeted the specific problem of associating layout components with logical labels, while less attention has been paid to the problem of extracting relationships the problem of detecting the reading order relationship between logical components, such as cross-references. Over 3 years of development, the company Foxit has demonstrated the layout recognition (LR) engine in revision 20601 to eager for the accuracy of the reading order. The bounding box of each paragraph can be obtained correctly by the Foxit LR engine, but the result of reading-order is not always correct for single-column, and two-column layout format due to the table issue, formula issue, and multiple mini separated bounding box and footer issue. Thus, the algorithm is developed to improve the accuracy of the reading order based on the Foxit LR structure. In this paper, a creative observation method (Here called the MESH method) is provided here to open a new chance in the research of the reading-order field. Here two important parameters are introduced, one parameter is the number of the bounding box on the right side of the present bounding box (NRight), and another parameter is the number of the bounding box under the present bounding box (Nunder). And the normalized x-value (x/the whole width), the normalized y-value (y/the whole height) of each bounding box, the x-, and y- position of each bounding box were also put into consideration. Initial experimental results of single column layout format demonstrate a 19.33% absolute improvement in accuracy of the reading-order over 7 PDF files (total 150 pages) using our proposed method based on the LR structure over the baseline method using the LR structure in 20601 revision, which its accuracy of the reading-order is 72%. And for two-column layout format, the preliminary results demonstrate a 44.44% absolute improvement in accuracy of the reading-order over 2 PDF files (total 18 pages) using our proposed method based on the LR structure over the baseline method using the LR structure in 20601 revision, which its accuracy of the reading-order is 0%. Until now, the footer issue and a part of multiple mini separated bounding box issue can be solved by using the MESH method. However, there are still three issues that cannot be solved, such as the table issue, formula issue, and the random multiple mini separated bounding boxes. But the detection of the table position and the recognition of the table structure are out of the scope in this paper, and there is needed another research. In the future, the tasks are chosen- how to detect the table position in the page and to extract the content of the table.

Keywords: document processing, reading order, observation method, layout recognition

Procedia PDF Downloads 159
13615 Vibration of Nanobeam Subjected to Constant Magnetic Field and Ramp-Type Thermal Loading under Non-Fourier Heat Conduction Law of Lord-Shulman

Authors: Hamdy M. Youssef

Abstract:

In this work, the usual Euler–Bernoulli nanobeam has been modeled in the context of Lord-Shulman thermoelastic theorem, which contains non-Fourier heat conduction law. The nanobeam has been subjected to a constant magnetic field and ramp-type thermal loading. The Laplace transform definition has been applied to the governing equations, and the solutions have been obtained by using a direct approach. The inversions of the Laplace transform have been calculated numerically by using Tzou approximation method. The solutions have been applied to a nanobeam made of silicon nitride. The distributions of the temperature increment, lateral deflection, strain, stress, and strain-energy density have been represented in figures with different values of the magnetic field intensity and ramp-time heat parameter. The value of the magnetic field intensity and ramp-time heat parameter have significant effects on all the studied functions, and they could be used as tuners to control the energy which has been generated through the nanobeam.

Keywords: nanobeam, vibration, constant magnetic field, ramp-type thermal loading, non-Fourier heat conduction law

Procedia PDF Downloads 118
13614 Quintic Spline Solution of Fourth-Order Parabolic Equations Arising in Beam Theory

Authors: Reza Mohammadi, Mahdieh Sahebi

Abstract:

We develop a method based on polynomial quintic spline for numerical solution of fourth-order non-homogeneous parabolic partial differential equation with variable coefficient. By using polynomial quintic spline in off-step points in space and finite difference in time directions, we obtained two three level implicit methods. Stability analysis of the presented method has been carried out. We solve four test problems numerically to validate the derived method. Numerical comparison with other methods shows the superiority of presented scheme.

Keywords: fourth-order parabolic equation, variable coefficient, polynomial quintic spline, off-step points

Procedia PDF Downloads 335
13613 Optimal Maintenance Policy for a Three-Unit System

Authors: A. Abbou, V. Makis, N. Salari

Abstract:

We study the condition-based maintenance (CBM) problem of a system subject to stochastic deterioration. The system is composed of three units (or modules): (i) Module 1 deterioration follows a Markov process with two operational states and one failure state. The operational states are partially observable through periodic condition monitoring. (ii) Module 2 deterioration follows a Gamma process with a known failure threshold. The deterioration level of this module is fully observable through periodic inspections. (iii) Only the operating age information is available of Module 3. The lifetime of this module has a general distribution. A CBM policy prescribes when to initiate a maintenance intervention and which modules to repair during intervention. Our objective is to determine the optimal CBM policy minimizing the long-run expected average cost of operating the system. This is achieved by formulating a Markov decision process (MDP) and developing the value iteration algorithm for solving the MDP. We provide numerical examples illustrating the cost-effectiveness of the optimal CBM policy through a comparison with heuristic policies commonly found in the literature.

Keywords: reliability, maintenance optimization, Markov decision process, heuristics

Procedia PDF Downloads 199
13612 Breast Cancer Early Recognition, New Methods of Screening, and Analysis

Authors: Sahar Heidary

Abstract:

Breast cancer is a main public common obstacle global. Additionally, it is the second top reason for tumor death across women. Considering breast cancer cure choices can aid private doctors in precaution for their patients through future cancer treatment. This article reviews usual management centered on stage, histology, and biomarkers. The growth of breast cancer is a multi-stage procedure including numerous cell kinds and its inhibition residues stimulating in the universe. Timely identification of breast cancer is one of the finest methods to stop this illness. Entirely chief therapeutic administrations mention screening mammography for women aged 40 years and older. Breast cancer metastasis interpretations for the mainstream of deaths from breast cancer. The discovery of breast cancer metastasis at the initial step is essential for managing and estimate of breast cancer development. Developing methods consuming the exploration of flowing cancer cells illustrate talented outcomes in forecasting and classifying the initial steps of breast cancer metastasis in patients. In public, mammography residues are the key screening implement though the efficiency of medical breast checks and self-checkup is less. Innovative screening methods are doubtful to exchange mammography in the close upcoming for screening the overall people.

Keywords: breast cancer, screening, metastasis, methods

Procedia PDF Downloads 144
13611 Failure Inference and Optimization for Step Stress Model Based on Bivariate Wiener Model

Authors: Soudabeh Shemehsavar

Abstract:

In this paper, we consider the situation under a life test, in which the failure time of the test units are not related deterministically to an observable stochastic time varying covariate. In such a case, the joint distribution of failure time and a marker value would be useful for modeling the step stress life test. The problem of accelerating such an experiment is considered as the main aim of this paper. We present a step stress accelerated model based on a bivariate Wiener process with one component as the latent (unobservable) degradation process, which determines the failure times and the other as a marker process, the degradation values of which are recorded at times of failure. Parametric inference based on the proposed model is discussed and the optimization procedure for obtaining the optimal time for changing the stress level is presented. The optimization criterion is to minimize the approximate variance of the maximum likelihood estimator of a percentile of the products’ lifetime distribution.

Keywords: bivariate normal, Fisher information matrix, inverse Gaussian distribution, Wiener process

Procedia PDF Downloads 305
13610 Fractional Order Controller Design for Vibration Attenuation in an Airplane Wing

Authors: Birs Isabela, Muresan Cristina, Folea Silviu, Prodan Ovidiu

Abstract:

The wing is one of the most important parts of an airplane because it ensures stability, sustenance and maneuverability of the airplane. Because of its shape, the airplane wing can be simplified to a smart beam. Active vibration suppression is realized using piezoelectric actuators that are mounted on the surface of the beam. This work presents a tuning procedure of fractional order controllers based on a graphical approach of the frequency domain representation. The efficacy of the method is proven by practically testing the controller on a laboratory scale experimental stand.

Keywords: fractional order control, piezoelectric actuators, smart beam, vibration suppression

Procedia PDF Downloads 301
13609 Syntax-Related Problems of Translation

Authors: Anna Kesoyan

Abstract:

The present paper deals with the syntax-related problems of translation from English into Armenian. Although Syntax is a part of grammar, syntax-related problems of translation are studied separately during the process of translation. Translation from one language to another is widely accepted as a challenging problem. This becomes even more challenging when the source and target languages are widely different in structure and style, as is the case with English and Armenian. Syntax-related problems of translation from English into Armenian are mainly connected with the syntactical structures of these languages, and particularly, with the word order of the sentence. The word order of the sentence of the Armenian language, which is a synthetic language, is usually characterized as “rather free”, and the word order of the English language, which is an analytical language, is characterized “fixed”. The following research examines the main translation means, particularly, syntactical transformations as the translator has to take real steps while trying to solve certain syntax-related problems. Most of the means of translation are based on the transformation of grammatical components of the sentence, without changing the main information of the text. There are several transformations that occur during translation such as word order of the sentence, transformations of certain grammatical constructions like Infinitive participial construction, Nominative with the Infinitive and Elliptical constructions which have been covered in the following research.

Keywords: elliptical constructions, nominative with the infinitive constructions, fixed and free word order, syntactic structures

Procedia PDF Downloads 431
13608 Quadrature Mirror Filter Bank Design Using Population Based Stochastic Optimization

Authors: Ju-Hong Lee, Ding-Chen Chung

Abstract:

The paper deals with the optimal design of two-channel linear-phase (LP) quadrature mirror filter (QMF) banks using a metaheuristic based optimization technique. Based on the theory of two-channel QMF banks using two recursive digital all-pass filters (DAFs), the design problem is appropriately formulated to result in an objective function which is a weighted sum of the group delay error of the designed QMF bank and the magnitude response error of the designed low-pass analysis filter. Through a frequency sampling and a weighted least squares approach, the optimization problem of the objective function can be solved by utilizing a particle swarm optimization algorithm. The resulting two-channel QMF banks can possess approximately LP response without magnitude distortion. Simulation results are presented for illustration and comparison.

Keywords: quadrature mirror filter bank, digital all-pass filter, weighted least squares algorithm, particle swarm optimization

Procedia PDF Downloads 501
13607 Design and Implementation of Pseudorandom Number Generator Using Android Sensors

Authors: Mochamad Beta Auditama, Yusuf Kurniawan

Abstract:

A smartphone or tablet require a strong randomness to establish secure encrypted communication, encrypt files, etc. Therefore, random number generation is one of the main keys to provide secrecy. Android devices are equipped with hardware-based sensors, such as accelerometer, gyroscope, etc. Each of these sensors provides a stochastic process which has a potential to be used as an extra randomness source, in addition to /dev/random and /dev/urandom pseudorandom number generators. Android sensors can provide randomness automatically. To obtain randomness from Android sensors, each one of Android sensors shall be used to construct an entropy source. After all entropy sources are constructed, output from these entropy sources are combined to provide more entropy. Then, a deterministic process is used to produces a sequence of random bits from the combined output. All of these processes are done in accordance with NIST SP 800-22 and the series of NIST SP 800-90. The operation conditions are done 1) on Android user-space, and 2) the Android device is placed motionless on a desk.

Keywords: Android hardware-based sensor, deterministic process, entropy source, random number generation/generators

Procedia PDF Downloads 359
13606 Scoring System for the Prognosis of Sepsis Patients in Intensive Care Units

Authors: Javier E. García-Gallo, Nelson J. Fonseca-Ruiz, John F. Duitama-Munoz

Abstract:

Sepsis is a syndrome that occurs with physiological and biochemical abnormalities induced by severe infection and carries a high mortality and morbidity, therefore the severity of its condition must be interpreted quickly. After patient admission in an intensive care unit (ICU), it is necessary to synthesize the large volume of information that is collected from patients in a value that represents the severity of their condition. Traditional severity of illness scores seeks to be applicable to all patient populations, and usually assess in-hospital mortality. However, the use of machine learning techniques and the data of a population that shares a common characteristic could lead to the development of customized mortality prediction scores with better performance. This study presents the development of a score for the one-year mortality prediction of the patients that are admitted to an ICU with a sepsis diagnosis. 5650 ICU admissions extracted from the MIMICIII database were evaluated, divided into two groups: 70% to develop the score and 30% to validate it. Comorbidities, demographics and clinical information of the first 24 hours after the ICU admission were used to develop a mortality prediction score. LASSO (least absolute shrinkage and selection operator) and SGB (Stochastic Gradient Boosting) variable importance methodologies were used to select the set of variables that make up the developed score; each of this variables was dichotomized and a cut-off point that divides the population into two groups with different mean mortalities was found; if the patient is in the group that presents a higher mortality a one is assigned to the particular variable, otherwise a zero is assigned. These binary variables are used in a logistic regression (LR) model, and its coefficients were rounded to the nearest integer. The resulting integers are the point values that make up the score when multiplied with each binary variables and summed. The one-year mortality probability was estimated using the score as the only variable in a LR model. Predictive power of the score, was evaluated using the 1695 admissions of the validation subset obtaining an area under the receiver operating characteristic curve of 0.7528, which outperforms the results obtained with Sequential Organ Failure Assessment (SOFA), Oxford Acute Severity of Illness Score (OASIS) and Simplified Acute Physiology Score II (SAPSII) scores on the same validation subset. Observed and predicted mortality rates within estimated probabilities deciles were compared graphically and found to be similar, indicating that the risk estimate obtained with the score is close to the observed mortality, it is also observed that the number of events (deaths) is indeed increasing as the outcome go from the decile with the lowest probabilities to the decile with the highest probabilities. Sepsis is a syndrome that carries a high mortality, 43.3% for the patients included in this study; therefore, tools that help clinicians to quickly and accurately predict a worse prognosis are needed. This work demonstrates the importance of customization of mortality prediction scores since the developed score provides better performance than traditional scoring systems.

Keywords: intensive care, logistic regression model, mortality prediction, sepsis, severity of illness, stochastic gradient boosting

Procedia PDF Downloads 200
13605 Recovery of Waste Acrylic Fibers for the Elimination of Basic Dyes

Authors: N. Ouslimani, M. T. Abadlia

Abstract:

Environment protection is a precondition for sustained growth and a better quality of life for all people on earth. Aqueous industrial effluents are the main sources of pollution. Among the compounds of these effluents, dyes are particularly resistant to discoloration by conventional methods, and discharges present many problems that must be supported. The scientific literature shows that synthetic organic dyes are compounds used in many industrial sectors. They are found in the chemical, car, paper industry and particularly the textile industry, where all the lines and grades of the chemical family are represented. The affinity between the fibers and dyes vary depending on the chemical structure of dyes and the type of materials to which they are applied. It is not uncommon to find that during the dyeing operation from 15 to 20 % of sulfur dyes, and sometimes up to 40 % of the reactants are discharged with the effluent. This study was conducted for the purpose of fading basics dyes from wastewater using as adsorbent fiber waste material. This technique presents an interesting alternative to usual treatment, as it allows the recovery of waste fibers, which can find uses as raw material for the manufacture of cleaning products or in other sectors In this study the results obtained by fading fiber waste are encouraging, given the rate of color removal which is about 90%.This method also helps to decrease BOD and suspended solids MES in an effective way.

Keywords: adsorption, dyes, fiber, valorization, wastewater

Procedia PDF Downloads 270
13604 Automation of Process Waste-Free Air Filtration in Production of Concrete, Reinforced with Basalt Fiber

Authors: Stanislav Perepechko

Abstract:

Industrial companies - one of the major sources of harmful substances to the atmosphere. The main cause of pollution on the concrete plants are cement dust emissions. All the cement silos, pneumatic transport, and ventilation systems equipped with filters, to avoid this. Today, many Russian companies have to decide on replacement morally and physically outdated filters and guided back to the electrostatic filters as usual equipment. The offered way of a cleaning of waste-free filtering of air differs in the fact that a filtering medium of the filter is used in concrete manufacture. Basalt is widespread and pollution-free material. In the course of cleaning, one part of basalt fiber and cement immediately goes to the mixer through flow-control units of initial basalt fiber and cement. Another part of basalt fiber goes to filters for purification of the air used in systems of an air lift, and ventilating emissions passes through them, and with trapped particles also goes to the mixer through flow-control units of the basalt fiber fulfilled in filters. At the same time, regulators are adjusted in such a way that total supply of basalt fiber and cement into the mixer remains invariable and corresponds to a given technological mode.

Keywords: waste-free air filtration, concrete, basalt fiber, building automation

Procedia PDF Downloads 417
13603 The Study of Intangible Assets at Various Firm States

Authors: Gulnara Galeeva, Yulia Kasperskaya

Abstract:

The study deals with the relevant problem related to the formation of the efficient investment portfolio of an enterprise. The structure of the investment portfolio is connected to the degree of influence of intangible assets on the enterprise’s income. This determines the importance of research on the content of intangible assets. However, intangible assets studies do not take into consideration how the enterprise state can affect the content and the importance of intangible assets for the enterprise`s income. This affects accurateness of the calculations. In order to study this problem, the research was divided into several stages. In the first stage, intangible assets were classified based on their synergies as the underlying intangibles and the additional intangibles. In the second stage, this classification was applied. It showed that the lifecycle model and the theory of abrupt development of the enterprise, that are taken into account while designing investment projects, constitute limit cases of a more general theory of bifurcations. The research identified that the qualitative content of intangible assets significant depends on how close the enterprise is to being in crisis. In the third stage, the author developed and applied the Wide Pairwise Comparison Matrix method. This allowed to establish that using the ratio of the standard deviation to the mean value of the elements of the vector of priority of intangible assets makes it possible to estimate the probability of a full-blown crisis of the enterprise. The author has identified a criterion, which allows making fundamental decisions on investment feasibility. The study also developed an additional rapid method of assessing the enterprise overall status based on using the questionnaire survey with its Director. The questionnaire consists only of two questions. The research specifically focused on the fundamental role of stochastic resonance in the emergence of bifurcation (crisis) in the economic development of the enterprise. The synergetic approach made it possible to describe the mechanism of the crisis start in details and also to identify a range of universal ways of overcoming the crisis. It was outlined that the structure of intangible assets transforms into a more organized state with the strengthened synchronization of all processes as a result of the impact of the sporadic (white) noise. Obtained results offer managers and business owners a simple and an affordable method of investment portfolio optimization, which takes into account how close the enterprise is to a state of a full-blown crisis.

Keywords: analytic hierarchy process, bifurcation, investment portfolio, intangible assets, wide matrix

Procedia PDF Downloads 191
13602 The Role and Importance of Genome Sequencing in Prediction of Cancer Risk

Authors: M. Sadeghi, H. Pezeshk, R. Tusserkani, A. Sharifi Zarchi, A. Malekpour, M. Foroughmand, S. Goliaei, M. Totonchi, N. Ansari–Pour

Abstract:

The role and relative importance of intrinsic and extrinsic factors in the development of complex diseases such as cancer still remains a controversial issue. Determining the amount of variation explained by these factors needs experimental data and statistical models. These models are nevertheless based on the occurrence and accumulation of random mutational events during stem cell division, thus rendering cancer development a stochastic outcome. We demonstrate that not only individual genome sequencing is uninformative in determining cancer risk, but also assigning a unique genome sequence to any given individual (healthy or affected) is not meaningful. Current whole-genome sequencing approaches are therefore unlikely to realize the promise of personalized medicine. In conclusion, since genome sequence differs from cell to cell and changes over time, it seems that determining the risk factor of complex diseases based on genome sequence is somewhat unrealistic, and therefore, the resulting data are likely to be inherently uninformative.

Keywords: cancer risk, extrinsic factors, genome sequencing, intrinsic factors

Procedia PDF Downloads 253
13601 Stochastic Analysis of Linux Operating System through Copula Distribution

Authors: Vijay Vir Singh

Abstract:

This work is focused studying the Linux operating system connected in a LAN (local area network). The STAR topology (to be called subsystem-1) and BUS topology (to be called subsystem-2) are taken into account, which are placed at two different locations and connected to a server through a hub. In the both topologies BUS topology and STAR topology, we have assumed n clients. The system has two types of failures i.e. partial failure and complete failure. Further, the partial failure has been categorized as minor and major partial failure. It is assumed that the minor partial failure degrades the sub-systems and the major partial failure make the subsystem break down mode. The system may completely fail due to failure of server hacking and blocking etc. The system is studied using supplementary variable technique and Laplace transform by using different types of failure and two types of repair. The various measures of reliability for example, availability of system, reliability of system, MTTF, profit function for different parametric values have been discussed.

Keywords: star topology, bus topology, blocking, hacking, Linux operating system, Gumbel-Hougaard family copula, supplementary variable

Procedia PDF Downloads 350
13600 A Privacy Protection Scheme Supporting Fuzzy Search for NDN Routing Cache Data Name

Authors: Feng Tao, Ma Jing, Guo Xian, Wang Jing

Abstract:

Named Data Networking (NDN) replaces IP address of traditional network with data name, and adopts dynamic cache mechanism. In the existing mechanism, however, only one-to-one search can be achieved because every data has a unique name corresponding to it. There is a certain mapping relationship between data content and data name, so if the data name is intercepted by an adversary, the privacy of the data content and user’s interest can hardly be guaranteed. In order to solve this problem, this paper proposes a one-to-many fuzzy search scheme based on order-preserving encryption to reduce the query overhead by optimizing the caching strategy. In this scheme, we use hash value to ensure the user’s query safe from each node in the process of search, so does the privacy of the requiring data content.

Keywords: NDN, order-preserving encryption, fuzzy search, privacy

Procedia PDF Downloads 466
13599 Nanoporous Metals Reinforced with Fullerenes

Authors: Deni̇z Ezgi̇ Gülmez, Mesut Kirca

Abstract:

Nanoporous (np) metals have attracted considerable attention owing to their cellular morphological features at atomistic scale which yield ultra-high specific surface area awarding a great potential to be employed in diverse applications such as catalytic, electrocatalytic, sensing, mechanical and optical. As one of the carbon based nanostructures, fullerenes are also another type of outstanding nanomaterials that have been extensively investigated due to their remarkable chemical, mechanical and optical properties. In this study, the idea of improving the mechanical behavior of nanoporous metals by inclusion of the fullerenes, which offers a new metal-carbon nanocomposite material, is examined and discussed. With this motivation, tensile mechanical behavior of nanoporous metals reinforced with carbon fullerenes is investigated by classical molecular dynamics (MD) simulations. Atomistic models of the nanoporous metals with ultrathin ligaments are obtained through a stochastic process simply based on the intersection of spherical volumes which has been used previously in literature. According to this technique, the atoms within the ensemble of intersecting spherical volumes is removed from the pristine solid block of the selected metal, which results in porous structures with spherical cells. Following this, fullerene units are added into the cellular voids to obtain final atomistic configurations for the numerical tensile tests. Several numerical specimens are prepared with different number of fullerenes per cell and with varied fullerene sizes. LAMMPS code was used to perform classical MD simulations to conduct uniaxial tension experiments on np models filled by fullerenes. The interactions between the metal atoms are modeled by using embedded atomic method (EAM) while adaptive intermolecular reactive empirical bond order (AIREBO) potential is employed for the interaction of carbon atoms. Furthermore, atomic interactions between the metal and carbon atoms are represented by Lennard-Jones potential with appropriate parameters. In conclusion, the ultimate goal of the study is to present the effects of fullerenes embedded into the cellular structure of np metals on the tensile response of the porous metals. The results are believed to be informative and instructive for the experimentalists to synthesize hybrid nanoporous materials with improved properties and multifunctional characteristics.

Keywords: fullerene, intersecting spheres, molecular dynamic, nanoporous metals

Procedia PDF Downloads 226
13598 Employing QR Code as an Effective Educational Tool for Quick Access to Sources of Kindergarten Concepts

Authors: Ahmed Amin Mousa, M. Abd El-Salam

Abstract:

This study discusses a simple solution for the problem of shortage in learning resources for kindergarten teachers. Occasionally, kindergarten teachers cannot access proper resources by usual search methods as libraries or search engines. Furthermore, these methods require a long time and efforts for preparing. The study is expected to facilitate accessing learning resources. Moreover, it suggests a potential direction for using QR code inside the classroom. The present work proposes that QR code can be used for digitizing kindergarten curriculums and accessing various learning resources. It investigates using QR code for saving information related to the concepts which kindergarten teachers use in the current educational situation. The researchers have established a guide for kindergarten teachers based on the Egyptian official curriculum. The guide provides different learning resources for each scientific and mathematical concept in the curriculum, and each learning resource is represented as a QR code image that contains its URL. Therefore, kindergarten teachers can use smartphone applications for reading QR codes and displaying the related learning resources for students immediately. The guide has been provided to a group of 108 teachers for using inside their classrooms. The results showed that the teachers approved the guide, and gave a good response.

Keywords: kindergarten, child, learning resources, QR code, smart phone, mobile

Procedia PDF Downloads 272
13597 Comparative Analysis of Effecting Factors on Fertility by Birth Order: A Hierarchical Approach

Authors: Ali Hesari, Arezoo Esmaeeli

Abstract:

Regarding to dramatic changes of fertility and higher order births during recent decades in Iran, access to knowledge about affecting factors on different birth orders has crucial importance. In this study, According to hierarchical structure of many of social sciences data and the effect of variables of different levels of social phenomena that determine different birth orders in 365 days ending to 1390 census have been explored by multilevel approach. In this paper, 2% individual row data for 1390 census is analyzed by HLM software. Three different hierarchical linear regression models are estimated for data analysis of the first and second, third, fourth and more birth order. Research results displays different outcomes for three models. Individual level variables entered in equation are; region of residence (rural/urban), age, educational level and labor participation status and province level variable is GDP per capita. Results show that individual level variables have different effects in these three models and in second level we have different random and fixed effects in these models.

Keywords: fertility, birth order, hierarchical approach, fixe effects, random effects

Procedia PDF Downloads 323
13596 Remote Video Supervision via DVB-H Channels

Authors: Hanen Ghabi, Youssef Oudhini, Hassen Mnif

Abstract:

By reference to recent publications dealing with the same problem, and as a follow-up to this research work already published, we propose in this article a new original idea of tele supervision exploiting the opportunities offered by the DVB-H system. The objective is to exploit the RF channels of the DVB-H network in order to insert digital remote monitoring images dedicated to a remote solar power plant. Indeed, the DVB-H (Digital Video Broadcast-Handheld) broadcasting system was designed and deployed for digital broadcasting on the same platform as the parent system, DVB-T. We claim to be able to exploit this approach in order to satisfy the operator of remote photovoltaic sites (and others) in order to remotely control the components of isolated installations by means of video surveillance.

Keywords: video surveillance, digital video broadcast-handheld, photovoltaic sites, AVC

Procedia PDF Downloads 167
13595 Optimal Design Solution in "The Small Module" Within the Possibilities of Ecology, Environmental Science/Engineering, and Economics

Authors: Hassan Wajid

Abstract:

We will commend accommodating an environmentally friendly architectural proposal that is extremely common/usual but whose features will make it a sustainable space. In this experiment, the natural and artificial built space is being proposed in such a way that deals with Environmental, Ecological, and Economic Criteria under different climatic conditions. Moreover, the criteria against ecology-environment-economics reflect in the different modules which are being experimented with and analyzed by multiple research groups. The ecological, environmental, and economic services are provided used as units of production side by side, resulting in local job creation and saving resources, for instance, conservation of rainwater, soil formation or protection, less energy consumption to achieve Net Zero, and a stable climate as a whole. The synthesized results from the collected data suggest several aspects to consider when designing buildings for beginning the design process under the supervision of instructors/directors who are responsible for developing curricula and sustainable goals. Hence, the results of the research and the suggestions will benefit the sustainable design through multiple results, heat analysis of different small modules, and comparisons. As a result, it is depleted as the resources are either consumed or the pollution contaminates the resources.

Keywords: optimization, ecology, environment, sustainable solution

Procedia PDF Downloads 48