Search results for: reduced order macro models
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 22885

Search results for: reduced order macro models

20275 Predicting Resistance of Commonly Used Antimicrobials in Urinary Tract Infections: A Decision Tree Analysis

Authors: Meera Tandan, Mohan Timilsina, Martin Cormican, Akke Vellinga

Abstract:

Background: In general practice, many infections are treated empirically without microbiological confirmation. Understanding susceptibility of antimicrobials during empirical prescribing can be helpful to reduce inappropriate prescribing. This study aims to apply a prediction model using a decision tree approach to predict the antimicrobial resistance (AMR) of urinary tract infections (UTI) based on non-clinical features of patients over 65 years. Decision tree models are a novel idea to predict the outcome of AMR at an initial stage. Method: Data was extracted from the database of the microbiological laboratory of the University Hospitals Galway on all antimicrobial susceptibility testing (AST) of urine specimens from patients over the age of 65 from January 2011 to December 2014. The primary endpoint was resistance to common antimicrobials (Nitrofurantoin, trimethoprim, ciprofloxacin, co-amoxiclav and amoxicillin) used to treat UTI. A classification and regression tree (CART) model was generated with the outcome ‘resistant infection’. The importance of each predictor (the number of previous samples, age, gender, location (nursing home, hospital, community) and causative agent) on antimicrobial resistance was estimated. Sensitivity, specificity, negative predictive (NPV) and positive predictive (PPV) values were used to evaluate the performance of the model. Seventy-five percent (75%) of the data were used as a training set and validation of the model was performed with the remaining 25% of the dataset. Results: A total of 9805 UTI patients over 65 years had their urine sample submitted for AST at least once over the four years. E.coli, Klebsiella, Proteus species were the most commonly identified pathogens among the UTI patients without catheter whereas Sertia, Staphylococcus aureus; Enterobacter was common with the catheter. The validated CART model shows slight differences in the sensitivity, specificity, PPV and NPV in between the models with and without the causative organisms. The sensitivity, specificity, PPV and NPV for the model with non-clinical predictors was between 74% and 88% depending on the antimicrobial. Conclusion: The CART models developed using non-clinical predictors have good performance when predicting antimicrobial resistance. These models predict which antimicrobial may be the most appropriate based on non-clinical factors. Other CART models, prospective data collection and validation and an increasing number of non-clinical factors will improve model performance. The presented model provides an alternative approach to decision making on antimicrobial prescribing for UTIs in older patients.

Keywords: antimicrobial resistance, urinary tract infection, prediction, decision tree

Procedia PDF Downloads 258
20274 The Purification of Waste Printing Developer with the Fixed Bed Adsorption Column

Authors: Kiurski S. Jelena, Ranogajec G. Jonjaua, Kecić S. Vesna, Oros B. Ivana

Abstract:

The present study investigates the effectiveness of newly designed clayey pellets (fired clay pellets diameter sizes of 5 and 8 mm, and unfired clay pellets with the diameter size of 15 mm) as the beds in the column adsorption process. The adsorption experiments in the batch mode were performed before the column experiment with the purpose to determine the order of adsorbent package in the column which was to be designed in the investigation. The column experiment was performed by using a known mass of the clayey beds and the volume of the waste printing developer, which was purified. The column was filled in the following order: fired clay pellets of the diameter size of 5 mm, fired clay pellets of the diameter size of 8 mm, and unfired clay pellets of the diameter size of 15 mm. The selected order of the adsorbents showed a high removal efficiency for zinc (97.8%) and copper (81.5%) ions. These efficiencies were better than those in the case of the already existing mode adsorption. The obtained experimental data present a good basis for the selection of an appropriate column fill, but further testing is necessary in order to obtain more accurate results.

Keywords: clay materials, fix bed adsorption column, metal ions, printing developer

Procedia PDF Downloads 327
20273 Preparation of New Organoclays and Applications for Adsorption of Telon Dyes in Aqueous Solutions

Authors: Benamar Makhoukhi

Abstract:

Clay ion-exchange using bismidazolium salts (MBIM) could provide organophilic clays materials that allow effective retention of polluting dyes. The present investigations deal with bentonite (Bt) modification using (ortho, meta and para) bisimidazolium cations and attempts to remove a synthetic textile dyes, such as (Telon-Orange, Telon-Red and Telon-Blue) by adsorption, from aqueous solutions. The surface modification of MBIM–Bt was examined using infrared spectroscopy (FTIR), X-ray diffraction (XRD) and thermogravimetric analysis (TGA). Adsorption tests applied to Telon dyes revealed a significant increase of the maximum adsorption capacity from ca. 21-28 to 88-108 mg.g-1 after intercalation. The highest adsorption level was noticed for Telon-Orange dye on the p-MBIM–Bt, presumably due higher interlayer space and better diffusion. The pseudo-first order rate equation was able to provide the best description of adsorption kinetics data for all three dyestuffs. The Langmuir and Freundlich adsorption models were applied to describe the equilibrium isotherms and the isotherm constants were also determined. The results show that MBIM–Bt could be employed as low-cost material for the removal of Telon dyes from effluents.

Keywords: Bentonite, Organoclay, Bisimidazolium, Dyes, Isotherms, Adsorption

Procedia PDF Downloads 448
20272 A Discourse Study of Multimodal Intertextuality in Egyptian Social Media Memes

Authors: Ola Hafez

Abstract:

This study examines the way selected Egyptian digitally mediated memes utilize intertextuality as a means of expression. It is motivated by the emerging digital socio-political humorous practice using various forms of political commentary in Egyptian social media. One of these forms involves the use of memes incorporating (often doctored) video frames taken from Egyptian plays, films and songs, and relocated in a different socio-political context, often with a caption that re-appropriates the frame for the purpose of critical commentary, thus juxtaposing the socio-political phenomena being addressed and the Egyptian artistic and cultural heritage. The paper presents a discourse study of a convenience sample of a recent social media campaign and carries out two levels of analysis. At the micro level, the study pinpoints the various modes of intertextuality employed, including verbal as well as visual intertextuality in the light of the work of social semiotics by Kress and van Leeuwen. At the macro level, the paper sheds light on the socio-political implications of such practice in the light of Political Discourse Analysis.

Keywords: digitally mediated discourse, discourse analysis, Egyptian Arabic, intertextuality, memes, multimodality, political discourse analysis

Procedia PDF Downloads 223
20271 Informed Urban Design: Minimizing Urban Heat Island Intensity via Stochastic Optimization

Authors: Luis Guilherme Resende Santos, Ido Nevat, Leslie Norford

Abstract:

The Urban Heat Island (UHI) is characterized by increased air temperatures in urban areas compared to undeveloped rural surrounding environments. With urbanization and densification, the intensity of UHI increases, bringing negative impacts on livability, health and economy. In order to reduce those effects, it is required to take into consideration design factors when planning future developments. Given design constraints such as population size and availability of area for development, non-trivial decisions regarding the buildings’ dimensions and their spatial distribution are required. We develop a framework for optimization of urban design in order to jointly minimize UHI intensity and buildings’ energy consumption. First, the design constraints are defined according to spatial and population limits in order to establish realistic boundaries that would be applicable in real life decisions. Second, the tools Urban Weather Generator (UWG) and EnergyPlus are used to generate outputs of UHI intensity and total buildings’ energy consumption, respectively. Those outputs are changed based on a set of variable inputs related to urban morphology aspects, such as building height, urban canyon width and population density. Lastly, an optimization problem is cast where the utility function quantifies the performance of each design candidate (e.g. minimizing a linear combination of UHI and energy consumption), and a set of constraints to be met is set. Solving this optimization problem is difficult, since there is no simple analytic form which represents the UWG and EnergyPlus models. We therefore cannot use any direct optimization techniques, but instead, develop an indirect “black box” optimization algorithm. To this end we develop a solution that is based on stochastic optimization method, known as the Cross Entropy method (CEM). The CEM translates the deterministic optimization problem into an associated stochastic optimization problem which is simple to solve analytically. We illustrate our model on a typical residential area in Singapore. Due to fast growth in population and built area and land availability generated by land reclamation, urban planning decisions are of the most importance for the country. Furthermore, the hot and humid climate in the country raises the concern for the impact of UHI. The problem presented is highly relevant to early urban design stages and the objective of such framework is to guide decision makers and assist them to include and evaluate urban microclimate and energy aspects in the process of urban planning.

Keywords: building energy consumption, stochastic optimization, urban design, urban heat island, urban weather generator

Procedia PDF Downloads 138
20270 The Influence of Students’ Learning Factor and Parents’ Involvement in Their Learning and Suspension: The Application of Big Data Analysis of Internet of Things Technology

Authors: Chih Ming Kung

Abstract:

This study is an empirical study examining the enrollment rate and dropout rate of students from the perspectives of students’ learning, parents’ involvement and the learning process. Methods: Using the data collected from the entry website of Internet of Things (IoT), parents’ participation and the installation pattern of exit poll website, an investigation was conducted. Results: This study discovered that in the aspect of the degree of involvement, the attractiveness of courses, self-performance and departmental loyalty exerts significant influences on the four aspects: psychological benefits, physical benefits, social benefits and educational benefits of learning benefits. Parents’ participation also exerts a significant influence on the learning benefits. A suitable tool on the cloud was designed to collect the dynamic big data of students’ learning process. Conclusion: This research’s results can be valuable references for the government when making and promoting related policies, with more macro view and consideration. It is also expected to be contributory to schools for the practical study of promotion for enrollment.

Keywords: students’ learning factor, parents’ involvement, involvement, technology

Procedia PDF Downloads 150
20269 Visualization of Interaction between Pochonia Chlamydosporia and Meloidogyne Incognita and Their Impact on Tomato Crop

Authors: Saifullah K., Muhammad Naziruddin Saifullah, Muhammad N.

Abstract:

The bio control potential and mechanism of P. chlamydosporia against Meloidogyne incognita was evaluated in the present study. Under invitro conditions, P. chlamydosporia was tested for parasitism of eggs and females of M. incognita. The results indicated that this fungus parasitized 87% eggs and 82% females. Culture filtrate (CF) of P. chlamydosporia was tested for its larvicide activity against M. incognita 2nd stage juvenile. The maximum mortality was 97.3% at 100% concentration of the culture filtrate while minimum mortality was 7.3% in 25% concentration after 24 hrs. The result of the pot experiment proved that P. chlamydosporia has reduced the incidence of RKN and improved all tested agronomic growth parameters. The treatment with inoculated M. incognita alone reduced plant height, fresh shoot, and fresh root weight by 44.7%, 29.8%, and 32.8% respectively over uninoculated healthy control. Histopathological studies on the interaction of Pochonia chlamydosporia and Meloidogyne incognita on tomato roots revealed anatomical changes among treatments. Less number of galls with small in size and scarcer abnormalities in the vascular cylinder was observed in plants inoculated with P. chlamydosporia and M. incognita than the plants treated with nematode only. The fungus was seen in in the intercellular spaces of cortical and epidermal cells while the vascular bundles of the plant remain intact, inoculated only with P. chlamydosporia. In the infected roots, many mature females were seen which feed on giant cells. The findings also revealed that control healthy plants were not affected and no histological changes were noted.

Keywords: histopathology, Pochonia chlamydosporia, Meloidogyne incognita, tomato

Procedia PDF Downloads 107
20268 Variability Management of Contextual Feature Model in Multi-Software Product Line

Authors: Muhammad Fezan Afzal, Asad Abbas, Imran Khan, Salma Imtiaz

Abstract:

Software Product Line (SPL) paradigm is used for the development of the family of software products that share common and variable features. Feature model is a domain of SPL that consists of common and variable features with predefined relationships and constraints. Multiple SPLs consist of a number of similar common and variable features, such as mobile phones and Tabs. Reusability of common and variable features from the different domains of SPL is a complex task due to the external relationships and constraints of features in the feature model. To increase the reusability of feature model resources from domain engineering, it is required to manage the commonality of features at the level of SPL application development. In this research, we have proposed an approach that combines multiple SPLs into a single domain and converts them to a common feature model. Extracting the common features from different feature models is more effective, less cost and time to market for the application development. For extracting features from multiple SPLs, the proposed framework consists of three steps: 1) find the variation points, 2) find the constraints, and 3) combine the feature models into a single feature model on the basis of variation points and constraints. By using this approach, reusability can increase features from the multiple feature models. The impact of this research is to reduce the development of cost, time to market and increase products of SPL.

Keywords: software product line, feature model, variability management, multi-SPLs

Procedia PDF Downloads 73
20267 Modeling Football Penalty Shootouts: How Improving Individual Performance Affects Team Performance and the Fairness of the ABAB Sequence

Authors: Pablo Enrique Sartor Del Giudice

Abstract:

Penalty shootouts often decide the outcome of important soccer matches. Although usually referred to as ”lotteries”, there is evidence that some national teams and clubs consistently perform better than others. The outcomes are therefore not explained just by mere luck, and therefore there are ways to improve the average performance of players, naturally at the expense of some sort of effort. In this article we study the payoff of player performance improvements in terms of the performance of the team as a whole. To do so we develop an analytical model with static individual performances, as well as Monte Carlo models that take into account the known influence of partial score and round number on individual performances. We find that within a range of usual values, the team performance improves above 70% faster than individual performances do. Using these models, we also estimate that the new ABBA penalty shootout ordering under test reduces almost all the known bias in favor of the first-shooting team under the current ABAB system.

Keywords: football, penalty shootouts, Montecarlo simulation, ABBA

Procedia PDF Downloads 168
20266 Revolving Ferrofluid Flow in Porous Medium with Rotating Disk

Authors: Paras Ram, Vikas Kumar

Abstract:

The transmission of Malaria with seasonal were studied through the use of mathematical models. The data from the annual number of Malaria cases reported to the Division of Epidemiology, Ministry of Public Health, Thailand during the period 1997-2011 were analyzed. The transmission of Malaria with seasonal was studied by formulating a mathematical model which had been modified to describe different situations encountered in the transmission of Malaria. In our model, the population was separated into two groups: the human and vector groups, and then constructed a system of nonlinear differential equations. Each human group was divided into susceptible, infectious in hot season, infectious in rainy season, infectious in cool season and recovered classes. The vector population was separated into two classes only: susceptible and infectious vectors. The analysis of the models was given by the standard dynamical modeling.

Keywords: ferrofluid, magnetic field, porous medium, rotating disk, Neuringer-Rosensweig Model

Procedia PDF Downloads 426
20265 Contextual Toxicity Detection with Data Augmentation

Authors: Julia Ive, Lucia Specia

Abstract:

Understanding and detecting toxicity is an important problem to support safer human interactions online. Our work focuses on the important problem of contextual toxicity detection, where automated classifiers are tasked with determining whether a short textual segment (usually a sentence) is toxic within its conversational context. We use “toxicity” as an umbrella term to denote a number of variants commonly named in the literature, including hate, abuse, offence, among others. Detecting toxicity in context is a non-trivial problem and has been addressed by very few previous studies. These previous studies have analysed the influence of conversational context in human perception of toxicity in controlled experiments and concluded that humans rarely change their judgements in the presence of context. They have also evaluated contextual detection models based on state-of-the-art Deep Learning and Natural Language Processing (NLP) techniques. Counterintuitively, they reached the general conclusion that computational models tend to suffer performance degradation in the presence of context. We challenge these empirical observations by devising better contextual predictive models that also rely on NLP data augmentation techniques to create larger and better data. In our study, we start by further analysing the human perception of toxicity in conversational data (i.e., tweets), in the absence versus presence of context, in this case, previous tweets in the same conversational thread. We observed that the conclusions of previous work on human perception are mainly due to data issues: The contextual data available does not provide sufficient evidence that context is indeed important (even for humans). The data problem is common in current toxicity datasets: cases labelled as toxic are either obviously toxic (i.e., overt toxicity with swear, racist, etc. words), and thus context does is not needed for a decision, or are ambiguous, vague or unclear even in the presence of context; in addition, the data contains labeling inconsistencies. To address this problem, we propose to automatically generate contextual samples where toxicity is not obvious (i.e., covert cases) without context or where different contexts can lead to different toxicity judgements for the same tweet. We generate toxic and non-toxic utterances conditioned on the context or on target tweets using a range of techniques for controlled text generation(e.g., Generative Adversarial Networks and steering techniques). On the contextual detection models, we posit that their poor performance is due to limitations on both of the data they are trained on (same problems stated above) and the architectures they use, which are not able to leverage context in effective ways. To improve on that, we propose text classification architectures that take the hierarchy of conversational utterances into account. In experiments benchmarking ours against previous models on existing and automatically generated data, we show that both data and architectural choices are very important. Our model achieves substantial performance improvements as compared to the baselines that are non-contextual or contextual but agnostic of the conversation structure.

Keywords: contextual toxicity detection, data augmentation, hierarchical text classification models, natural language processing

Procedia PDF Downloads 176
20264 A Comprehensive Metamodel of an Urbanized Information System: Experimental Case

Authors: Leila Trabelsi

Abstract:

The urbanization of Information Systems (IS) is an effective approach to master the complexity of the organization. It strengthens the coherence of IS and aligns it with the business strategy. Moreover, this approach has significant advantages such as reducing Information Technologies (IT) costs, enhancing the IS position in a competitive environment and ensuring the scalability of the IS through the integration of technological innovations. Therefore, the urbanization is considered as a business strategic decision. Thus, its embedding becomes a necessity in order to improve the IS practice. However, there is a lack of experimental cases studying meta-modelling of Urbanized Information System (UIS). The aim of this paper addresses new urbanization content meta-model which permits modelling, testing and taking into consideration organizational aspects. This methodological framework is structured according to two main abstraction levels, a conceptual level and an operational level. For each of these levels, different models are proposed and presented. The proposed model for has been empirically tested on company. The findings of this paper present an experimental study of urbanization meta-model. The paper points out the significant relationships between dimensions and their evolution.

Keywords: urbanization, information systems, enterprise architecture, meta-model

Procedia PDF Downloads 443
20263 Anticancer Potentials of Aqueous Tinospora cordifolia and Its Bioactive Polysaccharide, Arabinogalactan on Benzo(a)Pyrene Induced Pulmonary Tumorigenesis: A Study with Relevance to Blood Based Biomarkers

Authors: Vandana Mohan, Ashwani Koul

Abstract:

Aim: To evaluate the potential of Aqueous Tinospora cordifolia stem extract (Aq.Tc) and Arabinogalactan (AG) on pulmonary carcinogenesis and associated tumor markers. Background: Lung cancer is one of the most frequent malignancy with high mortality rate due to limitation of early detection resulting in low cure rates. Current research effort focuses on identifying some blood-based biomarkers like CEA, ctDNA and LDH which may have potential to detect cancer at an early stage, evaluation of therapeutic response and its recurrence. Medicinal plants and their active components have been widely investigated for their anticancer potentials. Aqueous preparation of T. Cordifolia extract is enriched in the polysaccharide fraction i.e., AG when compared with other types of extract. Moreover, reports are available of polysaccharide fraction of T. Cordifolia in in vitro lung cancer models which showed profound anti-metastatic activity against these cell lines. However, not much has been explored about its effect in in vivo lung cancer models and the underlying mechanism involved. Experimental Design: Mice were randomly segregated into six groups. Group I animals served as control. Group II animals were administered with Aq. Tc extract (200 mg/kg b.w.) p.o.on the alternate days. Group III animals were fed with AG (7.5 mg/kg b.w.) p.o. on the alternate days (thrice a week). Group IV animals were installed with Benzo(a)pyrene (50 mg/kg b.w.), i.p. twice within an interval of two weeks. Group V animals received Aq. Tc extract as in group II along with it B(a)P was installed after two weeks of Aq. Tc administration following the same protocol as for group IV. Group VI animals received AG as in group III along with it B(a)P was installed after two weeks of AG administration. Results: Administration of B(a)P to mice resulted in increased tumor incidence, multiplicity and pulmonary somatic index with concomitant increase in serum/plasma markers like CEA, ctDNA, LDH and TNF-α.Aq.Tc and AG supplementation significantly attenuated these alterations at different stages of tumorigenesis thereby showing potent anti-cancer effect in lung cancer. A pronounced decrease in serum/plasma markers were observed in animals treated with Aq.Tc as compared to those fed with AG. Also, extensive hyperproliferation of alveolar epithelium was prominent in B(a)P induced lung tumors. However, treatment of Aq.Tc and AG to lung tumor bearing mice exhibited reduced alveolar damage evident from decreased number of hyperchromatic irregular nuclei. A direct correlation between the concentration of tumor markers and the intensity of lung cancer was observed in animals bearing cancer co-treated with Aq.Tc and AG. Conclusion: These findings substantiate the chemopreventive potential of Aq.Tc and AG against lung tumorigenesis. Interestingly, Aq.Tc was found to be more effective in modulating the cancer as reflected by various observations which may be attributed to the synergism offered by various components of Aq.Tc. Further studies are in progress to understand the underlined mechanism in inhibiting lung tumorigenesis by Aq.Tc and AG.

Keywords: Arabinogalactan, Benzo(a)pyrene B(a)P, carcinoembryonic antigen (CEA), circulating tumor DNA (ctDNA), lactate dehydrogenase (LDH), Tinospora cordifolia

Procedia PDF Downloads 188
20262 A Simulation Study of E-Glass Reinforced Polyurethane Footbed and Investigation of Parameters Effecting Elastic Behaviour of Footbed Material

Authors: Berkay Ergene, Çağın Bolat

Abstract:

In this study, we mainly focused on a simulation study regarding composite footbed in order to contribute to shoe industry. As a footbed, e-glass fiber reinforced polyurethane was determined since polyurethane based materials are already used for footbed in shoe manufacturing frequently. Flat, elliptical and rectangular grooved shoe soles were modeled and analyzed separately as TPU, 10% glass fiber reinforced, 30% glass fiber reinforced and 50% glass fiber reinforced materials according to their properties under three point bending and compression situations to determine the relationship between model, material type and mechanical behaviours of composite model. ANSYS 14.0 APDL mechanical structural module is utilized in all simulations and analyzed stress and strain distributions for different footbed models and materials. Furthermore, materials constants like young modulus, shear modulus, Poisson ratio and density of the composites were calculated theoretically by using composite mixture rule and interpreted for mechanical aspects.

Keywords: composite, elastic behaviour, footbed, simulation

Procedia PDF Downloads 271
20261 Influence of the Refractory Period on Neural Networks Based on the Recognition of Neural Signatures

Authors: José Luis Carrillo-Medina, Roberto Latorre

Abstract:

Experimental evidence has revealed that different living neural systems can sign their output signals with some specific neural signature. Although experimental and modeling results suggest that neural signatures can have an important role in the activity of neural networks in order to identify the source of the information or to contextualize a message, the functional meaning of these neural fingerprints is still unclear. The existence of cellular mechanisms to identify the origin of individual neural signals can be a powerful information processing strategy for the nervous system. We have recently built different models to study the ability of a neural network to process information based on the emission and recognition of specific neural fingerprints. In this paper we further analyze the features that can influence on the information processing ability of this kind of networks. In particular, we focus on the role that the duration of a refractory period in each neuron after emitting a signed message can play in the network collective dynamics.

Keywords: neural signature, neural fingerprint, processing based on signal identification, self-organizing neural network

Procedia PDF Downloads 497
20260 Modeling Waiting and Service Time for Patients: A Case Study of Matawale Health Centre, Zomba, Malawi

Authors: Moses Aron, Elias Mwakilama, Jimmy Namangale

Abstract:

Spending more time on long queues for a basic service remains a common challenge to most developing countries, including Malawi. For health sector in particular, Out-Patient Department (OPD) experiences long queues. This puts the lives of patients at risk. However, using queuing analysis to under the nature of the problems and efficiency of service systems, such problems can be abated. Based on a kind of service, literature proposes different possible queuing models. However, unlike using generalized assumed models proposed by literature, use of real time case study data can help in deeper understanding the particular problem model and how such a model can vary from one day to the other and also from each case to another. As such, this study uses data obtained from one urban HC for BP, Pediatric and General OPD cases to investigate an average queuing time for patients within the system. It seeks to highlight the proper queuing model by investigating the kind of distributions functions over patient’s arrival time, inter-arrival time, waiting time and service time. Comparable with the standard set values by WHO, the study found that patients at this HC spend more waiting times than service times. On model investigation, different days presented different models ranging from an assumed M/M/1, M/M/2 to M/Er/2. As such, through sensitivity analysis, in general, a commonly assumed M/M/1 model failed to fit the data but rather an M/Er/2 demonstrated to fit well. An M/Er/3 model seemed to be good in terms of measuring resource utilization, proposing a need to increase medical personnel at this HC. However, an M/Er/4 showed to cause more idleness of human resources.

Keywords: health care, out-patient department, queuing model, sensitivity analysis

Procedia PDF Downloads 439
20259 Modelling and Simulation Efforts in Scale-Up and Characterization of Semi-Solid Dosage Forms

Authors: Saurav S. Rath, Birendra K. David

Abstract:

Generic pharmaceutical industry has to operate in strict timelines of product development and scale-up from lab to plant. Hence, detailed product & process understanding and implementation of appropriate mechanistic modelling and Quality-by-design (QbD) approaches are imperative in the product life cycle. This work provides example cases of such efforts in topical dosage products. Topical products are typically in the form of emulsions, gels, thick suspensions or even simple solutions. The efficacy of such products is determined by characteristics like rheology and morphology. Defining, and scaling up the right manufacturing process with a given set of ingredients, to achieve the right product characteristics presents as a challenge to the process engineer. For example, the non-Newtonian rheology varies not only with CPPs and CMAs but also is an implicit function of globule size (CQA). Hence, this calls for various mechanistic models, to help predict the product behaviour. This paper focusses on such models obtained from computational fluid dynamics (CFD) coupled with population balance modelling (PBM) and constitutive models (like shear, energy density). In a special case of the use of high shear homogenisers (HSHs) for the manufacture of thick emulsions/gels, this work presents some findings on (i) scale-up algorithm for HSH using shear strain, a novel scale-up parameter for estimating mixing parameters, (ii) non-linear relationship between viscosity and shear imparted into the system, (iii) effect of hold time on rheology of product. Specific examples of how this approach enabled scale-up across 1L, 10L, 200L, 500L and 1000L scales will be discussed.

Keywords: computational fluid dynamics, morphology, quality-by-design, rheology

Procedia PDF Downloads 271
20258 Forecasting Stock Indexes Using Bayesian Additive Regression Tree

Authors: Darren Zou

Abstract:

Forecasting the stock market is a very challenging task. Various economic indicators such as GDP, exchange rates, interest rates, and unemployment have a substantial impact on the stock market. Time series models are the traditional methods used to predict stock market changes. In this paper, a machine learning method, Bayesian Additive Regression Tree (BART) is used in predicting stock market indexes based on multiple economic indicators. BART can be used to model heterogeneous treatment effects, and thereby works well when models are misspecified. It also has the capability to handle non-linear main effects and multi-way interactions without much input from financial analysts. In this research, BART is proposed to provide a reliable prediction on day-to-day stock market activities. By comparing the analysis results from BART and with time series method, BART can perform well and has better prediction capability than the traditional methods.

Keywords: BART, Bayesian, predict, stock

Procedia PDF Downloads 136
20257 Validation of the Career Motivation Scale among Chinese University and Vocational College Teachers

Authors: Wei Zhang, Lifen Zhao

Abstract:

The present study aims to translate and validate the Career Motivation Scale among Chinese university and vocational college teachers. Exploratory factor analysis supported a three-factor structure that was consistent with the original structure of career motivation: career insight, career identity, and career resilience. Confirmatory factor analysis showed that a second-order three-factor model with correlated measurement errors best fit the data. Configural, metric, and scalar invariance models were tested, demonstrating that the Chinese version of the Career Motivation Scale did not differ across groups of school type, educational level, and working years in current institutions. The concurrent validity of the Chinese Career Motivation Scale was confirmed by its significant correlations with work engagement, career adaptability, career satisfaction, job crafting, and intention to quit. The results of the study indicated that the Chinese Career Motivation Scale was a valid and reliable measure of career motivation among university and vocational college teachers in China.

Keywords: career motivation scale, Chinese University, vocational college teachers, measurement invariance, validation

Procedia PDF Downloads 136
20256 Effect of Molybdenum Addition to Aluminum Grain Refined by Titanium Plus Boron on Its Grain Size and Mechanical Characteristics in the Cast and After Pressing by the Equal Channel Angular Pressing Conditions

Authors: A. I. O. Zaid, A. M. Attieh, S. M. A. Al Qawabah

Abstract:

Aluminum and its alloys solidify in columnar structure with large grain size which tends to reduce their mechanical strength and surface quality. They are, therefore, grain refined by addition of either titanium or titanium plus boron to their melt before solidification. Equal channel angular pressing, ECAP, process is a recent forming method for producing heavy plastic deformation in materials. In this paper, the effect of molybdenum addition to aluminum grain refined by Ti+B on its metallurgical and mechanical characteristics are investigated in the as cast condition and after pressing by the ECAP process. It was found that addition of Mo or Ti+B alone or together to aluminum resulted in grain refining of its microstructure in the as cast condition, as the average grain size was reduced from 139 micron to 46 micron when Mo and Ti+B are added together. Pressing by the ECAP process resulted in further refinement of the microstructure where 32 micron of average grain size was achieved in Al and the Al-Mo microalloy. Regarding the mechanical strength, addition of Mo or Ti+B alone to Al resulted in deterioration of its mechanical behavior but resulted in enhancement of its mechanical behavior when added together, increase of 10% in flow stress was achieved at 20% strain. However, pressing by ECAP addition of Mo or Ti+B alone to Al resulted in enhancement of its mechanical strength but reduced its strength when added together.

Keywords: ECAP, aluminum, cast, mechanical characteristics, Mo grain refiner

Procedia PDF Downloads 475
20255 Effect of Realistic Lubricant Properties on Thermal Electrohydrodynamic Lubrication Behavior in Circular Contacts

Authors: Puneet Katyal, Punit Kumar

Abstract:

A great deal of efforts has been done in the field of thermal effects in electrohydrodynamic lubrication (TEHL) during the last five decades. The focus was primarily on the development of an efficient numerical scheme to deal with the computational challenges involved in the solution of TEHL model; however, some important aspects related to the accurate description of lubricant properties such as viscosity, rheology and thermal conductivity in EHL point contact analysis remain largely neglected. A few studies available in this regard are based upon highly complex mathematical models difficult to formulate and execute. Using a simplified thermal EHL model for point contacts, this work sheds some light on the importance of accurate characterization of the lubricant properties and demonstrates that the computed TEHL characteristics are highly sensitive to lubricant properties. It also emphasizes the use of appropriate mathematical models with experimentally determined parameters to account for correct lubricant behaviour.

Keywords: TEHL, shear thinning, rheology, conductivity

Procedia PDF Downloads 204
20254 Orthogonal Metal Cutting Simulation of Steel AISI 1045 via Smoothed Particle Hydrodynamic Method

Authors: Seyed Hamed Hashemi Sohi, Gerald Jo Denoga

Abstract:

Machining or metal cutting is one of the most widely used production processes in industry. The quality of the process and the resulting machined product depends on parameters like tool geometry, material, and cutting conditions. However, the relationships of these parameters to the cutting process are often based mostly on empirical knowledge. In this study, computer modeling and simulation using LS-DYNA software and a Smoothed Particle Hydrodynamic (SPH) methodology, was performed on the orthogonal metal cutting process to analyze three-dimensional deformation of AISI 1045 medium carbon steel during machining. The simulation was performed using the following constitutive models: the Power Law model, the Johnson-Cook model, and the Zerilli-Armstrong models (Z-A). The outcomes were compared against the simulated results obtained by Cenk Kiliçaslan using the Finite Element Method (FEM) and the empirical results of Jaspers and Filice. The analysis shows that the SPH method combined with the Zerilli-Armstrong constitutive model is a viable alternative to simulating the metal cutting process. The tangential force was overestimated by 7%, and the normal force was underestimated by 16% when compared with empirical values. The simulation values for flow stress versus strain at various temperatures were also validated against empirical values. The SPH method using the Z-A model has also proven to be robust against issues of time-scaling. Experimental work was also done to investigate the effects of friction, rake angle and tool tip radius on the simulation.

Keywords: metal cutting, smoothed particle hydrodynamics, constitutive models, experimental, cutting forces analyses

Procedia PDF Downloads 264
20253 Passive Control of Elliptic Jet by Using Triangular and Truncated Tabs

Authors: Saif Akram, E. Rathakrishnan

Abstract:

The mixing promoting efficiency of two identical sharp and truncated vertex triangular tabs offering geometrical blockage of 2.5% each, placed at the exit of a Mach 1.5 elliptic nozzle was studied experimentally. The effectiveness of both the tabs in enhancing the mixing of jets with the ambient air are determined by measuring the Pitot pressure along the jet axis and the jet spread in both the minor and major axes of the elliptic nozzle, covering marginally overexpanded to moderately underexpanded levels at the nozzle exit. The results reveal that both the tabs enhance mixing characteristics of the uncontrolled elliptic jet when placed at minor axis. A core length reduction of 67% is achieved at NPR 3 which is the overexpanded state. Similarly, the core length is reduced by about 67%, 50% and 57% at NPRs of 4, 5 and 6 (underexpanded states) respectively. However, unlike the considerable increment in mixing promoting efficiency by the use of truncated vertex tabs for axisymmetric jets, the effect is not much pronounced for the case of supersonic elliptic jets. The CPD plots for both the cases almost overlap, especially when tabs are placed at minor axis, at all the pressure conditions. While, when the tabs are used at major axis, in the case of overexpanded condition, the sharp vertex triangular tabs act as a better mixing enhancer for the supersonic elliptic jets. For the jet controlled with truncated vertex triangular tabs, the core length reductions are of the same order as those for the sharp vertex triangular tabs. The jet mixing is hardly influenced by the tip effect in case of supersonic elliptic jet.

Keywords: elliptic jet, tabs, truncated, triangular

Procedia PDF Downloads 388
20252 Short Life Cycle Time Series Forecasting

Authors: Shalaka Kadam, Dinesh Apte, Sagar Mainkar

Abstract:

The life cycle of products is becoming shorter and shorter due to increased competition in market, shorter product development time and increased product diversity. Short life cycles are normal in retail industry, style business, entertainment media, and telecom and semiconductor industry. The subject of accurate forecasting for demand of short lifecycle products is of special enthusiasm for many researchers and organizations. Due to short life cycle of products the amount of historical data that is available for forecasting is very minimal or even absent when new or modified products are launched in market. The companies dealing with such products want to increase the accuracy in demand forecasting so that they can utilize the full potential of the market at the same time do not oversupply. This provides the challenge to develop a forecasting model that can forecast accurately while handling large variations in data and consider the complex relationships between various parameters of data. Many statistical models have been proposed in literature for forecasting time series data. Traditional time series forecasting models do not work well for short life cycles due to lack of historical data. Also artificial neural networks (ANN) models are very time consuming to perform forecasting. We have studied the existing models that are used for forecasting and their limitations. This work proposes an effective and powerful forecasting approach for short life cycle time series forecasting. We have proposed an approach which takes into consideration different scenarios related to data availability for short lifecycle products. We then suggest a methodology which combines statistical analysis with structured judgement. Also the defined approach can be applied across domains. We then describe the method of creating a profile from analogous products. This profile can then be used for forecasting products with historical data of analogous products. We have designed an application which combines data, analytics and domain knowledge using point-and-click technology. The forecasting results generated are compared using MAPE, MSE and RMSE error scores. Conclusion: Based on the results it is observed that no one approach is sufficient for short life-cycle forecasting and we need to combine two or more approaches for achieving the desired accuracy.

Keywords: forecast, short life cycle product, structured judgement, time series

Procedia PDF Downloads 363
20251 Incorporating Lexical-Semantic Knowledge into Convolutional Neural Network Framework for Pediatric Disease Diagnosis

Authors: Xiaocong Liu, Huazhen Wang, Ting He, Xiaozheng Li, Weihan Zhang, Jian Chen

Abstract:

The utilization of electronic medical record (EMR) data to establish the disease diagnosis model has become an important research content of biomedical informatics. Deep learning can automatically extract features from the massive data, which brings about breakthroughs in the study of EMR data. The challenge is that deep learning lacks semantic knowledge, which leads to impracticability in medical science. This research proposes a method of incorporating lexical-semantic knowledge from abundant entities into a convolutional neural network (CNN) framework for pediatric disease diagnosis. Firstly, medical terms are vectorized into Lexical Semantic Vectors (LSV), which are concatenated with the embedded word vectors of word2vec to enrich the feature representation. Secondly, the semantic distribution of medical terms serves as Semantic Decision Guide (SDG) for the optimization of deep learning models. The study evaluate the performance of LSV-SDG-CNN model on four kinds of Chinese EMR datasets. Additionally, CNN, LSV-CNN, and SDG-CNN are designed as baseline models for comparison. The experimental results show that LSV-SDG-CNN model outperforms baseline models on four kinds of Chinese EMR datasets. The best configuration of the model yielded an F1 score of 86.20%. The results clearly demonstrate that CNN has been effectively guided and optimized by lexical-semantic knowledge, and LSV-SDG-CNN model improves the disease classification accuracy with a clear margin.

Keywords: convolutional neural network, electronic medical record, feature representation, lexical semantics, semantic decision

Procedia PDF Downloads 129
20250 Evaluation of Natural Waste Materials for Ammonia Removal in Biofilters

Authors: R. F. Vieira, D. Lopes, I. Baptista, S. A. Figueiredo, V. F. Domingues, R. Jorge, C. Delerue-matos, O. M. Freitas

Abstract:

Odours are generated in municipal solid wastes management plants as a result of decomposition of organic matter, especially when anaerobic degradation occurs. Information was collected about the substances and respective concentration in the surrounding atmosphere of some management plants. The main components which are associated with these unpleasant odours were identified: ammonia, hydrogen sulfide and mercaptans. The first is the most common and the one that presents the highest concentrations, reaching values of 700 mg/m3. Biofiltration, which involves simultaneously biodegradation, absorption and adsorption processes, is a sustainable technology for the treatment of these odour emissions when a natural packing material is used. The packing material should ideally be cheap, durable, and allow the maximum microbiological activity and adsorption/absorption. The presence of nutrients and water is required for biodegradation processes. Adsorption and absorption are enhanced by high specific surface area, high porosity and low density. The main purpose of this work is the exploitation of natural waste materials, locally available, as packing media: heather (Erica lusitanica), chestnut bur (from Castanea sativa), peach pits (from Prunus persica) and eucalyptus bark (from Eucalyptus globulus). Preliminary batch tests of ammonia removal were performed in order to select the most interesting materials for biofiltration, which were then characterized. The following physical and chemical parameters were evaluated: density, moisture, pH, buffer and water retention capacity. The determination of equilibrium isotherms and the adjustment to Langmuir and Freundlich models was also performed. Both models can fit the experimental results. Based both in the material performance as adsorbent and in its physical and chemical characteristics, eucalyptus bark was considered the best material. It presents a maximum adsorption capacity of 0.78±0.45 mol/kg for ammonia. The results from its characterization are: 121 kg/m3 density, 9.8% moisture, pH equal to 5.7, buffer capacity of 0.370 mmol H+/kg of dry matter and water retention capacity of 1.4 g H2O/g of dry matter. The application of natural materials locally available, with little processing, in biofiltration is an economic and sustainable alternative that should be explored.

Keywords: ammonia removal, biofiltration, natural materials, odour control

Procedia PDF Downloads 373
20249 Fault Diagnosis and Fault-Tolerant Control of Bilinear-Systems: Application to Heating, Ventilation, and Air Conditioning Systems in Multi-Zone Buildings

Authors: Abderrhamane Jarou, Dominique Sauter, Christophe Aubrun

Abstract:

Over the past decade, the growing demand for energy efficiency in buildings has attracted the attention of the control community. Failures in HVAC (heating, ventilation and air conditioning) systems in buildings can have a significant impact on the desired and expected energy performance of buildings and on the user's comfort as well. FTC is a recent technology area that studies the adaptation of control algorithms to faulty operating conditions of a system. The application of Fault-Tolerant Control (FTC) in HVAC systems has gained attention in the last two decades. The objective is to maintain the variations in system performance due to faults within an acceptable range with respect to the desired nominal behavior. This paper considers the so-called active approach, which is based on fault and identification scheme combined with a control reconfiguration algorithm that consists in determining a new set of control parameters so that the reconfigured performance is "as close as possible, "in some sense, to the nominal performance. Thermal models of buildings and their HVAC systems are described by non-linear (usually bi-linear) equations. Most of the works carried out so far in FDI (fault diagnosis and isolation) or FTC consider a linearized model of the studied system. However, this model is only valid in a reduced range of variation. This study presents a new fault diagnosis (FD) algorithm based on a bilinear observer for the detection and accurate estimation of the magnitude of the HVAC system failure. The main contribution of the proposed FD algorithm is that instead of using specific linearized models, the algorithm inherits the structure of the actual bilinear model of the building thermal dynamics. As an immediate consequence, the algorithm is applicable to a wide range of unpredictable operating conditions, i.e., weather dynamics, outdoor air temperature, zone occupancy profile. A bilinear fault detection observer is proposed for a bilinear system with unknown inputs. The residual vector in the observer design is decoupled from the unknown inputs and, under certain conditions, is made sensitive to all faults. Sufficient conditions are given for the existence of the observer and results are given for the explicit computation of observer design matrices. Dedicated observer schemes (DOS) are considered for sensor FDI while unknown input bilinear observers are considered for actuator or system components FDI. The proposed strategy for FTC works as follows: At a first level, FDI algorithms are implemented, making it also possible to estimate the magnitude of the fault. Once the fault is detected, the fault estimation is then used to feed the second level and reconfigure the control low so that that expected performances are recovered. This paper is organized as follows. A general structure for fault-tolerant control of buildings is first presented and the building model under consideration is introduced. Then, the observer-based design for Fault Diagnosis of bilinear systems is studied. The FTC approach is developed in Section IV. Finally, a simulation example is given in Section V to illustrate the proposed method.

Keywords: bilinear systems, fault diagnosis, fault-tolerant control, multi-zones building

Procedia PDF Downloads 176
20248 A Study on the New Weapon Requirements Analytics Using Simulations and Big Data

Authors: Won Il Jung, Gene Lee, Luis Rabelo

Abstract:

Since many weapon systems are getting more complex and diverse, various problems occur in terms of the acquisition cost, time, and performance limitation. As a matter of fact, the experiment execution in real world is costly, dangerous, and time-consuming to obtain Required Operational Characteristics (ROC) for a new weapon acquisition although enhancing the fidelity of experiment results. Also, until presently most of the research contained a large amount of assumptions so therefore a bias is present in the experiment results. At this moment, the new methodology is proposed to solve these problems without a variety of assumptions. ROC of the new weapon system is developed through the new methodology, which is a way to analyze big data generated by simulating various scenarios based on virtual and constructive models which are involving 6 Degrees of Freedom (6DoF). The new methodology enables us to identify unbiased ROC on new weapons by reducing assumptions and provide support in terms of the optimal weapon systems acquisition.

Keywords: big data, required operational characteristics (ROC), virtual and constructive models, weapon acquisition

Procedia PDF Downloads 292
20247 Instability Index Method and Logistic Regression to Assess Landslide Susceptibility in County Route 89, Taiwan

Authors: Y. H. Wu, Ji-Yuan Lin, Yu-Ming Liou

Abstract:

This study aims to set up the landslide susceptibility map of County Route 89 at Ren-Ai Township in Nantou County using the Instability Index Method and Logistic regression. Seven susceptibility factors including Slope Angle, Aspect, Elevation, Distance to fold, Distance to River, Distance to Road and Accumulated Rainfall were obtained by GIS based on the Typhoon Toraji landslide area identified by Industrial Technology Research Institute in 2001. To calculate the landslide percentage of each factor and acquire the weight and grade the grid by means of Instability Index Method. In this study, landslide susceptibility can be classified into four grades: high, medium high, medium low and low, in order to determine the advantages and disadvantages of the two models. The precision of this model is verified by classification error matrix and SRC curve. These results suggest that the logistic regression model is a preferred method than instability index in the assessment of landslide susceptibility. It is suitable for the landslide prediction and precaution in this area in the future.

Keywords: instability index method, logistic regression, landslide susceptibility, SRC curve

Procedia PDF Downloads 296
20246 Numerical Investigation of AL₂O₃ Nanoparticle Effect on a Boiling Forced Swirl Flow Field

Authors: Ataollah Rabiee1, Amir Hossein Kamalinia, Alireza Atf

Abstract:

One of the most important issues in the design of nuclear fusion power plants is the heat removal from the hottest region at the diverter. Various methods could be employed in order to improve the heat transfer efficiency, such as generating turbulent flow and injection of nanoparticles in the host fluid. In the current study, Water/AL₂O₃ nanofluid forced swirl flow boiling has been investigated by using a homogeneous thermophysical model within the Eulerian-Eulerian framework through a twisted tape tube, and the boiling phenomenon was modeled using the Rensselaer Polytechnic Institute (RPI) approach. In addition to comparing the results with the experimental data and their reasonable agreement, it was evidenced that higher flow mixing results in more uniform bulk temperature and lower wall temperature along the twisted tape tube. The presence of AL₂O₃ nanoparticles in the boiling flow field showed that increasing the nanoparticle concentration leads to a reduced vapor volume fraction and wall temperature. The Computational fluid dynamics (CFD) results show that the average heat transfer coefficient in the tube increases both by increasing the nanoparticle concentration and the insertion of twisted tape, which significantly affects the thermal field of the boiling flow.

Keywords: nanoparticle, boiling, CFD, two phase flow, alumina, ITER

Procedia PDF Downloads 130