Search results for: drug release model
16916 Sound Performance of a Composite Acoustic Coating With Embedded Parallel Plates Under Hydrostatic Pressure
Authors: Bo Hu, Shibo Wang, Haoyang Zhang, Jie Shi
Abstract:
With the development of sonar detection technology, the acoustic stealth technology of underwater vehicles is facing severe challenges. The underwater acoustic coating is developing towards the direction of low-frequency absorption capability and broad absorption frequency bandwidth. In this paper, an acoustic model of underwater acoustic coating of composite material embedded with periodical steel structure is presented. The model has multiple high absorption peaks in the frequency range of 1kHz-8kHz, where achieves high sound absorption and broad bandwidth performance. It is found that the frequencies of the absorption peaks are related to the classic half-wavelength transmission principle. The sound absorption performance of the acoustic model is investigated by the finite element method using COMSOL software. The sound absorption mechanism of the proposed model is explained by the distributions of the displacement vector field. The influence of geometric parameters of periodical steel structure, including thickness and distance, on the sound absorption ability of the proposed model are further discussed. The acoustic model proposed in this study provides an idea for the design of underwater low-frequency broadband acoustic coating, and the results shows the possibility and feasibility for practical underwater application.Keywords: acoustic coating, composite material, broad frequency bandwidth, sound absorption performance
Procedia PDF Downloads 17416915 LGG Architecture for Brain Tumor Segmentation Using Convolutional Neural Network
Authors: Sajeeha Ansar, Asad Ali Safi, Sheikh Ziauddin, Ahmad R. Shahid, Faraz Ahsan
Abstract:
The most aggressive form of brain tumor is called glioma. Glioma is kind of tumor that arises from glial tissue of the brain and occurs quite often. A fully automatic 2D-CNN model for brain tumor segmentation is presented in this paper. We performed pre-processing steps to remove noise and intensity variances using N4ITK and standard intensity correction, respectively. We used Keras open-source library with Theano as backend for fast implementation of CNN model. In addition, we used BRATS 2015 MRI dataset to evaluate our proposed model. Furthermore, we have used SimpleITK open-source library in our proposed model to analyze images. Moreover, we have extracted random 2D patches for proposed 2D-CNN model for efficient brain segmentation. Extracting 2D patched instead of 3D due to less dimensional information present in 2D which helps us in reducing computational time. Dice Similarity Coefficient (DSC) is used as performance measure for the evaluation of the proposed method. Our method achieved DSC score of 0.77 for complete, 0.76 for core, 0.77 for enhanced tumor regions. However, these results are comparable with methods already implemented 2D CNN architecture.Keywords: brain tumor segmentation, convolutional neural networks, deep learning, LGG
Procedia PDF Downloads 18216914 Machine Learning Data Architecture
Authors: Neerav Kumar, Naumaan Nayyar, Sharath Kashyap
Abstract:
Most companies see an increase in the adoption of machine learning (ML) applications across internal and external-facing use cases. ML applications vend output either in batch or real-time patterns. A complete batch ML pipeline architecture comprises data sourcing, feature engineering, model training, model deployment, model output vending into a data store for downstream application. Due to unclear role expectations, we have observed that scientists specializing in building and optimizing models are investing significant efforts into building the other components of the architecture, which we do not believe is the best use of scientists’ bandwidth. We propose a system architecture created using AWS services that bring industry best practices to managing the workflow and simplifies the process of model deployment and end-to-end data integration for an ML application. This narrows down the scope of scientists’ work to model building and refinement while specialized data engineers take over the deployment, pipeline orchestration, data quality, data permission system, etc. The pipeline infrastructure is built and deployed as code (using terraform, cdk, cloudformation, etc.) which makes it easy to replicate and/or extend the architecture to other models that are used in an organization.Keywords: data pipeline, machine learning, AWS, architecture, batch machine learning
Procedia PDF Downloads 6416913 Data-Driven Dynamic Overbooking Model for Tour Operators
Authors: Kannapha Amaruchkul
Abstract:
We formulate a dynamic overbooking model for a tour operator, in which most reservations contain at least two people. The cancellation rate and the timing of the cancellation may depend on the group size. We propose two overbooking policies, namely economic- and service-based. In an economic-based policy, we want to minimize the expected oversold and underused cost, whereas, in a service-based policy, we ensure that the probability of an oversold situation does not exceed the pre-specified threshold. To illustrate the applicability of our approach, we use tour package data in 2016-2018 from a tour operator in Thailand to build a data-driven robust optimization model, and we tested the proposed overbooking policy in 2019. We also compare the data-driven approach to the conventional approach of fitting data into a probability distribution.Keywords: applied stochastic model, data-driven robust optimization, overbooking, revenue management, tour operator
Procedia PDF Downloads 13416912 PM Air Quality of Windsor Regional Scale Transport’s Impact and Climate Change
Authors: Moustafa Osman Mohammed
Abstract:
This paper is mapping air quality model to engineering the industrial system that ultimately utilized in extensive range of energy systems, distribution resources, and end-user technologies. The model is determining long-range transport patterns contribution as area source can either traced from 48 hrs backward trajectory model or remotely described from background measurements data in those days. The trajectory model will be run within stable conditions and quite constant parameters of the atmospheric pressure at the most time of the year. Air parcel trajectory is necessary for estimating the long-range transport of pollutants and other chemical species. It provides a better understanding of airflow patterns. Since a large amount of meteorological data and a great number of calculations are required to drive trajectory, it will be very useful to apply HYPSLIT model to locate areas and boundaries influence air quality at regional location of Windsor. 2–days backward trajectories model at high and low concentration measurements below and upward the benchmark which was areas influence air quality measurement levels. The benchmark level will be considered as 30 (μg/m3) as the moderate level for Ontario region. Thereby, air quality model is incorporating a midpoint concept between biotic and abiotic components to broaden the scope of quantification impact. The later outcomes’ theories of environmental obligation suggest either a recommendation or a decision of what is a legislative should be achieved in mitigation measures of air emission impact ultimately.Keywords: air quality, management systems, environmental impact assessment, industrial ecology, climate change
Procedia PDF Downloads 24716911 Modeling and Statistical Analysis of a Soap Production Mix in Bejoy Manufacturing Industry, Anambra State, Nigeria
Authors: Okolie Chukwulozie Paul, Iwenofu Chinwe Onyedika, Sinebe Jude Ebieladoh, M. C. Nwosu
Abstract:
The research work is based on the statistical analysis of the processing data. The essence is to analyze the data statistically and to generate a design model for the production mix of soap manufacturing products in Bejoy manufacturing company Nkpologwu, Aguata Local Government Area, Anambra state, Nigeria. The statistical analysis shows the statistical analysis and the correlation of the data. T test, Partial correlation and bi-variate correlation were used to understand what the data portrays. The design model developed was used to model the data production yield and the correlation of the variables show that the R2 is 98.7%. However, the results confirm that the data is fit for further analysis and modeling. This was proved by the correlation and the R-squared.Keywords: General Linear Model, correlation, variables, pearson, significance, T-test, soap, production mix and statistic
Procedia PDF Downloads 44516910 The Comparison of Chromium Ions Release for Stainless Steel between Artificial Saliva and Breadfruit Leaf Extracts
Authors: Mirna Febriani
Abstract:
The use of stainless steel wires in the field of dentistry is widely used, especially for orthodontic and prosthodontic treatment using stainless steel wire. The oral cavity is the ideal environment for corrosion, which can be caused by saliva. Prevention of corrosion on stainless steel wires can be done by using an organic or non-organic corrosion inhibitor. One of the organic inhibitors that can be used to prevent corrosion is the leaves of breadfruit. The method used for this research using Atomic Absorption Spectrophotometric test. The results showed that the difference of chromium ion releases on soaking in saliva and breadfruit leaf extracts on days 1, 3, 7 and 14. Statically calculation with independent T-test with p < 0,05 showed the significant difference. The conclusion of this study shows that breadfruit leaf extract can inhibit the corrosion rate of stainless steel wires.Keywords: chromium ion, stainless steel, artificial saliva, breadfruit leaf
Procedia PDF Downloads 17016909 Utilization of Informatics to Transform Clinical Data into a Simplified Reporting System to Examine the Analgesic Prescribing Practices of a Single Urban Hospital’s Emergency Department
Authors: Rubaiat S. Ahmed, Jemer Garrido, Sergey M. Motov
Abstract:
Clinical informatics (CI) enables the transformation of data into a systematic organization that improves the quality of care and the generation of positive health outcomes.Innovative technology through informatics that compiles accurate data on analgesic utilization in the emergency department can enhance pain management in this important clinical setting. We aim to establish a simplified reporting system through CI to examine and assess the analgesic prescribing practices in the EDthrough executing a U.S. federal grant project on opioid reduction initiatives. Queried data points of interest from a level-one trauma ED’s electronic medical records were used to create data sets and develop informational/visual reporting dashboards (on Microsoft Excel and Google Sheets) concerning analgesic usage across several pre-defined parameters and performance metrics using CI. The data was then qualitatively analyzed to evaluate ED analgesic prescribing trends by departmental clinicians and leadership. During a 12-month reporting period (Dec. 1, 2020 – Nov. 30, 2021) for the ongoing project, about 41% of all ED patient visits (N = 91,747) were for pain conditions, of which 81.6% received analgesics in the ED and at discharge (D/C). Of those treated with analgesics, 24.3% received opioids compared to 75.7% receiving opioid alternatives in the ED and at D/C, including non-pharmacological modalities. Demographics showed among patients receiving analgesics, 56.7% were aged between 18-64, 51.8% were male, 51.7% were white, and 66.2% had government funded health insurance. Ninety-one percent of all opioids prescribed were in the ED, with intravenous (IV) morphine, IV fentanyl, and morphine sulfate immediate release (MSIR) tablets accounting for 88.0% of ED dispensed opioids. With 9.3% of all opioids prescribed at D/C, MSIR was dispensed 72.1% of the time. Hydrocodone, oxycodone, and tramadol usage to only 10-15% of the time, and hydromorphone at 0%. Of opioid alternatives, non-steroidal anti-inflammatory drugs were utilized 60.3% of the time, 23.5% with local anesthetics and ultrasound-guided nerve blocks, and 7.9% with acetaminophen as the primary non-opioid drug categories prescribed by ED providers. Non-pharmacological analgesia included virtual reality and other modalities. An average of 18.5 ED opioid orders and 1.9 opioid D/C prescriptions per 102.4 daily ED patient visits was observed for the period. Compared to other specialties within our institution, 2.0% of opioid D/C prescriptions are given by ED providers, compared to the national average of 4.8%. Opioid alternatives accounted for 69.7% and 30.3% usage, versus 90.7% and 9.3% for opioids in the ED and D/C, respectively.There is a pressing need for concise, relevant, and reliable clinical data on analgesic utilization for ED providers and leadership to evaluate prescribing practices and make data-driven decisions. Basic computer software can be used to create effective visual reporting dashboards with indicators that convey relevant and timely information in an easy-to-digest manner. We accurately examined our ED's analgesic prescribing practices using CI through dashboard reporting. Such reporting tools can quickly identify key performance indicators and prioritize data to enhance pain management and promote safe prescribing practices in the emergency setting.Keywords: clinical informatics, dashboards, emergency department, health informatics, healthcare informatics, medical informatics, opioids, pain management, technology
Procedia PDF Downloads 14416908 Machine Learning Analysis of Student Success in Introductory Calculus Based Physics I Course
Authors: Chandra Prayaga, Aaron Wade, Lakshmi Prayaga, Gopi Shankar Mallu
Abstract:
This paper presents the use of machine learning algorithms to predict the success of students in an introductory physics course. Data having 140 rows pertaining to the performance of two batches of students was used. The lack of sufficient data to train robust machine learning models was compensated for by generating synthetic data similar to the real data. CTGAN and CTGAN with Gaussian Copula (Gaussian) were used to generate synthetic data, with the real data as input. To check the similarity between the real data and each synthetic dataset, pair plots were made. The synthetic data was used to train machine learning models using the PyCaret package. For the CTGAN data, the Ada Boost Classifier (ADA) was found to be the ML model with the best fit, whereas the CTGAN with Gaussian Copula yielded Logistic Regression (LR) as the best model. Both models were then tested for accuracy with the real data. ROC-AUC analysis was performed for all the ten classes of the target variable (Grades A, A-, B+, B, B-, C+, C, C-, D, F). The ADA model with CTGAN data showed a mean AUC score of 0.4377, but the LR model with the Gaussian data showed a mean AUC score of 0.6149. ROC-AUC plots were obtained for each Grade value separately. The LR model with Gaussian data showed consistently better AUC scores compared to the ADA model with CTGAN data, except in two cases of the Grade value, C- and A-.Keywords: machine learning, student success, physics course, grades, synthetic data, CTGAN, gaussian copula CTGAN
Procedia PDF Downloads 4416907 Model of Monitoring and Evaluation of Student’s Learning Achievement: Application of Value-Added Assessment
Authors: Jatuphum Ketchatturat
Abstract:
Value-added assessment has been used for developing the model of monitoring and evaluation of student's learning achievement. The steps of model development consist of 1) study and analyisis of the school and the district report system of student achievement and progress, 2) collecting the data of student achievement to develop the value added indicator, 3) developing the system of value-added assessment by participatory action research approach, 4) putting the system of value-added assessment into the educational district of secondary school, 5) determining the quality of the developed system of value-added assessment. The components of the developed model consist of 1) the database of value-added assessment of student's learning achievement, 2) the process of monitoring and evaluation the student's learning achievement, and 3) the reporting system of value-added assessment of student's learning achievement.Keywords: learning achievement, monitoring and evaluation, value-added assessment
Procedia PDF Downloads 42416906 Application of Random Forest Model in The Prediction of River Water Quality
Authors: Turuganti Venkateswarlu, Jagadeesh Anmala
Abstract:
Excessive runoffs from various non-point source land uses, and other point sources are rapidly contaminating the water quality of streams in the Upper Green River watershed, Kentucky, USA. It is essential to maintain the stream water quality as the river basin is one of the major freshwater sources in this province. It is also important to understand the water quality parameters (WQPs) quantitatively and qualitatively along with their important features as stream water is sensitive to climatic events and land-use practices. In this paper, a model was developed for predicting one of the significant WQPs, Fecal Coliform (FC) from precipitation, temperature, urban land use factor (ULUF), agricultural land use factor (ALUF), and forest land-use factor (FLUF) using Random Forest (RF) algorithm. The RF model, a novel ensemble learning algorithm, can even find out advanced feature importance characteristics from the given model inputs for different combinations. This model’s outcomes showed a good correlation between FC and climate events and land use factors (R2 = 0.94) and precipitation and temperature are the primary influencing factors for FC.Keywords: water quality, land use factors, random forest, fecal coliform
Procedia PDF Downloads 19716905 An Adaptive Hybrid Surrogate-Assisted Particle Swarm Optimization Algorithm for Expensive Structural Optimization
Authors: Xiongxiong You, Zhanwen Niu
Abstract:
Choosing an appropriate surrogate model plays an important role in surrogates-assisted evolutionary algorithms (SAEAs) since there are many types and different kernel functions in the surrogate model. In this paper, an adaptive selection of the best suitable surrogate model method is proposed to solve different kinds of expensive optimization problems. Firstly, according to the prediction residual error sum of square (PRESS) and different model selection strategies, the excellent individual surrogate models are integrated into multiple ensemble models in each generation. Then, based on the minimum root of mean square error (RMSE), the best suitable surrogate model is selected dynamically. Secondly, two methods with dynamic number of models and selection strategies are designed, which are used to show the influence of the number of individual models and selection strategy. Finally, some compared studies are made to deal with several commonly used benchmark problems, as well as a rotor system optimization problem. The results demonstrate the accuracy and robustness of the proposed method.Keywords: adaptive selection, expensive optimization, rotor system, surrogates assisted evolutionary algorithms
Procedia PDF Downloads 14116904 Discovering Event Outliers for Drug as Commercial Products
Authors: Arunas Burinskas, Aurelija Burinskiene
Abstract:
On average, ten percent of drugs - commercial products are not available in pharmacies due to shortage. The shortage event disbalance sales and requires a recovery period, which is too long. Therefore, one of the critical issues that pharmacies do not record potential sales transactions during shortage and recovery periods. The authors suggest estimating outliers during shortage and recovery periods. To shorten the recovery period, the authors suggest using average sales per sales day prediction, which helps to protect the data from being downwards or upwards. Authors use the outlier’s visualization method across different drugs and apply the Grubbs test for significance evaluation. The researched sample is 100 drugs in a one-month time frame. The authors detected that high demand variability products had outliers. Among analyzed drugs, which are commercial products i) High demand variability drugs have a one-week shortage period, and the probability of facing a shortage is equal to 69.23%. ii) Mid demand variability drugs have three days shortage period, and the likelihood to fall into deficit is equal to 34.62%. To avoid shortage events and minimize the recovery period, real data must be set up. Even though there are some outlier detection methods for drug data cleaning, they have not been used for the minimization of recovery period once a shortage has occurred. The authors use Grubbs’ test real-life data cleaning method for outliers’ adjustment. In the paper, the outliers’ adjustment method is applied with a confidence level of 99%. In practice, the Grubbs’ test was used to detect outliers for cancer drugs and reported positive results. The application of the Grubbs’ test is used to detect outliers which exceed boundaries of normal distribution. The result is a probability that indicates the core data of actual sales. The application of the outliers’ test method helps to represent the difference of the mean of the sample and the most extreme data considering the standard deviation. The test detects one outlier at a time with different probabilities from a data set with an assumed normal distribution. Based on approximation data, the authors constructed a framework for scaling potential sales and estimating outliers with Grubbs’ test method. The suggested framework is applicable during the shortage event and recovery periods. The proposed framework has practical value and could be used for the minimization of the recovery period required after the shortage of event occurrence.Keywords: drugs, Grubbs' test, outlier, shortage event
Procedia PDF Downloads 13416903 Investigation of the Bioactivity and Efficacy of Personal Care Products Formulated Using Extracts of Azadirachta indica A. Juss
Authors: Ade O. Oyewole, Sunday O. Okoh, Ruth O. Ishola, Adenike D. Odusote, Chima C. Igwe, Gloria N. Elemo, Anthony I. Okoh
Abstract:
Azadirachta indica (Neem tree) also referred to as an all-purpose tree is used in a wide range of medical preparations in tropical and subtropical countries for prevention and management of various livestock, crops products and human diseases. In Nigeria however, the potentials of this plant have not been fully exploited thus it causes an environmental nuisance during the fruiting season. With a rise in the demand for herbal personal care products globally extracts from different parts of the neem plant were used as the bio-active ingredients in the formulation of personal care products. In this study, formulated neem soap, body cream, lotion, toothpaste and shampoo are analyzed to determine their antibacterial, antifungal, and toxicity properties. The efficacies of these products for management of infectious diseases, both oral and dermal, were also investigated in vitro. Oil from the neem seeds obtained using a mechanical press and acetone extracts of both the neem bark and leaves obtained by the maceration method were used in the formulation and production of the neem personal care products. The antimicrobial and toxicity properties of these products were investigated by agar diffusion, and haemolytic methods respectively. The five neem products (NPs) exhibited strong antibacterial activities against four multi–drug resistant pathogenic and three none pathogenic bacterial strains (Escherichia coli (180), Listeria ivanovii, Staphylococcus aureus, Enterobacter cloacae, Vibro spp., Streptococcus uberis, Mycobacterium smegmatis), except the neem lotion with insignificant activity against E. coli and S. aureus. The minimum inhibitory concentration (MIC) range was between 0.20-0.40 mg/ mL. The 5 NPs demonstrated moderate activity against three clinical dermatophytes isolates (Tinea corporis, Tinea capitis, and Tinea cruiz) as well as one fungal strain (Candida albican) with the MIC ranging between 0.30 - 0.50 mg/ mL and 0.550 mg/mL respectively. The soap and shampoo were the most active against test bacteria and fungi. The haemolytic analysis results on the 5 NPs indicated none toxicity at 0.50 mg/ mL in sheep red blood cells (SRBC).Keywords: antimicrobial, Azadirachta indica, multi–drug resistant pathogenic bacteria, personal care products
Procedia PDF Downloads 27016902 Hybrid Inventory Model Optimization under Uncertainties: A Case Study in a Manufacturing Plant
Authors: E. Benga, T. Tengen, A. Alugongo
Abstract:
Periodic and continuous inventory models are the two classical management tools used to handle inventories. These models have advantages and disadvantages. The implementation of both continuous (r,Q) inventory and periodic (R, S) inventory models in most manufacturing plants comes with higher cost. Such high inventory costs are due to the fact that most manufacturing plants are not flexible enough. Since demand and lead-time are two important variables of every inventory models, their effect on the flexibility of the manufacturing plant matter most. Unfortunately, these effects are not clearly understood by managers. The reason is that the decision parameters of the continuous (r, Q) inventory and periodic (R, S) inventory models are not designed to effectively deal with the issues of uncertainties such as poor manufacturing performances, delivery performance supplies performances. There is, therefore, a need to come up with a predictive and hybrid inventory model that can combine in some sense the feature of the aforementioned inventory models. A linear combination technique is used to hybridize both continuous (r, Q) inventory and periodic (R, S) inventory models. The behavior of such hybrid inventory model is described by a differential equation and then optimized. From the results obtained after simulation, the continuous (r, Q) inventory model is more effective than the periodic (R, S) inventory models in the short run, but this difference changes as time goes by. Because the hybrid inventory model is more cost effective than the continuous (r,Q) inventory and periodic (R, S) inventory models in long run, it should be implemented for strategic decisions.Keywords: periodic inventory, continuous inventory, hybrid inventory, optimization, manufacturing plant
Procedia PDF Downloads 38216901 A Fishery Regulation Model: Bargaining over Fishing Pressure
Authors: Duplan Yves Jamont Junior
Abstract:
The Diamond-Mortensen-Pissarides model widely used in labor economics is tailored to fishery. By this way, a fishing function is defined to depict the fishing technology, and Bellman equations are established to describe the behaviors of fishermen and conservationists. On this basis, a negotiation takes place as a Nash-bargaining over the upper limit of the fishing pressure between both political representative groups of fishermen and conservationists. The existence and uniqueness conditions of the Nash-bargained fishing pressure are established. Given the biomass evolution equation, the dynamics of the model variables (fishing pressure, biomass, fish need) is studied.Keywords: conservation, fishery, fishing, Nash bargaining
Procedia PDF Downloads 26016900 Model for Remanufacture of Medical Equipment in Cross Border Collaboration
Authors: Kingsley Oturu, Winifred Ijomah, Wale Coker, Chibueze Achi
Abstract:
With the impact of BREXIT and the need for cross-border collaboration, this international research investigated the use of a conceptual model for remanufacturing medical equipment (with a focus on anesthetic machines and baby incubators). Early findings of the research suggest that contextual factors need to be taken into consideration, as well as an emphasis on cleaning (e.g., sterilization) during the process of remanufacturing medical equipment. For example, copper tubings may be more important in the remanufacturing of anesthetic equipment in tropical climates than in cold climates.Keywords: medical equipment remanufacture, sustainability, circular business models, remanufacture process model
Procedia PDF Downloads 17216899 An Investigation about Rate Of Evaporation from the Water Surface and LNG Pool
Authors: Farokh Alipour, Ali Falavand, Neda Beit Saeid
Abstract:
The calculation of the effect of accidental releases of flammable materials such as LNG requires the use of a suitable consequence model. This study is due to providing a planning advice for developments in the vicinity of LNG sites and other sites handling flammable materials. In this paper, an applicable algorithm that is able to model pool fires on water is presented and applied to estimate pool fire damage zone. This procedure can be used to model pool fires on land and could be helpful in consequence modeling and domino effect zone measurements of flammable materials which is needed in site selection and plant layout.Keywords: LNG, pool fire, spill, radiation
Procedia PDF Downloads 40216898 Caspase-11 and AIM2 Inflammasome are Involved in Smoking-Induced COPD and Lung Adenocarcinoma
Authors: Chiara Colarusso, Michela Terlizzi, Aldo Pinto, Rosalinda Sorrentino
Abstract:
Cigarette smoking is the main cause and the most common risk factor for both COPD and lung cancer. In our previous studies, we proved that caspase-11 in mice and its human analogue, caspase-4, are involved in lung carcinogenesis and that AIM2 inflammasome might play a pro-cancerous role in lung cancer. Therefore, the aim of this study was to investigate potential crosstalk between COPD and lung cancer, focusing on AIM2 and caspase-11-dependent inflammasome signaling pathway. To mimic COPD, we took advantage of an experimental first-hand smoking mouse model and, to confirm what was observed in mice, we used human samples of lung adenocarcinoma patients stratified according to the smoking and COPD status. We demonstrated that smoke exposure led to emphysema-like features, bronchial tone impairment, and release of IL-1-like cytokines (IL-1α, IL-1β, IL-33, IL-18) in a caspase-1 independent manner in C57Bl/6N. Rather, a dysfunctional caspase-11 in smoke-exposed 129Sv mice was associated to lower bronchial inflammation, collagen deposition, and IL-1-like inflammation. In addition, for the first time, we found that AIM2 inflammasome is involved in lung inflammation in smoking and COPD, in that its expression was higher in smoke-exposed C57Bl/6N compared to 129Sv smoking mice, who instead did not show any alteration of AIM2 in both macrophages and dendritic cells. Moreover, we found that AIM2 expression in the cancerous tissue, albeit higher than non-cancerous tissue, was not statistically different according to the COPD and smoking status. Instead, the higher expression of AIM2 in non-cancerous tissue of smoker COPD patients than smokers who did not have COPD was correlated to a higher hazard ratio of poor survival rate than patients who presented lower levels of AIM2. In conclusion, our data highlight that caspase-11 in mice is associated to smoke-induced lung latent inflammation which could drive the establishment of lung cancer, and that AIM2 inflammasome plays a role at the crosstalk between smoking/COPD and lung adenocarcinoma in that its higher presence is correlated to lower survival rate of smoker COPD adenocarcinoma.Keywords: COPD, inflammasome, lung cancer, lung inflammation, smoke
Procedia PDF Downloads 15616897 The Rapid Industrialization Model
Authors: Fredrick Etyang
Abstract:
This paper presents a Rapid Industrialization Model (RIM) designed to support existing industrialization policies, strategies and industrial development plans at National, Regional and Constituent level in Africa. The model will reinforce efforts to attainment of inclusive and sustainable industrialization of Africa by state and non-state actors. The overall objective of this model is to serve as a framework for rapid industrialization in developing economies and the specific objectives range from supporting rapid industrialization development to promoting a structural change in the economy, a balanced regional industrial growth, achievement of local, regional and international competitiveness in areas of clear comparative advantage in industrial exports and ultimately, the RIM will serve as a step-by-step guideline for the industrialization of African Economies. This model is a product of a scientific research process underpinned by desk research through the review of African countries development plans, strategies, datasets, industrialization efforts and consultation with key informants. The rigorous research process unearthed multi-directional and renewed efforts towards industrialization of Africa premised on collective commitment of individual states, regional economic communities and the African union commission among other strategic stakeholders. It was further, established that the inputs into industrialization of Africa outshine the levels of industrial development on the continent. The RIM comes in handy to serve as step-by-step framework for African countries to follow in their industrial development efforts of transforming inputs into tangible outputs and outcomes in the short, intermediate and long-run. This model postulates three stages of industrialization and three phases toward rapid industrialization of African economies, the model is simple to understand, easily implementable and contextualizable with high return on investment for each unit invested into industrialization supported by the model. Therefore, effective implementation of the model will result into inclusive and sustainable rapid industrialization of Africa.Keywords: economic development, industrialization, economic efficiency, exports and imports
Procedia PDF Downloads 8416896 Self-Compacting White Concrete Mix Design Using the Particle Matrix Model
Authors: Samindi Samarakoon, Ørjan Sletbakk Vie, Remi Kleiven Fjelldal
Abstract:
White concrete facade elements are widely used in construction industry. It is challenging to achieve the desired workability in casting of white concrete elements. Particle Matrix model was used for proportioning the self-compacting white concrete (SCWC) to control segregation and bleeding and to improve workability. The paper presents how to reach the target slump flow while controlling bleeding and segregation in SCWC. The amount of aggregates, binders and mixing water, as well as type and dosage of superplasticizer (SP) to be used are the major factors influencing the properties of SCWC. Slump flow and compressive strength tests were carried out to examine the performance of SCWC, and the results indicate that the particle matrix model could produce successfully SCWC controlling segregation and bleeding.Keywords: white concrete, particle matrix model, mix design, construction industry
Procedia PDF Downloads 27016895 CFD Studies on Forced Convection Nanofluid Flow Inside a Circular Conduit
Authors: M. Khalid, W. Rashmi, L. L. Kwan
Abstract:
This work provides an overview on the experimental and numerical simulations of various nanofluids and their flow and heat transfer behavior. It was further extended to study the effect of nanoparticle concentration, fluid flow rates and thermo-physical properties on the heat transfer enhancement of Al2O3/water nanofluid in a turbulent flow circular conduit using ANSYS FLUENT™ 14.0. Single-phase approximation (homogeneous model) and two-phase (mixture and Eulerian) models were used to simulate the nanofluid flow behavior in the 3-D horizontal pipe. The numerical results were further validated with experimental correlations reported in the literature. It was found that heat transfer of nanofluids increases with increasing particle volume concentration and Reynolds number, respectively. Results showed good agreement (~9% deviation) with the experimental correlations, especially for a single-phase model with constant properties. Among two-phase models, mixture model (~14% deviation) showed better prediction compared to Eulerian-dispersed model (~18% deviation) when temperature independent properties were used. Non-drag forces were also employed in the Eulerian two-phase model. However, the two-phase mixture model with temperature dependent nanofluid properties gave slightly closer agreement (~12% deviation).Keywords: nanofluid, CFD, heat transfer, forced convection, circular conduit
Procedia PDF Downloads 52316894 Nowcasting Indonesian Economy
Authors: Ferry Kurniawan
Abstract:
In this paper, we nowcast quarterly output growth in Indonesia by exploiting higher frequency data (monthly indicators) using a mixed-frequency factor model and exploiting both quarterly and monthly data. Nowcasting quarterly GDP in Indonesia is particularly relevant for the central bank of Indonesia which set the policy rate in the monthly Board of Governors Meeting; whereby one of the important step is the assessment of the current state of the economy. Thus, having an accurate and up-to-date quarterly GDP nowcast every time new monthly information becomes available would clearly be of interest for central bank of Indonesia, for example, as the initial assessment of the current state of the economy -including nowcast- will be used as input for longer term forecast. We consider a small scale mixed-frequency factor model to produce nowcasts. In particular, we specify variables as year-on-year growth rates thus the relation between quarterly and monthly data is expressed in year-on-year growth rates. To assess the performance of the model, we compare the nowcasts with two other approaches: autoregressive model –which is often difficult when forecasting output growth- and Mixed Data Sampling (MIDAS) regression. In particular, both mixed frequency factor model and MIDAS nowcasts are produced by exploiting the same set of monthly indicators. Hence, we compare the nowcasts performance of the two approaches directly. To preview the results, we find that by exploiting monthly indicators using mixed-frequency factor model and MIDAS regression we improve the nowcast accuracy over a benchmark simple autoregressive model that uses only quarterly frequency data. However, it is not clear whether the MIDAS or mixed-frequency factor model is better. Neither set of nowcasts encompasses the other; suggesting that both nowcasts are valuable in nowcasting GDP but neither is sufficient. By combining the two individual nowcasts, we find that the nowcast combination not only increases the accuracy - relative to individual nowcasts- but also lowers the risk of the worst performance of the individual nowcasts.Keywords: nowcasting, mixed-frequency data, factor model, nowcasts combination
Procedia PDF Downloads 33116893 Software Reliability Prediction Model Analysis
Authors: Lela Mirtskhulava, Mariam Khunjgurua, Nino Lomineishvili, Koba Bakuria
Abstract:
Software reliability prediction gives a great opportunity to measure the software failure rate at any point throughout system test. A software reliability prediction model provides with the technique for improving reliability. Software reliability is very important factor for estimating overall system reliability, which depends on the individual component reliabilities. It differs from hardware reliability in that it reflects the design perfection. Main reason of software reliability problems is high complexity of software. Various approaches can be used to improve the reliability of software. We focus on software reliability model in this article, assuming that there is a time redundancy, the value of which (the number of repeated transmission of basic blocks) can be an optimization parameter. We consider given mathematical model in the assumption that in the system may occur not only irreversible failures, but also a failure that can be taken as self-repairing failures that significantly affect the reliability and accuracy of information transfer. Main task of the given paper is to find a time distribution function (DF) of instructions sequence transmission, which consists of random number of basic blocks. We consider the system software unreliable; the time between adjacent failures has exponential distribution.Keywords: exponential distribution, conditional mean time to failure, distribution function, mathematical model, software reliability
Procedia PDF Downloads 46416892 Parametric Modeling for Survival Data with Competing Risks Using the Generalized Gompertz Distribution
Authors: Noora Al-Shanfari, M. Mazharul Islam
Abstract:
The cumulative incidence function (CIF) is a fundamental approach for analyzing survival data in the presence of competing risks, which estimates the marginal probability for each competing event. Parametric modeling of CIF has the advantage of fitting various shapes of CIF and estimates the impact of covariates with maximum efficiency. To calculate the total CIF's covariate influence using a parametric model., it is essential to parametrize the baseline of the CIF. As the CIF is an improper function by nature, it is necessary to utilize an improper distribution when applying parametric models. The Gompertz distribution, which is an improper distribution, is limited in its applicability as it only accounts for monotone hazard shapes. The generalized Gompertz distribution, however, can adapt to a wider range of hazard shapes, including unimodal, bathtub, and monotonic increasing or decreasing hazard shapes. In this paper, the generalized Gompertz distribution is used to parametrize the baseline of the CIF, and the parameters of the proposed model are estimated using the maximum likelihood approach. The proposed model is compared with the existing Gompertz model using the Akaike information criterion. Appropriate statistical test procedures and model-fitting criteria will be used to test the adequacy of the model. Both models are applied to the ‘colon’ dataset, which is available in the “biostat3” package in R.Keywords: competing risks, cumulative incidence function, improper distribution, parametric modeling, survival analysis
Procedia PDF Downloads 10416891 The Photovoltaic Panel at End of Life: Experimental Study of Metals Release
Authors: M. Tammaro, S. Manzo, J. Rimauro, A. Salluzzo, S. Schiavo
Abstract:
The solar photovoltaic (PV) modules are considered to have a negligible environmental impact compared to the fossil energy. Therefore also the waste management and the corresponding potential environmental hazard needs to be considered. The case of the photovoltaic panel is unique because the time lag from the manufacturing to the decommissioning as waste usually takes 25-30 years. Then the environmental hazard associated with end life of PV panels has been largely related to their metal contents. The principal concern regards the presence of heavy metals as Cd in thin film (TF) modules or Pb and Cr in crystalline silicon (c-Si) panels. At the end of life of PV panels, these dangerous substances could be released in the environment, if special requirements for their disposal are not adopted. Nevertheless, in literature, only a few experimental study about metal emissions from silicon crystalline/thin film panels and the corresponding environmental effect are present. As part of a study funded by the Italian national consortium for the waste collection and recycling (COBAT), the present work was aimed to analyze experimentally the potential release into the environment of hazardous elements, particularly metals, from PV waste. In this paper, for the first time, eighteen releasable metals a large number of photovoltaic panels, by c-Si and TF, manufactured in the last 30 years, together with the environmental effects by a battery of ecotoxicological tests, were investigated. Leaching tests are conducted on the crushed samples of PV module. The test is conducted according to Italian and European Standard procedure for hazard assessment of the granular waste and of the sludge. The sample material is shaken for 24 hours in HDPE bottles with an overhead mixer Rotax 6.8 VELP at indoor temperature and using pure water (18 MΩ resistivity) as leaching solution. The liquid-to-solid ratio was 10 (L/S=10, i.e. 10 liters of water per kg of solid). The ecotoxicological tests were performed in the subsequent 24 hours. A battery of toxicity test with bacteria (Vibrio fisheri), algae (Pseudochirneriella subcapitata) and crustacea (Daphnia magna) was carried out on PV panel leachates obtained as previously described and immediately stored in dark and at 4°C until testing (in the next 24 hours). For understand the actual pollution load, a comparison with the current European and Italian benchmark limits was performed. The trend of leachable metal amount from panels in relation to manufacturing years was then highlighted in order to assess the environmental sustainability of PV technology over time. The experimental results were very heterogeneous and show that the photovoltaic panels could represent an environmental hazard. The experimental results showed that the amounts of some hazardous metals (Pb, Cr, Cd, Ni), for c-Si and TF, exceed the law limits and they are a clear indication of the potential environmental risk of photovoltaic panels "as a waste" without a proper management.Keywords: photovoltaic panel, environment, ecotoxicity, metals emission
Procedia PDF Downloads 26016890 A Systemic Maturity Model
Authors: Emir H. Pernet, Jeimy J. Cano
Abstract:
Maturity models, used descriptively to explain changes in reality or normatively to guide managers to make interventions to make organizations more effective and efficient, are based on the principles of statistical quality control promulgated by Shewhart in the years 30, and on the principles of PDCA continuous improvement (Plan, Do, Check, Act) developed by Deming and Juran. Some frameworks developed over the concept of maturity models includes COBIT, CMM, and ITIL. This paper presents some limitations of traditional maturity models, most of them based on points of reflection and analysis done by some authors. Almost all limitations are related to the mechanistic and reductionist approach of the principles over those models are built. As Systems Theory helps the understanding of the dynamics of organizations and organizational change, the development of a systemic maturity model can help to overcome some of those limitations. This document proposes a systemic maturity model, based on a systemic conceptualization of organizations, focused on the study of the functioning of the parties, the relationships among them, and their behavior as a whole. The concept of maturity from the system theory perspective is conceptually defined as an emergent property of the organization, which arises from as a result of the degree of alignment and integration of their processes. This concept is operationalized through a systemic function that measures the maturity of an organization, and finally validated by the measuring of maturity in organizations. For its operationalization and validation, the model was applied to measure the maturity of organizational Governance, Risk and Compliance (GRC) processes.Keywords: GRC, maturity model, systems theory, viable system model
Procedia PDF Downloads 31216889 Effect of Gamma Radiation on Bromophenol Blue Dyed Films as Dosimeter
Authors: Priyanka R. Oberoi, Chandra B. Maurya, Prakash A. Mahanwar
Abstract:
Ionizing radiation can cause a drastic change in the physical and chemical properties of the material exposed. Numerous medical devices are sterilized by ionizing radiation. In the current research paper, an attempt was made to develop precise and inexpensive polymeric film dosimeter which can be used for controlling radiation dosage. Polymeric film containing (pH sensitive dye) indicator dye Bromophenol blue (BPB) was casted to check the effect of Gamma radiation on its optical and physical properties. The film was exposed to gamma radiation at 4 kGy/hr in the range of 0 to 300 kGy at an interval of 50 kGy. Release of vinyl acetate from an emulsion on high radiation reacts with the BPB fading the color of the film from blue to light blue and then finally colorless, indicating a change in pH from basic to acidic form. The change was characterized by using CIE l*a*b*, ultra-violet spectroscopy and FT-IR respectively.Keywords: bromophenol blue, dosimeter, gamma radiation, polymer
Procedia PDF Downloads 29016888 Mathematical Modeling of Skin Condensers for Domestic Refrigerator
Authors: Nitin Ghule, S. G. Taji
Abstract:
A mathematical model of hot-wall condensers used in refrigerators is presented. The model predicts the heat transfer characteristics of condenser and the effects of various design and operating parameters on condenser tube length and capacity. A finite element approach was used to model the condenser. The condenser tube is divided into elemental units, with each element consisting of adhesive tape, refrigerant tube and outer metal sheet. The heat transfer characteristics of each section are then analyzed by considering the heat transfer through the tube wall, tape and the outer sheet. Variations in inner heat transfer coefficient and pressure drop are considered depending on temperature, fluid phase, type of flow and orientation of tube. Variation in outer heat transfer coefficient is also taken into account. Various materials were analysed for the tube, tape and outer sheet.Keywords: condenser, domestic refrigerator, heat transfer, mathematical model
Procedia PDF Downloads 45216887 Evaluation of Weather Risk Insurance for Agricultural Products Using a 3-Factor Pricing Model
Authors: O. Benabdeljelil, A. Karioun, S. Amami, R. Rouger, M. Hamidine
Abstract:
A model for preventing the risks related to climate conditions in the agricultural sector is presented. It will determine the yearly optimum premium to be paid by a producer in order to reach his required turnover. The model is based on both climatic stability and 'soft' responses of usually grown species to average climate variations at the same place and inside a safety ball which can be determined from past meteorological data. This allows the use of linear regression expression for dependence of production result in terms of driving meteorological parameters, the main ones of which are daily average sunlight, rainfall and temperature. By simple best parameter fit from the expert table drawn with professionals, optimal representation of yearly production is determined from records of previous years, and yearly payback is evaluated from minimum yearly produced turnover. The model also requires accurate pricing of commodity at N+1. Therefore, a pricing model is developed using 3 state variables, namely the spot price, the difference between the mean-term and the long-term forward price, and the long-term structure of the model. The use of historical data enables to calibrate the parameters of state variables, and allows the pricing of commodity. Application to beet sugar underlines pricer precision. Indeed, the percentage of accuracy between computed result and real world is 99,5%. Optimal premium is then deduced and gives the producer a useful bound for negotiating an offer by insurance companies to effectively protect its harvest. The application to beet production in French Oise department illustrates the reliability of present model with as low as 6% difference between predicted and real data. The model can be adapted to almost any agricultural field by changing state parameters and calibrating their associated coefficients.Keywords: agriculture, production model, optimal price, meteorological factors, 3-factor model, parameter calibration, forward price
Procedia PDF Downloads 376