Search results for: Vector Error Correction Model (VECM)
16071 Invasive Ranges of Gorse (Ulex europaeus) in South Australia and Sri Lanka Using Species Distribution Modelling
Authors: Champika S. Kariyawasam
Abstract:
The distribution of gorse (Ulex europaeus) plants in South Australia has been modelled using 126 presence-only location data as a function of seven climate parameters. The predicted range of U. europaeus is mainly along the Mount Lofty Ranges in the Adelaide Hills and on Kangaroo Island. Annual precipitation and yearly average aridity index appeared to be the highest contributing variables to the final model formulation. The Jackknife procedure was employed to identify the contribution of different variables to gorse model outputs and response curves were used to predict changes with changing environmental variables. Based on this analysis, it was revealed that the combined effect of one or more variables could make a completely different impact to the original variables on their own to the model prediction. This work also demonstrates the need for a careful approach when selecting environmental variables for projecting correlative models to climatically distinct area. Maxent acts as a robust model when projecting the fitted species distribution model to another area with changing climatic conditions, whereas the generalized linear model, bioclim, and domain models to be less robust in this regard. These findings are important not only for predicting and managing invasive alien gorse in South Australia and Sri Lanka but also in other countries of the invasive range.Keywords: invasive species, Maxent, species distribution modelling, Ulex europaeus
Procedia PDF Downloads 13416070 Elastoplastic and Ductile Damage Model Calibration of Steels for Bolt-Sphere Joints Used in China’s Space Structure Construction
Authors: Huijuan Liu, Fukun Li, Hao Yuan
Abstract:
The bolted spherical node is a common type of joint in space steel structures. The bolt-sphere joint portion almost always controls the bearing capacity of the bolted spherical node. The investigation of the bearing performance and progressive failure in service often requires high-fidelity numerical models. This paper focuses on the constitutive models of bolt steel and sphere steel used in China’s space structure construction. The elastoplastic model is determined by a standard tensile test and calibrated Voce saturated hardening rule. The ductile damage is found dominant based on the fractography analysis. Then Rice-Tracey ductile fracture rule is selected and the model parameters are calibrated based on tensile tests of notched specimens. These calibrated material models can benefit research or engineering work in similar fields.Keywords: bolt-sphere joint, steel, constitutive model, ductile damage, model calibration
Procedia PDF Downloads 13916069 Magnetic Properties of Nickel Oxide Nanoparticles in Superparamagnetic State
Authors: Navneet Kaur, S. D. Tiwari
Abstract:
Superparamagnetism is an interesting phenomenon and observed in small particles of magnetic materials. It arises due to a reduction in particle size. In the superparamagnetic state, as the thermal energy overcomes magnetic anisotropy energy, the magnetic moment vector of particles flip their magnetization direction between states of minimum energy. Superparamagnetic nanoparticles have been attracting the researchers due to many applications such as information storage, magnetic resonance imaging, biomedical applications, and sensors. For information storage, thermal fluctuations lead to loss of data. So that nanoparticles should have high blocking temperature. And to achieve this, nanoparticles should have a higher magnetic moment and magnetic anisotropy constant. In this work, the magnetic anisotropy constant of the antiferromagnetic nanoparticles system is determined. Magnetic studies on nanoparticles of NiO (nickel oxide) are reported well. This antiferromagnetic nanoparticle system has high blocking temperature and magnetic anisotropy constant of order 105 J/m3. The magnetic study of NiO nanoparticles in the superparamagnetic region is presented. NiO particles of two different sizes, i.e., 6 and 8 nm, are synthesized using the chemical route. These particles are characterized by an x-ray diffractometer, transmission electron microscope, and superconducting quantum interference device magnetometry. The magnetization vs. applied magnetic field and temperature data for both samples confirm their superparamagnetic nature. The blocking temperature for 6 and 8 nm particles is found to be 200 and 172 K, respectively. Magnetization vs. applied magnetic field data of NiO is fitted to an appropriate magnetic expression using a non-linear least square fit method. The role of particle size distribution and magnetic anisotropy is taken in to account in magnetization expression. The source code is written in Python programming language. This fitting provides us the magnetic anisotropy constant for NiO and other magnetic fit parameters. The particle size distribution estimated matches well with the transmission electron micrograph. The value of magnetic anisotropy constants for 6 and 8 nm particles is found to be 1.42 X 105 and 1.20 X 105 J/m3, respectively. The obtained magnetic fit parameters are verified using the Neel model. It is concluded that the effect of magnetic anisotropy should not be ignored while studying the magnetization process of nanoparticles.Keywords: anisotropy, superparamagnetic, nanoparticle, magnetization
Procedia PDF Downloads 13516068 Advancements in Predicting Diabetes Biomarkers: A Machine Learning Epigenetic Approach
Authors: James Ladzekpo
Abstract:
Background: The urgent need to identify new pharmacological targets for diabetes treatment and prevention has been amplified by the disease's extensive impact on individuals and healthcare systems. A deeper insight into the biological underpinnings of diabetes is crucial for the creation of therapeutic strategies aimed at these biological processes. Current predictive models based on genetic variations fall short of accurately forecasting diabetes. Objectives: Our study aims to pinpoint key epigenetic factors that predispose individuals to diabetes. These factors will inform the development of an advanced predictive model that estimates diabetes risk from genetic profiles, utilizing state-of-the-art statistical and data mining methods. Methodology: We have implemented a recursive feature elimination with cross-validation using the support vector machine (SVM) approach for refined feature selection. Building on this, we developed six machine learning models, including logistic regression, k-Nearest Neighbors (k-NN), Naive Bayes, Random Forest, Gradient Boosting, and Multilayer Perceptron Neural Network, to evaluate their performance. Findings: The Gradient Boosting Classifier excelled, achieving a median recall of 92.17% and outstanding metrics such as area under the receiver operating characteristics curve (AUC) with a median of 68%, alongside median accuracy and precision scores of 76%. Through our machine learning analysis, we identified 31 genes significantly associated with diabetes traits, highlighting their potential as biomarkers and targets for diabetes management strategies. Conclusion: Particularly noteworthy were the Gradient Boosting Classifier and Multilayer Perceptron Neural Network, which demonstrated potential in diabetes outcome prediction. We recommend future investigations to incorporate larger cohorts and a wider array of predictive variables to enhance the models' predictive capabilities.Keywords: diabetes, machine learning, prediction, biomarkers
Procedia PDF Downloads 5816067 Normalized Compression Distance Based Scene Alteration Analysis of a Video
Authors: Lakshay Kharbanda, Aabhas Chauhan
Abstract:
In this paper, an application of Normalized Compression Distance (NCD) to detect notable scene alterations occurring in videos is presented. Several research groups have been developing methods to perform image classification using NCD, a computable approximation to Normalized Information Distance (NID) by studying the degree of similarity in images. The timeframes where significant aberrations between the frames of a video have occurred have been identified by obtaining a threshold NCD value, using two compressors: LZMA and BZIP2 and defining scene alterations using Pixel Difference Percentage metrics.Keywords: image compression, Kolmogorov complexity, normalized compression distance, root mean square error
Procedia PDF Downloads 34116066 Predicting Groundwater Areas Using Data Mining Techniques: Groundwater in Jordan as Case Study
Authors: Faisal Aburub, Wael Hadi
Abstract:
Data mining is the process of extracting useful or hidden information from a large database. Extracted information can be used to discover relationships among features, where data objects are grouped according to logical relationships; or to predict unseen objects to one of the predefined groups. In this paper, we aim to investigate four well-known data mining algorithms in order to predict groundwater areas in Jordan. These algorithms are Support Vector Machines (SVMs), Naïve Bayes (NB), K-Nearest Neighbor (kNN) and Classification Based on Association Rule (CBA). The experimental results indicate that the SVMs algorithm outperformed other algorithms in terms of classification accuracy, precision and F1 evaluation measures using the datasets of groundwater areas that were collected from Jordanian Ministry of Water and Irrigation.Keywords: classification, data mining, evaluation measures, groundwater
Procedia PDF Downloads 28116065 Machine Learning Data Architecture
Authors: Neerav Kumar, Naumaan Nayyar, Sharath Kashyap
Abstract:
Most companies see an increase in the adoption of machine learning (ML) applications across internal and external-facing use cases. ML applications vend output either in batch or real-time patterns. A complete batch ML pipeline architecture comprises data sourcing, feature engineering, model training, model deployment, model output vending into a data store for downstream application. Due to unclear role expectations, we have observed that scientists specializing in building and optimizing models are investing significant efforts into building the other components of the architecture, which we do not believe is the best use of scientists’ bandwidth. We propose a system architecture created using AWS services that bring industry best practices to managing the workflow and simplifies the process of model deployment and end-to-end data integration for an ML application. This narrows down the scope of scientists’ work to model building and refinement while specialized data engineers take over the deployment, pipeline orchestration, data quality, data permission system, etc. The pipeline infrastructure is built and deployed as code (using terraform, cdk, cloudformation, etc.) which makes it easy to replicate and/or extend the architecture to other models that are used in an organization.Keywords: data pipeline, machine learning, AWS, architecture, batch machine learning
Procedia PDF Downloads 6516064 Data-Driven Dynamic Overbooking Model for Tour Operators
Authors: Kannapha Amaruchkul
Abstract:
We formulate a dynamic overbooking model for a tour operator, in which most reservations contain at least two people. The cancellation rate and the timing of the cancellation may depend on the group size. We propose two overbooking policies, namely economic- and service-based. In an economic-based policy, we want to minimize the expected oversold and underused cost, whereas, in a service-based policy, we ensure that the probability of an oversold situation does not exceed the pre-specified threshold. To illustrate the applicability of our approach, we use tour package data in 2016-2018 from a tour operator in Thailand to build a data-driven robust optimization model, and we tested the proposed overbooking policy in 2019. We also compare the data-driven approach to the conventional approach of fitting data into a probability distribution.Keywords: applied stochastic model, data-driven robust optimization, overbooking, revenue management, tour operator
Procedia PDF Downloads 13516063 PM Air Quality of Windsor Regional Scale Transport’s Impact and Climate Change
Authors: Moustafa Osman Mohammed
Abstract:
This paper is mapping air quality model to engineering the industrial system that ultimately utilized in extensive range of energy systems, distribution resources, and end-user technologies. The model is determining long-range transport patterns contribution as area source can either traced from 48 hrs backward trajectory model or remotely described from background measurements data in those days. The trajectory model will be run within stable conditions and quite constant parameters of the atmospheric pressure at the most time of the year. Air parcel trajectory is necessary for estimating the long-range transport of pollutants and other chemical species. It provides a better understanding of airflow patterns. Since a large amount of meteorological data and a great number of calculations are required to drive trajectory, it will be very useful to apply HYPSLIT model to locate areas and boundaries influence air quality at regional location of Windsor. 2–days backward trajectories model at high and low concentration measurements below and upward the benchmark which was areas influence air quality measurement levels. The benchmark level will be considered as 30 (μg/m3) as the moderate level for Ontario region. Thereby, air quality model is incorporating a midpoint concept between biotic and abiotic components to broaden the scope of quantification impact. The later outcomes’ theories of environmental obligation suggest either a recommendation or a decision of what is a legislative should be achieved in mitigation measures of air emission impact ultimately.Keywords: air quality, management systems, environmental impact assessment, industrial ecology, climate change
Procedia PDF Downloads 24916062 Modeling and Statistical Analysis of a Soap Production Mix in Bejoy Manufacturing Industry, Anambra State, Nigeria
Authors: Okolie Chukwulozie Paul, Iwenofu Chinwe Onyedika, Sinebe Jude Ebieladoh, M. C. Nwosu
Abstract:
The research work is based on the statistical analysis of the processing data. The essence is to analyze the data statistically and to generate a design model for the production mix of soap manufacturing products in Bejoy manufacturing company Nkpologwu, Aguata Local Government Area, Anambra state, Nigeria. The statistical analysis shows the statistical analysis and the correlation of the data. T test, Partial correlation and bi-variate correlation were used to understand what the data portrays. The design model developed was used to model the data production yield and the correlation of the variables show that the R2 is 98.7%. However, the results confirm that the data is fit for further analysis and modeling. This was proved by the correlation and the R-squared.Keywords: General Linear Model, correlation, variables, pearson, significance, T-test, soap, production mix and statistic
Procedia PDF Downloads 44616061 Fundamentals of Islamic Resistive Economy and Practical Solutions: A Study from Perspective of Infallible Imams
Authors: Abolfazl Alishahi Ghalehjoughi
Abstract:
Economic independence and security of Islamic world is the top priority. Economic dependence of Muslim countries on economies of non-Muslim imperialist countries results in political and cultural dependencies, and such dependencies will jeopardize the noble Islamic culture; because the will of a dependent country to implements the noble teachings of Islam would be faced with challenges. Solidarity of Muslim countries to achieve a uniformed and resistive economy-based Islamic economic system can improve ability of Islamic world to resist and counteract economic shocks produced by imperialists. Islam is the most complete religion in every aspect, from ideological and epistemological, to legislative and ethical, and economic aspect is no exception. Islam provides solutions to develop a flourishing economy for the whole Islamic nation. Knowledge of such solutions and identification of mechanisms to operationalise them in Islamic communities can highly contributed to establishment of the superior Islamic economy. Encourage of hard working, achievement and knowledge production, correction of consumption patterns, optimized management of import and export, avoiding Islamically prohibited income, economic discipline and equity, and promotion of interest free loan and the like are among the most important solutions to realize such resistive economy.Keywords: resistive economy, cultural independence, Islam, solidarity
Procedia PDF Downloads 39516060 Machine Learning Analysis of Student Success in Introductory Calculus Based Physics I Course
Authors: Chandra Prayaga, Aaron Wade, Lakshmi Prayaga, Gopi Shankar Mallu
Abstract:
This paper presents the use of machine learning algorithms to predict the success of students in an introductory physics course. Data having 140 rows pertaining to the performance of two batches of students was used. The lack of sufficient data to train robust machine learning models was compensated for by generating synthetic data similar to the real data. CTGAN and CTGAN with Gaussian Copula (Gaussian) were used to generate synthetic data, with the real data as input. To check the similarity between the real data and each synthetic dataset, pair plots were made. The synthetic data was used to train machine learning models using the PyCaret package. For the CTGAN data, the Ada Boost Classifier (ADA) was found to be the ML model with the best fit, whereas the CTGAN with Gaussian Copula yielded Logistic Regression (LR) as the best model. Both models were then tested for accuracy with the real data. ROC-AUC analysis was performed for all the ten classes of the target variable (Grades A, A-, B+, B, B-, C+, C, C-, D, F). The ADA model with CTGAN data showed a mean AUC score of 0.4377, but the LR model with the Gaussian data showed a mean AUC score of 0.6149. ROC-AUC plots were obtained for each Grade value separately. The LR model with Gaussian data showed consistently better AUC scores compared to the ADA model with CTGAN data, except in two cases of the Grade value, C- and A-.Keywords: machine learning, student success, physics course, grades, synthetic data, CTGAN, gaussian copula CTGAN
Procedia PDF Downloads 4416059 Model of Monitoring and Evaluation of Student’s Learning Achievement: Application of Value-Added Assessment
Authors: Jatuphum Ketchatturat
Abstract:
Value-added assessment has been used for developing the model of monitoring and evaluation of student's learning achievement. The steps of model development consist of 1) study and analyisis of the school and the district report system of student achievement and progress, 2) collecting the data of student achievement to develop the value added indicator, 3) developing the system of value-added assessment by participatory action research approach, 4) putting the system of value-added assessment into the educational district of secondary school, 5) determining the quality of the developed system of value-added assessment. The components of the developed model consist of 1) the database of value-added assessment of student's learning achievement, 2) the process of monitoring and evaluation the student's learning achievement, and 3) the reporting system of value-added assessment of student's learning achievement.Keywords: learning achievement, monitoring and evaluation, value-added assessment
Procedia PDF Downloads 42616058 Application of Random Forest Model in The Prediction of River Water Quality
Authors: Turuganti Venkateswarlu, Jagadeesh Anmala
Abstract:
Excessive runoffs from various non-point source land uses, and other point sources are rapidly contaminating the water quality of streams in the Upper Green River watershed, Kentucky, USA. It is essential to maintain the stream water quality as the river basin is one of the major freshwater sources in this province. It is also important to understand the water quality parameters (WQPs) quantitatively and qualitatively along with their important features as stream water is sensitive to climatic events and land-use practices. In this paper, a model was developed for predicting one of the significant WQPs, Fecal Coliform (FC) from precipitation, temperature, urban land use factor (ULUF), agricultural land use factor (ALUF), and forest land-use factor (FLUF) using Random Forest (RF) algorithm. The RF model, a novel ensemble learning algorithm, can even find out advanced feature importance characteristics from the given model inputs for different combinations. This model’s outcomes showed a good correlation between FC and climate events and land use factors (R2 = 0.94) and precipitation and temperature are the primary influencing factors for FC.Keywords: water quality, land use factors, random forest, fecal coliform
Procedia PDF Downloads 19916057 An Efficient Machine Learning Model to Detect Metastatic Cancer in Pathology Scans Using Principal Component Analysis Algorithm, Genetic Algorithm, and Classification Algorithms
Authors: Bliss Singhal
Abstract:
Machine learning (ML) is a branch of Artificial Intelligence (AI) where computers analyze data and find patterns in the data. The study focuses on the detection of metastatic cancer using ML. Metastatic cancer is the stage where cancer has spread to other parts of the body and is the cause of approximately 90% of cancer-related deaths. Normally, pathologists spend hours each day to manually classifying whether tumors are benign or malignant. This tedious task contributes to mislabeling metastasis being over 60% of the time and emphasizes the importance of being aware of human error and other inefficiencies. ML is a good candidate to improve the correct identification of metastatic cancer, saving thousands of lives and can also improve the speed and efficiency of the process, thereby taking fewer resources and time. So far, the deep learning methodology of AI has been used in research to detect cancer. This study is a novel approach to determining the potential of using preprocessing algorithms combined with classification algorithms in detecting metastatic cancer. The study used two preprocessing algorithms: principal component analysis (PCA) and the genetic algorithm, to reduce the dimensionality of the dataset and then used three classification algorithms: logistic regression, decision tree classifier, and k-nearest neighbors to detect metastatic cancer in the pathology scans. The highest accuracy of 71.14% was produced by the ML pipeline comprising of PCA, the genetic algorithm, and the k-nearest neighbor algorithm, suggesting that preprocessing and classification algorithms have great potential for detecting metastatic cancer.Keywords: breast cancer, principal component analysis, genetic algorithm, k-nearest neighbors, decision tree classifier, logistic regression
Procedia PDF Downloads 8416056 Hybrid Inventory Model Optimization under Uncertainties: A Case Study in a Manufacturing Plant
Authors: E. Benga, T. Tengen, A. Alugongo
Abstract:
Periodic and continuous inventory models are the two classical management tools used to handle inventories. These models have advantages and disadvantages. The implementation of both continuous (r,Q) inventory and periodic (R, S) inventory models in most manufacturing plants comes with higher cost. Such high inventory costs are due to the fact that most manufacturing plants are not flexible enough. Since demand and lead-time are two important variables of every inventory models, their effect on the flexibility of the manufacturing plant matter most. Unfortunately, these effects are not clearly understood by managers. The reason is that the decision parameters of the continuous (r, Q) inventory and periodic (R, S) inventory models are not designed to effectively deal with the issues of uncertainties such as poor manufacturing performances, delivery performance supplies performances. There is, therefore, a need to come up with a predictive and hybrid inventory model that can combine in some sense the feature of the aforementioned inventory models. A linear combination technique is used to hybridize both continuous (r, Q) inventory and periodic (R, S) inventory models. The behavior of such hybrid inventory model is described by a differential equation and then optimized. From the results obtained after simulation, the continuous (r, Q) inventory model is more effective than the periodic (R, S) inventory models in the short run, but this difference changes as time goes by. Because the hybrid inventory model is more cost effective than the continuous (r,Q) inventory and periodic (R, S) inventory models in long run, it should be implemented for strategic decisions.Keywords: periodic inventory, continuous inventory, hybrid inventory, optimization, manufacturing plant
Procedia PDF Downloads 38316055 A Fishery Regulation Model: Bargaining over Fishing Pressure
Authors: Duplan Yves Jamont Junior
Abstract:
The Diamond-Mortensen-Pissarides model widely used in labor economics is tailored to fishery. By this way, a fishing function is defined to depict the fishing technology, and Bellman equations are established to describe the behaviors of fishermen and conservationists. On this basis, a negotiation takes place as a Nash-bargaining over the upper limit of the fishing pressure between both political representative groups of fishermen and conservationists. The existence and uniqueness conditions of the Nash-bargained fishing pressure are established. Given the biomass evolution equation, the dynamics of the model variables (fishing pressure, biomass, fish need) is studied.Keywords: conservation, fishery, fishing, Nash bargaining
Procedia PDF Downloads 26116054 Model for Remanufacture of Medical Equipment in Cross Border Collaboration
Authors: Kingsley Oturu, Winifred Ijomah, Wale Coker, Chibueze Achi
Abstract:
With the impact of BREXIT and the need for cross-border collaboration, this international research investigated the use of a conceptual model for remanufacturing medical equipment (with a focus on anesthetic machines and baby incubators). Early findings of the research suggest that contextual factors need to be taken into consideration, as well as an emphasis on cleaning (e.g., sterilization) during the process of remanufacturing medical equipment. For example, copper tubings may be more important in the remanufacturing of anesthetic equipment in tropical climates than in cold climates.Keywords: medical equipment remanufacture, sustainability, circular business models, remanufacture process model
Procedia PDF Downloads 17316053 An Investigation about Rate Of Evaporation from the Water Surface and LNG Pool
Authors: Farokh Alipour, Ali Falavand, Neda Beit Saeid
Abstract:
The calculation of the effect of accidental releases of flammable materials such as LNG requires the use of a suitable consequence model. This study is due to providing a planning advice for developments in the vicinity of LNG sites and other sites handling flammable materials. In this paper, an applicable algorithm that is able to model pool fires on water is presented and applied to estimate pool fire damage zone. This procedure can be used to model pool fires on land and could be helpful in consequence modeling and domino effect zone measurements of flammable materials which is needed in site selection and plant layout.Keywords: LNG, pool fire, spill, radiation
Procedia PDF Downloads 40416052 Texture-Based Image Forensics from Video Frame
Authors: Li Zhou, Yanmei Fang
Abstract:
With current technology, images and videos can be obtained more easily than ever. It is so easy to manipulate these digital multimedia information when obtained, and that the content or source of the image and video could be easily tampered. In this paper, we propose to identify the image and video frame by the texture-based approach, e.g. Markov Transition Probability (MTP), which is in space domain, DCT domain and DWT domain, respectively. In the experiment, image and video frame database is constructed, and is used to train and test the classifier Support Vector Machine (SVM). Experiment results show that the texture-based approach has good performance. In order to verify the experiment result, and testify the universality and robustness of algorithm, we build a random testing dataset, the random testing result is in keeping with above experiment.Keywords: multimedia forensics, video frame, LBP, MTP, SVM
Procedia PDF Downloads 42916051 Vitamin D Status in Relation to Body Mass Index: Population of Carpathian Region
Authors: Vladyslav Povoroznyuk, Ivan Pankiv
Abstract:
The present research has attempted to link a higher body weight with a lower vitamin D status. Objective: Vitamin D status of Carpathian region population in Ukraine was studied to examine whether serum levels of 25-hydroxyvitamin D [25(OH)D] are associated with body mass index (BMI). Methods: Data collected from 302 adults (18–84 years) were analyzed. Variables measured included serum 25(OH)D, weight and height used to determine BMI status. Results: Mean 25(OH)D level was 23.2 ± 8.1 ng/mL for the group; 26.3 ± 8.4 ng/mL and 22.8 ± 9.1 ng/mL for males and females, respectively. Based on BMI, 3.6% were underweight, 21.2% had a normal weight, 46.4% were overweight and 28.8% obese. Only in 28 cases (9.3%), content of 25(ОН)D in the serum of blood was within the normal limits, and there were vitamin D deficiency and insufficiency observed in other cases (90.7%). Thus, severe vitamin D deficiency was revealed in 1.7% of the inspected. A significant interrelation between levels of 25(OH)D in blood and BMI was found among persons with BMI 25-29.9 kg/m2. Mean value of 25(OH)D levels among persons with obesity did not differ to a significant extent from indexes in persons with normal body weight. Conclusion: Status of vitamin D among the population of Carpathian region remains far from optimal and requires urgent measures in correction and prevention. Results confirmed a poor inverse relationship between vitamin D status and BMI. Intercommunication between maintenance of vitamin D and BMI requires further investigations.Keywords: body mass index, Carpathian region, obesity, vitamin D
Procedia PDF Downloads 39216050 The Rapid Industrialization Model
Authors: Fredrick Etyang
Abstract:
This paper presents a Rapid Industrialization Model (RIM) designed to support existing industrialization policies, strategies and industrial development plans at National, Regional and Constituent level in Africa. The model will reinforce efforts to attainment of inclusive and sustainable industrialization of Africa by state and non-state actors. The overall objective of this model is to serve as a framework for rapid industrialization in developing economies and the specific objectives range from supporting rapid industrialization development to promoting a structural change in the economy, a balanced regional industrial growth, achievement of local, regional and international competitiveness in areas of clear comparative advantage in industrial exports and ultimately, the RIM will serve as a step-by-step guideline for the industrialization of African Economies. This model is a product of a scientific research process underpinned by desk research through the review of African countries development plans, strategies, datasets, industrialization efforts and consultation with key informants. The rigorous research process unearthed multi-directional and renewed efforts towards industrialization of Africa premised on collective commitment of individual states, regional economic communities and the African union commission among other strategic stakeholders. It was further, established that the inputs into industrialization of Africa outshine the levels of industrial development on the continent. The RIM comes in handy to serve as step-by-step framework for African countries to follow in their industrial development efforts of transforming inputs into tangible outputs and outcomes in the short, intermediate and long-run. This model postulates three stages of industrialization and three phases toward rapid industrialization of African economies, the model is simple to understand, easily implementable and contextualizable with high return on investment for each unit invested into industrialization supported by the model. Therefore, effective implementation of the model will result into inclusive and sustainable rapid industrialization of Africa.Keywords: economic development, industrialization, economic efficiency, exports and imports
Procedia PDF Downloads 8716049 Setting Control Limits For Inaccurate Measurements
Authors: Ran Etgar
Abstract:
The process of rounding off measurements in continuous variables is commonly encountered. Although it usually has minor effects, sometimes it can lead to poor outcomes in statistical process control using X ̅-chart. The traditional control limits can cause incorrect conclusions if applied carelessly. This study looks into the limitations of classical control limits, particularly the impact of asymmetry. An approach to determining the distribution function of the measured parameter (Y ̅) is presented, resulting in a more precise method to establish the upper and lower control limits. The proposed method, while slightly more complex than Shewhart's original idea, is still user-friendly and accurate and only requires the use of two straightforward tables.Keywords: quality control, process control, round-off, measurement, rounding error
Procedia PDF Downloads 10116048 Self-Compacting White Concrete Mix Design Using the Particle Matrix Model
Authors: Samindi Samarakoon, Ørjan Sletbakk Vie, Remi Kleiven Fjelldal
Abstract:
White concrete facade elements are widely used in construction industry. It is challenging to achieve the desired workability in casting of white concrete elements. Particle Matrix model was used for proportioning the self-compacting white concrete (SCWC) to control segregation and bleeding and to improve workability. The paper presents how to reach the target slump flow while controlling bleeding and segregation in SCWC. The amount of aggregates, binders and mixing water, as well as type and dosage of superplasticizer (SP) to be used are the major factors influencing the properties of SCWC. Slump flow and compressive strength tests were carried out to examine the performance of SCWC, and the results indicate that the particle matrix model could produce successfully SCWC controlling segregation and bleeding.Keywords: white concrete, particle matrix model, mix design, construction industry
Procedia PDF Downloads 27116047 CFD Studies on Forced Convection Nanofluid Flow Inside a Circular Conduit
Authors: M. Khalid, W. Rashmi, L. L. Kwan
Abstract:
This work provides an overview on the experimental and numerical simulations of various nanofluids and their flow and heat transfer behavior. It was further extended to study the effect of nanoparticle concentration, fluid flow rates and thermo-physical properties on the heat transfer enhancement of Al2O3/water nanofluid in a turbulent flow circular conduit using ANSYS FLUENT™ 14.0. Single-phase approximation (homogeneous model) and two-phase (mixture and Eulerian) models were used to simulate the nanofluid flow behavior in the 3-D horizontal pipe. The numerical results were further validated with experimental correlations reported in the literature. It was found that heat transfer of nanofluids increases with increasing particle volume concentration and Reynolds number, respectively. Results showed good agreement (~9% deviation) with the experimental correlations, especially for a single-phase model with constant properties. Among two-phase models, mixture model (~14% deviation) showed better prediction compared to Eulerian-dispersed model (~18% deviation) when temperature independent properties were used. Non-drag forces were also employed in the Eulerian two-phase model. However, the two-phase mixture model with temperature dependent nanofluid properties gave slightly closer agreement (~12% deviation).Keywords: nanofluid, CFD, heat transfer, forced convection, circular conduit
Procedia PDF Downloads 52516046 Nowcasting Indonesian Economy
Authors: Ferry Kurniawan
Abstract:
In this paper, we nowcast quarterly output growth in Indonesia by exploiting higher frequency data (monthly indicators) using a mixed-frequency factor model and exploiting both quarterly and monthly data. Nowcasting quarterly GDP in Indonesia is particularly relevant for the central bank of Indonesia which set the policy rate in the monthly Board of Governors Meeting; whereby one of the important step is the assessment of the current state of the economy. Thus, having an accurate and up-to-date quarterly GDP nowcast every time new monthly information becomes available would clearly be of interest for central bank of Indonesia, for example, as the initial assessment of the current state of the economy -including nowcast- will be used as input for longer term forecast. We consider a small scale mixed-frequency factor model to produce nowcasts. In particular, we specify variables as year-on-year growth rates thus the relation between quarterly and monthly data is expressed in year-on-year growth rates. To assess the performance of the model, we compare the nowcasts with two other approaches: autoregressive model –which is often difficult when forecasting output growth- and Mixed Data Sampling (MIDAS) regression. In particular, both mixed frequency factor model and MIDAS nowcasts are produced by exploiting the same set of monthly indicators. Hence, we compare the nowcasts performance of the two approaches directly. To preview the results, we find that by exploiting monthly indicators using mixed-frequency factor model and MIDAS regression we improve the nowcast accuracy over a benchmark simple autoregressive model that uses only quarterly frequency data. However, it is not clear whether the MIDAS or mixed-frequency factor model is better. Neither set of nowcasts encompasses the other; suggesting that both nowcasts are valuable in nowcasting GDP but neither is sufficient. By combining the two individual nowcasts, we find that the nowcast combination not only increases the accuracy - relative to individual nowcasts- but also lowers the risk of the worst performance of the individual nowcasts.Keywords: nowcasting, mixed-frequency data, factor model, nowcasts combination
Procedia PDF Downloads 33216045 Software Reliability Prediction Model Analysis
Authors: Lela Mirtskhulava, Mariam Khunjgurua, Nino Lomineishvili, Koba Bakuria
Abstract:
Software reliability prediction gives a great opportunity to measure the software failure rate at any point throughout system test. A software reliability prediction model provides with the technique for improving reliability. Software reliability is very important factor for estimating overall system reliability, which depends on the individual component reliabilities. It differs from hardware reliability in that it reflects the design perfection. Main reason of software reliability problems is high complexity of software. Various approaches can be used to improve the reliability of software. We focus on software reliability model in this article, assuming that there is a time redundancy, the value of which (the number of repeated transmission of basic blocks) can be an optimization parameter. We consider given mathematical model in the assumption that in the system may occur not only irreversible failures, but also a failure that can be taken as self-repairing failures that significantly affect the reliability and accuracy of information transfer. Main task of the given paper is to find a time distribution function (DF) of instructions sequence transmission, which consists of random number of basic blocks. We consider the system software unreliable; the time between adjacent failures has exponential distribution.Keywords: exponential distribution, conditional mean time to failure, distribution function, mathematical model, software reliability
Procedia PDF Downloads 46616044 Parametric Modeling for Survival Data with Competing Risks Using the Generalized Gompertz Distribution
Authors: Noora Al-Shanfari, M. Mazharul Islam
Abstract:
The cumulative incidence function (CIF) is a fundamental approach for analyzing survival data in the presence of competing risks, which estimates the marginal probability for each competing event. Parametric modeling of CIF has the advantage of fitting various shapes of CIF and estimates the impact of covariates with maximum efficiency. To calculate the total CIF's covariate influence using a parametric model., it is essential to parametrize the baseline of the CIF. As the CIF is an improper function by nature, it is necessary to utilize an improper distribution when applying parametric models. The Gompertz distribution, which is an improper distribution, is limited in its applicability as it only accounts for monotone hazard shapes. The generalized Gompertz distribution, however, can adapt to a wider range of hazard shapes, including unimodal, bathtub, and monotonic increasing or decreasing hazard shapes. In this paper, the generalized Gompertz distribution is used to parametrize the baseline of the CIF, and the parameters of the proposed model are estimated using the maximum likelihood approach. The proposed model is compared with the existing Gompertz model using the Akaike information criterion. Appropriate statistical test procedures and model-fitting criteria will be used to test the adequacy of the model. Both models are applied to the ‘colon’ dataset, which is available in the “biostat3” package in R.Keywords: competing risks, cumulative incidence function, improper distribution, parametric modeling, survival analysis
Procedia PDF Downloads 10916043 A Systemic Maturity Model
Authors: Emir H. Pernet, Jeimy J. Cano
Abstract:
Maturity models, used descriptively to explain changes in reality or normatively to guide managers to make interventions to make organizations more effective and efficient, are based on the principles of statistical quality control promulgated by Shewhart in the years 30, and on the principles of PDCA continuous improvement (Plan, Do, Check, Act) developed by Deming and Juran. Some frameworks developed over the concept of maturity models includes COBIT, CMM, and ITIL. This paper presents some limitations of traditional maturity models, most of them based on points of reflection and analysis done by some authors. Almost all limitations are related to the mechanistic and reductionist approach of the principles over those models are built. As Systems Theory helps the understanding of the dynamics of organizations and organizational change, the development of a systemic maturity model can help to overcome some of those limitations. This document proposes a systemic maturity model, based on a systemic conceptualization of organizations, focused on the study of the functioning of the parties, the relationships among them, and their behavior as a whole. The concept of maturity from the system theory perspective is conceptually defined as an emergent property of the organization, which arises from as a result of the degree of alignment and integration of their processes. This concept is operationalized through a systemic function that measures the maturity of an organization, and finally validated by the measuring of maturity in organizations. For its operationalization and validation, the model was applied to measure the maturity of organizational Governance, Risk and Compliance (GRC) processes.Keywords: GRC, maturity model, systems theory, viable system model
Procedia PDF Downloads 31316042 Mathematical Modeling of Skin Condensers for Domestic Refrigerator
Authors: Nitin Ghule, S. G. Taji
Abstract:
A mathematical model of hot-wall condensers used in refrigerators is presented. The model predicts the heat transfer characteristics of condenser and the effects of various design and operating parameters on condenser tube length and capacity. A finite element approach was used to model the condenser. The condenser tube is divided into elemental units, with each element consisting of adhesive tape, refrigerant tube and outer metal sheet. The heat transfer characteristics of each section are then analyzed by considering the heat transfer through the tube wall, tape and the outer sheet. Variations in inner heat transfer coefficient and pressure drop are considered depending on temperature, fluid phase, type of flow and orientation of tube. Variation in outer heat transfer coefficient is also taken into account. Various materials were analysed for the tube, tape and outer sheet.Keywords: condenser, domestic refrigerator, heat transfer, mathematical model
Procedia PDF Downloads 454