Search results for: generalized linear model
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 18913

Search results for: generalized linear model

15673 Pushover Analysis of Reinforced Concrete Buildings Using Full Jacket Technics: A Case Study on an Existing Old Building in Madinah

Authors: Tarek M. Alguhane, Ayman H. Khalil, M. N. Fayed, Ayman M. Ismail

Abstract:

The retrofitting of existing buildings to resist the seismic loads is very important to avoid losing lives or financial disasters. The aim at retrofitting processes is increasing total structure strength by increasing stiffness or ductility ratio. In addition, the response modification factors (R) have to satisfy the code requirements for suggested retrofitting types. In this study, two types of jackets are used, i.e. full reinforced concrete jackets and surrounding steel plate jackets. The study is carried out on an existing building in Madinah by performing static pushover analysis before and after retrofitting the columns. The selected model building represents nearly all-typical structure lacks structure built before 30 years ago in Madina City, KSA. The comparison of the results indicates a good enhancement of the structure respect to the applied seismic forces. Also, the response modification factor of the RC building is evaluated for the studied cases before and after retrofitting. The design of all vertical elements (columns) is given. The results show that the design of retrofitted columns satisfied the code's design stress requirements. However, for some retrofitting types, the ductility requirements represented by response modification factor do not satisfy KSA design code (SBC- 301).

Keywords: concrete jackets, steel jackets, RC buildings, pushover analysis, non-Linear analysis

Procedia PDF Downloads 359
15672 An Experimental Study on Some Conventional and Hybrid Models of Fuzzy Clustering

Authors: Jeugert Kujtila, Kristi Hoxhalli, Ramazan Dalipi, Erjon Cota, Ardit Murati, Erind Bedalli

Abstract:

Clustering is a versatile instrument in the analysis of collections of data providing insights of the underlying structures of the dataset and enhancing the modeling capabilities. The fuzzy approach to the clustering problem increases the flexibility involving the concept of partial memberships (some value in the continuous interval [0, 1]) of the instances in the clusters. Several fuzzy clustering algorithms have been devised like FCM, Gustafson-Kessel, Gath-Geva, kernel-based FCM, PCM etc. Each of these algorithms has its own advantages and drawbacks, so none of these algorithms would be able to perform superiorly in all datasets. In this paper we will experimentally compare FCM, GK, GG algorithm and a hybrid two-stage fuzzy clustering model combining the FCM and Gath-Geva algorithms. Firstly we will theoretically dis-cuss the advantages and drawbacks for each of these algorithms and we will describe the hybrid clustering model exploiting the advantages and diminishing the drawbacks of each algorithm. Secondly we will experimentally compare the accuracy of the hybrid model by applying it on several benchmark and synthetic datasets.

Keywords: fuzzy clustering, fuzzy c-means algorithm (FCM), Gustafson-Kessel algorithm, hybrid clustering model

Procedia PDF Downloads 506
15671 The Development of Directed-Project Based Learning as Language Learning Model to Improve Students' English Achievement

Authors: Tri Pratiwi, Sufyarma Marsidin, Hermawati Syarif, Yahya

Abstract:

The 21st-century skills being highly promoted today are Creativity and Innovation, Critical Thinking and Problem Solving, Communication and Collaboration. Communication Skill is one of the essential skills that should be mastered by the students. To master Communication Skills, students must first master their Language Skills. Language Skills is one of the main supporting factors in improving Communication Skills of a person because by learning Language Skills students are considered capable of communicating well and correctly so that the message or how to deliver the message to the listener can be conveyed clearly and easily understood. However, it cannot be denied that English output or learning outcomes which are less optimal is the problem which is frequently found in the implementation of the learning process. This research aimed to improve students’ language skills by developing learning model in English subject for VIII graders of SMP N 1 Uram Jaya through Directed-Project Based Learning (DPjBL) implementation. This study is designed in Research and Development (R & D) using ADDIE model development. The researcher collected data through observation, questionnaire, interview, test, and documentation which were then analyzed qualitatively and quantitatively. The results showed that DPjBL is effective to use, it is seen from the difference in value between the pretest and posttest of the control class and the experimental class. From the results of a questionnaire filled in general, the students and teachers agreed to DPjBL learning model. This learning model can increase the students' English achievement.

Keywords: language skills, learning model, Directed-Project Based Learning (DPjBL), English achievement

Procedia PDF Downloads 162
15670 MHD Boundary Layer Flow of a Nanofluid Past a Wedge Shaped Wick in Heat Pipe

Authors: Ziya Uddin

Abstract:

This paper deals with the theoretical and numerical investigation of magneto-hydrodynamic boundary layer flow of a nano fluid past a wedge shaped wick in heat pipe used for the cooling of electronic components and different type of machines. To incorporate the effect of nanoparticle diameter, concentration of nanoparticles in the pure fluid, nano thermal layer formed around the nanoparticle and Brownian motion of nano particles etc., appropriate models are used for the effective thermal and physical properties of nano fluids. To model the rotation of nano particles inside the base fluid, microfluidics theory is used. In this investigation ethylene glycol (EG) based nanofluids, are taken into account. The non-linear equations governing the flow and heat transfer are solved by using a very effective particle swarm optimization technique along with Runge-Kutta method. The values of heat transfer coefficient are found for different parameters involved in the formulation viz. nanoparticle concentration, nanoparticle size, magnetic field and wedge angle etc. It is found that the wedge angle, presence of magnetic field, nanoparticle size and nanoparticle concentration etc. have prominent effects on fluid flow and heat transfer characteristics for the considered configuration.

Keywords: nanofluids, wedge shaped wick, heat pipe, numerical modeling, particle swarm optimization, nanofluid applications, Heat transfer

Procedia PDF Downloads 380
15669 Gas Pressure Evaluation through Radial Velocity Measurement of Fluid Flow Modeled by Drift Flux Model

Authors: Aicha Rima Cheniti, Hatem Besbes, Joseph Haggege, Christophe Sintes

Abstract:

In this paper, we consider a drift flux mixture model of the blood flow. The mixture consists of gas phase which is carbon dioxide and liquid phase which is an aqueous carbon dioxide solution. This model was used to determine the distributions of the mixture velocity, the mixture pressure, and the carbon dioxide pressure. These theoretical data are used to determine a measurement method of mean gas pressure through the determination of radial velocity distribution. This method can be applicable in experimental domain.

Keywords: mean carbon dioxide pressure, mean mixture pressure, mixture velocity, radial velocity

Procedia PDF Downloads 316
15668 A Predictive Model of Supply and Demand in the State of Jalisco, Mexico

Authors: M. Gil, R. Montalvo

Abstract:

Business Intelligence (BI) has become a major source of competitive advantages for firms around the world. BI has been defined as the process of data visualization and reporting for understanding what happened and what is happening. Moreover, BI has been studied for its predictive capabilities in the context of trade and financial transactions. The current literature has identified that BI permits managers to identify market trends, understand customer relations, and predict demand for their products and services. This last capability of BI has been of special concern to academics. Specifically, due to its power to build predictive models adaptable to specific time horizons and geographical regions. However, the current literature of BI focuses on predicting specific markets and industries because the impact of such predictive models was relevant to specific industries or organizations. Currently, the existing literature has not developed a predictive model of BI that takes into consideration the whole economy of a geographical area. This paper seeks to create a predictive model of BI that would show the bigger picture of a geographical area. This paper uses a data set from the Secretary of Economic Development of the state of Jalisco, Mexico. Such data set includes data from all the commercial transactions that occurred in the state in the last years. By analyzing such data set, it will be possible to generate a BI model that predicts supply and demand from specific industries around the state of Jalisco. This research has at least three contributions. Firstly, a methodological contribution to the BI literature by generating the predictive supply and demand model. Secondly, a theoretical contribution to BI current understanding. The model presented in this paper incorporates the whole picture of the economic field instead of focusing on a specific industry. Lastly, a practical contribution might be relevant to local governments that seek to improve their economic performance by implementing BI in their policy planning.

Keywords: business intelligence, predictive model, supply and demand, Mexico

Procedia PDF Downloads 120
15667 Market Integration in the ECCAS Sub-Region

Authors: Mouhamed Mbouandi Njikam

Abstract:

This work assesses the trade potential of countries in the Economic Community of Central Africa States (ECCAS). The gravity model of trade is used to evaluate the trade flows of member countries, and to compute the trade potential index of ECCAS during 1995-2010. The focus is on the removal of tariffs and non-tariff barriers in the sub-region. Estimates from the gravity model are used for the calculation of the sub-region’s commercial potential. Its three main findings are: (i) the background research shows a low level of integration in the sub-region and open economies; (ii) a low level of industrialization and diversification are the main factors reducing trade potential in the sub-region; (iii) the trade creation predominate on the deflections of trade between member countries.

Keywords: gravity model, ECCAS, trade flows, trade potential, regional cooperation

Procedia PDF Downloads 416
15666 Development of Time Series Forecasting Model for Dengue Cases in Nakhon Si Thammarat, Southern Thailand

Authors: Manit Pollar

Abstract:

Identifying the dengue epidemic periods early would be helpful to take necessary actions to prevent the dengue outbreaks. Providing an accurate prediction on dengue epidemic seasons will allow sufficient time to take the necessary decisions and actions to safeguard the situation for local authorities. This study aimed to develop a forecasting model on number of dengue incidences in Nakhon Si Thammarat Province, Southern Thailand using time series analysis. We develop Seasonal Autoregressive Moving Average (SARIMA) models on the monthly data collected between 2003-2011 and validated the models using data collected between January-September 2012. The result of this study revealed that the SARIMA(1,1,0)(1,2,1)12 model closely described the trends and seasons of dengue incidence and confirmed the existence of dengue fever cases in Nakhon Si Thammarat for the years between 2003-2011. The study showed that the one-step approach for predicting dengue incidences provided significantly more accurate predictions than the twelve-step approach. The model, even if based purely on statistical data analysis, can provide a useful basis for allocation of resources for disease prevention.

Keywords: SARIMA, time series model, dengue cases, Thailand

Procedia PDF Downloads 353
15665 Direct Electrical Communication of Redox Enzyme Based on 3-Dimensional Crosslinked Redox Enzyme/Carbon Nanotube on a Thiol-Modified Au Surface

Authors: A. K. M. Kafi, S. N. Nina, Mashitah M. Yusoff

Abstract:

In this work, we have described a new 3-dimensional (3D) network of crosslinked Horseradish Peroxidase/Carbon Nanotube (HRP/CNT) on a thiol-modified Au surface in order to build up the effective electrical wiring of the enzyme units with the electrode. This was achieved by the electropolymerization of aniline-functionalized carbon nanotubes (CNTs) and 4-aminothiophenol -modified-HRP on a 4-aminothiophenol monolayer-modified Au electrode. The synthesized 3D HRP/CNT networks were characterized with cyclic voltammetry and amperometry, resulting the establishment direct electron transfer between the redox active unit of HRP and the Au surface. Electrochemical measurements reveal that the immobilized HRP exhibits high biological activity and stability and a quasi-reversible redox peak of the redox center of HRP was observed at about −0.355 and −0.275 V vs. Ag/AgCl. The electron transfer rate constant, KS and electron transfer co-efficient were found to be 0.57 s-1 and 0.42, respectively. Based on the electrocatalytic process by direct electrochemistry of HRP, a biosensor for detecting H2O2 was developed. The developed biosensor exhibits excellent electrocatalytic activity for the reduction of H2O2. The proposed biosensor modified with HRP/CNT 3D network displays a broader linear range and a lower detection limit for H2O2 determination. The linear range is from 1.0×10−7 to 1.2×10−4M with a detection limit of 2.2.0×10−8M at 3σ. Moreover, this biosensor exhibits very high sensitivity, good reproducibility and long-time stability. In summary, ease of fabrication, a low cost, fast response and high sensitivity are the main advantages of the new biosensor proposed in this study. These obvious advantages would really help for the real analytical applicability of the proposed biosensor.

Keywords: biosensor, nanomaterials, redox enzyme, thiol-modified Au surface

Procedia PDF Downloads 323
15664 Structural Analysis and Detail Design of APV Module Structure Using Topology Optimization Design

Authors: Hyun Kyu Cho, Jun Soo Kim, Young Hoon Lee, Sang Hoon Kang, Young Chul Park

Abstract:

In the study, structure for one of offshore drilling system APV(Air Pressure Vessle) modules was designed by using topology optimum design and performed structural safety evaluation according to DNV rules. 3D model created base on design area and non-design area separated by using topology optimization for the environmental loads. This model separated 17 types for wind loads and dynamic loads and performed structural analysis evaluation for each model. As a result, the maximum stress occurred 181.25MPa.

Keywords: APV, topology optimum design, DNV, structural analysis, stress

Procedia PDF Downloads 418
15663 Developing Integrated Model for Building Design and Evacuation Planning

Authors: Hao-Hsi Tseng, Hsin-Yun Lee

Abstract:

In the process of building design, the designers have to complete the spatial design and consider the evacuation performance at the same time. It is usually difficult to combine the two planning processes and it results in the gap between spatial design and evacuation performance. Then the designers cannot complete an integrated optimal design solution. In addition, the evacuation routing models proposed by previous researchers is different from the practical evacuation decisions in the real field. On the other hand, more and more building design projects are executed by Building Information Modeling (BIM) in which the design content is formed by the object-oriented framework. Thus, the integration of BIM and evacuation simulation can make a significant contribution for designers. Therefore, this research plan will establish a model that integrates spatial design and evacuation planning. The proposed model will provide the support for the spatial design modifications and optimize the evacuation planning. The designers can complete the integrated design solution in BIM. Besides, this research plan improves the evacuation routing method to make the simulation results more practical. The proposed model will be applied in a building design project for evaluation and validation when it will provide the near-optimal design suggestion. By applying the proposed model, the integration and efficiency of the design process are improved and the evacuation plan is more useful. The quality of building spatial design will be better.

Keywords: building information modeling, evacuation, design, floor plan

Procedia PDF Downloads 450
15662 An Optimization Model for Waste Management in Demolition Works

Authors: Eva Queheille, Franck Taillandier, Nadia Saiyouri

Abstract:

Waste management has become a major issue in demolition works, because of its environmental impact (energy consumption, resource consumption, pollution…). However, improving waste management requires to take also into account the overall demolition process and to consider demolition main objectives (e.g. cost, delay). Establishing a strategy with these conflicting objectives (economic and environment) remains complex. In order to provide a decision-support for demolition companies, a multi-objective optimization model was developed. In this model, a demolition strategy is computed from a set of 80 decision variables (worker team composition, machines, treatment for each type of waste, choice of treatment platform…), which impacts the demolition objectives. The model has experimented on a real-case study (demolition of several buildings in France). To process the optimization, different optimization algorithms (NSGA2, MOPSO, DBEA…) were tested. Results allow the engineer in charge of this case, to build a sustainable demolition strategy without affecting cost or delay.

Keywords: deconstruction, life cycle assessment, multi-objective optimization, waste management

Procedia PDF Downloads 146
15661 Geospatial Curve Fitting Methods for Disease Mapping of Tuberculosis in Eastern Cape Province, South Africa

Authors: Davies Obaromi, Qin Yongsong, James Ndege

Abstract:

To interpolate scattered or regularly distributed data, there are imprecise or exact methods. However, there are some of these methods that could be used for interpolating data in a regular grid and others in an irregular grid. In spatial epidemiology, it is important to examine how a disease prevalence rates are distributed in space, and how they relate with each other within a defined distance and direction. In this study, for the geographic and graphic representation of the disease prevalence, linear and biharmonic spline methods were implemented in MATLAB, and used to identify, localize and compare for smoothing in the distribution patterns of tuberculosis (TB) in Eastern Cape Province. The aim of this study is to produce a more “smooth” graphical disease map for TB prevalence patterns by a 3-D curve fitting techniques, especially the biharmonic splines that can suppress noise easily, by seeking a least-squares fit rather than exact interpolation. The datasets are represented generally as a 3D or XYZ triplets, where X and Y are the spatial coordinates and Z is the variable of interest and in this case, TB counts in the province. This smoothing spline is a method of fitting a smooth curve to a set of noisy observations using a spline function, and it has also become the conventional method for its high precision, simplicity and flexibility. Surface and contour plots are produced for the TB prevalence at the provincial level for 2012 – 2015. From the results, the general outlook of all the fittings showed a systematic pattern in the distribution of TB cases in the province and this is consistent with some spatial statistical analyses carried out in the province. This new method is rarely used in disease mapping applications, but it has a superior advantage to be assessed at subjective locations rather than only on a rectangular grid as seen in most traditional GIS methods of geospatial analyses.

Keywords: linear, biharmonic splines, tuberculosis, South Africa

Procedia PDF Downloads 235
15660 Neural Synchronization - The Brain’s Transfer of Sensory Data

Authors: David Edgar

Abstract:

To understand how the brain’s subconscious and conscious functions, we must conquer the physics of Unity, which leads to duality’s algorithm. Where the subconscious (bottom-up) and conscious (top-down) processes function together to produce and consume intelligence, we use terms like ‘time is relative,’ but we really do understand the meaning. In the brain, there are different processes and, therefore, different observers. These different processes experience time at different rates. A sensory system such as the eyes cycles measurement around 33 milliseconds, the conscious process of the frontal lobe cycles at 300 milliseconds, and the subconscious process of the thalamus cycle at 5 milliseconds. Three different observers experience time differently. To bridge observers, the thalamus, which is the fastest of the processes, maintains a synchronous state and entangles the different components of the brain’s physical process. The entanglements form a synchronous cohesion between the brain components allowing them to share the same state and execute in the same measurement cycle. The thalamus uses the shared state to control the firing sequence of the brain’s linear subconscious process. Sharing state also allows the brain to cheat on the amount of sensory data that must be exchanged between components. Only unpredictable motion is transferred through the synchronous state because predictable motion already exists in the shared framework. The brain’s synchronous subconscious process is entirely based on energy conservation, where prediction regulates energy usage. So, the eyes every 33 milliseconds dump their sensory data into the thalamus every day. The thalamus is going to perform a motion measurement to identify the unpredictable motion in the sensory data. Here is the trick. The thalamus conducts its measurement based on the original observation time of the sensory system (33 ms), not its own process time (5 ms). This creates a data payload of synchronous motion that preserves the original sensory observation. Basically, a frozen moment in time (Flat 4D). The single moment in time can then be processed through the single state maintained by the synchronous process. Other processes, such as consciousness (300 ms), can interface with the synchronous state to generate awareness of that moment. Now, synchronous data traveling through a separate faster synchronous process creates a theoretical time tunnel where observation time is tunneled through the synchronous process and is reproduced on the other side in the original time-relativity. The synchronous process eliminates time dilation by simply removing itself from the equation so that its own process time does not alter the experience. To the original observer, the measurement appears to be instantaneous, but in the thalamus, a linear subconscious process generating sensory perception and thought production is being executed. It is all just occurring in the time available because other observation times are slower than thalamic measurement time. For life to exist in the physical universe requires a linear measurement process, it just hides by operating at a faster time relativity. What’s interesting is time dilation is not the problem; it’s the solution. Einstein said there was no universal time.

Keywords: neural synchronization, natural intelligence, 99.95% IoT data transmission savings, artificial subconscious intelligence (ASI)

Procedia PDF Downloads 119
15659 Application of Public Access Two-Dimensional Hydrodynamic and Distributed Hydrological Models for Flood Forecasting in Ungauged Basins

Authors: Ahmad Shayeq Azizi, Yuji Toda

Abstract:

In Afghanistan, floods are the most frequent and recurrent events among other natural disasters. On the other hand, lack of monitoring data is a severe problem, which increases the difficulty of making the appropriate flood countermeasures of flood forecasting. This study is carried out to simulate the flood inundation in Harirud River Basin by application of distributed hydrological model, Integrated Flood Analysis System (IFAS) and 2D hydrodynamic model, International River Interface Cooperative (iRIC) based on satellite rainfall combined with historical peak discharge and global accessed data. The results of the simulation can predict the inundation area, depth and velocity, and the hardware countermeasures such as the impact of levee installation can be discussed by using the present method. The methodology proposed in this study is suitable for the area where hydrological and geographical data including river survey data are poorly observed.

Keywords: distributed hydrological model, flood inundation, hydrodynamic model, ungauged basins

Procedia PDF Downloads 160
15658 Numerical Modeling of Flow in USBR II Stilling Basin with End Adverse Slope

Authors: Hamidreza Babaali, Alireza Mojtahedi, Nasim Soori, Saba Soori

Abstract:

Hydraulic jump is one of the effective ways of energy dissipation in stilling basins that the ‎energy is highly dissipated by jumping. Adverse slope surface at the end stilling basin is ‎caused to increase energy dissipation and stability of the hydraulic jump. In this study, the adverse slope ‎has been added to end of United States Bureau of Reclamation (USBR) II stilling basin in hydraulic model of Nazloochay dam with scale 1:40, and flow simulated into stilling basin using Flow-3D ‎software. The numerical model is verified by experimental data of water depth in ‎stilling basin. Then, the parameters of water level profile, Froude Number, pressure, air ‎entrainment and turbulent dissipation investigated for discharging 300 m3/s using K-Ɛ and Re-Normalization Group (RNG) turbulence ‎models. The results showed a good agreement between numerical and experimental model‎ as ‎numerical model can be used to optimize of stilling basins.‎

Keywords: experimental and numerical modelling, end adverse slope, flow ‎parameters, USBR II stilling basin

Procedia PDF Downloads 172
15657 Synthesis of Pd@ Cu Core−Shell Nanowires by Galvanic Displacement of Cu by Pd²⁺ Ions as a Modified Glassy Carbon Electrode for the Simultaneous Determination of Dihydroxybenzene Isomers Speciation

Authors: Majid Farsadrouh Rashti, Parisa Jahani, Amir Shafiee, Mehrdad Mofidi

Abstract:

The dihydroxybenzene isomers, hydroquinone (HQ), catechol (CC) and resorcinol (RS) have been widely recognized as important environmental pollutants due to their toxicity and low degradability in the ecological environment. Speciation of HQ, CC and RS is very important for environmental analysis because they co-exist of these isomers in environmental samples and are too difficult to degrade as an environmental contaminant with high toxicity. There are many analytical methods have been reported for detecting these isomers, such as spectrophotometry, fluorescence, High-performance liquid chromatography (HPLC) and electrochemical methods. These methods have attractive advantages such as simple and fast response, low maintenance costs, wide linear analysis range, high efficiency, excellent selectivity and high sensitivity. A novel modified glassy carbon electrode (GCE) with Pd@ Cu/CNTs core−shell nanowires for the simultaneous determination of hydroquinone (HQ), catechol (CC) and resorcinol (RS) is described. A detailed investigation by field emission scanning electron microscopy and electrochemistry was performed in order to elucidate the preparation process and properties of the GCE/ Pd/CuNWs-CNTs. The electrochemical response characteristic of the modified GPE/LFOR toward HQ, CC and RS were investigated by cyclic voltammetry, differential pulse voltammetry (DPV) and Chronoamperometry. Under optimum conditions, the calibrations curves were linear up to 228 µM for each with detection limits of 0.4, 0.6 and 0.8 µM for HQ, CC and RS, respectively. The diffusion coefficient for the oxidation of HQ, CC and RS at the modified electrode was calculated as 6.5×10⁻⁵, 1.6 ×10⁻⁵ and 8.5 ×10⁻⁵ cm² s⁻¹, respectively. DPV was used for the simultaneous determination of HQ, CC and RS at the modified electrode and the relative standard deviations were 2.1%, 1.9% and 1.7% for HQ, CC and RS, respectively. Moreover, GCE/Pd/CuNWs-CNTs was successfully used for determination of HQ, CC and RS in real samples.

Keywords: dihydroxybenzene isomers, galvanized copper nanowires, electrochemical sensor, Palladium, speciation

Procedia PDF Downloads 125
15656 An Investigation on Designing and Enhancing the Performance of H-Darrieus Wind Turbine of 10KW at the Medium Range of Wind Speed in Vietnam

Authors: Ich Long Ngo, Dinh Tai Dang, Ngoc Tu Nguyen, Minh Duc Nguyen

Abstract:

This paper describes an investigation on designing and enhancing the performance of H-Darrieus wind turbine (HDWT) of 10kW at the medium wind speed. The aerodynamic characteristics of this turbine were investigated by both theoretical and numerical approaches. The optimal design procedure was first proposed to enhance the power coefficient under various effects, such as airfoil type, number of blades, solidity, aspect ratio, and tip speed ratio. As a result, the overall design of the 10kW HDWT was well achieved, and the power characteristic of this turbine was found by numerical approach. Additionally, the maximum power coefficient predicted is up to 0.41 at the tip speed ratio of 3.7 and wind speed of 8 m/s. Particularly, a generalized correlation of power coefficient with tip speed ratio and wind speed is first proposed. These results obtained are very useful for enhancing the performance of the HDWTs placed in a country with high wind power potential like Vietnam.

Keywords: computational fluid dynamics, double multiple stream tube, h-darrieus wind turbine, renewable energy

Procedia PDF Downloads 107
15655 A Novel Machining Method and Tool-Path Generation for Bent Mandrel

Authors: Hong Lu, Yongquan Zhang, Wei Fan, Xiangang Su

Abstract:

Bent mandrel has been widely used as precise mould in automobile industry, shipping industry and aviation industry. To improve the versatility and efficiency of turning method of bent mandrel with fixed rotational center, an instantaneous machining model based on cutting parameters and machine dimension is prospered in this paper. The spiral-like tool path generation approach in non-axisymmetric turning process of bent mandrel is developed as well to deal with the error of part-to-part repeatability in existed turning model. The actual cutter-location points are calculated by cutter-contact points, which are obtained from the approach of spiral sweep process using equal-arc-length segment principle in polar coordinate system. The tool offset is set to avoid the interference between tool and work piece is also considered in the machining model. Depend on the spindle rotational angle, synchronization control of X-axis, Z-axis and C-axis is adopted to generate the tool-path of the turning process. The simulation method is developed to generate NC program according to the presented model, which includes calculation of cutter-location points and generation of tool-path of cutting process. With the approach of a bent mandrel taken as an example, the maximum offset of center axis is 4mm in the 3D space. Experiment results verify that the machining model and turning method are appropriate for the characteristics of bent mandrel.

Keywords: bent mandrel, instantaneous machining model, simulation method, tool-path generation

Procedia PDF Downloads 331
15654 Assessing Effects of an Intervention on Bottle-Weaning and Reducing Daily Milk Intake from Bottles in Toddlers Using Two-Part Random Effects Models

Authors: Yungtai Lo

Abstract:

Two-part random effects models have been used to fit semi-continuous longitudinal data where the response variable has a point mass at 0 and a continuous right-skewed distribution for positive values. We review methods proposed in the literature for analyzing data with excess zeros. A two-part logit-log-normal random effects model, a two-part logit-truncated normal random effects model, a two-part logit-gamma random effects model, and a two-part logit-skew normal random effects model were used to examine effects of a bottle-weaning intervention on reducing bottle use and daily milk intake from bottles in toddlers aged 11 to 13 months in a randomized controlled trial. We show in all four two-part models that the intervention promoted bottle-weaning and reduced daily milk intake from bottles in toddlers drinking from a bottle. We also show that there are no differences in model fit using either the logit link function or the probit link function for modeling the probability of bottle-weaning in all four models. Furthermore, prediction accuracy of the logit or probit link function is not sensitive to the distribution assumption on daily milk intake from bottles in toddlers not off bottles.

Keywords: two-part model, semi-continuous variable, truncated normal, gamma regression, skew normal, Pearson residual, receiver operating characteristic curve

Procedia PDF Downloads 343
15653 First Principle Calculations of the Structural and Optoelectronic Properties of Cubic Perovskite CsSrF3

Authors: Meriem Harmel, Houari Khachai

Abstract:

We have investigated the structural, electronic and optical properties of a compound perovskite CsSrF3 using the full-potential linearized augmented plane wave (FP-LAPW) method within density functional theory (DFT). In this approach, both the local density approximation (LDA) and the generalized gradient approximation (GGA) were used for exchange-correlation potential calculation. The ground state properties such as lattice parameter, bulk modulus and its pressure derivative were calculated and the results are compared whit experimental and theoretical data. Electronic and bonding properties are discussed from the calculations of band structure, density of states and electron charge density, where the fundamental energy gap is direct under ambient conditions. The contribution of the different bands was analyzed from the total and partial density of states curves. The optical properties (namely: the real and the imaginary parts of the dielectric function ε(ω), the refractive index n(ω) and the extinction coefficient k(ω)) were calculated for radiation up to 35.0 eV. This is the first quantitative theoretical prediction of the optical properties for the investigated compound and still awaits experimental confirmations.

Keywords: DFT, fluoroperovskite, electronic structure, optical properties

Procedia PDF Downloads 457
15652 DenseNet and Autoencoder Architecture for COVID-19 Chest X-Ray Image Classification and Improved U-Net Lung X-Ray Segmentation

Authors: Jonathan Gong

Abstract:

Purpose AI-driven solutions are at the forefront of many pathology and medical imaging methods. Using algorithms designed to better the experience of medical professionals within their respective fields, the efficiency and accuracy of diagnosis can improve. In particular, X-rays are a fast and relatively inexpensive test that can diagnose diseases. In recent years, X-rays have not been widely used to detect and diagnose COVID-19. The under use of Xrays is mainly due to the low diagnostic accuracy and confounding with pneumonia, another respiratory disease. However, research in this field has expressed a possibility that artificial neural networks can successfully diagnose COVID-19 with high accuracy. Models and Data The dataset used is the COVID-19 Radiography Database. This dataset includes images and masks of chest X-rays under the labels of COVID-19, normal, and pneumonia. The classification model developed uses an autoencoder and a pre-trained convolutional neural network (DenseNet201) to provide transfer learning to the model. The model then uses a deep neural network to finalize the feature extraction and predict the diagnosis for the input image. This model was trained on 4035 images and validated on 807 separate images from the ones used for training. The images used to train the classification model include an important feature: the pictures are cropped beforehand to eliminate distractions when training the model. The image segmentation model uses an improved U-Net architecture. This model is used to extract the lung mask from the chest X-ray image. The model is trained on 8577 images and validated on a validation split of 20%. These models are calculated using the external dataset for validation. The models’ accuracy, precision, recall, f1-score, IOU, and loss are calculated. Results The classification model achieved an accuracy of 97.65% and a loss of 0.1234 when differentiating COVID19-infected, pneumonia-infected, and normal lung X-rays. The segmentation model achieved an accuracy of 97.31% and an IOU of 0.928. Conclusion The models proposed can detect COVID-19, pneumonia, and normal lungs with high accuracy and derive the lung mask from a chest X-ray with similarly high accuracy. The hope is for these models to elevate the experience of medical professionals and provide insight into the future of the methods used.

Keywords: artificial intelligence, convolutional neural networks, deep learning, image processing, machine learning

Procedia PDF Downloads 123
15651 The Impact of the Composite Expanded Graphite PCM on the PV Panel Whole Year Electric Output: Case Study Milan

Authors: Hasan A Al-Asadi, Ali Samir, Afrah Turki Awad, Ali Basem

Abstract:

Integrating the phase change material (PCM) with photovoltaic (PV) panels is one of the effective techniques to minimize the PV panel temperature and increase their electric output. In order to investigate the impact of the PCM on the electric output of the PV panels for a whole year, a lumped-distributed parameter model for the PV-PCM module has been developed. This development has considered the impact of the PCM density variation between the solid phase and liquid phase. This contribution will increase the assessment accuracy of the electric output of the PV-PCM module. The second contribution is to assess the impact of the expanded composite graphite-PCM on the PV electric output in Milan for a whole year. The novel one-dimensional model has been solved using MATLAB software. The results of this model have been validated against literature experiment work. The weather and the solar radiation data have been collected. The impact of expanded graphite-PCM on the electric output of the PV panel for a whole year has been investigated. The results indicate this impact has an enhancement rate of 2.39% for the electric output of the PV panel in Milan for a whole year.

Keywords: PV panel efficiency, PCM, numerical model, solar energy

Procedia PDF Downloads 162
15650 Analytical Solution for Stellar Distance Based on Photon Dominated Cosmic Expansion Model

Authors: Xiaoyun Li, Suoang Longzhou

Abstract:

This paper derives the analytical solution of stellar distance according to its redshift based on the photon-dominated universe expansion model. Firstly, it calculates stellar separation speed and the farthest distance of observable stars via simulation. Then the analytical solution of stellar distance according to its redshift is derived. It shows that when the redshift is large, the stellar distance (and its separation speed) is not proportional to its redshift due to the relativity effect. It also reveals the relationship between stellar age and its redshift. The correctness of the analytical solution is verified by the latest astronomic observations of Ia supernovas in 2020.

Keywords: redshift, cosmic expansion model, analytical solution, stellar distance

Procedia PDF Downloads 156
15649 Knowledge Audit Model for Requirement Elicitation Process

Authors: Laleh Taheri, Noraini C. Pa, Rusli Abdullah, Salfarina Abdullah

Abstract:

Knowledge plays an important role to the success of any organization. Software development organizations are highly knowledge-intensive organizations especially in their Requirement Elicitation Process (REP). There are several problems regarding communicating and using the knowledge in REP such as misunderstanding, being out of scope, conflicting information and changes of requirements. All of these problems occurred in transmitting the requirements knowledge during REP. Several researches have been done in REP in order to solve the problem towards requirements. Knowledge Audit (KA) approaches were proposed in order to solve managing knowledge in human resources, financial, and manufacturing. There is lack of study applying the KA in requirements elicitation process. Therefore, this paper proposes a KA model for REP in supporting to acquire good requirements.

Keywords: knowledge audit, requirement elicitation process, KA model, knowledge in requirement elicitation

Procedia PDF Downloads 339
15648 Preference for Housing Services and Rational House Price Bubbles

Authors: Stefanie Jeanette Huber

Abstract:

This paper explores the relevance and implications of preferences for housing services on house price fluctuations through the lens of an overlapping generation’s model. The model implies that an economy whose agents have lower preferences for housing services is characterized with lower expenditure shares on housing services and will tend to experience more frequent and more volatile housing bubbles. These model predictions are tested empirically in the companion paper Housing Booms and Busts - Convergences and Divergences across OECD countries. Between 1970 - 2013, countries who spend less on housing services as a share of total income experienced significantly more housing cycles and the associated housing boom-bust cycles were more violent. Finally, the model is used to study the impact of rental subsidies and help-to-buy schemes on rational housing bubbles. Rental subsidies are found to contribute to the control of housing bubbles, whereas help-to- buy scheme makes the economy more bubble-prone.

Keywords: housing bubbles, housing booms and busts, preference for housing services, expenditure shares for housing services, rental and purchase subsidies

Procedia PDF Downloads 288
15647 Autonomous Quantum Competitive Learning

Authors: Mohammed A. Zidan, Alaa Sagheer, Nasser Metwally

Abstract:

Real-time learning is an important goal that most of artificial intelligence researches try to achieve it. There are a lot of problems and applications which require low cost learning such as learn a robot to be able to classify and recognize patterns in real time and real-time recall. In this contribution, we suggest a model of quantum competitive learning based on a series of quantum gates and additional operator. The proposed model enables to recognize any incomplete patterns, where we can increase the probability of recognizing the pattern at the expense of the undesired ones. Moreover, these undesired ones could be utilized as new patterns for the system. The proposed model is much better compared with classical approaches and more powerful than the current quantum competitive learning approaches.

Keywords: competitive learning, quantum gates, quantum gates, winner-take-all

Procedia PDF Downloads 462
15646 Predicting Indonesia External Debt Crisis: An Artificial Neural Network Approach

Authors: Riznaldi Akbar

Abstract:

In this study, we compared the performance of the Artificial Neural Network (ANN) model with back-propagation algorithm in correctly predicting in-sample and out-of-sample external debt crisis in Indonesia. We found that exchange rate, foreign reserves, and exports are the major determinants to experiencing external debt crisis. The ANN in-sample performance provides relatively superior results. The ANN model is able to classify correctly crisis of 89.12 per cent with reasonably low false alarms of 7.01 per cent. In out-of-sample, the prediction performance fairly deteriorates compared to their in-sample performances. It could be explained as the ANN model tends to over-fit the data in the in-sample, but it could not fit the out-of-sample very well. The 10-fold cross-validation has been used to improve the out-of-sample prediction accuracy. The results also offer policy implications. The out-of-sample performance could be very sensitive to the size of the samples, as it could yield a higher total misclassification error and lower prediction accuracy. The ANN model could be used to identify past crisis episodes with some accuracy, but predicting crisis outside the estimation sample is much more challenging because of the presence of uncertainty.

Keywords: debt crisis, external debt, artificial neural network, ANN

Procedia PDF Downloads 433
15645 Failure Inference and Optimization for Step Stress Model Based on Bivariate Wiener Model

Authors: Soudabeh Shemehsavar

Abstract:

In this paper, we consider the situation under a life test, in which the failure time of the test units are not related deterministically to an observable stochastic time varying covariate. In such a case, the joint distribution of failure time and a marker value would be useful for modeling the step stress life test. The problem of accelerating such an experiment is considered as the main aim of this paper. We present a step stress accelerated model based on a bivariate Wiener process with one component as the latent (unobservable) degradation process, which determines the failure times and the other as a marker process, the degradation values of which are recorded at times of failure. Parametric inference based on the proposed model is discussed and the optimization procedure for obtaining the optimal time for changing the stress level is presented. The optimization criterion is to minimize the approximate variance of the maximum likelihood estimator of a percentile of the products’ lifetime distribution.

Keywords: bivariate normal, Fisher information matrix, inverse Gaussian distribution, Wiener process

Procedia PDF Downloads 315
15644 The Effects of Different Parameters of Wood Floating Debris on Scour Rate Around Bridge Piers

Authors: Muhanad Al-Jubouri

Abstract:

A local scour is the most important of the several scours impacting bridge performance and security. Even though scour is widespread in bridges, especially during flood seasons, the experimental tests could not be applied to many standard highway bridges. A computational fluid dynamics numerical model was used to solve the problem of calculating local scouring and deposition for non-cohesive silt and clear water conditions near single and double cylindrical piers with the effect of floating debris. When FLOW-3D software is employed with the Rang turbulence model, the Nilsson bed-load transfer equation and fine mesh size are considered. The numerical findings of single cylindrical piers correspond pretty well with the physical model's results. Furthermore, after parameter effectiveness investigates the range of outcomes based on predicted user inputs such as the bed-load equation, mesh cell size, and turbulence model, the final numerical predictions are compared to experimental data. When the findings are compared, the error rate for the deepest point of the scour is equivalent to 3.8% for the single pier example.

Keywords: local scouring, non-cohesive, clear water, computational fluid dynamics, turbulence model, bed-load equation, debris

Procedia PDF Downloads 65