Search results for: equivalent linear model
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 19027

Search results for: equivalent linear model

15847 The Impact of Bequest Taxation on Human Capital Accumulation

Authors: Maciej Dudek, Robert Kruszewski, Janusz Kudla, Konrad Walczyk

Abstract:

In this paper, we study how taxation of bequests affects human capital formation in the long term and short term horizon. Our underlying model is an overlapping generation model (OLG) with some degree of altruism on the part of the ancestors' generation towards their descendants. We ask the question in three separate frameworks. First, we study a simple one-sector model where a proxy of human capital is wage income. It the steady-state -for CRRA utility function and human capital produced with non-decreasing returns -the taxation of bequests is neutral to the accumulation of human capital. In the second framework, neutrality applies to the growth rates of human capital, physical capital, and consumption. In this case, taxation increases the level of bequests, leading to a lower value of current consumption. Finally in we consider two periods model instead of infinite horizon model as long as the tax revenue is at least partially rebated back to the public, the fraction of human capital engaged in the process of formation of human capital increases with the tax rate on bequests. In other words, taxation of bequests is partially offset by an increase in human capital formation. Higher human capital allows the future generation to earn higher wages, and today's generation can find it optimal to endow the future generation with more human capital when taxation is imposed on physical capital transferred to the next generation.

Keywords: taxation, bequests, policy, human capital

Procedia PDF Downloads 154
15846 Prediction of Phonon Thermal Conductivity of F.C.C. Al by Molecular Dynamics Simulation

Authors: Leila Momenzadeh, Alexander V. Evteev, Elena V. Levchenko, Tanvir Ahmed, Irina Belova, Graeme Murch

Abstract:

In this work, the phonon thermal conductivity of f.c.c. Al is investigated in detail in the temperature range 100 – 900 K within the framework of equilibrium molecular dynamics simulations making use of the Green-Kubo formalism and one of the most reliable embedded-atom method potentials. It is found that the heat current auto-correlation function of the f.c.c. Al model demonstrates a two-stage temporal decay similar to the previously observed for f.c.c Cu model. After the first stage of decay, the heat current auto-correlation function of the f.c.c. Al model demonstrates a peak in the temperature range 100-800 K. The intensity of the peak decreases as the temperature increases. At 900 K, it transforms to a shoulder. To describe the observed two-stage decay of the heat current auto-correlation function of the f.c.c. Al model, we employ decomposition model recently developed for phonon-mediated thermal transport in a monoatomic lattice. We found that the electronic contribution to the total thermal conductivity of f.c.c. Al dominates over the whole studied temperature range. However, the phonon contribution to the total thermal conductivity of f.c.c. Al increases as temperature decreases. It is about 1.05% at 900 K and about 12.5% at 100 K.

Keywords: aluminum, gGreen-Kubo formalism, molecular dynamics, phonon thermal conductivity

Procedia PDF Downloads 403
15845 The Aesthetics of Time in Thus Spoke Zarathustra: A Reappraisal of the Eternal Recurrence of the Same

Authors: Melanie Tang

Abstract:

According to Nietzsche, the eternal recurrence is his most important idea. However, it is perhaps his most cryptic and difficult to interpret. Early readings considered it as a cosmological hypothesis about the cyclic nature of time. However, following Nehamas’s ‘Life as Literature’ (1985), it has become a widespread interpretation that the eternal recurrence never really had any theoretical dimensions, and is not actually a philosophy of time, but a practical thought experiment intended to measure the extent to which we have mastered and perfected our lives. This paper endeavours to challenge this line of thought becoming scholarly consensus, and to carry out a more complex analysis of the eternal recurrence as it is presented in Thus Spoke Zarathustra. In its wider scope, this research proposes that Thus Spoke Zarathustra — as opposed to The Birth of Tragedy — be taken as the primary source for a study of Nietzsche’s Aesthetics, due to its more intrinsic aesthetic qualities and expressive devices. The eternal recurrence is the central philosophy of a work that communicates its ideas in unprecedentedly experimental and aesthetic terms, and a more in-depth understanding of why Nietzsche chooses to present his conception of time in aesthetic terms is warranted. Through hermeneutical analysis of Thus Spoke Zarathustra and engagement with secondary sources such as those by Nehamas, Karl Löwith, and Jill Marsden, the present analysis challenges the ethics of self-perfection upon which current interpretations of the recurrence are based, as well as their reliance upon a linear conception of time. Instead, it finds the recurrence to be a cyclic interplay between the self and the world, rather than a metric pertaining solely to the self. In this interpretation, time is found to be composed of an intertemporal rather than linear multitude of will to power, which structures itself through tensional cycles into an experience of circular time that can be seen to have aesthetic dimensions. In putting forth this understanding of the eternal recurrence, this research hopes to reopen debate on this key concept in the field of Nietzsche studies.

Keywords: Nietzsche, eternal recurrence, Zarathustra, aesthetics, time

Procedia PDF Downloads 136
15844 Structural Damage Detection Using Modal Data Employing Teaching Learning Based Optimization

Authors: Subhajit Das, Nirjhar Dhang

Abstract:

Structural damage detection is a challenging work in the field of structural health monitoring (SHM). The damage detection methods mainly focused on the determination of the location and severity of the damage. Model updating is a well known method to locate and quantify the damage. In this method, an error function is defined in terms of difference between the signal measured from ‘experiment’ and signal obtained from undamaged finite element model. This error function is minimised with a proper algorithm, and the finite element model is updated accordingly to match the measured response. Thus, the damage location and severity can be identified from the updated model. In this paper, an error function is defined in terms of modal data viz. frequencies and modal assurance criteria (MAC). MAC is derived from Eigen vectors. This error function is minimized by teaching-learning-based optimization (TLBO) algorithm, and the finite element model is updated accordingly to locate and quantify the damage. Damage is introduced in the model by reduction of stiffness of the structural member. The ‘experimental’ data is simulated by the finite element modelling. The error due to experimental measurement is introduced in the synthetic ‘experimental’ data by adding random noise, which follows Gaussian distribution. The efficiency and robustness of this method are explained through three examples e.g., one truss, one beam and one frame problem. The result shows that TLBO algorithm is efficient to detect the damage location as well as the severity of damage using modal data.

Keywords: damage detection, finite element model updating, modal assurance criteria, structural health monitoring, teaching learning based optimization

Procedia PDF Downloads 203
15843 Machine Learning Techniques for Estimating Ground Motion Parameters

Authors: Farid Khosravikia, Patricia Clayton

Abstract:

The main objective of this study is to evaluate the advantages and disadvantages of various machine learning techniques in forecasting ground-motion intensity measures given source characteristics, source-to-site distance, and local site condition. Intensity measures such as peak ground acceleration and velocity (PGA and PGV, respectively) as well as 5% damped elastic pseudospectral accelerations at different periods (PSA), are indicators of the strength of shaking at the ground surface. Estimating these variables for future earthquake events is a key step in seismic hazard assessment and potentially subsequent risk assessment of different types of structures. Typically, linear regression-based models, with pre-defined equations and coefficients, are used in ground motion prediction. However, due to the restrictions of the linear regression methods, such models may not capture more complex nonlinear behaviors that exist in the data. Thus, this study comparatively investigates potential benefits from employing other machine learning techniques as a statistical method in ground motion prediction such as Artificial Neural Network, Random Forest, and Support Vector Machine. The algorithms are adjusted to quantify event-to-event and site-to-site variability of the ground motions by implementing them as random effects in the proposed models to reduce the aleatory uncertainty. All the algorithms are trained using a selected database of 4,528 ground-motions, including 376 seismic events with magnitude 3 to 5.8, recorded over the hypocentral distance range of 4 to 500 km in Oklahoma, Kansas, and Texas since 2005. The main reason of the considered database stems from the recent increase in the seismicity rate of these states attributed to petroleum production and wastewater disposal activities, which necessities further investigation in the ground motion models developed for these states. Accuracy of the models in predicting intensity measures, generalization capability of the models for future data, as well as usability of the models are discussed in the evaluation process. The results indicate the algorithms satisfy some physically sound characteristics such as magnitude scaling distance dependency without requiring pre-defined equations or coefficients. Moreover, it is shown that, when sufficient data is available, all the alternative algorithms tend to provide more accurate estimates compared to the conventional linear regression-based method, and particularly, Random Forest outperforms the other algorithms. However, the conventional method is a better tool when limited data is available.

Keywords: artificial neural network, ground-motion models, machine learning, random forest, support vector machine

Procedia PDF Downloads 112
15842 Free Vibration Analysis of Timoshenko Beams at Higher Modes with Central Concentrated Mass Using Coupled Displacement Field Method

Authors: K. Meera Saheb, K. Krishna Bhaskar

Abstract:

Complex structures used in many fields of engineering are made up of simple structural elements like beams, plates etc. These structural elements, sometimes carry concentrated masses at discrete points, and when subjected to severe dynamic environment tend to vibrate with large amplitudes. The frequency amplitude relationship is very much essential in determining the response of these structural elements subjected to the dynamic loads. For Timoshenko beams, the effects of shear deformation and rotary inertia are to be considered to evaluate the fundamental linear and nonlinear frequencies. A commonly used method for solving vibration problem is energy method, or a finite element analogue of the same. In the present Coupled Displacement Field method the number of undetermined coefficients is reduced to half when compared to the famous Rayleigh Ritz method, which significantly simplifies the procedure to solve the vibration problem. This is accomplished by using a coupling equation derived from the static equilibrium of the shear flexible structural element. The prime objective of the present paper here is to study, in detail, the effect of a central concentrated mass on the large amplitude free vibrations of uniform shear flexible beams. Accurate closed form expressions for linear frequency parameter for uniform shear flexible beams with a central concentrated mass was developed and the results are presented in digital form.

Keywords: coupled displacement field, coupling equation, large amplitude vibrations, moderately thick plates

Procedia PDF Downloads 212
15841 Increasing of Gain in Unstable Thin Disk Resonator

Authors: M. Asl. Dehghan, M. H. Daemi, S. Radmard, S. H. Nabavi

Abstract:

Thin disk lasers are engineered for efficient thermal cooling and exhibit superior performance for this task. However the disk thickness and large pumped area make the use of this gain format in a resonator difficult when constructing a single-mode laser. Choosing an unstable resonator design is beneficial for this purpose. On the other hand, the low gain medium restricts the application of unstable resonators to low magnifications and therefore to a poor beam quality. A promising idea to enable the application of unstable resonators to wide aperture, low gain lasers is to couple a fraction of the out coupled radiation back into the resonator. The output coupling gets dependent on the ratio of the back reflection and can be adjusted independently from the magnification. The excitation of the converging wave can be done by the use of an external reflector. The resonator performance is numerically predicted. First of all the threshold condition of linear, V and 2V shape resonator is investigated. Results show that the maximum magnification is 1.066 that is very low for high quality purposes. Inserting an additional reflector covers the low gain. The reflectivity and the related magnification of a 350 micron Yb:YAG disk are calculated. The theoretical model was based on the coupled Kirchhoff integrals and solved numerically by the Fox and Li algorithm. Results show that with back reflection mechanism in combination with increasing the number of beam incidents on disk, high gain and high magnification can occur.

Keywords: unstable resonators, thin disk lasers, gain, external reflector

Procedia PDF Downloads 399
15840 Microwave Dielectric Relaxation Study of Diethanolamine with Triethanolamine from 10 MHz-20 GHz

Authors: A. V. Patil

Abstract:

The microwave dielectric relaxation study of diethanolamine with triethanolamine binary mixture have been determined over the frequency range of 10 MHz to 20 GHz, at various temperatures using time domain reflectometry (TDR) method for 11 concentrations of the system. The present work reveals molecular interaction between same multi-functional groups [−OH and –NH2] of the alkanolamines (diethanolamine and triethanolamine) using different models such as Debye model, Excess model, and Kirkwood model. The dielectric parameters viz. static dielectric constant (ε0) and relaxation time (τ) have been obtained with Debye equation characterized by a single relaxation time without relaxation time distribution by the least squares fit method.

Keywords: diethanolamine, excess properties, kirkwood properties, time domain reflectometry, triethanolamine

Procedia PDF Downloads 285
15839 Performance Evaluation of the Classic seq2seq Model versus a Proposed Semi-supervised Long Short-Term Memory Autoencoder for Time Series Data Forecasting

Authors: Aswathi Thrivikraman, S. Advaith

Abstract:

The study is aimed at designing encoders for deciphering intricacies in time series data by redescribing the dynamics operating on a lower-dimensional manifold. A semi-supervised LSTM autoencoder is devised and investigated to see if the latent representation of the time series data can better forecast the data. End-to-end training of the LSTM autoencoder, together with another LSTM network that is connected to the latent space, forces the hidden states of the encoder to represent the most meaningful latent variables relevant for forecasting. Furthermore, the study compares the predictions with those of a traditional seq2seq model.

Keywords: LSTM, autoencoder, forecasting, seq2seq model

Procedia PDF Downloads 137
15838 Hierarchical Tree Long Short-Term Memory for Sentence Representations

Authors: Xiuying Wang, Changliang Li, Bo Xu

Abstract:

A fixed-length feature vector is required for many machine learning algorithms in NLP field. Word embeddings have been very successful at learning lexical information. However, they cannot capture the compositional meaning of sentences, which prevents them from a deeper understanding of language. In this paper, we introduce a novel hierarchical tree long short-term memory (HTLSTM) model that learns vector representations for sentences of arbitrary syntactic type and length. We propose to split one sentence into three hierarchies: short phrase, long phrase and full sentence level. The HTLSTM model gives our algorithm the potential to fully consider the hierarchical information and long-term dependencies of language. We design the experiments on both English and Chinese corpus to evaluate our model on sentiment analysis task. And the results show that our model outperforms several existing state of the art approaches significantly.

Keywords: deep learning, hierarchical tree long short-term memory, sentence representation, sentiment analysis

Procedia PDF Downloads 339
15837 'Gender' and 'Gender Equalities': Conceptual Issues

Authors: Moustafa Ali

Abstract:

The aim of this paper is to discuss and question some of the widely accepted concepts within the conceptual framework of gender from terminological, scientific, and Muslim cultural perspectives, and to introduce a new definition and a model of gender in the Arab and Muslim societies. This paper, therefore, uses a generic methodology and document analysis and comes in three sections and a conclusion. The first section discusses some of the terminological issues in the conceptual framework of gender. The second section highlights scientific issues, introduces a definition and a model of gender, whereas the third section offers Muslim cultural perspectives on some issues related to gender in the Muslim world. The paper, then, concludes with findings and recommendations reached so far.

Keywords: gender definition, gender equalities, sex-gender separability, fairness-based model of gender

Procedia PDF Downloads 120
15836 Logistics Model for Improving Quality in Railway Transport

Authors: Eva Nedeliakova, Juraj Camaj, Jaroslav Masek

Abstract:

This contribution is focused on the methodology for identifying levels of quality and improving quality through new logistics model in railway transport. It is oriented on the application of dynamic quality models, which represent an innovative method of evaluation quality services. Through this conception, time factor, expected, and perceived quality in each moment of the transportation process within logistics chain can be taken into account. Various models describe the improvement of the quality which emphases the time factor throughout the whole transportation logistics chain. Quality of services in railway transport can be determined by the existing level of service quality, by detecting the causes of dissatisfaction employees but also customers, to uncover strengths and weaknesses. This new logistics model is able to recognize critical processes in logistic chain. It includes service quality rating that must respect its specific properties, which are unrepeatability, impalpability, their use right at the time they are provided and particularly changeability, which is significant factor in the conditions of rail transport as well. These peculiarities influence the quality of service regarding the constantly increasing requirements and that result in new ways of finding progressive attitudes towards the service quality rating.

Keywords: logistics model, quality, railway transport

Procedia PDF Downloads 548
15835 Simple Multiple-Attribute Rating Technique for Optimal Decision-Making Model on Selecting Best Spiker of World Grand Prix

Authors: Chen Chih-Cheng, Chen I-Cheng, Lee Yung-Tan, Kuo Yen-Whea, Yu Chin-Hung

Abstract:

The purpose of this study is to construct a model for best spike player selection in a top volleyball tournament of the world. Data consisted of the records of 2013 World Grand Prix declared by International Volleyball Federation (FIVB). Simple Multiple-Attribute Rating Technique (SMART) was used for optimal decision-making model on the best spike player selection. The research results showed that the best spike player ranking by SMART is different than the ranking by FIVB. The results demonstrated the effectiveness and feasibility of the proposed model.

Keywords: simple multiple-attribute rating technique, World Grand Prix, best spike player, International Volleyball Federation

Procedia PDF Downloads 458
15834 The Relationship between Employee Commitment, Job Satisfaction and External Market Orientation in Vietnamese Joint-Stock Commercial Banks

Authors: Nguyen Ngoc Que Tran

Abstract:

Purpose: The purpose of this paper is to investigate the relationship between internal market orientation, external market orientation, employee commitment and job satisfaction. Design/methodology/approach: This study collected data through a survey and utilized simple linear regression and multiple regression analysis to determine if there was any support for the research hypotheses as presented in the previous chapter. Findings: Using data from 256 employees of four leading joint stock banks in Vietnam, the empirical results indicates that employee commitment is positively related with external market orientation, job satisfaction is positively related to employee commitment, and employee commitment and job satisfaction are positively related to external market orientation. However, job satisfaction has no significant positive effect on external market orientation. Theoretical contribution: The primary contribution to marketing theory arising from this study is the integration of job satisfaction, employee commitment, and external market orientation in a single research model. Practical implications: The major contribution to practice is an external market oriented bank has to respond rapidly to the future needs and preferences of its customers. This could result in high levels of commitment to the service process and in doing so provide Vietnamese joint-stock commercial banks with a competitive advantage. The finding is important for the banking service sector in general and the Vietnamese banking industry in particular.

Keywords: employee commitment, job satisfaction and external market orientation, vietnam, bank

Procedia PDF Downloads 396
15833 An Experimental Study on Some Conventional and Hybrid Models of Fuzzy Clustering

Authors: Jeugert Kujtila, Kristi Hoxhalli, Ramazan Dalipi, Erjon Cota, Ardit Murati, Erind Bedalli

Abstract:

Clustering is a versatile instrument in the analysis of collections of data providing insights of the underlying structures of the dataset and enhancing the modeling capabilities. The fuzzy approach to the clustering problem increases the flexibility involving the concept of partial memberships (some value in the continuous interval [0, 1]) of the instances in the clusters. Several fuzzy clustering algorithms have been devised like FCM, Gustafson-Kessel, Gath-Geva, kernel-based FCM, PCM etc. Each of these algorithms has its own advantages and drawbacks, so none of these algorithms would be able to perform superiorly in all datasets. In this paper we will experimentally compare FCM, GK, GG algorithm and a hybrid two-stage fuzzy clustering model combining the FCM and Gath-Geva algorithms. Firstly we will theoretically dis-cuss the advantages and drawbacks for each of these algorithms and we will describe the hybrid clustering model exploiting the advantages and diminishing the drawbacks of each algorithm. Secondly we will experimentally compare the accuracy of the hybrid model by applying it on several benchmark and synthetic datasets.

Keywords: fuzzy clustering, fuzzy c-means algorithm (FCM), Gustafson-Kessel algorithm, hybrid clustering model

Procedia PDF Downloads 497
15832 The Development of Directed-Project Based Learning as Language Learning Model to Improve Students' English Achievement

Authors: Tri Pratiwi, Sufyarma Marsidin, Hermawati Syarif, Yahya

Abstract:

The 21st-century skills being highly promoted today are Creativity and Innovation, Critical Thinking and Problem Solving, Communication and Collaboration. Communication Skill is one of the essential skills that should be mastered by the students. To master Communication Skills, students must first master their Language Skills. Language Skills is one of the main supporting factors in improving Communication Skills of a person because by learning Language Skills students are considered capable of communicating well and correctly so that the message or how to deliver the message to the listener can be conveyed clearly and easily understood. However, it cannot be denied that English output or learning outcomes which are less optimal is the problem which is frequently found in the implementation of the learning process. This research aimed to improve students’ language skills by developing learning model in English subject for VIII graders of SMP N 1 Uram Jaya through Directed-Project Based Learning (DPjBL) implementation. This study is designed in Research and Development (R & D) using ADDIE model development. The researcher collected data through observation, questionnaire, interview, test, and documentation which were then analyzed qualitatively and quantitatively. The results showed that DPjBL is effective to use, it is seen from the difference in value between the pretest and posttest of the control class and the experimental class. From the results of a questionnaire filled in general, the students and teachers agreed to DPjBL learning model. This learning model can increase the students' English achievement.

Keywords: language skills, learning model, Directed-Project Based Learning (DPjBL), English achievement

Procedia PDF Downloads 155
15831 Gas Pressure Evaluation through Radial Velocity Measurement of Fluid Flow Modeled by Drift Flux Model

Authors: Aicha Rima Cheniti, Hatem Besbes, Joseph Haggege, Christophe Sintes

Abstract:

In this paper, we consider a drift flux mixture model of the blood flow. The mixture consists of gas phase which is carbon dioxide and liquid phase which is an aqueous carbon dioxide solution. This model was used to determine the distributions of the mixture velocity, the mixture pressure, and the carbon dioxide pressure. These theoretical data are used to determine a measurement method of mean gas pressure through the determination of radial velocity distribution. This method can be applicable in experimental domain.

Keywords: mean carbon dioxide pressure, mean mixture pressure, mixture velocity, radial velocity

Procedia PDF Downloads 306
15830 A Predictive Model of Supply and Demand in the State of Jalisco, Mexico

Authors: M. Gil, R. Montalvo

Abstract:

Business Intelligence (BI) has become a major source of competitive advantages for firms around the world. BI has been defined as the process of data visualization and reporting for understanding what happened and what is happening. Moreover, BI has been studied for its predictive capabilities in the context of trade and financial transactions. The current literature has identified that BI permits managers to identify market trends, understand customer relations, and predict demand for their products and services. This last capability of BI has been of special concern to academics. Specifically, due to its power to build predictive models adaptable to specific time horizons and geographical regions. However, the current literature of BI focuses on predicting specific markets and industries because the impact of such predictive models was relevant to specific industries or organizations. Currently, the existing literature has not developed a predictive model of BI that takes into consideration the whole economy of a geographical area. This paper seeks to create a predictive model of BI that would show the bigger picture of a geographical area. This paper uses a data set from the Secretary of Economic Development of the state of Jalisco, Mexico. Such data set includes data from all the commercial transactions that occurred in the state in the last years. By analyzing such data set, it will be possible to generate a BI model that predicts supply and demand from specific industries around the state of Jalisco. This research has at least three contributions. Firstly, a methodological contribution to the BI literature by generating the predictive supply and demand model. Secondly, a theoretical contribution to BI current understanding. The model presented in this paper incorporates the whole picture of the economic field instead of focusing on a specific industry. Lastly, a practical contribution might be relevant to local governments that seek to improve their economic performance by implementing BI in their policy planning.

Keywords: business intelligence, predictive model, supply and demand, Mexico

Procedia PDF Downloads 105
15829 Market Integration in the ECCAS Sub-Region

Authors: Mouhamed Mbouandi Njikam

Abstract:

This work assesses the trade potential of countries in the Economic Community of Central Africa States (ECCAS). The gravity model of trade is used to evaluate the trade flows of member countries, and to compute the trade potential index of ECCAS during 1995-2010. The focus is on the removal of tariffs and non-tariff barriers in the sub-region. Estimates from the gravity model are used for the calculation of the sub-region’s commercial potential. Its three main findings are: (i) the background research shows a low level of integration in the sub-region and open economies; (ii) a low level of industrialization and diversification are the main factors reducing trade potential in the sub-region; (iii) the trade creation predominate on the deflections of trade between member countries.

Keywords: gravity model, ECCAS, trade flows, trade potential, regional cooperation

Procedia PDF Downloads 408
15828 Water Equivalent from the Point of View of Fast Neutron Removal Cross-Section

Authors: Mohammed Alrajhi

Abstract:

Radiological properties of gel dosimeters and phantom materials are often evaluated in terms of effective atomic number, electron density, photon mass attenuation coefficient, photon mass energy absorption coefficient and total stopping power of electrons. To evaluate the water equivalence of such materials for fast neutron attenuation 19 different types of gel dosimeters and phantom materials were considered. Macroscopic removal cross-sections for fast neutrons (ΣR cm-1) have been calculated for a range of ferrous-sulphate and polymeric gel dosimeters using Nxcom Program. The study showed that the value of ΣR/ρ (cm2.g-1) for all polymer gels were in close agreement (1.5- 2.8%) with that of water. As such, the slight differences in ΣR/ρ between water and gels are small and may be considered negligible. Also, the removal cross-section of the studied phantom materials were very close (~ ±1.5%) to that of water except bone (cortical) which had about 38% variation. Finally, the variation of removal cross-section with hydrogen content was studied.

Keywords: cross-section, neutron, photon, coefficient, mathematics

Procedia PDF Downloads 357
15827 Development of Time Series Forecasting Model for Dengue Cases in Nakhon Si Thammarat, Southern Thailand

Authors: Manit Pollar

Abstract:

Identifying the dengue epidemic periods early would be helpful to take necessary actions to prevent the dengue outbreaks. Providing an accurate prediction on dengue epidemic seasons will allow sufficient time to take the necessary decisions and actions to safeguard the situation for local authorities. This study aimed to develop a forecasting model on number of dengue incidences in Nakhon Si Thammarat Province, Southern Thailand using time series analysis. We develop Seasonal Autoregressive Moving Average (SARIMA) models on the monthly data collected between 2003-2011 and validated the models using data collected between January-September 2012. The result of this study revealed that the SARIMA(1,1,0)(1,2,1)12 model closely described the trends and seasons of dengue incidence and confirmed the existence of dengue fever cases in Nakhon Si Thammarat for the years between 2003-2011. The study showed that the one-step approach for predicting dengue incidences provided significantly more accurate predictions than the twelve-step approach. The model, even if based purely on statistical data analysis, can provide a useful basis for allocation of resources for disease prevention.

Keywords: SARIMA, time series model, dengue cases, Thailand

Procedia PDF Downloads 342
15826 Structural Analysis and Detail Design of APV Module Structure Using Topology Optimization Design

Authors: Hyun Kyu Cho, Jun Soo Kim, Young Hoon Lee, Sang Hoon Kang, Young Chul Park

Abstract:

In the study, structure for one of offshore drilling system APV(Air Pressure Vessle) modules was designed by using topology optimum design and performed structural safety evaluation according to DNV rules. 3D model created base on design area and non-design area separated by using topology optimization for the environmental loads. This model separated 17 types for wind loads and dynamic loads and performed structural analysis evaluation for each model. As a result, the maximum stress occurred 181.25MPa.

Keywords: APV, topology optimum design, DNV, structural analysis, stress

Procedia PDF Downloads 408
15825 Developing Integrated Model for Building Design and Evacuation Planning

Authors: Hao-Hsi Tseng, Hsin-Yun Lee

Abstract:

In the process of building design, the designers have to complete the spatial design and consider the evacuation performance at the same time. It is usually difficult to combine the two planning processes and it results in the gap between spatial design and evacuation performance. Then the designers cannot complete an integrated optimal design solution. In addition, the evacuation routing models proposed by previous researchers is different from the practical evacuation decisions in the real field. On the other hand, more and more building design projects are executed by Building Information Modeling (BIM) in which the design content is formed by the object-oriented framework. Thus, the integration of BIM and evacuation simulation can make a significant contribution for designers. Therefore, this research plan will establish a model that integrates spatial design and evacuation planning. The proposed model will provide the support for the spatial design modifications and optimize the evacuation planning. The designers can complete the integrated design solution in BIM. Besides, this research plan improves the evacuation routing method to make the simulation results more practical. The proposed model will be applied in a building design project for evaluation and validation when it will provide the near-optimal design suggestion. By applying the proposed model, the integration and efficiency of the design process are improved and the evacuation plan is more useful. The quality of building spatial design will be better.

Keywords: building information modeling, evacuation, design, floor plan

Procedia PDF Downloads 440
15824 An Optimization Model for Waste Management in Demolition Works

Authors: Eva Queheille, Franck Taillandier, Nadia Saiyouri

Abstract:

Waste management has become a major issue in demolition works, because of its environmental impact (energy consumption, resource consumption, pollution…). However, improving waste management requires to take also into account the overall demolition process and to consider demolition main objectives (e.g. cost, delay). Establishing a strategy with these conflicting objectives (economic and environment) remains complex. In order to provide a decision-support for demolition companies, a multi-objective optimization model was developed. In this model, a demolition strategy is computed from a set of 80 decision variables (worker team composition, machines, treatment for each type of waste, choice of treatment platform…), which impacts the demolition objectives. The model has experimented on a real-case study (demolition of several buildings in France). To process the optimization, different optimization algorithms (NSGA2, MOPSO, DBEA…) were tested. Results allow the engineer in charge of this case, to build a sustainable demolition strategy without affecting cost or delay.

Keywords: deconstruction, life cycle assessment, multi-objective optimization, waste management

Procedia PDF Downloads 139
15823 Numerical Investigations on Group Piles’ Lateral Bearing Capacity Considering Interaction of Soil and Structure

Authors: Mahdi Sadeghian, Mahmoud Hassanlourad, Alireza Ardakani, Reza Dinarvand

Abstract:

In this research, the behavior of monopiles, under lateral loads, was investigated with vertical and oblique piles by Finite Element Method. In engineering practice when soil-pile interaction comes to the picture some simplifications are applied to reduce the design time. As a simplified replacement of soil and pile interaction analysis, pile could be replaced by a column. The height of the column would be equal to the free length of the pile plus a portion of the embedded length of it. One of the important factors studied in this study was that columns with an equivalent length (free length plus a part of buried depth) could be used instead of soil and pile modeling. The results of the analysis show that the more internal friction angle of the soil increases, the more the bearing capacity of the soil is achieved. This additional length is 6 to 11 times of the pile diameter in dense soil although in loose sandy soil this range might increase.

Keywords: Depth of fixity, Lateral bearing capacity, Oblique pile, Pile group, Soil-structure interaction

Procedia PDF Downloads 209
15822 Application of Public Access Two-Dimensional Hydrodynamic and Distributed Hydrological Models for Flood Forecasting in Ungauged Basins

Authors: Ahmad Shayeq Azizi, Yuji Toda

Abstract:

In Afghanistan, floods are the most frequent and recurrent events among other natural disasters. On the other hand, lack of monitoring data is a severe problem, which increases the difficulty of making the appropriate flood countermeasures of flood forecasting. This study is carried out to simulate the flood inundation in Harirud River Basin by application of distributed hydrological model, Integrated Flood Analysis System (IFAS) and 2D hydrodynamic model, International River Interface Cooperative (iRIC) based on satellite rainfall combined with historical peak discharge and global accessed data. The results of the simulation can predict the inundation area, depth and velocity, and the hardware countermeasures such as the impact of levee installation can be discussed by using the present method. The methodology proposed in this study is suitable for the area where hydrological and geographical data including river survey data are poorly observed.

Keywords: distributed hydrological model, flood inundation, hydrodynamic model, ungauged basins

Procedia PDF Downloads 150
15821 Numerical Modeling of Flow in USBR II Stilling Basin with End Adverse Slope

Authors: Hamidreza Babaali, Alireza Mojtahedi, Nasim Soori, Saba Soori

Abstract:

Hydraulic jump is one of the effective ways of energy dissipation in stilling basins that the ‎energy is highly dissipated by jumping. Adverse slope surface at the end stilling basin is ‎caused to increase energy dissipation and stability of the hydraulic jump. In this study, the adverse slope ‎has been added to end of United States Bureau of Reclamation (USBR) II stilling basin in hydraulic model of Nazloochay dam with scale 1:40, and flow simulated into stilling basin using Flow-3D ‎software. The numerical model is verified by experimental data of water depth in ‎stilling basin. Then, the parameters of water level profile, Froude Number, pressure, air ‎entrainment and turbulent dissipation investigated for discharging 300 m3/s using K-Ɛ and Re-Normalization Group (RNG) turbulence ‎models. The results showed a good agreement between numerical and experimental model‎ as ‎numerical model can be used to optimize of stilling basins.‎

Keywords: experimental and numerical modelling, end adverse slope, flow ‎parameters, USBR II stilling basin

Procedia PDF Downloads 160
15820 A Novel Machining Method and Tool-Path Generation for Bent Mandrel

Authors: Hong Lu, Yongquan Zhang, Wei Fan, Xiangang Su

Abstract:

Bent mandrel has been widely used as precise mould in automobile industry, shipping industry and aviation industry. To improve the versatility and efficiency of turning method of bent mandrel with fixed rotational center, an instantaneous machining model based on cutting parameters and machine dimension is prospered in this paper. The spiral-like tool path generation approach in non-axisymmetric turning process of bent mandrel is developed as well to deal with the error of part-to-part repeatability in existed turning model. The actual cutter-location points are calculated by cutter-contact points, which are obtained from the approach of spiral sweep process using equal-arc-length segment principle in polar coordinate system. The tool offset is set to avoid the interference between tool and work piece is also considered in the machining model. Depend on the spindle rotational angle, synchronization control of X-axis, Z-axis and C-axis is adopted to generate the tool-path of the turning process. The simulation method is developed to generate NC program according to the presented model, which includes calculation of cutter-location points and generation of tool-path of cutting process. With the approach of a bent mandrel taken as an example, the maximum offset of center axis is 4mm in the 3D space. Experiment results verify that the machining model and turning method are appropriate for the characteristics of bent mandrel.

Keywords: bent mandrel, instantaneous machining model, simulation method, tool-path generation

Procedia PDF Downloads 323
15819 Assessing Effects of an Intervention on Bottle-Weaning and Reducing Daily Milk Intake from Bottles in Toddlers Using Two-Part Random Effects Models

Authors: Yungtai Lo

Abstract:

Two-part random effects models have been used to fit semi-continuous longitudinal data where the response variable has a point mass at 0 and a continuous right-skewed distribution for positive values. We review methods proposed in the literature for analyzing data with excess zeros. A two-part logit-log-normal random effects model, a two-part logit-truncated normal random effects model, a two-part logit-gamma random effects model, and a two-part logit-skew normal random effects model were used to examine effects of a bottle-weaning intervention on reducing bottle use and daily milk intake from bottles in toddlers aged 11 to 13 months in a randomized controlled trial. We show in all four two-part models that the intervention promoted bottle-weaning and reduced daily milk intake from bottles in toddlers drinking from a bottle. We also show that there are no differences in model fit using either the logit link function or the probit link function for modeling the probability of bottle-weaning in all four models. Furthermore, prediction accuracy of the logit or probit link function is not sensitive to the distribution assumption on daily milk intake from bottles in toddlers not off bottles.

Keywords: two-part model, semi-continuous variable, truncated normal, gamma regression, skew normal, Pearson residual, receiver operating characteristic curve

Procedia PDF Downloads 335
15818 DenseNet and Autoencoder Architecture for COVID-19 Chest X-Ray Image Classification and Improved U-Net Lung X-Ray Segmentation

Authors: Jonathan Gong

Abstract:

Purpose AI-driven solutions are at the forefront of many pathology and medical imaging methods. Using algorithms designed to better the experience of medical professionals within their respective fields, the efficiency and accuracy of diagnosis can improve. In particular, X-rays are a fast and relatively inexpensive test that can diagnose diseases. In recent years, X-rays have not been widely used to detect and diagnose COVID-19. The under use of Xrays is mainly due to the low diagnostic accuracy and confounding with pneumonia, another respiratory disease. However, research in this field has expressed a possibility that artificial neural networks can successfully diagnose COVID-19 with high accuracy. Models and Data The dataset used is the COVID-19 Radiography Database. This dataset includes images and masks of chest X-rays under the labels of COVID-19, normal, and pneumonia. The classification model developed uses an autoencoder and a pre-trained convolutional neural network (DenseNet201) to provide transfer learning to the model. The model then uses a deep neural network to finalize the feature extraction and predict the diagnosis for the input image. This model was trained on 4035 images and validated on 807 separate images from the ones used for training. The images used to train the classification model include an important feature: the pictures are cropped beforehand to eliminate distractions when training the model. The image segmentation model uses an improved U-Net architecture. This model is used to extract the lung mask from the chest X-ray image. The model is trained on 8577 images and validated on a validation split of 20%. These models are calculated using the external dataset for validation. The models’ accuracy, precision, recall, f1-score, IOU, and loss are calculated. Results The classification model achieved an accuracy of 97.65% and a loss of 0.1234 when differentiating COVID19-infected, pneumonia-infected, and normal lung X-rays. The segmentation model achieved an accuracy of 97.31% and an IOU of 0.928. Conclusion The models proposed can detect COVID-19, pneumonia, and normal lungs with high accuracy and derive the lung mask from a chest X-ray with similarly high accuracy. The hope is for these models to elevate the experience of medical professionals and provide insight into the future of the methods used.

Keywords: artificial intelligence, convolutional neural networks, deep learning, image processing, machine learning

Procedia PDF Downloads 114