Search results for: modeling and optimization
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 6826

Search results for: modeling and optimization

3376 Using Discrete Event Simulation Approach to Reduce Waiting Times in Computed Tomography Radiology Department

Authors: Mwafak Shakoor

Abstract:

The purpose of this study was to reduce patient waiting times, improve system throughput and improve resources utilization in radiology department. A discrete event simulation model was developed using Arena simulation software to investigate different alternatives to improve the overall system delivery based on adding resource scenarios due to the linkage between patient waiting times and resource availability. The study revealed that there is no addition investment need to procure additional scanner but hospital management deploy managerial tactics to enhance machine utilization and reduce the long waiting time in the department.

Keywords: discrete event simulation, radiology department, arena, waiting time, healthcare modeling, computed tomography

Procedia PDF Downloads 592
3375 Statistical Comparison of Ensemble Based Storm Surge Forecasting Models

Authors: Amin Salighehdar, Ziwen Ye, Mingzhe Liu, Ionut Florescu, Alan F. Blumberg

Abstract:

Storm surge is an abnormal water level caused by a storm. Accurate prediction of a storm surge is a challenging problem. Researchers developed various ensemble modeling techniques to combine several individual forecasts to produce an overall presumably better forecast. There exist some simple ensemble modeling techniques in literature. For instance, Model Output Statistics (MOS), and running mean-bias removal are widely used techniques in storm surge prediction domain. However, these methods have some drawbacks. For instance, MOS is based on multiple linear regression and it needs a long period of training data. To overcome the shortcomings of these simple methods, researchers propose some advanced methods. For instance, ENSURF (Ensemble SURge Forecast) is a multi-model application for sea level forecast. This application creates a better forecast of sea level using a combination of several instances of the Bayesian Model Averaging (BMA). An ensemble dressing method is based on identifying best member forecast and using it for prediction. Our contribution in this paper can be summarized as follows. First, we investigate whether the ensemble models perform better than any single forecast. Therefore, we need to identify the single best forecast. We present a methodology based on a simple Bayesian selection method to select the best single forecast. Second, we present several new and simple ways to construct ensemble models. We use correlation and standard deviation as weights in combining different forecast models. Third, we use these ensembles and compare with several existing models in literature to forecast storm surge level. We then investigate whether developing a complex ensemble model is indeed needed. To achieve this goal, we use a simple average (one of the simplest and widely used ensemble model) as benchmark. Predicting the peak level of Surge during a storm as well as the precise time at which this peak level takes place is crucial, thus we develop a statistical platform to compare the performance of various ensemble methods. This statistical analysis is based on root mean square error of the ensemble forecast during the testing period and on the magnitude and timing of the forecasted peak surge compared to the actual time and peak. In this work, we analyze four hurricanes: hurricanes Irene and Lee in 2011, hurricane Sandy in 2012, and hurricane Joaquin in 2015. Since hurricane Irene developed at the end of August 2011 and hurricane Lee started just after Irene at the beginning of September 2011, in this study we consider them as a single contiguous hurricane event. The data set used for this study is generated by the New York Harbor Observing and Prediction System (NYHOPS). We find that even the simplest possible way of creating an ensemble produces results superior to any single forecast. We also show that the ensemble models we propose generally have better performance compared to the simple average ensemble technique.

Keywords: Bayesian learning, ensemble model, statistical analysis, storm surge prediction

Procedia PDF Downloads 309
3374 Behaviour of Reinforced Concrete Infilled Frames under Seismic Loads

Authors: W. Badla

Abstract:

A significant portion of the buildings constructed in Algeria is structural frames with infill panels which are usually considered as non structural components and are neglected in the analysis. However, these masonry panels tend to influence the structural response. Thus, these structures can be regarded as seismic risk buildings, although in the Algerian seismic code there is little guidance on the seismic evaluation of infilled frame buildings. In this study, three RC frames with 2, 4, and 8 story and subjected to three recorded Algerian accelerograms are studied. The diagonal strut approach is adopted for modeling the infill panels and a fiber model is used to model RC members. This paper reports on the seismic evaluation of RC frames with brick infill panels. The results obtained show that the masonry panels enhance the load lateral capacity of the buildings and the infill panel configuration influences the response of the structures.

Keywords: seismic design, RC frames, infill panels, non linear dynamic analysis

Procedia PDF Downloads 546
3373 A Meso Macro Model Prediction of Laminated Composite Damage Elastic Behaviour

Authors: A. Hocine, A. Ghouaoula, S. M. Medjdoub, M. Cherifi

Abstract:

The present paper proposed a meso–macro model describing the mechanical behaviour composite laminates of staking sequence [+θ/-θ]s under tensil loading. The behaviour of a layer is ex-pressed through elasticity coupled to damage. The elastic strain is due to the elasticity of the layer and can be modeled by using the classical laminate theory, and the laminate is considered as an orthotropic material. This means that no coupling effect between strain and curvature is considered. In the present work, the damage is associated to cracking of the matrix and parallel to the fibers and it being taken into account by the changes in the stiffness of the layers. The anisotropic damage is completely described by a single scalar variable and its evolution law is specified from the principle of maximum dissipation. The stress/strain relationship is investigated in plane stress loading.

Keywords: damage, behavior modeling, meso-macro model, composite laminate, membrane loading

Procedia PDF Downloads 476
3372 Aliasing Free and Additive Error in Spectra for Alpha Stable Signals

Authors: R. Sabre

Abstract:

This work focuses on the symmetric alpha stable process with continuous time frequently used in modeling the signal with indefinitely growing variance, often observed with an unknown additive error. The objective of this paper is to estimate this error from discrete observations of the signal. For that, we propose a method based on the smoothing of the observations via Jackson polynomial kernel and taking into account the width of the interval where the spectral density is non-zero. This technique allows avoiding the “Aliasing phenomenon” encountered when the estimation is made from the discrete observations of a process with continuous time. We have studied the convergence rate of the estimator and have shown that the convergence rate improves in the case where the spectral density is zero at the origin. Thus, we set up an estimator of the additive error that can be subtracted for approaching the original signal without error.

Keywords: spectral density, stable processes, aliasing, non parametric

Procedia PDF Downloads 130
3371 A Comparative Study of Approaches in User-Centred Health Information Retrieval

Authors: Harsh Thakkar, Ganesh Iyer

Abstract:

In this paper, we survey various user-centered or context-based biomedical health information retrieval systems. We present and discuss the performance of systems submitted in CLEF eHealth 2014 Task 3 for this purpose. We classify and focus on comparing the two most prevalent retrieval models in biomedical information retrieval namely: Language Model (LM) and Vector Space Model (VSM). We also report on the effectiveness of using external medical resources and ontologies like MeSH, Metamap, UMLS, etc. We observed that the LM based retrieval systems outperform VSM based systems on various fronts. From the results we conclude that the state-of-art system scores for MAP was 0.4146, P@10 was 0.7560 and NDCG@10 was 0.7445, respectively. All of these score were reported by systems built on language modeling approaches.

Keywords: clinical document retrieval, concept-based information retrieval, query expansion, language models, vector space models

Procedia PDF Downloads 320
3370 Stock Prediction and Portfolio Optimization Thesis

Authors: Deniz Peksen

Abstract:

This thesis aims to predict trend movement of closing price of stock and to maximize portfolio by utilizing the predictions. In this context, the study aims to define a stock portfolio strategy from models created by using Logistic Regression, Gradient Boosting and Random Forest. Recently, predicting the trend of stock price has gained a significance role in making buy and sell decisions and generating returns with investment strategies formed by machine learning basis decisions. There are plenty of studies in the literature on the prediction of stock prices in capital markets using machine learning methods but most of them focus on closing prices instead of the direction of price trend. Our study differs from literature in terms of target definition. Ours is a classification problem which is focusing on the market trend in next 20 trading days. To predict trend direction, fourteen years of data were used for training. Following three years were used for validation. Finally, last three years were used for testing. Training data are between 2002-06-18 and 2016-12-30 Validation data are between 2017-01-02 and 2019-12-31 Testing data are between 2020-01-02 and 2022-03-17 We determine Hold Stock Portfolio, Best Stock Portfolio and USD-TRY Exchange rate as benchmarks which we should outperform. We compared our machine learning basis portfolio return on test data with return of Hold Stock Portfolio, Best Stock Portfolio and USD-TRY Exchange rate. We assessed our model performance with the help of roc-auc score and lift charts. We use logistic regression, Gradient Boosting and Random Forest with grid search approach to fine-tune hyper-parameters. As a result of the empirical study, the existence of uptrend and downtrend of five stocks could not be predicted by the models. When we use these predictions to define buy and sell decisions in order to generate model-based-portfolio, model-based-portfolio fails in test dataset. It was found that Model-based buy and sell decisions generated a stock portfolio strategy whose returns can not outperform non-model portfolio strategies on test dataset. We found that any effort for predicting the trend which is formulated on stock price is a challenge. We found same results as Random Walk Theory claims which says that stock price or price changes are unpredictable. Our model iterations failed on test dataset. Although, we built up several good models on validation dataset, we failed on test dataset. We implemented Random Forest, Gradient Boosting and Logistic Regression. We discovered that complex models did not provide advantage or additional performance while comparing them with Logistic Regression. More complexity did not lead us to reach better performance. Using a complex model is not an answer to figure out the stock-related prediction problem. Our approach was to predict the trend instead of the price. This approach converted our problem into classification. However, this label approach does not lead us to solve the stock prediction problem and deny or refute the accuracy of the Random Walk Theory for the stock price.

Keywords: stock prediction, portfolio optimization, data science, machine learning

Procedia PDF Downloads 80
3369 Selecting the Best Software Product Using Analytic Hierarchy Process and Fuzzy-Analytic Hierarchy Process Modules

Authors: Anas Hourani, Batool Ahmad

Abstract:

Software applications play an important role inside any institute. They are employed to manage all processes and store entities-related data in the computer. Therefore, choosing the right software product that meets institute requirements is not an easy decision in view of considering multiple criteria, different points of views, and many standards. As a case study, Mutah University, located in Jordan, is in essential need of customized software, and several companies presented their software products which are very similar in quality. In this regard, an analytic hierarchy process (AHP) and a fuzzy analytic hierarchy process (Fuzzy-AHP) models are proposed in this research to identify the most suitable and best-fit software product that meets the institute requirements. The results indicate that both modules are able to help the decision-makers to make a decision, especially in complex decision problems.

Keywords: analytic hierarchy process, decision modeling, fuzzy analytic hierarchy process, software product

Procedia PDF Downloads 392
3368 Design of Non-uniform Circular Antenna Arrays Using Firefly Algorithm for Side Lobe Level Reduction

Authors: Gopi Ram, Durbadal Mandal, Rajib Kar, Sakti Prasad Ghoshal

Abstract:

A design problem of non-uniform circular antenna arrays for maximum reduction of both the side lobe level (SLL) and first null beam width (FNBW) is dealt with. This problem is modeled as a simple optimization problem. The method of Firefly algorithm (FFA) is used to determine an optimal set of current excitation weights and antenna inter-element separations that provide radiation pattern with maximum SLL reduction and much improvement on FNBW as well. Circular array antenna laid on x-y plane is assumed. FFA is applied on circular arrays of 8-, 10-, and 12- elements. Various simulation results are presented and hence performances of side lobe and FNBW are analyzed. Experimental results show considerable reductions of both the SLL and FNBW with respect to those of the uniform case and some standard algorithms GA, PSO, and SA applied to the same problem.

Keywords: circular arrays, first null beam width, side lobe level, FFA

Procedia PDF Downloads 259
3367 Cryptocurrencies: Business Students’ Awareness and Universities’ Adoption Readiness and Compatibility of Use Considering the Mediation of Attitudes

Authors: Eric S. Parilla, Marc Edward Abadilla

Abstract:

The study aims to determine the effect of awareness of business students towards cryptocurrencies and the readiness of universities and colleges to accept cryptocurrencies as a medium of exchange, considering the mediation of business students’ attitudes. The research used partial least squares structural equation modeling (PLS-SEM) and deployed a questionnaire attuned to the awareness and attitudes of business students towards cryptocurrencies and readiness and compatibility of use in universities and colleges in Ilocos Norte. The output of the investigation revealed that awareness of business students is not correlated to the readiness of universities and colleges, which means that even though students understand cryptocurrencies, it is not an assurance that universities and colleges are ready to accept them as the medium of exchange. The study proposes that training and seminars for business students and professionals should be conducted to expand understanding and acceptance of cryptocurrencies.

Keywords: cryptocurrencies, awareness, readiness, attitudes

Procedia PDF Downloads 225
3366 Simulation for the Magnetized Plasma Compression Study

Authors: Victor V. Kuzenov, Sergei V. Ryzhkov

Abstract:

Ongoing experimental and theoretical studies on magneto-inertial confinement fusion (Angara, C-2, CJS-100, General Fusion, MagLIF, MAGPIE, MC-1, YG-1, Omega) and new constructing facilities (Baikal, C-2W, Z300 and Z800) require adequate modeling and description of the physical processes occurring in high-temperature dense plasma in a strong magnetic field. This paper presents a mathematical model, numerical method, and results of the computer analysis of the compression process and the energy transfer in the target plasma, used in magneto-inertial fusion (MIF). The computer simulation of the compression process of the magnetized target by the high-power laser pulse and the high-speed plasma jets is presented. The characteristic patterns of the two methods of the target compression are being analysed.

Keywords: magnetized target, magneto-inertial fusion, mathematical model, plasma and laser beams

Procedia PDF Downloads 296
3365 Optimized and Secured Digital Watermarking Using Fuzzy Entropy, Bezier Curve and Visual Cryptography

Authors: R. Rama Kishore, Sunesh

Abstract:

Recent development in the usage of internet for different purposes creates a great threat for the copyright protection of the digital images. Digital watermarking can be used to address the problem. This paper presents detailed review of the different watermarking techniques, latest trends in the field of secured, robust and imperceptible watermarking. It also discusses the different optimization techniques used in the field of watermarking in order to improve the robustness and imperceptibility of the method. Different measures are discussed to evaluate the performance of the watermarking algorithm. At the end, this paper proposes a watermarking algorithm using (2, 2) share visual cryptography and Bezier curve based algorithm to improve the security of the watermark. The proposed method uses fractional transformation to improve the robustness of the copyright protection of the method. The algorithm is optimized using fuzzy entropy for better results.

Keywords: digital watermarking, fractional transform, visual cryptography, Bezier curve, fuzzy entropy

Procedia PDF Downloads 366
3364 Force Feedback Enabled Syringe for Aspiration and Biopsy

Authors: Pelin Su Firat, Sohyung Cho

Abstract:

Biopsy or aspiration procedures are known to be complicated as they involve the penetration of a needle through human tissues, including vital organs. This research presents the design of a force sensor-guided device to be used with syringes and needles for aspiration and biopsy. The development of the device was aimed to help accomplish accurate needle placement and increase the performance of the surgeon in navigating the tool and tracking the target. Specifically, a prototype for a force-sensor embedded syringe has been created using 3D (3-Dimensional) modeling and printing techniques in which two different force sensors were used to provide significant force feedback to users during the operations when needles pernitrate different tissues. From the extensive tests using synthetic tissues, it is shown that the proposed syringe design has accomplished the desired accuracy, efficiency, repeatability, and effectiveness. Further development is desirable through usability tests.

Keywords: biopsy, syringe, force sensors, haptic feedback

Procedia PDF Downloads 69
3363 Self-Efficacy of Preschool Teachers and Their Perception of Excellent Preschools

Authors: Yael Fisher

Abstract:

Little is known about perceived self-efficacy of public preschool teachers, their perception of preschool excellence, or the relations between the two. There were three purposes for this research: defining the professional self-efficacy of preschool teachers (PTSE); defining preschool teachers' perception of preschool excellence (PTPPE); and investigating the relationship between the two. Scales for PTSE and PTPPE were developed especially for this study. Public preschool teachers (N = 202) participated during the 2013 school year. Structural Equation Modeling was performed to test the fit between the research model and the obtained data. PTPSE scale (α = 0.91) was comprised of three subscales: pedagogy (α=0.84), organization (α = 0.85) and staff (α = 0.72). The PTPPE scale (α = 0.92) is also composed of three subscales: organization and pedagogy (α = 0.88), staff (α = 0.84) and parents (α = 0.83). The goodness of fit measures were RMSEA = 0.045, CFI = 0.97, NFI = 0.89, df = 173, χ²=242.94, p= .000, showing GFI = 1.4 (< 3) as a good fit. Understanding self-efficacy of preschool teachers, preschool could and should lead to better professional development (in-service training) of preschool teachers.

Keywords: self-efficacy, public pre schools, preschool excellence, SEM

Procedia PDF Downloads 130
3362 Mean Velocity Modeling of Open-Channel Flow with Submerged Vegetation

Authors: Mabrouka Morri, Amel Soualmia, Philippe Belleudy

Abstract:

Vegetation affects the mean and turbulent flow structure. It may increase flood risks and sediment transport. Therefore, it is important to develop analytical approaches for the bed shear stress on vegetated bed, to predict resistance caused by vegetation. In the recent years, experimental and numerical models have both been developed to model the effects of submerged vegetation on open-channel flow. In this paper, different analytic models are compared and tested using the criteria of deviation, to explore their capacity for predicting the mean velocity and select the suitable one that will be applied in real case of rivers. The comparison between the measured data in vegetated flume and simulated mean velocities indicated, a good performance, in the case of rigid vegetation, whereas, Huthoff model shows the best agreement with a high coefficient of determination (R2=80%) and the smallest error in the prediction of the average velocities.

Keywords: analytic models, comparison, mean velocity, vegetation

Procedia PDF Downloads 276
3361 Performance Prediction Methodology of Slow Aging Assets

Authors: M. Ben Slimene, M.-S. Ouali

Abstract:

Asset management of urban infrastructures faces a multitude of challenges that need to be overcome to obtain a reliable measurement of performances. Predicting the performance of slowly aging systems is one of those challenges, which helps the asset manager to investigate specific failure modes and to undertake the appropriate maintenance and rehabilitation interventions to avoid catastrophic failures as well as to optimize the maintenance costs. This article presents a methodology for modeling the deterioration of slowly degrading assets based on an operating history. It consists of extracting degradation profiles by grouping together assets that exhibit similar degradation sequences using an unsupervised classification technique derived from artificial intelligence. The obtained clusters are used to build the performance prediction models. This methodology is applied to a sample of a stormwater drainage culvert dataset.

Keywords: artificial Intelligence, clustering, culvert, regression model, slow degradation

Procedia PDF Downloads 112
3360 Recovery of Fried Soybean Oil Using Bentonite as an Adsorbent: Optimization, Isotherm and Kinetics Studies

Authors: Prakash Kumar Nayak, Avinash Kumar, Uma Dash, Kalpana Rayaguru

Abstract:

Soybean oil is one of the most widely consumed cooking oils, worldwide. Deep-fat frying of foods at higher temperatures adds unique flavour, golden brown colour and crispy texture to foods. But it brings in various changes like hydrolysis, oxidation, hydrogenation and thermal alteration to oil. The presence of Peroxide value (PV) is one of the most important factors affecting the quality of the deep-fat fried oil. Using bentonite as an adsorbent, the PV can be reduced, thereby improving the quality of the soybean oil. In this study, operating parameters like heating time of oil (10, 15, 20, 25 & 30 h), contact time ( 5, 10, 15, 20, 25 h) and concentration of adsorbent (0.25, 0.5, 0.75, 1.0 and 1.25 g/ 100 ml of oil) have been optimized by response surface methodology (RSM) considering percentage reduction of PV as a response. Adsorption data were analysed by fitting with Langmuir and Freundlich isotherm model. The results show that the Langmuir model shows the best fit compared to the Freundlich model. The adsorption process was also found to follow a pseudo-second-order kinetic model.

Keywords: bentonite, Langmuir isotherm, peroxide value, RSM, soybean oil

Procedia PDF Downloads 375
3359 Optimization of Machining Parameters in AlSi/10%AlN Metal Matrix Composite Material by TiN Coating Insert

Authors: Nurul Na'imy Wan, Mohamad Sazali Said, Jaharah Ab. Ghani, Rusli Othman

Abstract:

This paper presents the surface roughness of the aluminium silicon alloy (AlSi) matrix composite which has been reinforced with aluminium nitride (AlN). Experiments were conducted at various cutting speeds, feed rates, and depths of cut, according to a standard orthogonal array L27 of Taguchi method using TiN coating tool of insert. The signal-to-noise (S/N) ratio and analysis of variance are applied to study the characteristic performance of cutting speeds, feed rates and depths of cut in measuring the surface roughness during the milling operation. The surface roughness was observed using Mitutoyo Formtracer CS-500 and analyzed using the Taguchi method. From the Taguchi analysis, it was found that cutting speed of 230 m/min, feed rate of 0.4 mm/tooth, depth of cut of 0.3 mm were the optimum machining parameters using TiN coating insert.

Keywords: AlSi/AlN metal matrix composite (MMC), surface roughness, Taguchi method, machining parameters

Procedia PDF Downloads 433
3358 Design and Implementation of an AI-Enabled Task Assistance and Management System

Authors: Arun Prasad Jaganathan

Abstract:

In today's dynamic industrial world, traditional task allocation methods often fall short in adapting to evolving operational conditions. This paper introduces an AI-enabled task assistance and management system designed to overcome the limitations of conventional approaches. By using artificial intelligence (AI) and machine learning (ML), the system intelligently interprets user instructions, analyzes tasks, and allocates resources based on real-time data and environmental factors. Additionally, geolocation tracking enables proactive identification of potential delays, ensuring timely interventions. With its transparent reporting mechanisms, the system provides stakeholders with clear insights into task progress, fostering accountability and informed decision-making. The paper presents a comprehensive overview of the system architecture, algorithm, and implementation, highlighting its potential to revolutionize task management across diverse industries.

Keywords: artificial intelligence, machine learning, task allocation, operational efficiency, resource optimization

Procedia PDF Downloads 59
3357 Marketing in the Age of Artificial Intelligence: Implications for Consumption Patterns of Halal Food

Authors: Djermani Farouk, Sri Rahayu Hijrah Hati, Fenitra Maminirin, Permata Wulandari

Abstract:

This study investigates the implications of Artificial Intelligence Marketing (AIM) marketing mix (PRD) Product, (PRC) Price, (PRM), Promotion and (PLC) Place on consumption patterns of halal food (CPHF). A quantitative approach was adopted in this study and responses were obtained from 350 Indonesian consumers. Using Partial Least Squares-Structural Equation Modeling (PLS-SEM), the results show that there is a direct support of marketing mix (PRD, PRC, PLC) to AIM and CPHF, while PRM does not play a significant role in CPHF. In addition, the findings reveal that AIM mediates significantly the relationship between PLC, PRC and PRM and CPHF, while AIM indicates no mediation between PRD and CPHF. Indonesian consumer’s exhibit serious concerns with consumption patterns of halal food. it is recommended that managers focus their attention on marketing strategies to predict consumer behavior in terms of consumption patterns of halal food through the integration of AIM.

Keywords: marketing mix, consumption patterns, artificial intelligence marketing, Halal food

Procedia PDF Downloads 34
3356 Cardiovascular Modeling Software Tools in Medicine

Authors: J. Fernandez, R. Fernandez de Canete, J. Perea-Paizal, J. C. Ramos-Diaz

Abstract:

The high prevalence of cardiovascular diseases has provoked a raising interest in the development of mathematical models in order to evaluate the cardiovascular function both under physiological and pathological conditions. In this paper, a physical model of the cardiovascular system with intrinsic regulation is presented and implemented by using the object-oriented Modelica simulation software tools.  For this task, a multi-compartmental system previously validated with physiological data has been built, based on the interconnection of cardiovascular elements such as resistances, capacitances and pumping among others, by following an electrohydraulic analogy. The results obtained under both physiological and pathological scenarios provide an easy interpretative key to analyze the hemodynamic behavior of the patient. The described approach represents a valuable tool in the teaching of physiology for graduate medical and nursing students among others.

Keywords: cardiovascular system, MODELICA simulation software, physical modelling, teaching tool

Procedia PDF Downloads 300
3355 Design and Optimization of a Customized External Fixation Device for Lower Limb Injuries

Authors: Mohammed S. Alqahtani, Paulo J. Bartolo

Abstract:

External fixation is a common technique for the treatment and stabilization of bone fractures. Different designs have been proposed by companies and research groups, but all of them present limitations such as high weight, not comfortable to use, and not customized to individual patients. This paper proposes a lightweight customized external fixator, overcoming some of these limitations. External fixators are designed using a set of techniques such as medical imaging, CAD modelling, finite element analysis, and full factorial design of experiments. Key design parameters are discussed, and the optimal set of parameters is used to design the final external fixator. Numerical simulations are used to validate design concepts. Results present an optimal external fixation design with weight reduction of 13% without compromising its stiffness and structural integrity. External fixators are also designed to be additively manufactured, allowing to develop a strategy for personalization.

Keywords: computer-aided design modelling, external fixation, finite element analysis, full factorial, personalization

Procedia PDF Downloads 160
3354 Modelling Railway Noise Over Large Areas, Assisted by GIS

Authors: Conrad Weber

Abstract:

The modelling of railway noise over large projects areas can be very time consuming in terms of preparing the noise models and calculation time. An open-source GIS program has been utilised to assist with the modelling of operational noise levels for 675km of railway corridor. A range of GIS algorithms were utilised to break up the noise model area into manageable calculation sizes. GIS was utilised to prepare and filter a range of noise modelling inputs, including building files, land uses and ground terrain. A spreadsheet was utilised to manage the accuracy of key input parameters, including train speeds, train types, curve corrections, bridge corrections and engine notch settings. GIS was utilised to present the final noise modelling results. This paper explains the noise modelling process and how the spreadsheet and GIS were utilised to accurately model this massive project efficiently.

Keywords: noise, modeling, GIS, rail

Procedia PDF Downloads 122
3353 Credit Risk Evaluation Using Genetic Programming

Authors: Ines Gasmi, Salima Smiti, Makram Soui, Khaled Ghedira

Abstract:

Credit risk is considered as one of the important issues for financial institutions. It provokes great losses for banks. To this objective, numerous methods for credit risk evaluation have been proposed. Many evaluation methods are black box models that cannot adequately reveal information hidden in the data. However, several works have focused on building transparent rules-based models. For credit risk assessment, generated rules must be not only highly accurate, but also highly interpretable. In this paper, we aim to build both, an accurate and transparent credit risk evaluation model which proposes a set of classification rules. In fact, we consider the credit risk evaluation as an optimization problem which uses a genetic programming (GP) algorithm, where the goal is to maximize the accuracy of generated rules. We evaluate our proposed approach on the base of German and Australian credit datasets. We compared our finding with some existing works; the result shows that the proposed GP outperforms the other models.

Keywords: credit risk assessment, rule generation, genetic programming, feature selection

Procedia PDF Downloads 353
3352 Analysis of Joint Source Channel LDPC Coding for Correlated Sources Transmission over Noisy Channels

Authors: Marwa Ben Abdessalem, Amin Zribi, Ammar Bouallègue

Abstract:

In this paper, a Joint Source Channel coding scheme based on LDPC codes is investigated. We consider two concatenated LDPC codes, one allows to compress a correlated source and the second to protect it against channel degradations. The original information can be reconstructed at the receiver by a joint decoder, where the source decoder and the channel decoder run in parallel by transferring extrinsic information. We investigate the performance of the JSC LDPC code in terms of Bit-Error Rate (BER) in the case of transmission over an Additive White Gaussian Noise (AWGN) channel, and for different source and channel rate parameters. We emphasize how JSC LDPC presents a performance tradeoff depending on the channel state and on the source correlation. We show that, the JSC LDPC is an efficient solution for a relatively low Signal-to-Noise Ratio (SNR) channel, especially with highly correlated sources. Finally, a source-channel rate optimization has to be applied to guarantee the best JSC LDPC system performance for a given channel.

Keywords: AWGN channel, belief propagation, joint source channel coding, LDPC codes

Procedia PDF Downloads 357
3351 Arsenite Remediation by Green Nano Zero Valent Iron

Authors: Ratthiwa Deewan, Visanu Tanboonchuy

Abstract:

The optimal conditions for green synthesis of zero-valent (G-NZVI) synthesis are investigated in this study using a Box Behnken design. The factors that were used in the study consisted of 3 factors as follows: the iron solution to mango peel extract ratio (1:1-1:3), feeding rate of mango peel extracts (1-5 mL/min), and agitation speed (300-30 rpm). The results showed that the optimization of conditions using the regression model was appropriate. The optimal conditions of the synthesis of G-NZVI for arsenate removal are the iron solution to mango peel extract ratio of 1:1, the feeding rate of mango peel extract at 5 mL/min, and the agitation speed rate of 300 rpm, which was able to arsenate removal of 100%.

Keywords: Box Behnken design, arsenate removal, green nano zero valent iron, arsenic

Procedia PDF Downloads 29
3350 Creative Self-efficacy and Innovation Speed of New Ventures: The Mediating Role of Entrepreneurial Bricolage

Authors: Yi-Wen Chen, Hsueh-Liang Fan

Abstract:

Evidence shows that start-ups success is positively correlated with innovation speed. However, new ventures are seldom able to acquire abundant resources for new product development (NPD), which means that entrepreneurs may depend on personal creativity instead of physical investments to achieve and accelerate speed of first product launch. This study accentuates the role of entrepreneurial bricolage, which defined as making do by applying combinations of the resources at hand to new problems and opportunities, in the relations of creative self-efficacy and innovation speed. This study uses structural equation modeling to test the hypotheses in a sample of 203 start-ups operating various creative markets. Results reveal that creative self-efficacy is positively and directly associated with innovation speed, whereas entrepreneurial bricolage plays a full mediator. These findings offer important theoretical and practical implications.

Keywords: creative self-efficacy, innovation speed, entrepreneurial bricolage, new ventures

Procedia PDF Downloads 530
3349 Ecosystem Modeling along the Western Bay of Bengal

Authors: A. D. Rao, Sachiko Mohanty, R. Gayathri, V. Ranga Rao

Abstract:

Modeling on coupled physical and biogeochemical processes of coastal waters is vital to identify the primary production status under different natural and anthropogenic conditions. About 7, 500 km length of Indian coastline is occupied with number of semi enclosed coastal bodies such as estuaries, inlets, bays, lagoons, and other near shore, offshore shelf waters, etc. This coastline is also rich in wide varieties of ecosystem flora and fauna. Directly/indirectly extensive domestic and industrial sewage enter into these coastal water bodies affecting the ecosystem character and create environment problems such as water quality degradation, hypoxia, anoxia, harmful algal blooms, etc. lead to decline in fishery and other related biological production. The present study is focused on the southeast coast of India, starting from Pulicat to Gulf of Mannar, which is rich in marine diversity such as lagoon, mangrove and coral ecosystem. Three dimensional Massachusetts Institute of Technology general circulation model (MITgcm) along with Darwin biogeochemical module is configured for the western Bay of Bengal (BoB) to study the biogeochemistry over this region. The biogeochemical module resolves the cycling of carbon, phosphorous, nitrogen, silica, iron and oxygen through inorganic, living, dissolved and particulate organic phases. The model domain extends from 4°N-16.5°N and 77°E-86°E with a horizontal resolution of 1 km. The bathymetry is derived from General Bathymetric Chart of the Oceans (GEBCO), which has a resolution of 30 sec. The model is initialized by using the temperature, salinity filed from the World Ocean Atlas (WOA2013) of National Oceanographic Data Centre with a resolution of 0.25°. The model is forced by the surface wind stress from ASCAT and the photosynthetically active radiation from the MODIS-Aqua satellite. Seasonal climatology of nutrients (phosphate, nitrate and silicate) for the southwest BoB region are prepared using available National Institute of Oceanography (NIO) in-situ data sets and compared with the WOA2013 seasonal climatology data. The model simulations with the two different initial conditions viz., WOA2013 and the generated NIO climatology, showed evident changes in the concentration and the evolution of the nutrients in the study region. It is observed that the availability of nutrients is more in NIO data compared to WOA in the model domain. The model simulated primary productivity is compared with the spatially distributed satellite derived chlorophyll data and at various locations with the in-situ data. The seasonal variability of the model simulated primary productivity is also studied.

Keywords: Bay of Bengal, Massachusetts Institute of Technology general circulation model, MITgcm, biogeochemistry, primary productivity

Procedia PDF Downloads 141
3348 The Influence of 3D Printing Course on Middle School Students' Spatial Thinking Ability

Authors: Wang Xingjuan, Qian Dongming

Abstract:

As a common thinking ability, spatial thinking ability plays an increasingly important role in the information age. The key to cultivating students' spatial thinking ability is to cultivate students' ability to process and transform graphics. The 3D printing course enables students to constantly touch the rotation and movement of objects during the modeling process and to understand spatial graphics from different views. To this end, this article combines the classic PSVT: R test to explore the impact of 3D printing courses on the spatial thinking ability of middle school students. The results of the study found that: (1) Through the study of the 3D printing course, the students' spatial ability test scores have been significantly improved, which indirectly reflects the improvement of the spatial thinking ability level. (2) The student's spatial thinking ability test results are influenced by the parent's occupation.

Keywords: 3D printing, middle school students, spatial thinking ability, influence

Procedia PDF Downloads 190
3347 Designing Sustainable and Energy-Efficient Urban Network: A Passive Architectural Approach with Solar Integration and Urban Building Energy Modeling (UBEM) Tools

Authors: A. Maghoul, A. Rostampouryasouri, MR. Maghami

Abstract:

The development of an urban design and power network planning has been gaining momentum in recent years. The integration of renewable energy with urban design has been widely regarded as an increasingly important solution leading to climate change and energy security. Through the use of passive strategies and solar integration with Urban Building Energy Modeling (UBEM) tools, architects and designers can create high-quality designs that meet the needs of clients and stakeholders. To determine the most effective ways of combining renewable energy with urban development, we analyze the relationship between urban form and renewable energy production. The procedure involved in this practice include passive solar gain (in building design and urban design), solar integration, location strategy, and 3D models with a case study conducted in Tehran, Iran. The study emphasizes the importance of spatial and temporal considerations in the development of sector coupling strategies for solar power establishment in arid and semi-arid regions. The substation considered in the research consists of two parallel transformers, 13 lines, and 38 connection points. Each urban load connection point is equipped with 500 kW of solar PV capacity and 1 kWh of battery Energy Storage (BES) to store excess power generated from solar, injecting it into the urban network during peak periods. The simulations and analyses have occurred in EnergyPlus software. Passive solar gain involves maximizing the amount of sunlight that enters a building to reduce the need for artificial lighting and heating. Solar integration involves integrating solar photovoltaic (PV) power into smart grids to reduce emissions and increase energy efficiency. Location strategy is crucial to maximize the utilization of solar PV in an urban distribution feeder. Additionally, 3D models are made in Revit, and they are keys component of decision-making in areas including climate change mitigation, urban planning, and infrastructure. we applied these strategies in this research, and the results show that it is possible to create sustainable and energy-efficient urban environments. Furthermore, demand response programs can be used in conjunction with solar integration to optimize energy usage and reduce the strain on the power grid. This study highlights the influence of ancient Persian architecture on Iran's urban planning system, as well as the potential for reducing pollutants in building construction. Additionally, the paper explores the advances in eco-city planning and development and the emerging practices and strategies for integrating sustainability goals.

Keywords: energy-efficient urban planning, sustainable architecture, solar energy, sustainable urban design

Procedia PDF Downloads 76