Search results for: stochastic approximation gradient
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 1614

Search results for: stochastic approximation gradient

474 Accurate Cortical Reconstruction in Narrow Sulci with Zero-Non-Zero Distance (ZNZD) Vector Field

Authors: Somojit Saha, Rohit K. Chatterjee, Sarit K. Das, Avijit Kar

Abstract:

A new force field is designed for propagation of the parametric contour into deep narrow cortical fold in the application of knowledge based reconstruction of cerebral cortex from MR image of brain. Designing of this force field is highly inspired by the Generalized Gradient Vector Flow (GGVF) model and markedly differs in manipulation of image information in order to determine the direction of propagation of the contour. While GGVF uses edge map as its main driving force, the newly designed force field uses the map of distance between zero valued pixels and their nearest non-zero valued pixel as its main driving force. Hence, it is called Zero-Non-Zero Distance (ZNZD) force field. The objective of this force field is forceful propagation of the contour beyond spurious convergence due to partial volume effect (PVE) in to narrow sulcal fold. Being function of the corresponding non-zero pixel value, the force field has got an inherent property to determine spuriousness of the edge automatically. It is effectively applied along with some morphological processing in the application of cortical reconstruction to breach the hindrance of PVE in narrow sulci where conventional GGVF fails.

Keywords: deformable model, external force field, partial volume effect, cortical reconstruction, MR image of brain

Procedia PDF Downloads 377
473 Understanding Surface Failures in Thick Asphalt Pavement: A 3-D Finite Element Model Analysis

Authors: Hana Gebremariam Liliso

Abstract:

This study investigates the factors contributing to the deterioration of thick asphalt pavements, such as rutting and cracking. We focus on the combined influence of traffic loads and pavement structure. This study uses a three-dimensional finite element model with a Mohr-Coulomb failure criterion to analyze the stress levels near the pavement's surface under realistic conditions. Our model considers various factors, including tire-pavement contact stresses, asphalt properties, moving loads, and dynamic analysis. This research suggests that cracking tends to occur between dual tires. Some key discoveries include the risk of cracking increases as temperatures rise; surface cracking at high temperatures is associated with distortional deformation; using a uniform contact stress distribution underestimates the risk of failure compared to realistic three-dimensional tire contact stress, particularly at high temperatures; the risk of failure is higher near the surface when there is a negative temperature gradient in the asphalt layer; and debonding beneath the surface layer leads to increased shear stress and premature failure around the interface.

Keywords: asphalt pavement, surface failure, 3d finite element model, multiaxial stress states, Mohr-Coulomb failure criterion

Procedia PDF Downloads 42
472 Customer Preference in the Textile Market: Fabric-Based Analysis

Authors: Francisca Margarita Ocran

Abstract:

Underwear, and more particularly bras and panties, are defined as intimate clothing. Strictly speaking, they enhance the place of women in the public or private satchel. Therefore, women's lingerie is a complex garment with a high involvement profile, motivating consumers to buy it not only by its functional utility but also by the multisensory experience it provides them. Customer behavior models are generally based on customer data mining, and each model is designed to answer questions at a specific time. Predicting the customer experience is uncertain and difficult. Thus, knowledge of consumers' tastes in lingerie deserves to be treated as an experiential product, where the dimensions of the experience motivating consumers to buy a lingerie product and to remain faithful to it must be analyzed in detail by the manufacturers and retailers to engage and retain consumers, which is why this research aims to identify the variables that push consumers to choose their lingerie product, based on an in-depth analysis of the types of fabrics used to make lingerie. The data used in this study comes from online purchases. Machine learning approach with the use of Python programming language and Pycaret gives us a precision of 86.34%, 85.98%, and 84.55% for the three algorithms to use concerning the preference of a buyer in front of a range of lingerie. Gradient Boosting, random forest, and K Neighbors were used in this study; they are very promising and rich in the classification of preference in the textile industry.

Keywords: consumer behavior, data mining, lingerie, machine learning, preference

Procedia PDF Downloads 68
471 Production of Hydroxy Marilone C as a Bioactive Compound from Streptomyces badius

Authors: Osama H. Elsayed, Mohsen M. S. Asker, Mahmoud A. Swelim, Ibrahim H. Abbas, Aziza I. Attwa, Mohamed E. El Awady

Abstract:

Hydroxy marilone C is a bioactive metabolite was produced from the culture broth of Streptomyces badius isolated from Egyptian soil. hydroxy marilone C was purified and fractionated by silica gel column with a gradient mobile phase dicloromethane (DCM) : Methanol then Sephadex LH-20 column using methanol as a mobile phase. It was subjected to many instruments as Infrared (IR), nuclear magnetic resonance (NMR), Mass spectroscopy (MS) and UV spectroscopy to the elucidation of its structure. It was evaluated for antioxidant, cytotoxicity against human alveolar basal epithelial cell line (A-549) and human breast adenocarcinoma cell line (MCF-7) and antiviral activities; showed that the maximum antioxidant activity was 78.8 % at 3000 µg/ml after 90 min. and the IC50 value against DPPH radical found about 1500 µg/ml after 60 min. By Using MTT assay the effect of the pure compound on the proliferation of A-549 cells and MCF-7 cells were 443 µg/ml and 147.9 µg/ml, respectively. While for detection of antiviral activity using Madin-Darby canine kidney (MDCK) cells the maximum cytotoxicity was at 27.9% and IC50 was 128.1µg/ml. The maximum concentration required for protecting 50% of the virus-infected cells against H1N1 viral cytopathogenicity (EC50) was 33.25% for 80 µg/ml. This results indicated that the hydroxy marilone C has a potential antitumor and antiviral activities.

Keywords: hydroxy marilone C, production, bioactive compound, Streptomyces badius

Procedia PDF Downloads 244
470 Second Order Optimality Conditions in Nonsmooth Analysis on Riemannian Manifolds

Authors: Seyedehsomayeh Hosseini

Abstract:

Much attention has been paid over centuries to understanding and solving the problem of minimization of functions. Compared to linear programming and nonlinear unconstrained optimization problems, nonlinear constrained optimization problems are much more difficult. Since the procedure of finding an optimizer is a search based on the local information of the constraints and the objective function, it is very important to develop techniques using geometric properties of the constraints and the objective function. In fact, differential geometry provides a powerful tool to characterize and analyze these geometric properties. Thus, there is clearly a link between the techniques of optimization on manifolds and standard constrained optimization approaches. Furthermore, there are manifolds that are not defined as constrained sets in R^n an important example is the Grassmann manifolds. Hence, to solve optimization problems on these spaces, intrinsic methods are used. In a nondifferentiable problem, the gradient information of the objective function generally cannot be used to determine the direction in which the function is decreasing. Therefore, techniques of nonsmooth analysis are needed to deal with such a problem. As a manifold, in general, does not have a linear structure, the usual techniques, which are often used in nonsmooth analysis on linear spaces, cannot be applied and new techniques need to be developed. This paper presents necessary and sufficient conditions for a strict local minimum of extended real-valued, nonsmooth functions defined on Riemannian manifolds.

Keywords: Riemannian manifolds, nonsmooth optimization, lower semicontinuous functions, subdifferential

Procedia PDF Downloads 348
469 The Analysis of Emergency Shutdown Valves Torque Data in Terms of Its Use as a Health Indicator for System Prognostics

Authors: Ewa M. Laskowska, Jorn Vatn

Abstract:

Industry 4.0 focuses on digital optimization of industrial processes. The idea is to use extracted data in order to build a decision support model enabling use of those data for real time decision making. In terms of predictive maintenance, the desired decision support tool would be a model enabling prognostics of system's health based on the current condition of considered equipment. Within area of system prognostics and health management, a commonly used health indicator is Remaining Useful Lifetime (RUL) of a system. Because the RUL is a random variable, it has to be estimated based on available health indicators. Health indicators can be of different types and come from different sources. They can be process variables, equipment performance variables, data related to number of experienced failures, etc. The aim of this study is the analysis of performance variables of emergency shutdown valves (ESV) used in oil and gas industry. ESV is inspected periodically, and at each inspection torque and time of valve operation are registered. The data will be analyzed by means of machine learning or statistical analysis. The purpose is to investigate whether the available data could be used as a health indicator for a prognostic purpose. The second objective is to examine what is the most efficient way to incorporate the data into predictive model. The idea is to check whether the data can be applied in form of explanatory variables in Markov process or whether other stochastic processes would be a more convenient to build an RUL model based on the information coming from registered data.

Keywords: emergency shutdown valves, health indicator, prognostics, remaining useful lifetime, RUL

Procedia PDF Downloads 74
468 A TgCNN-Based Surrogate Model for Subsurface Oil-Water Phase Flow under Multi-Well Conditions

Authors: Jian Li

Abstract:

The uncertainty quantification and inversion problems of subsurface oil-water phase flow usually require extensive repeated forward calculations for new runs with changed conditions. To reduce the computational time, various forms of surrogate models have been built. Related research shows that deep learning has emerged as an effective surrogate model, while most surrogate models with deep learning are purely data-driven, which always leads to poor robustness and abnormal results. To guarantee the model more consistent with the physical laws, a coupled theory-guided convolutional neural network (TgCNN) based surrogate model is built to facilitate computation efficiency under the premise of satisfactory accuracy. The model is a convolutional neural network based on multi-well reservoir simulation. The core notion of this proposed method is to bridge two separate blocks on top of an overall network. They underlie the TgCNN model in a coupled form, which reflects the coupling nature of pressure and water saturation in the two-phase flow equation. The model is driven by not only labeled data but also scientific theories, including governing equations, stochastic parameterization, boundary, and initial conditions, well conditions, and expert knowledge. The results show that the TgCNN-based surrogate model exhibits satisfactory accuracy and efficiency in subsurface oil-water phase flow under multi-well conditions.

Keywords: coupled theory-guided convolutional neural network, multi-well conditions, surrogate model, subsurface oil-water phase

Procedia PDF Downloads 66
467 Bayesian Inference for High Dimensional Dynamic Spatio-Temporal Models

Authors: Sofia M. Karadimitriou, Kostas Triantafyllopoulos, Timothy Heaton

Abstract:

Reduced dimension Dynamic Spatio-Temporal Models (DSTMs) jointly describe the spatial and temporal evolution of a function observed subject to noise. A basic state space model is adopted for the discrete temporal variation, while a continuous autoregressive structure describes the continuous spatial evolution. Application of such a DSTM relies upon the pre-selection of a suitable reduced set of basic functions and this can present a challenge in practice. In this talk, we propose an online estimation method for high dimensional spatio-temporal data based upon DSTM and we attempt to resolve this issue by allowing the basis to adapt to the observed data. Specifically, we present a wavelet decomposition in order to obtain a parsimonious approximation of the spatial continuous process. This parsimony can be achieved by placing a Laplace prior distribution on the wavelet coefficients. The aim of using the Laplace prior, is to filter wavelet coefficients with low contribution, and thus achieve the dimension reduction with significant computation savings. We then propose a Hierarchical Bayesian State Space model, for the estimation of which we offer an appropriate particle filter algorithm. The proposed methodology is illustrated using real environmental data.

Keywords: multidimensional Laplace prior, particle filtering, spatio-temporal modelling, wavelets

Procedia PDF Downloads 412
466 Flame Volume Prediction and Validation for Lean Blowout of Gas Turbine Combustor

Authors: Ejaz Ahmed, Huang Yong

Abstract:

The operation of aero engines has a critical importance in the vicinity of lean blowout (LBO) limits. Lefebvre’s model of LBO based on empirical correlation has been extended to flame volume concept by the authors. The flame volume takes into account the effects of geometric configuration, the complex spatial interaction of mixing, turbulence, heat transfer and combustion processes inside the gas turbine combustion chamber. For these reasons, flame volume based LBO predictions are more accurate. Although LBO prediction accuracy has improved, it poses a challenge associated with Vf estimation in real gas turbine combustors. This work extends the approach of flame volume prediction previously based on fuel iterative approximation with cold flow simulations to reactive flow simulations. Flame volume for 11 combustor configurations has been simulated and validated against experimental data. To make prediction methodology robust as required in the preliminary design stage, reactive flow simulations were carried out with the combination of probability density function (PDF) and discrete phase model (DPM) in FLUENT 15.0. The criterion for flame identification was defined. Two important parameters i.e. critical injection diameter (Dp,crit) and critical temperature (Tcrit) were identified, and their influence on reactive flow simulation was studied for Vf estimation. Obtained results exhibit ±15% error in Vf estimation with experimental data.

Keywords: CFD, combustion, gas turbine combustor, lean blowout

Procedia PDF Downloads 250
465 Trace Analysis of Genotoxic Impurity Pyridine in Sitagliptin Drug Material Using UHPLC-MS

Authors: Bashar Al-Sabti, Jehad Harbali

Abstract:

Background: Pyridine is a reactive base that might be used in preparing sitagliptin. International Agency for Research on Cancer classifies pyridine in group 2B; this classification means that pyridine is possibly carcinogenic to humans. Therefore, pyridine should be monitored at the allowed limit in sitagliptin pharmaceutical ingredients. Objective: The aim of this study was to develop a novel ultra high performance liquid chromatography mass spectrometry (UHPLC-MS) method to estimate the quantity of pyridine impurity in sitagliptin pharmaceutical ingredients. Methods: The separation was performed on C8 shim-pack (150 mm X 4.6 mm, 5 µm) in reversed phase mode using a mobile phase of water-methanol-acetonitrile containing 4 mM ammonium acetate in gradient mode. Pyridine was detected by mass spectrometer using selected ionization monitoring mode at m/z = 80. The flow rate of the method was 0.75 mL/min. Results: The method showed excellent sensitivity with a quantitation limit of 1.5 ppm of pyridine relative to sitagliptin. The linearity of the method was excellent at the range of 1.5-22.5 ppm with a correlation coefficient of 0.9996. Recoveries values were between 93.59-103.55%. Conclusions: The results showed good linearity, precision, accuracy, sensitivity, selectivity, and robustness. The studied method was applied to test three batches of sitagliptin raw materials. Highlights: This method is useful for monitoring pyridine in sitagliptin during its synthesis and testing sitagliptin raw materials before using them in the production of pharmaceutical products.

Keywords: genotoxic impurity, pyridine, sitagliptin, UHPLC -MS

Procedia PDF Downloads 79
464 Artificial Neural Network Modeling and Genetic Algorithm Based Optimization of Hydraulic Design Related to Seepage under Concrete Gravity Dams on Permeable Soils

Authors: Muqdad Al-Juboori, Bithin Datta

Abstract:

Hydraulic structures such as gravity dams are classified as essential structures, and have the vital role in providing strong and safe water resource management. Three major aspects must be considered to achieve an effective design of such a structure: 1) The building cost, 2) safety, and 3) accurate analysis of seepage characteristics. Due to the complexity and non-linearity relationships of the seepage process, many approximation theories have been developed; however, the application of these theories results in noticeable errors. The analytical solution, which includes the difficult conformal mapping procedure, could be applied for a simple and symmetrical problem only. Therefore, the objectives of this paper are to: 1) develop a surrogate model based on numerical simulated data using SEEPW software to approximately simulate seepage process related to a hydraulic structure, 2) develop and solve a linked simulation-optimization model based on the developed surrogate model to describe the seepage occurring under a concrete gravity dam, in order to obtain optimum and safe design at minimum cost. The result shows that the linked simulation-optimization model provides an efficient and optimum design of concrete gravity dams.

Keywords: artificial neural network, concrete gravity dam, genetic algorithm, seepage analysis

Procedia PDF Downloads 209
463 Approximation of PE-MOCVD to ALD for TiN Concerning Resistivity and Chemical Composition

Authors: D. Geringswald, B. Hintze

Abstract:

The miniaturization of circuits is advancing. During chip manufacturing, structures are filled for example by metal organic chemical vapor deposition (MOCVD). Since this process reaches its limits in case of very high aspect ratios, the use of alternatives such as the atomic layer deposition (ALD) is possible, requiring the extension of existing coating systems. However, it is an unsolved question to what extent MOCVD can achieve results similar as an ALD process. In this context, this work addresses the characterization of a metal organic vapor deposition of titanium nitride. Based on the current state of the art, the film properties coating thickness, sheet resistance, resistivity, stress and chemical composition are considered. The used setting parameters are temperature, plasma gas ratio, plasma power, plasma treatment time, deposition time, deposition pressure, number of cycles and TDMAT flow. The derived process instructions for unstructured wafers and inside a structure with high aspect ratio include lowering the process temperature and increasing the number of cycles, the deposition and the plasma treatment time as well as the plasma gas ratio of hydrogen to nitrogen (H2:N2). In contrast to the current process configuration, the deposited titanium nitride (TiN) layer is more uniform inside the entire test structure. Consequently, this paper provides approaches to employ the MOCVD for structures with increasing aspect ratios.

Keywords: ALD, high aspect ratio, PE-MOCVD, TiN

Procedia PDF Downloads 285
462 Numerical Analysis of Wire Laser Additive Manufacturing for Low Carbon Steels+

Authors: Juan Manuel Martinez Alvarez, Michele Chiumenti

Abstract:

This work explores the benefit of the thermo-metallurgical simulation to tackle the Wire Laser Additive Manufacturing (WLAM) of low-carbon steel components. The Finite Element Analysis is calibrated by process monitoring via thermal imaging and thermocouples measurements, to study the complex thermo-metallurgical behavior inherent to the WLAM process of low carbon steel parts.A critical aspect is the analysis of the heterogeneity in the resulting microstructure. This heterogeneity depends on both the thermal history and the residual stresses experienced during the WLAM process. Because of low carbon grades are highly sensitive to quenching, a high-gradient microstructure often arises due to the layer-by-layer metal deposition in WLAM. The different phases have been identified by scanning electron microscope. A clear influence of the heterogeneities on the final mechanical performance has been established by the subsequent mechanical characterization. The thermo-metallurgical analysis has been used to determine the actual thermal history and the corresponding thermal gradients during the printing process. The correlation between the thermos-mechanical evolution, the printing parameters and scanning sequence has been established. Therefore, an enhanced printing strategy, including optimized process window has been used to minimize the microstructure heterogeneity at ArcelorMittal.

Keywords: additive manufacturing, numerical simulation, metallurgy, steel

Procedia PDF Downloads 51
461 Enhancement Method of Network Traffic Anomaly Detection Model Based on Adversarial Training With Category Tags

Authors: Zhang Shuqi, Liu Dan

Abstract:

For the problems in intelligent network anomaly traffic detection models, such as low detection accuracy caused by the lack of training samples, poor effect with small sample attack detection, a classification model enhancement method, F-ACGAN(Flow Auxiliary Classifier Generative Adversarial Network) which introduces generative adversarial network and adversarial training, is proposed to solve these problems. Generating adversarial data with category labels could enhance the training effect and improve classification accuracy and model robustness. FACGAN consists of three steps: feature preprocess, which includes data type conversion, dimensionality reduction and normalization, etc.; A generative adversarial network model with feature learning ability is designed, and the sample generation effect of the model is improved through adversarial iterations between generator and discriminator. The adversarial disturbance factor of the gradient direction of the classification model is added to improve the diversity and antagonism of generated data and to promote the model to learn from adversarial classification features. The experiment of constructing a classification model with the UNSW-NB15 dataset shows that with the enhancement of FACGAN on the basic model, the classification accuracy has improved by 8.09%, and the score of F1 has improved by 6.94%.

Keywords: data imbalance, GAN, ACGAN, anomaly detection, adversarial training, data augmentation

Procedia PDF Downloads 86
460 Stochastic Optimization of a Vendor-Managed Inventory Problem in a Two-Echelon Supply Chain

Authors: Bita Payami-Shabestari, Dariush Eslami

Abstract:

The purpose of this paper is to develop a multi-product economic production quantity model under vendor management inventory policy and restrictions including limited warehouse space, budget, and number of orders, average shortage time and maximum permissible shortage. Since the “costs” cannot be predicted with certainty, it is assumed that data behave under uncertain environment. The problem is first formulated into the framework of a bi-objective of multi-product economic production quantity model. Then, the problem is solved with three multi-objective decision-making (MODM) methods. Then following this, three methods had been compared on information on the optimal value of the two objective functions and the central processing unit (CPU) time with the statistical analysis method and the multi-attribute decision-making (MADM). The results are compared with statistical analysis method and the MADM. The results of the study demonstrate that augmented-constraint in terms of optimal value of the two objective functions and the CPU time perform better than global criteria, and goal programming. Sensitivity analysis is done to illustrate the effect of parameter variations on the optimal solution. The contribution of this research is the use of random costs data in developing a multi-product economic production quantity model under vendor management inventory policy with several constraints.

Keywords: economic production quantity, random cost, supply chain management, vendor-managed inventory

Procedia PDF Downloads 112
459 Optimizing Groundwater Pumping for a Complex Groundwater/Surface Water System

Authors: Emery A. Coppola Jr., Suna Cinar, Ferenc Szidarovszky

Abstract:

Over-pumping of groundwater resources is a serious problem world-wide. In addition to depleting this valuable resource, hydraulically connected sensitive ecological resources like wetlands and surface water bodies are often impacted and even destroyed by over-pumping. Effectively managing groundwater in a way that satisfy human demand while preserving natural resources is a daunting challenge that will only worsen with growing human populations and climate change. As presented in this paper, a numerical flow model developed for a hypothetical but realistic groundwater/surface water system was combined with formal optimization. Response coefficients were used in an optimization management model to maximize groundwater pumping in a complex, multi-layered aquifer system while protecting against groundwater over-draft, streamflow depletion, and wetland impacts. Pumping optimization was performed for different constraint sets that reflect different resource protection preferences, yielding significantly different optimal pumping solutions. A sensitivity analysis on the optimal solutions was performed on select response coefficients to identify differences between wet and dry periods. Stochastic optimization was also performed, where uncertainty associated with changing irrigation demand due to changing weather conditions are accounted for. One of the strengths of this optimization approach is that it can efficiently and accurately identify superior management strategies that minimize risk and adverse environmental impacts associated with groundwater pumping under different hydrologic conditions.

Keywords: numerical groundwater flow modeling, water management optimization, groundwater overdraft, streamflow depletion

Procedia PDF Downloads 217
458 Rising Velocity of a Non-Newtonian Liquids in Capillary Tubes

Authors: Reza Sabbagh, Linda Hasanovich, Aleksey Baldygin, David S. Nobes, Prashant R. Waghmare

Abstract:

The capillary filling process is significantly important to study for numerous applications such as the under filling of the material in electronic packaging or liquid hydrocarbons seepage through porous structure. The approximation of the fluid being Newtonian, i.e., linear relationship between the shear stress and deformation rate cannot be justified in cases where the extent of non-Newtonian behavior of liquid governs the surface driven transport, i.e., capillarity action. In this study, the capillary action of a non-Newtonian fluid is not only analyzed, but also the modified generalized theoretical analysis for the capillary transport is proposed. The commonly observed three regimes: surface forces dominant (travelling air-liquid interface), developing flow (viscous force dominant), and developed regimes (interfacial, inertial and viscous forces are comparable) are identified. The velocity field along each regime is quantified with Newtonian and non-Newtonian fluid in square shaped vertically oriented channel. Theoretical understanding of capillary imbibition process, particularly in the case of Newtonian fluids, is relied on the simplified assumption of a fully developed velocity profile which has been revisited for developing a modified theory for the capillary transport of non-Newtonian fluids. Furthermore, the development of the velocity profile from the entrance regime to the developed regime, for different power law fluids, is also investigated theoretically and experimentally.

Keywords: capillary, non-Newtonian flow, shadowgraphy, rising velocity

Procedia PDF Downloads 190
457 Portfolio Optimization with Reward-Risk Ratio Measure Based on the Mean Absolute Deviation

Authors: Wlodzimierz Ogryczak, Michal Przyluski, Tomasz Sliwinski

Abstract:

In problems of portfolio selection, the reward-risk ratio criterion is optimized to search for a risky portfolio with the maximum increase of the mean return in proportion to the risk measure increase when compared to the risk-free investments. In the classical model, following Markowitz, the risk is measured by the variance thus representing the Sharpe ratio optimization and leading to the quadratic optimization problems. Several Linear Programming (LP) computable risk measures have been introduced and applied in portfolio optimization. In particular, the Mean Absolute Deviation (MAD) measure has been widely recognized. The reward-risk ratio optimization with the MAD measure can be transformed into the LP formulation with the number of constraints proportional to the number of scenarios and the number of variables proportional to the total of the number of scenarios and the number of instruments. This may lead to the LP models with huge number of variables and constraints in the case of real-life financial decisions based on several thousands scenarios, thus decreasing their computational efficiency and making them hardly solvable by general LP tools. We show that the computational efficiency can be then dramatically improved by an alternative model based on the inverse risk-reward ratio minimization and by taking advantages of the LP duality. In the introduced LP model the number of structural constraints is proportional to the number of instruments thus not affecting seriously the simplex method efficiency by the number of scenarios and therefore guaranteeing easy solvability. Moreover, we show that under natural restriction on the target value the MAD risk-reward ratio optimization is consistent with the second order stochastic dominance rules.

Keywords: portfolio optimization, reward-risk ratio, mean absolute deviation, linear programming

Procedia PDF Downloads 389
456 Radial Distortion Correction Based on the Concept of Verifying the Planarity of a Specimen

Authors: Shih-Heng Tung, Ming-Hsiang Shih, Wen-Pei Sung

Abstract:

Because of the rapid development of digital camera and computer, digital image correlation method has drawn lots of attention recently and has been applied to a variety of fields. However, the image distortion is inevitable when the image is captured through a lens. This image distortion problem can result in an innegligible error while using digital image correlation method. There are already many different ways to correct the image distortion, and most of them require specific image patterns or precise control points. A new distortion correction method is proposed in this study. The proposed method is based on the fact that a flat surface should keep flat when it is measured using three-dimensional (3D) digital image measurement technique. Lens distortion can be divided into radial distortion, decentering distortion and thin prism distortion. Because radial distortion has a more noticeable influence than the other types of distortions, this method deals only with radial distortion. The simplified 3D digital image measurement technique is adopted to measure the surface coordinates of a flat specimen. Then the gradient method is applied to find the best correction parameters. A few experiments are carried out in this study to verify the correctness of this method. The results show that this method can achieve a good accuracy and it is suitable for both large and small distortion conditions. The most important advantage is that it requires neither mark with specific pattern nor precise control points.

Keywords: 3D DIC, radial distortion, distortion correction, planarity

Procedia PDF Downloads 537
455 Evolving Convolutional Filter Using Genetic Algorithm for Image Classification

Authors: Rujia Chen, Ajit Narayanan

Abstract:

Convolutional neural networks (CNN), as typically applied in deep learning, use layer-wise backpropagation (BP) to construct filters and kernels for feature extraction. Such filters are 2D or 3D groups of weights for constructing feature maps at subsequent layers of the CNN and are shared across the entire input. BP as a gradient descent algorithm has well-known problems of getting stuck at local optima. The use of genetic algorithms (GAs) for evolving weights between layers of standard artificial neural networks (ANNs) is a well-established area of neuroevolution. In particular, the use of crossover techniques when optimizing weights can help to overcome problems of local optima. However, the application of GAs for evolving the weights of filters and kernels in CNNs is not yet an established area of neuroevolution. In this paper, a GA-based filter development algorithm is proposed. The results of the proof-of-concept experiments described in this paper show the proposed GA algorithm can find filter weights through evolutionary techniques rather than BP learning. For some simple classification tasks like geometric shape recognition, the proposed algorithm can achieve 100% accuracy. The results for MNIST classification, while not as good as possible through standard filter learning through BP, show that filter and kernel evolution warrants further investigation as a new subarea of neuroevolution for deep architectures.

Keywords: neuroevolution, convolutional neural network, genetic algorithm, filters, kernels

Procedia PDF Downloads 165
454 The Impact of Window Opening Occupant Behavior Models on Building Energy Performance

Authors: Habtamu Tkubet Ebuy

Abstract:

Purpose Conventional dynamic energy simulation tools go beyond the static dimension of simplified methods by providing better and more accurate prediction of building performance. However, their ability to forecast actual performance is undermined by a low representation of human interactions. The purpose of this study is to examine the potential benefits of incorporating information on occupant diversity into occupant behavior models used to simulate building performance. The co-simulation of the stochastic behavior of the occupants substantially increases the accuracy of the simulation. Design/methodology/approach In this article, probabilistic models of the "opening and closing" behavior of the window of inhabitants have been developed in a separate multi-agent platform, SimOcc, and implemented in the building simulation, TRNSYS, in such a way that the behavior of the window with the interconnectivity can be reflected in the simulation analysis of the building. Findings The results of the study prove that the application of complex behaviors is important to research in predicting actual building performance. The results aid in the identification of the gap between reality and existing simulation methods. We hope this study and its results will serve as a guide for researchers interested in investigating occupant behavior in the future. Research limitations/implications Further case studies involving multi-user behavior for complex commercial buildings need to more understand the impact of the occupant behavior on building performance. Originality/value This study is considered as a good opportunity to achieve the national strategy by showing a suitable tool to help stakeholders in the design phase of new or retrofitted buildings to improve the performance of office buildings.

Keywords: occupant behavior, co-simulation, energy consumption, thermal comfort

Procedia PDF Downloads 84
453 Hydrogen Sulfide Removal from Biogas Using Biofilm on Packed Bed of Salak Fruit Seeds

Authors: Retno A. S. Lestari, Wahyudi B. Sediawan, Siti Syamsiah, Sarto

Abstract:

Sulfur-oxidizing bacteria were isolated and then grown on snakefruits seeds forming biofilm. Their performance in sulfide removal were experimentally observed. Snakefruit seeds were then used as packing material in a cylindrical tube. Biological treatment of hydrogen sulfide from biogas was investigated using biofilm on packed bed of snakefruits seeds. Biogas containing 27,9512 ppm of hydrogen sulfide was flown through the bed. Then the hydrogen sulfide concentrations in the outlet at various times were analyzed. A set of simple kinetics model for the rate of the sulfide removal and the bacterial growth was proposed. The axial sulfide concentration gradient in the flowing liquid are assumed to be steady-state. Mean while the biofilm grows on the surface of the seeds and the oxidation takes place in the biofilm. Since the biofilm is very thin, the sulfide concentration in the biofilm is assumed to be uniform. The simultaneous ordinary differential equations obtained were then solved numerically using Runge-Kutta method. The acuracy of the model proposed was tested by comparing the calcultion results using the model with the experimental data obtained. It turned out that the model proposed can be applied to describe the removal of sulfide liquid using bio-filter in packed bed. The values of the parameters were also obtained by curve-fitting. The biofilter could remove 89,83 % of the inlet of hydrogen sulfide from biogas for 2.5 h, and optimum loading of 8.33 ml/h.

Keywords: Sulfur-oxidizing bacteria, snakefruits seeds, biofilm, packing material, biogas

Procedia PDF Downloads 391
452 Prone Positioning and Clinical Outcomes of Mechanically Ventilated Patients with Severe Acute Respiratory Distress Syndrome

Authors: Maha Salah Abdullah Ismail, Mahmoud M. Alsagheir, Mohammed Salah Abd Allah

Abstract:

Acute respiratory distress syndrome (ARDS) is characterized by permeability pulmonary edema and refractory hypoxemia. Lung-protective ventilation is still the key of better outcome in ARDS. Prone position reduces the trans-pulmonary pressure gradient, recruiting collapsed regions of the lung without increasing airway pressure or hyperinflation. Prone ventilation showed improved oxygenation and improved outcomes in severe hypoxemic patients with ARDS. This study evaluates the effect of prone positioning on mechanically ventilated patients with ARDS. A quasi-experimental design was carried out at Critical Care Units, on 60 patients. Two tools were utilized to collect data; Socio demographic, medical and clinical outcomes data sheet. Results of the present study indicated that prone position improves oxygenation in patients with severe respiratory distress syndrome. The study recommended that use prone position in patients with severe ARDS, as early as possible and for long sessions. Also, replication of this study on larger probability sample at the different geographical location is highly recommended.

Keywords: acute respiratory distress syndrome, critical care, mechanical ventilation, prone position

Procedia PDF Downloads 521
451 The Utilization of FSI Technique and Two-Way Particle Coupling System on Particle Dynamics in the Human Alveoli

Authors: Hassan Athari, Abdurrahim Bolukbasi, Dogan Ciloglu

Abstract:

This study represented the respiratory alveoli system, and determined the trajectory of inhaled particles more accurately using the modified three-dimensional model with deformable walls of alveoli. The study also considered the tissue tension in the model to demonstrate the effect of lung. Tissue tensions are transferred by the lung parenchyma and produce the pressure gradient. This load expands the alveoli and establishes a sub-ambient (vacuum) pressure within the lungs. Thus, at the alveolar level, the flow field and movement of alveoli wall lead to an integrated effect. In this research, we assume that the three-dimensional alveolus has a visco-elastic tissue (walls). For accurate investigation of pulmonary tissue mechanical properties on particle transport and alveolar flow field, the actual relevance between tissue movement and airflow is solved by two-way FSI (Fluid Structure Interaction) simulation technique in the alveolus. Therefore, the essence of real simulation of pulmonary breathing mechanics can be achieved by developing a coupled FSI computational model. We, therefore conduct a series of FSI simulations over a range of tissue models and breathing rates. As a result, the fluid flows and streamlines have changed during present flexible model against the rigid models and also the two-way coupling particle trajectories have changed against the one-way particle coupling.

Keywords: FSI, two-way particle coupling, alveoli, CDF

Procedia PDF Downloads 234
450 Transmission Line Matrix (TLM) Modelling of Microstrip Circular Antenna

Authors: Jugoslav Jokovic, Tijana Dimitrijevic, Nebojsa Doncov

Abstract:

The goal of this paper is to investigate the possibilities and effectiveness of the TLM (Transmission Line Matrix) method for modelling of up-to-date microstrip antennas with circular geometry that have significant application in modern wireless communication systems. The coaxially fed microstrip antenna configurations with circular patch are analyzed by using the in-house 3DTLMcyl_cw solver based on computational electromagnetic TLM method adapted to the cylindrical grid and enhanced with the compact wire model. Opposed to the widely used rectangular TLM mesh, where a staircase approximation has to be used to describe curved boundaries, precise modelling of circular boundaries can be accomplished in the cylindrical grid irrespective of the mesh resolution. Using the compact wire model incorporated in cylindrical mesh, it is possible to model coaxial feed and include the influence of the real excitation in the antenna model. The conventional and inverted configuration of a coaxially fed circular patch antenna are considered, comparing the resonances obtained using TLM cylindrical model with results reached by the corresponding model in a rectangular grid as well as with experimental ones. Bearing in mind that accuracy of simulated results depends on a relevantly created model, besides structure geometry and dimensions, it is important to consider additional modelling issues, regarding appropriate mesh resolution and a relevant extension of a mesh around the considered structure that would provide convergence of the results.

Keywords: computational electromagnetic, coaxial feed, microstrip antenna, TLM modelling

Procedia PDF Downloads 267
449 Deep Reinforcement Learning Model for Autonomous Driving

Authors: Boumaraf Malak

Abstract:

The development of intelligent transportation systems (ITS) and artificial intelligence (AI) are spurring us to pave the way for the widespread adoption of autonomous vehicles (AVs). This is open again opportunities for smart roads, smart traffic safety, and mobility comfort. A highly intelligent decision-making system is essential for autonomous driving around dense, dynamic objects. It must be able to handle complex road geometry and topology, as well as complex multiagent interactions, and closely follow higher-level commands such as routing information. Autonomous vehicles have become a very hot research topic in recent years due to their significant ability to reduce traffic accidents and personal injuries. Using new artificial intelligence-based technologies handles important functions in scene understanding, motion planning, decision making, vehicle control, social behavior, and communication for AV. This paper focuses only on deep reinforcement learning-based methods; it does not include traditional (flat) planar techniques, which have been the subject of extensive research in the past because reinforcement learning (RL) has become a powerful learning framework now capable of learning complex policies in high dimensional environments. The DRL algorithm used so far found solutions to the four main problems of autonomous driving; in our paper, we highlight the challenges and point to possible future research directions.

Keywords: deep reinforcement learning, autonomous driving, deep deterministic policy gradient, deep Q-learning

Procedia PDF Downloads 64
448 Theoretical Analysis of the Optical and Solid State Properties of Thin Film

Authors: E. I. Ugwu

Abstract:

Theoretical analysis of the optical and Solid State properties of ZnS thin film using beam propagation technique in which a scalar wave is propagated through the material thin film deposited on a substrate with the assumption that the dielectric medium is section into a homogenous reference dielectric constant term, and a perturbed dielectric term, representing the deposited thin film medium is presented in this work. These two terms, constitute arbitrary complex dielectric function that describes dielectric perturbation imposed by the medium of for the system. This is substituted into a defined scalar wave equation in which the appropriate Green’s Function was defined on it and solved using series technique. The green’s value obtained from Green’s Function was used in Dyson’s and Lippmann Schwinger equations in conjunction with Born approximation method in computing the propagated field for different input regions of field wavelength during which the influence of the dielectric constants and mesh size of the thin film on the propagating field were depicted. The results obtained from the computed field were used in turn to generate the data that were used to compute the band gaps, solid state and optical properties of the thin film such as reflectance, Transmittance and reflectance with which the band gap obtained was found to be in close approximate to that of experimental value.

Keywords: scalar wave, optical and solid state properties, thin film, dielectric medium, perturbation, Lippmann Schwinger equations, Green’s Function, propagation

Procedia PDF Downloads 424
447 Grain Growth Behavior of High Carbon Microalloyed Steels Containing Very Low Amounts of Niobium

Authors: Huseyin Zengin, Muhammet Emre Turan, Yunus Turen, Hayrettin Ahlatci, Yavuz Sun

Abstract:

This study aimed for understanding the effects of dilute Nb additions on the austenite microstructure of microalloyed steels at five different reheating temperatures from 950 °C to 1300 °C. Four microalloyed high-carbon steels having 0.8 %wt C were examined in which three of them had varying Nb concentrations from 0.005 wt% to 0.02 wt% and one of them had no Nb concentration. The quantitative metallographic techniques were used to measure the average prior austenite grain size in order to compare the grain growth pinning effects of Nb precipitates as a function of reheating temperature. Due to the higher stability of the precipitates with increasing Nb concentrations, the grain coarsening temperature that resulted in inefficient grain growth impediment and a bimodal grain distribution in the microstructure, showed an increase with increasing Nb concentration. The respective grain coarsening temperatures (T_GC) in an ascending order for the steels having 0.005 wt% Nb, 0.01 wt% Nb and 0.02 wt% Nb were 950 °C, 1050 °C and 1150 °C. According to these observed grain coarsening temperatures, an approximation was made considering the complete dissolution temperature (T_DISS) of second phase particles as T_GC=T_DISS-300. On the other hand, the plain carbon steel did not show abnormal grain growth behaviour due to the absence of second phase particles. It was also observed that the higher the Nb concentration, the smaller the average prior austenite grain size although the small increments in Nb concenration did not change the average grain size considerably.

Keywords: microalloyed steels, prior austenite grains, second phase particles, grain coarsening temperature

Procedia PDF Downloads 247
446 Using Machine Learning to Classify Human Fetal Health and Analyze Feature Importance

Authors: Yash Bingi, Yiqiao Yin

Abstract:

Reduction of child mortality is an ongoing struggle and a commonly used factor in determining progress in the medical field. The under-5 mortality number is around 5 million around the world, with many of the deaths being preventable. In light of this issue, Cardiotocograms (CTGs) have emerged as a leading tool to determine fetal health. By using ultrasound pulses and reading the responses, CTGs help healthcare professionals assess the overall health of the fetus to determine the risk of child mortality. However, interpreting the results of the CTGs is time-consuming and inefficient, especially in underdeveloped areas where an expert obstetrician is hard to come by. Using a support vector machine (SVM) and oversampling, this paper proposed a model that classifies fetal health with an accuracy of 99.59%. To further explain the CTG measurements, an algorithm based on Randomized Input Sampling for Explanation ((RISE) of Black-box Models was created, called Feature Alteration for explanation of Black Box Models (FAB), and compared the findings to Shapley Additive Explanations (SHAP) and Local Interpretable Model Agnostic Explanations (LIME). This allows doctors and medical professionals to classify fetal health with high accuracy and determine which features were most influential in the process.

Keywords: machine learning, fetal health, gradient boosting, support vector machine, Shapley values, local interpretable model agnostic explanations

Procedia PDF Downloads 126
445 Implementation of a Lattice Boltzmann Method for Pulsatile Flow with Moment Based Boundary Condition

Authors: Zainab A. Bu Sinnah, David I. Graham

Abstract:

The Lattice Boltzmann Method has been developed and used to simulate both steady and unsteady fluid flow problems such as turbulent flows, multiphase flow and flows in the vascular system. As an example, the study of blood flow and its properties can give a greater understanding of atherosclerosis and the flow parameters which influence this phenomenon. The blood flow in the vascular system is driven by a pulsating pressure gradient which is produced by the heart. As a very simple model of this, we simulate plane channel flow under periodic forcing. This pulsatile flow is essentially the standard Poiseuille flow except that the flow is driven by the periodic forcing term. Moment boundary conditions, where various moments of the particle distribution function are specified, are applied at solid walls. We used a second-order single relaxation time model and investigated grid convergence using two distinct approaches. In the first approach, we fixed both Reynolds and Womersley numbers and varied relaxation time with grid size. In the second approach, we fixed the Womersley number and relaxation time. The expected second-order convergence was obtained for the second approach. For the first approach, however, the numerical method converged, but not necessarily to the appropriate analytical result. An explanation is given for these observations.

Keywords: Lattice Boltzmann method, single relaxation time, pulsatile flow, moment based boundary condition

Procedia PDF Downloads 216