Search results for: characteristic method
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 19963

Search results for: characteristic method

18373 Neuro-Epigenetic Changes on Diabetes Induced-Synaptic Fidelity in Brain

Authors: Valencia Fernandes, Dharmendra Kumar Khatri, Shashi Bala Singh

Abstract:

Background and Aim: Epigenetics are the inaudible signatures of several pathological processes in the brain. This study understands the influence of DNA methylation, a major epigenetic modification, in the prefrontal cortex and hippocampus of the diabetic brain and its notable effect on the cellular chaperones and synaptic proteins. Method: Chronic high fat diet and STZ-induced diabetic mice were studied for cognitive dysfunction, and global DNA methylation, as well as DNA methyltransferase (DNMT) activity, were assessed. Further, the cellular chaperones and synaptic proteins were examined using DNMT inhibitor, 5-aza-2′-deoxycytidine (5-aza-dC)-via intracerebroventricular injection. Moreover, % methylation of these synaptic proteins were also studied so as to correlate its epigenetic involvement. Computationally, its interaction with the DNMT enzyme were also studied using bioinformatic tools. Histological studies for morphological alterations and neuronal degeneration were also studied. Neurogenesis, a characteristic marker for new learning and memory formation, was also assessed via the BrdU staining. Finally, the most important behavioral studies, including the Morris water maze, Y maze, passive avoidance, and Novel object recognition test, were performed to study its cognitive functions. Results: Altered global DNA methylation and increased levels of DNMTs within the nucleus were confirmed in the cortex and hippocampus of the diseased mice, suggesting hypermethylation at a genetic level. Treatment with AzadC, a global DNA demethylating agent, ameliorated the protein and gene expression of the cellular chaperones and synaptic fidelity. Furthermore, the methylation analysis profile showed hypermethylation of the hsf1 protein, a master regulator for chaperones and thus, confirmed the epigenetic involvement in the diseased brain. Morphological improvements and decreased neurodegeneration, along with enhanced neurogenesis in the treatment group, suggest that epigenetic modulations do participate in learning and memory. This is supported by the improved behavioral test battery seen in the treatment group. Conclusion: DNA methylation could possibly accord in dysregulating the memory-associated proteins at chronic stages in type 2 diabetes. This could suggest a substantial contribution to the underlying pathophysiology of several metabolic syndromes like insulin resistance, obesity and also participate in transitioning this damage centrally, such as cognitive dysfunction.

Keywords: epigenetics, cognition, chaperones, DNA methylation

Procedia PDF Downloads 205
18372 Land Subsidence Monitoring in Semarang and Demak Coastal Area Using Persistent Scatterer Interferometric Synthetic Aperture Radar

Authors: Reyhan Azeriansyah, Yudo Prasetyo, Bambang Darmo Yuwono

Abstract:

Land subsidence is one of the problems that occur in the coastal areas of Java Island, one of which is the Semarang and Demak areas located in the northern region of Central Java. The impact of sea erosion, rising sea levels, soil structure vulnerable and economic development activities led to both these areas often occurs on land subsidence. To know how much land subsidence that occurred in the region needs to do the monitoring carried out by remote sensing methods such as PS-InSAR method. PS-InSAR is a remote sensing technique that is the development of the DInSAR method that can monitor the movement of the ground surface that allows users to perform regular measurements and monitoring of fixed objects on the surface of the earth. PS InSAR processing is done using Standford Method of Persistent Scatterers (StaMPS). Same as the recent analysis technique, Persistent Scatterer (PS) InSAR addresses both the decorrelation and atmospheric problems of conventional InSAR. StaMPS identify and extract the deformation signal even in the absence of bright scatterers. StaMPS is also applicable in areas undergoing non-steady deformation, with no prior knowledge of the variations in deformation rate. In addition, this method can also cover a large area so that the decline in the face of the land can cover all coastal areas of Semarang and Demak. From the PS-InSAR method can be known the impact on the existing area in Semarang and Demak region per year. The PS-InSAR results will also be compared with the GPS monitoring data to determine the difference in land decline that occurs between the two methods. By utilizing remote sensing methods such as PS-InSAR method, it is hoped that the PS-InSAR method can be utilized in monitoring the land subsidence and can assist other survey methods such as GPS surveys and the results can be used in policy determination in the affected coastal areas of Semarang and Demak.

Keywords: coastal area, Demak, land subsidence, PS-InSAR, Semarang, StaMPS

Procedia PDF Downloads 266
18371 Rating and Generating Sudoku Puzzles Based on Constraint Satisfaction Problems

Authors: Bahare Fatemi, Seyed Mehran Kazemi, Nazanin Mehrasa

Abstract:

Sudoku is a logic-based combinatorial puzzle game which people in different ages enjoy playing it. The challenging and addictive nature of this game has made it a ubiquitous game. Most magazines, newspapers, puzzle books, etc. publish lots of Sudoku puzzles every day. These puzzles often come in different levels of difficulty so that all people, from beginner to expert, can play the game and enjoy it. Generating puzzles with different levels of difficulty is a major concern of Sudoku designers. There are several works in the literature which propose ways of generating puzzles having a desirable level of difficulty. In this paper, we propose a method based on constraint satisfaction problems to evaluate the difficulty of the Sudoku puzzles. Then, we propose a hill climbing method to generate puzzles with different levels of difficulty. Whereas other methods are usually capable of generating puzzles with only few number of difficulty levels, our method can be used to generate puzzles with arbitrary number of different difficulty levels. We test our method by generating puzzles with different levels of difficulty and having a group of 15 people solve all the puzzles and recording the time they spend for each puzzle.

Keywords: constraint satisfaction problem, generating Sudoku puzzles, hill climbing

Procedia PDF Downloads 402
18370 Enhanced Retrieval-Augmented Generation (RAG) Method with Knowledge Graph and Graph Neural Network (GNN) for Automated QA Systems

Authors: Zhihao Zheng, Zhilin Wang, Linxin Liu

Abstract:

In the research of automated knowledge question-answering systems, accuracy and efficiency are critical challenges. This paper proposes a knowledge graph-enhanced Retrieval-Augmented Generation (RAG) method, combined with a Graph Neural Network (GNN) structure, to automatically determine the correctness of knowledge competition questions. First, a domain-specific knowledge graph was constructed from a large corpus of academic journal literature, with key entities and relationships extracted using Natural Language Processing (NLP) techniques. Then, the RAG method's retrieval module was expanded to simultaneously query both text databases and the knowledge graph, leveraging the GNN to further extract structured information from the knowledge graph. During answer generation, contextual information provided by the knowledge graph and GNN is incorporated to improve the accuracy and consistency of the answers. Experimental results demonstrate that the knowledge graph and GNN-enhanced RAG method perform excellently in determining the correctness of questions, achieving an accuracy rate of 95%. Particularly in cases involving ambiguity or requiring contextual information, the structured knowledge provided by the knowledge graph and GNN significantly enhances the RAG method's performance. This approach not only demonstrates significant advantages in improving the accuracy and efficiency of automated knowledge question-answering systems but also offers new directions and ideas for future research and practical applications.

Keywords: knowledge graph, graph neural network, retrieval-augmented generation, NLP

Procedia PDF Downloads 40
18369 Relationship between Teachers' Empowerment and Personality Traits, Case Study: Tehran Public Schools of Region 5

Authors: Alireza Ladan Moghaddam, Hadi Rezghi Shirsavar, Panteha Pirayandeh

Abstract:

This study aimed to investigate the Relationship between Teachers' Empowerment and Personality Traits (Case Study: Tehran Public Schools of region 5). To achieve this objective, a descriptive research in correlation type has been used. The statistical population of this research includes all teachers and administrators in Tehran Public Schools of region 5. In this study, a five factor model of personality and a 49-item questionnaire to measure empowerment of teachers have been used to assess personality traits of administrators and the teachers' empowerment, respectively. The research hypotheses test has been done using SPSS and LISREL software. The results show that in general there is no significant relationship between personality traits of administrators and teachers' empowerment, and among the 5 dimensions of personality, there is only significant relationship between the characteristic of administrators' agreeableness and teachers' empowerment. The results suggested a way to improve knowledge and skills of teachers a top priority administrator consider. In addition, the performance of teachers affected by the performance of the executive directors, so it is necessary to improve their yield towering notice.

Keywords: personality traits, five factor model of personality, teacher, empowerment

Procedia PDF Downloads 366
18368 A Fourier Method for Risk Quantification and Allocation of Credit Portfolios

Authors: Xiaoyu Shen, Fang Fang, Chujun Qiu

Abstract:

Herewith we present a Fourier method for credit risk quantification and allocation in the factor-copula model framework. The key insight is that, compared to directly computing the cumulative distribution function of the portfolio loss via Monte Carlo simulation, it is, in fact, more efficient to calculate the transformation of the distribution function in the Fourier domain instead and inverting back to the real domain can be done in just one step and semi-analytically, thanks to the popular COS method (with some adjustments). We also show that the Euler risk allocation problem can be solved in the same way since it can be transformed into the problem of evaluating a conditional cumulative distribution function. Once the conditional or unconditional cumulative distribution function is known, one can easily calculate various risk metrics. The proposed method not only fills the niche in literature, to the best of our knowledge, of accurate numerical methods for risk allocation but may also serve as a much faster alternative to the Monte Carlo simulation method for risk quantification in general. It can cope with various factor-copula model choices, which we demonstrate via examples of a two-factor Gaussian copula and a two-factor Gaussian-t hybrid copula. The fast error convergence is proved mathematically and then verified by numerical experiments, in which Value-at-Risk, Expected Shortfall, and conditional Expected Shortfall are taken as examples of commonly used risk metrics. The calculation speed and accuracy are tested to be significantly superior to the MC simulation for real-sized portfolios. The computational complexity is, by design, primarily driven by the number of factors instead of the number of obligors, as in the case of Monte Carlo simulation. The limitation of this method lies in the "curse of dimension" that is intrinsic to multi-dimensional numerical integration, which, however, can be relaxed with the help of dimension reduction techniques and/or parallel computing, as we will demonstrate in a separate paper. The potential application of this method has a wide range: from credit derivatives pricing to economic capital calculation of the banking book, default risk charge and incremental risk charge computation of the trading book, and even to other risk types than credit risk.

Keywords: credit portfolio, risk allocation, factor copula model, the COS method, Fourier method

Procedia PDF Downloads 167
18367 Expansive-Restrictive Style: Conceptualizing Knowledge Workers

Authors: Ram Manohar Singh, Meenakshi Gupta

Abstract:

Various terms such as ‘learning style’, ‘cognitive style’, ‘conceptual style’, ‘thinking style’, ‘intellectual style’ are used in literature to refer to an individual’s characteristic and consistent approach to organizing and processing information. However, style concepts are criticized for mutually overlapping definitions and confusing classification. This confusion should be addressed at the conceptual as well as empirical level. This paper is an attempt to bridge this gap in literature by proposing a new concept: expansive-restrictive intellectual style based on phenomenological analysis of an auto-ethnography and interview of 26 information technology (IT) professionals working in knowledge intensive organizations (KIOs) in India. Expansive style is an individual’s preference to expand his/her horizon of knowledge and understanding by gaining real meaning and structure of his/her work. On the contrary restrictive style is characterized by an individual’s preference to take minimalist approach at work reflected in executing a job efficiently without an attempt to understand the real meaning and structure of the work. The analysis suggests that expansive-restrictive style has three dimensions: (1) field dependence-independence (2) cognitive involvement and (3) epistemological beliefs.

Keywords: expansive, knowledge workers, restrictive, style

Procedia PDF Downloads 424
18366 Computationally Efficient Stacking Sequence Blending for Composite Structures with a Large Number of Design Regions Using Cellular Automata

Authors: Ellen Van Den Oord, Julien Marie Jan Ferdinand Van Campen

Abstract:

This article introduces a computationally efficient method for stacking sequence blending of composite structures. The computational efficiency makes the presented method especially interesting for composite structures with a large number of design regions. Optimization of composite structures with an unequal load distribution may lead to locally optimized thicknesses and ply orientations that are incompatible with one another. Blending constraints can be enforced to achieve structural continuity. In literature, many methods can be found to implement structural continuity by means of stacking sequence blending in one way or another. The complexity of the problem makes the blending of a structure with a large number of adjacent design regions, and thus stacking sequences, prohibitive. In this work the local stacking sequence optimization is preconditioned using a method found in the literature that couples the mechanical behavior of the laminate, in the form of lamination parameters, to blending constraints, yielding near-optimal easy-to-blend designs. The preconditioned design is then fed to the scheme using cellular automata that have been developed by the authors. The method is applied to the benchmark 18-panel horseshoe blending problem to demonstrate its performance. The computational efficiency of the proposed method makes it especially suited for composite structures with a large number of design regions.

Keywords: composite, blending, optimization, lamination parameters

Procedia PDF Downloads 228
18365 A New Conjugate Gradient Method with Guaranteed Descent

Authors: B. Sellami, M. Belloufi

Abstract:

Conjugate gradient methods are an important class of methods for unconstrained optimization, especially for large-scale problems. Recently, they have been much studied. In this paper, we propose a new two-parameter family of conjugate gradient methods for unconstrained optimization. The two-parameter family of methods not only includes the already existing three practical nonlinear conjugate gradient methods, but also has other family of conjugate gradient methods as subfamily. The two-parameter family of methods with the Wolfe line search is shown to ensure the descent property of each search direction. Some general convergence results are also established for the two-parameter family of methods. The numerical results show that this method is efficient for the given test problems. In addition, the methods related to this family are uniformly discussed.

Keywords: unconstrained optimization, conjugate gradient method, line search, global convergence

Procedia PDF Downloads 452
18364 Experimental Study of Water Injection into Manifold on Engine Performance and Emissions in Compression Ignition Engine

Authors: N. Rajmohan, M. R. Swaminathan

Abstract:

The performance of a diesel engine depends mainly on mixing of the fuel and air in the combustion chamber. The diesel engine suffers from significant generation of nitric oxide and particulate matter emission due to incomplete combustion. As the fuel is injected directly into the combustion chamber in conventional diesel engines, spatial distributions of air-fuel ratio vary widely from rich to lean in combustion chamber. The NOx is formed in stoichiometric zone and smoke is generated during diffusion combustion period where the combustion rate becomes slower. One of the effective methods to reduce oxides of nitrogen and particulate matter emissions simultaneously is to reduce the intake charge temperature in diesel engines. Therefore, in the present study, the effect of water injection into intake air on performance and emission characteristic of single cylinder CI engine are carried out at different load and constant speed, with variable water to diesel ratio by mass. The water is injected into intake air by an elementary carburetor.

Keywords: engine emission control, oxides of nitrogen, diesel engine, ignition engine

Procedia PDF Downloads 362
18363 Periodic Topology and Size Optimization Design of Tower Crane Boom

Authors: Wu Qinglong, Zhou Qicai, Xiong Xiaolei, Zhang Richeng

Abstract:

In order to achieve the layout and size optimization of the web members of tower crane boom, a truss topology and cross section size optimization method based on continuum is proposed considering three typical working conditions. Firstly, the optimization model is established by replacing web members with web plates. And the web plates are divided into several sub-domains so that periodic soft kill option (SKO) method can be carried out for topology optimization of the slender boom. After getting the optimized topology of web plates, the optimized layout of web members is formed through extracting the principal stress distribution. Finally, using the web member radius as design variable, the boom compliance as objective and the material volume of the boom as constraint, the cross section size optimization mathematical model is established. The size optimization criterion is deduced from the mathematical model by Lagrange multiplier method and Kuhn-Tucker condition. By comparing the original boom with the optimal boom, it is identified that this optimization method can effectively lighten the boom and improve its performance.

Keywords: tower crane boom, topology optimization, size optimization, periodic, SKO, optimization criterion

Procedia PDF Downloads 554
18362 Influence of Salbutamol (Beta Adrenergic Agonist) on Carcass Characteristics and Same Blood Parameters in Male Broiler Chicken

Authors: Seyyed Naeim Saber, Javad Usefi

Abstract:

This study was conducted to determine the effect of salbutamol (beta-adrenergic agonist) on carcass characteristics and some blood parameters in male broiler chicks. Four hundred and twenty day-old (Coob-500) male broiler were used in this experiment for six weeks. All birds were randomly divided into 7 treatment groups with 4 replicates of 15 birds per pen. Treatment groups included: control, 5, 10, and 15 mg salbutamol per liter water and 10, 20 and 30 mg salbutamol per kg diet. The data obtained from this study indicated that supplementation of salbutamol in water and diets have significant effect on live body weight, abdominal fat, and gizzard weight (p<0.05). Also adding salbutamol in broiler water and feed did not have significant effect on thigh and breast dry matter, thigh and breast crude protein, and thigh and breast crude fat (p>0.05). The results from this study demonstrated that salbutamol has significant (p<0.05) effect on hemoglobin content and RV/TV but it does not have significant effect (p>0.05) on hematocrit amount.

Keywords: salbutamol, beta-adrenergic agonist, broiler chicks, carcass characteristic

Procedia PDF Downloads 240
18361 Seismic Vulnerability of Structures Designed in Accordance with the Allowable Stress Design and Load Resistant Factor Design Methods

Authors: Mohammadreza Vafaei, Amirali Moradi, Sophia C. Alih

Abstract:

The method selected for the design of structures not only can affect their seismic vulnerability but also can affect their construction cost. For the design of steel structures, two distinct methods have been introduced by existing codes, namely allowable stress design (ASD) and load resistant factor design (LRFD). This study investigates the effect of using the aforementioned design methods on the seismic vulnerability and construction cost of steel structures. Specifically, a 20-story building equipped with special moment resisting frame and an eccentrically braced system was selected for this study. The building was designed for three different intensities of peak ground acceleration including 0.2 g, 0.25 g, and 0.3 g using the ASD and LRFD methods. The required sizes of beams, columns, and braces were obtained using response spectrum analysis. Then, the designed frames were subjected to nine natural earthquake records which were scaled to the designed response spectrum. For each frame, the base shear, story shears, and inter-story drifts were calculated and then were compared. Results indicated that the LRFD method led to a more economical design for the frames. In addition, the LRFD method resulted in lower base shears and larger inter-story drifts when compared with the ASD method. It was concluded that the application of the LRFD method not only reduced the weights of structural elements but also provided a higher safety margin against seismic actions when compared with the ASD method.

Keywords: allowable stress design, load resistant factor design, nonlinear time history analysis, seismic vulnerability, steel structures

Procedia PDF Downloads 269
18360 An Online Mastery Learning Method Based on a Dynamic Formative Evaluation

Authors: Jeongim Kang, Moon Hee Kim, Seong Baeg Kim

Abstract:

This paper proposes a novel e-learning model that is based on a dynamic formative evaluation. On evaluating the existing format of e-learning, conditions regarding repetitive learning to achieve mastery, causes issues for learners to lose tension and become neglectful of learning. The dynamic formative evaluation proposed is able to supplement limitation of the existing approaches. Since a repetitive learning method does not provide a perfect feedback, this paper puts an emphasis on the dynamic formative evaluation that is able to maximize learning achievement. Through the dynamic formative evaluation, the instructor is able to refer to the evaluation result when making estimation about the learner. To show the flow chart of learning, based on the dynamic formative evaluation, the model proves its effectiveness and validity.

Keywords: online learning, dynamic formative evaluation, mastery learning, repetitive learning method, learning achievement

Procedia PDF Downloads 511
18359 Age–Related Changes of the Sella Turcica Morphometry in Adults Older Than 20-25 Years

Authors: Yu. I. Pigolkin, M. A. Garcia Corro

Abstract:

Age determination of unknown dead bodies in forensic personal identification is a complicated process which involves the application of numerous methods and techniques. Skeletal remains are less exposed to influences of environmental factors. In order to enhance the accuracy of forensic age estimation additional properties of bones correlating with age are required to be revealed. Material and Methods: Dimensional examination of the sella turcica was carried out on cadavers with the cranium opened by a circular vibrating saw. The sample consisted of a total of 90 Russian subjects, ranging in age from two months and 87 years. Results: The tendency of dimensional variations throughout life was detected. There were no observed gender differences in the morphometry of the sella turcica. The shared use of the sella turcica depth and length values revealed the possibility to categorize an examined sample in a certain age period. Conclusions: Based on the results of existing methods of age determination, the morphometry of the sella turcica can be an additional characteristic, amplifying the received values, and accordingly, increasing the accuracy of forensic biological age diagnosis.

Keywords: age–related changes in bone structures, forensic personal identification, sella turcica morphometry, body identification

Procedia PDF Downloads 275
18358 Temporally Coherent 3D Animation Reconstruction from RGB-D Video Data

Authors: Salam Khalifa, Naveed Ahmed

Abstract:

We present a new method to reconstruct a temporally coherent 3D animation from single or multi-view RGB-D video data using unbiased feature point sampling. Given RGB-D video data, in form of a 3D point cloud sequence, our method first extracts feature points using both color and depth information. In the subsequent steps, these feature points are used to match two 3D point clouds in consecutive frames independent of their resolution. Our new motion vectors based dynamic alignment method then fully reconstruct a spatio-temporally coherent 3D animation. We perform extensive quantitative validation using novel error functions to analyze the results. We show that despite the limiting factors of temporal and spatial noise associated to RGB-D data, it is possible to extract temporal coherence to faithfully reconstruct a temporally coherent 3D animation from RGB-D video data.

Keywords: 3D video, 3D animation, RGB-D video, temporally coherent 3D animation

Procedia PDF Downloads 373
18357 The Estimation Method of Inter-Story Drift for Buildings Based on Evolutionary Learning

Authors: Kyu Jin Kim, Byung Kwan Oh, Hyo Seon Park

Abstract:

The seismic responses-based structural health monitoring system has been performed to reduce seismic damage. The inter-story drift ratio which is the major index of the seismic capacity assessment is employed for estimating the seismic damage of buildings. Meanwhile, seismic response analysis to estimate the structural responses of building demands significantly high computational cost due to increasing number of high-rise and large buildings. To estimate the inter-story drift ratio of buildings from the earthquake efficiently, this paper suggests the estimation method of inter-story drift for buildings using an artificial neural network (ANN). In the method, the radial basis function neural network (RBFNN) is integrated with optimization algorithm to optimize the variable through evolutionary learning that refers to evolutionary radial basis function neural network (ERBFNN). The estimation method estimates the inter-story drift without seismic response analysis when the new earthquakes are subjected to buildings. The effectiveness of the estimation method is verified through a simulation using multi-degree of freedom system.

Keywords: structural health monitoring, inter-story drift ratio, artificial neural network, radial basis function neural network, genetic algorithm

Procedia PDF Downloads 327
18356 Estimating 3D-Position of a Stationary Random Acoustic Source Using Bispectral Analysis of 4-Point Detected Signals

Authors: Katsumi Hirata

Abstract:

To develop the useful acoustic environmental recognition system, the method of estimating 3D-position of a stationary random acoustic source using bispectral analysis of 4-point detected signals is proposed. The method uses information about amplitude attenuation and propagation delay extracted from amplitude ratios and angles of auto- and cross-bispectra of the detected signals. It is expected that using bispectral analysis affects less influence of Gaussian noises than using conventional power spectral one. In this paper, the basic principle of the method is mentioned first, and its validity and features are considered from results of the fundamental experiments assumed ideal circumstances.

Keywords: 4-point detection, a stationary random acoustic source, auto- and cross-bispectra, estimation of 3D-position

Procedia PDF Downloads 359
18355 Mine Project Evaluations in the Rising of Uncertainty: Real Options Analysis

Authors: I. Inthanongsone, C. Drebenstedt, J. C. Bongaerts, P. Sontamino

Abstract:

The major concern in evaluating the value of mining projects related to the deficiency of the traditional discounted cash flow (DCF) method. This method does not take uncertainties into account and, hence it does not allow for an economic assessment of managerial flexibility and operational adaptability, which are increasingly determining long-term corporate success. Such an assessment can be performed with the real options valuation (ROV) approach, since it allows for a comparative evaluation of unforeseen uncertainties in a project life cycle. This paper presents an economic evaluation model for open pit mining projects based on real options valuation approach. Uncertainties in the model are caused by metal prices and cost uncertainties and the system dynamics (SD) modeling method is used to structure and solve the real options model. The model is applied to a case study. It can be shown that that managerial flexibility reacting to uncertainties may create additional value to a mining project in comparison to the outcomes of a DCF method. One important insight for management dealing with uncertainty is seen in choosing the optimal time to exercise strategic options.

Keywords: DCF methods, ROV approach, system dynamics modeling methods, uncertainty

Procedia PDF Downloads 501
18354 Optimization of Process Parameters by Using Taguchi Method for Bainitic Steel Machining

Authors: Vinay Patil, Swapnil Kekade, Ashish Supare, Vinayak Pawar, Shital Jadhav, Rajkumar Singh

Abstract:

In recent days, bainitic steel is used in automobile and non-automobile sectors due to its high strength. Bainitic steel is difficult to machine because of its high hardness, hence in this paper machinability of bainitic steel is studied by using Taguchi design of experiments (DOE) approach. Convectional turning experiments were done by using L16 orthogonal array for three input parameters viz. cutting speed, depth of cut and feed. The Taguchi method is applied to study the performance characteristics of machining parameters with surface roughness (Ra), cutting force and tool wear rate. By using Taguchi analysis, optimized process parameters for best surface finish and minimum cutting forces were analyzed.

Keywords: conventional turning, Taguchi method, S/N ratio, bainitic steel machining

Procedia PDF Downloads 331
18353 Use of Artificial Intelligence and Two Object-Oriented Approaches (k-NN and SVM) for the Detection and Characterization of Wetlands in the Centre-Val de Loire Region, France

Authors: Bensaid A., Mostephaoui T., Nedjai R.

Abstract:

Nowadays, wetlands are the subject of contradictory debates opposing scientific, political and administrative meanings. Indeed, given their multiple services (drinking water, irrigation, hydrological regulation, mineral, plant and animal resources...), wetlands concentrate many socio-economic and biodiversity issues. In some regions, they can cover vast areas (>100 thousand ha) of the landscape, such as the Camargue area in the south of France, inside the Rhone delta. The high biological productivity of wetlands, the strong natural selection pressures and the diversity of aquatic environments have produced many species of plants and animals that are found nowhere else. These environments are tremendous carbon sinks and biodiversity reserves depending on their age, composition and surrounding environmental conditions, wetlands play an important role in global climate projections. Covering more than 3% of the earth's surface, wetlands have experienced since the beginning of the 1990s a tremendous revival of interest, which has resulted in the multiplication of inventories, scientific studies and management experiments. The geographical and physical characteristics of the wetlands of the central region conceal a large number of natural habitats that harbour a great biological diversity. These wetlands, one of the natural habitats, are still influenced by human activities, especially agriculture, which affects its layout and functioning. In this perspective, decision-makers need to delimit spatial objects (natural habitats) in a certain way to be able to take action. Thus, wetlands are no exception to this rule even if it seems to be a difficult exercise to delimit a type of environment as whose main characteristic is often to occupy the transition between aquatic and terrestrial environment. However, it is possible to map wetlands with databases, derived from the interpretation of photos and satellite images, such as the European database Corine Land cover, which allows quantifying and characterizing for each place the characteristic wetland types. Scientific studies have shown limitations when using high spatial resolution images (SPOT, Landsat, ASTER) for the identification and characterization of small wetlands (1 hectare). To address this limitation, it is important to note that these wetlands generally represent spatially complex features. Indeed, the use of very high spatial resolution images (>3m) is necessary to map small and large areas. However, with the recent evolution of artificial intelligence (AI) and deep learning methods for satellite image processing have shown a much better performance compared to traditional processing based only on pixel structures. Our research work is also based on spectral and textural analysis on THR images (Spot and IRC orthoimage) using two object-oriented approaches, the nearest neighbour approach (k-NN) and the Super Vector Machine approach (SVM). The k-NN approach gave good results for the delineation of wetlands (wet marshes and moors, ponds, artificial wetlands water body edges, ponds, mountain wetlands, river edges and brackish marshes) with a kappa index higher than 85%.

Keywords: land development, GIS, sand dunes, segmentation, remote sensing

Procedia PDF Downloads 72
18352 Attachment Systems and Psychotherapy: An Internal Secure Caregiver to Heal and Protect the Parts of Our Clients: InCorporer Method

Authors: Julien Baillet

Abstract:

In light of 30 years of scientific research, InCorporer Method was created in 2019 as a new approach to heal traumatic, developmental, and dissociative injuries. Following natural nervous system functions, InCorporer aims to heal, develop, and update the different defensive mammalian subsystems: fight, flight, freeze, feign death, cry for help, & energy regulator. The dimensions taken into account are: (i) Heal the traumatic injuries who are still bleeding, (ii) Develop the systems that never received the security, attention, and affection they needed. (iii) Update the parts that stayed stuck in the past, ignoring for too long that they are out of danger now. Through the Present Part and its caregiving skills, InCorporer method enables a balanced, soothed, and collaborative personality system. To be as integrative as possible, InCorporer method has been designed according to several fields of research, such as structural dissociation theory, attachment theory, and information processing theory. In this paper, the author presents how the internal caregiver is developed and trained to heal all the different parts/subsystems of our clients through mindful attention and reflex movement integration.

Keywords: PTSD, attachment, dissociation, part work

Procedia PDF Downloads 77
18351 Efficient Method for Inducing Embryos from Isolated Microspores of Durum Wheat

Authors: Zelikha Labbani

Abstract:

Durum wheat represents an attractive species to study androgenesis via isolated microspore culture in order to increase the efficiency of androgenic yield in recalcitrant species such as in induction embryogenesis. We describe here an efficient method for inducing embryos from isolated microspores of durum wheat. It is shown that this method, associated with cold alone or cold plus mannitol pretreatment, or mannitol alone of the spikes kept within their sheath leaves during different times, has significant positive effects on embryo production. The aim of this study was, therefore, to test the effect of mannitol 0,3M and cold pretreatment on the quality and quantity of embryos produced from microspore culture from wheat cultivars.

Keywords: in vitro embryogenesis, isolated microspores culture, durum wheat, pretreatments, mannitol 0.3m, cold pretreatment

Procedia PDF Downloads 57
18350 Hybrid Artificial Bee Colony and Least Squares Method for Rule-Based Systems Learning

Authors: Ahcene Habbi, Yassine Boudouaoui

Abstract:

This paper deals with the problem of automatic rule generation for fuzzy systems design. The proposed approach is based on hybrid artificial bee colony (ABC) optimization and weighted least squares (LS) method and aims to find the structure and parameters of fuzzy systems simultaneously. More precisely, two ABC based fuzzy modeling strategies are presented and compared. The first strategy uses global optimization to learn fuzzy models, the second one hybridizes ABC and weighted least squares estimate method. The performances of the proposed ABC and ABC-LS fuzzy modeling strategies are evaluated on complex modeling problems and compared to other advanced modeling methods.

Keywords: automatic design, learning, fuzzy rules, hybrid, swarm optimization

Procedia PDF Downloads 437
18349 Rationalized Haar Transforms Approach to Design of Observer for Control Systems with Unknown Inputs

Authors: Joon-Hoon Park

Abstract:

The fundamental concept of observability is important in both theoretical and practical points of modern control systems. In modern control theory, a control system has criteria for determining the design solution exists for the system parameters and design objectives. The idea of observability relates to the condition of observing or estimating the state variables from the output variables that is generally measurable. To design closed-loop control system, the practical problems of implementing the feedback of the state variables must be considered and implementing state feedback control problem has been existed in this case. All the state variables are not available, so it is requisite to design and implement an observer that will estimate the state variables form the output parameters. However sometimes unknown inputs are presented in control systems as practical cases. This paper presents a design method and algorithm for observer of control system with unknown input parameters based on Rationalized Haar transform. The proposed method is more advantageous than the other numerical method.

Keywords: orthogonal functions, rationalized Haar transforms, control system observer, algebraic method

Procedia PDF Downloads 370
18348 Optimization of Moisture Content for Highest Tensile Strength of Instant Soluble Milk Tablet and Flowability of Milk Powder

Authors: Siddharth Vishwakarma, Danie Shajie A., Mishra H. N.

Abstract:

Milk powder becomes very useful in the low milk supply area but the exact amount to add for one glass of milk and the handling is difficult. So, the idea of instant soluble milk tablet comes into existence for its high solubility and easy handling. The moisture content of milk tablets is increased by the direct addition of water with no additives for binding. The variation of the tensile strength of instant soluble milk tablets and the flowability of milk powder with the moisture content is analyzed and optimized for the highest tensile strength of instant soluble milk tablets and flowability, above a particular value of milk powder using response surface methodology. The flowability value is necessary for ease in quantifying the milk powder, as a feed, in the designed tablet making machine. The instant soluble nature of milk tablets purely depends upon the disintegration characteristic of tablets in water whose study is under progress. Conclusions: The optimization results are very useful in the commercialization of milk tablets.

Keywords: flowability, milk powder, response surface methodology, tablet making machine, tensile strength

Procedia PDF Downloads 182
18347 System Identification in Presence of Outliers

Authors: Chao Yu, Qing-Guo Wang, Dan Zhang

Abstract:

The outlier detection problem for dynamic systems is formulated as a matrix decomposition problem with low-rank, sparse matrices and further recast as a semidefinite programming (SDP) problem. A fast algorithm is presented to solve the resulting problem while keeping the solution matrix structure and it can greatly reduce the computational cost over the standard interior-point method. The computational burden is further reduced by proper construction of subsets of the raw data without violating low rank property of the involved matrix. The proposed method can make exact detection of outliers in case of no or little noise in output observations. In case of significant noise, a novel approach based on under-sampling with averaging is developed to denoise while retaining the saliency of outliers and so-filtered data enables successful outlier detection with the proposed method while the existing filtering methods fail. Use of recovered “clean” data from the proposed method can give much better parameter estimation compared with that based on the raw data.

Keywords: outlier detection, system identification, matrix decomposition, low-rank matrix, sparsity, semidefinite programming, interior-point methods, denoising

Procedia PDF Downloads 307
18346 An Estimating Parameter of the Mean in Normal Distribution by Maximum Likelihood, Bayes, and Markov Chain Monte Carlo Methods

Authors: Autcha Araveeporn

Abstract:

This paper is to compare the parameter estimation of the mean in normal distribution by Maximum Likelihood (ML), Bayes, and Markov Chain Monte Carlo (MCMC) methods. The ML estimator is estimated by the average of data, the Bayes method is considered from the prior distribution to estimate Bayes estimator, and MCMC estimator is approximated by Gibbs sampling from posterior distribution. These methods are also to estimate a parameter then the hypothesis testing is used to check a robustness of the estimators. Data are simulated from normal distribution with the true parameter of mean 2, and variance 4, 9, and 16 when the sample sizes is set as 10, 20, 30, and 50. From the results, it can be seen that the estimation of MLE, and MCMC are perceivably different from the true parameter when the sample size is 10 and 20 with variance 16. Furthermore, the Bayes estimator is estimated from the prior distribution when mean is 1, and variance is 12 which showed the significant difference in mean with variance 9 at the sample size 10 and 20.

Keywords: Bayes method, Markov chain Monte Carlo method, maximum likelihood method, normal distribution

Procedia PDF Downloads 356
18345 Ice Load Measurements on Known Structures Using Image Processing Methods

Authors: Azam Fazelpour, Saeed R. Dehghani, Vlastimil Masek, Yuri S. Muzychka

Abstract:

This study employs a method based on image analyses and structure information to detect accumulated ice on known structures. The icing of marine vessels and offshore structures causes significant reductions in their efficiency and creates unsafe working conditions. Image processing methods are used to measure ice loads automatically. Most image processing methods are developed based on captured image analyses. In this method, ice loads on structures are calculated by defining structure coordinates and processing captured images. A pyramidal structure is designed with nine cylindrical bars as the known structure of experimental setup. Unsymmetrical ice accumulated on the structure in a cold room represents the actual case of experiments. Camera intrinsic and extrinsic parameters are used to define structure coordinates in the image coordinate system according to the camera location and angle. The thresholding method is applied to capture images and detect iced structures in a binary image. The ice thickness of each element is calculated by combining the information from the binary image and the structure coordinate. Averaging ice diameters from different camera views obtains ice thicknesses of structure elements. Comparison between ice load measurements using this method and the actual ice loads shows positive correlations with an acceptable range of error. The method can be applied to complex structures defining structure and camera coordinates.

Keywords: camera calibration, ice detection, ice load measurements, image processing

Procedia PDF Downloads 368
18344 Collaborative Implementation of Master Plans in Afghanistan's Context Considering Land Readjustment as Case Study

Authors: Ahmad Javid Habib, Tetsuo Kidokoro

Abstract:

There is an increasing demand for developing urban land to provide better living conditions for all citizens in Afghanistan. Most of the development will involve the acquisition of land. And the current land acquisition method practiced by central government is expropriation, which is a cash-based transaction method that imposes heavy fiscal burden on local municipalities and central government, and it does not protect ownership rights and social equity of landowners besides it relocates the urban poor to remote areas with limited access to jobs and public services. The questionnaire analysis, backed by observations of different case studies in countries where land readjustment is used as a collaborative land development tool indicates that the method plays a key role in valuing landowners’ rights, giving other community members and stakeholders the opportunity to collaboratively implement urban development projects. The practice of the method is reducing the heavy fiscal burden on the local and central governments and is a better option to deal with the current development challenges in Afghanistan.

Keywords: collaboration, land readjustment, master plan, expropriation

Procedia PDF Downloads 295