Search results for: evolutionary optimization techniques
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 9536

Search results for: evolutionary optimization techniques

8696 The Effects of Applying Linguistic Principles and Teaching Techniques in Teaching English at Secondary School in Thailand

Authors: Wannakarn Likitrattanaporn

Abstract:

The purposes of this investigation were to investigate the effects of applying linguistic principles and teaching techniques in teaching English through experimenting the Adapted English Lessons and to determine the teachers’ opinions as well as students’ opinions towards the Adapted Lessons. The subjects of the study were 5 Thai teachers, who teach English, and 85 Grade 10 mixed-ability students at Triamudom Suksa Pattanakarn Ratchada School, Bangkok, Thailand. The research instruments included the Adapted English Lessons, questionnaires asking teachers’ and students’ opinions towards the Adapted Lessons and the informal interview. The data from the research instruments was collected and analyzed concerning the teachers’ and students’ opinions towards adapting linguistic principles and teaching techniques. Linguistic principles of minimal pair and articulatory phonetics and teaching techniques of mimicry-memorization; vocabulary substitution drills, language pattern drills, reading comprehension exercise, practicing listening, speaking and writing skill and communicative activities; informal talk and free writing are applied. The data was statistically compiled according to an arithmetic percentage. The results showed that the teachers and students have very highly positive opinions towards adapting linguistic principles for teaching and learning phonological accuracy. Teaching techniques provided in the Adapted English Lessons can be used efficiently in the classroom. The teachers and students have positive opinions towards them too.

Keywords: applying linguistic principles and teaching techniques, teachers’ and students’ opinions, teaching English, the adapted English lessons

Procedia PDF Downloads 468
8695 An Integrated Approach to the Carbonate Reservoir Modeling: Case Study of the Eastern Siberia Field

Authors: Yana Snegireva

Abstract:

Carbonate reservoirs are known for their heterogeneity, resulting from various geological processes such as diagenesis and fracturing. These complexities may cause great challenges in understanding fluid flow behavior and predicting the production performance of naturally fractured reservoirs. The investigation of carbonate reservoirs is crucial, as many petroleum reservoirs are naturally fractured, which can be difficult due to the complexity of their fracture networks. This can lead to geological uncertainties, which are important for global petroleum reserves. The problem outlines the key challenges in carbonate reservoir modeling, including the accurate representation of fractures and their connectivity, as well as capturing the impact of fractures on fluid flow and production. Traditional reservoir modeling techniques often oversimplify fracture networks, leading to inaccurate predictions. Therefore, there is a need for a modern approach that can capture the complexities of carbonate reservoirs and provide reliable predictions for effective reservoir management and production optimization. The modern approach to carbonate reservoir modeling involves the utilization of the hybrid fracture modeling approach, including the discrete fracture network (DFN) method and implicit fracture network, which offer enhanced accuracy and reliability in characterizing complex fracture systems within these reservoirs. This study focuses on the application of the hybrid method in the Nepsko-Botuobinskaya anticline of the Eastern Siberia field, aiming to prove the appropriateness of this method in these geological conditions. The DFN method is adopted to model the fracture network within the carbonate reservoir. This method considers fractures as discrete entities, capturing their geometry, orientation, and connectivity. But the method has significant disadvantages since the number of fractures in the field can be very high. Due to limitations in the amount of main memory, it is very difficult to represent these fractures explicitly. By integrating data from image logs (formation micro imager), core data, and fracture density logs, a discrete fracture network (DFN) model can be constructed to represent fracture characteristics for hydraulically relevant fractures. The results obtained from the DFN modeling approaches provide valuable insights into the East Siberia field's carbonate reservoir behavior. The DFN model accurately captures the fracture system, allowing for a better understanding of fluid flow pathways, connectivity, and potential production zones. The analysis of simulation results enables the identification of zones of increased fracturing and optimization opportunities for reservoir development with the potential application of enhanced oil recovery techniques, which were considered in further simulations on the dual porosity and dual permeability models. This approach considers fractures as separate, interconnected flow paths within the reservoir matrix, allowing for the characterization of dual-porosity media. The case study of the East Siberia field demonstrates the effectiveness of the hybrid model method in accurately representing fracture systems and predicting reservoir behavior. The findings from this study contribute to improved reservoir management and production optimization in carbonate reservoirs with the use of enhanced and improved oil recovery methods.

Keywords: carbonate reservoir, discrete fracture network, fracture modeling, dual porosity, enhanced oil recovery, implicit fracture model, hybrid fracture model

Procedia PDF Downloads 66
8694 Investigating the Effect of VR, Time Study and Ergonomics on the Design of Industrial Workstations

Authors: Aydin Azizi, Poorya Ghafoorpoor Yazdi

Abstract:

This paper presents the review of the studies on the ergonomics, virtual reality, and work measurement (time study) at the industrial workstations because each of these three individual techniques can be used to improve the design of workstations and task position. The objective of this paper is to give an overall literature review that if there is any relation between these three different techniques. Therefore, it is so important to review the scientific studies to find a better and effective way for improving design of workstations. On the other hand, manufacturers found that instead of using one of the approaches, utilizing the combination of these individual techniques are more effective to reduce the cost and production time.

Keywords: ergonomics, time study, virtual reality, workplace

Procedia PDF Downloads 114
8693 A Review on Artificial Neural Networks in Image Processing

Authors: B. Afsharipoor, E. Nazemi

Abstract:

Artificial neural networks (ANNs) are powerful tool for prediction which can be trained based on a set of examples and thus, it would be useful for nonlinear image processing. The present paper reviews several paper regarding applications of ANN in image processing to shed the light on advantage and disadvantage of ANNs in this field. Different steps in the image processing chain including pre-processing, enhancement, segmentation, object recognition, image understanding and optimization by using ANN are summarized. Furthermore, results on using multi artificial neural networks are presented.

Keywords: neural networks, image processing, segmentation, object recognition, image understanding, optimization, MANN

Procedia PDF Downloads 390
8692 Obtaining Constants of Johnson-Cook Material Model Using a Combined Experimental, Numerical Simulation and Optimization Method

Authors: F. Rahimi Dehgolan, M. Behzadi, J. Fathi Sola

Abstract:

In this article, the Johnson-Cook material model’s constants for structural steel ST.37 have been determined by a method which integrates experimental tests, numerical simulation, and optimization. In the first step, a quasi-static test was carried out on a plain specimen. Next, the constants were calculated for it by minimizing the difference between the results acquired from the experiment and numerical simulation. Then, a quasi-static tension test was performed on three notched specimens with different notch radii. At last, in order to verify the results, they were used in numerical simulation of notched specimens and it was observed that experimental and simulation results are in good agreement. Changing the diameter size of the plain specimen in the necking area was set as the objective function in the optimization step. For final validation of the proposed method, diameter variation was considered as a parameter and its sensitivity to a change in any of the model constants was examined and the results were completely corroborating.

Keywords: constants, Johnson-Cook material model, notched specimens, quasi-static test, sensitivity

Procedia PDF Downloads 293
8691 Modelling and Optimisation of Floating Drum Biogas Reactor

Authors: L. Rakesh, T. Y. Heblekar

Abstract:

This study entails the development and optimization of a mathematical model for a floating drum biogas reactor from first principles using thermal and empirical considerations. The model was derived on the basis of mass conservation, lumped mass heat transfer formulations and empirical biogas formation laws. The treatment leads to a system of coupled nonlinear ordinary differential equations whose solution mapped four-time independent controllable parameters to five output variables which adequately serve to describe the reactor performance. These equations were solved numerically using fourth order Runge-Kutta method for a range of input parameter values. Using the data so obtained an Artificial Neural Network with a single hidden layer was trained using Levenberg-Marquardt Damped Least Squares (DLS) algorithm. This network was then fine-tuned for optimal mapping by varying hidden layer size. This fast forward model was then employed as a health score generator in the Bacterial Foraging Optimization code. The optimal operating state of the simplified Biogas reactor was thus obtained.

Keywords: biogas, floating drum reactor, neural network model, optimization

Procedia PDF Downloads 132
8690 Multi-Criteria Decision Making Approaches for Facility Planning Problem Evaluation: A Survey

Authors: Ahmed M. El-Araby, Ibrahim Sabry, Ahmed El-Assal

Abstract:

The relationships between the industrial facilities, the capacity available for these facilities, and the costs involved are the main factors in deciding the correct selection of a facility layout. In general, an issue of facility layout is considered to be an unstructured problem of decision-making. The objective of this work is to provide a survey that describes the techniques by which a facility planning problem can be solved and also the effect of these techniques on the efficiency of the layout. The multi-criteria decision making (MCDM) techniques can be classified according to the previous researches into three categories which are the use of single MCDM, combining two or more MCDM, and the integration of MCDM with another technique such as genetic algorithms (GA). This paper presents a review of different multi-criteria decision making (MCDM) techniques that have been proposed in the literature to pick the most suitable layout design. These methods are particularly suitable to deal with complex situations, including various criteria and conflicting goals which need to be optimized simultaneously.

Keywords: facility layout, MCDM, GA, literature review

Procedia PDF Downloads 191
8689 Enhancing Model Interoperability and Reuse by Designing and Developing a Unified Metamodel Standard

Authors: Arash Gharibi

Abstract:

Mankind has always used models to solve problems. Essentially, models are simplified versions of reality, whose need stems from having to deal with complexity; many processes or phenomena are too complex to be described completely. Thus a fundamental model requirement is that it contains the characteristic features that are essential in the context of the problem to be solved or described. Models are used in virtually every scientific domain to deal with various problems. During the recent decades, the number of models has increased exponentially. Publication of models as part of original research has traditionally been in in scientific periodicals, series, monographs, agency reports, national journals and laboratory reports. This makes it difficult for interested groups and communities to stay informed about the state-of-the-art. During the modeling process, many important decisions are made which impact the final form of the model. Without a record of these considerations, the final model remains ill-defined and open to varying interpretations. Unfortunately, the details of these considerations are often lost or in case there is any existing information about a model, it is likely to be written intuitively in different layouts and in different degrees of detail. In order to overcome these issues, different domains have attempted to implement their own approaches to preserve their models’ information in forms of model documentation. The most frequently cited model documentation approaches show that they are domain specific, not to applicable to the existing models and evolutionary flexibility and intrinsic corrections and improvements are not possible with the current approaches. These issues are all because of a lack of unified standards for model documentation. As a way forward, this research will propose a new standard for capturing and managing models’ information in a unified way so that interoperability and reusability of models become possible. This standard will also be evolutionary, meaning members of modeling realm could contribute to its ongoing developments and improvements. In this paper, the current 3 of the most common metamodels are reviewed and according to pros and cons of each, a new metamodel is proposed.

Keywords: metamodel, modeling, interoperability, reuse

Procedia PDF Downloads 188
8688 A Mixture Vine Copula Structures Model for Dependence Wind Speed among Wind Farms and Its Application in Reactive Power Optimization

Authors: Yibin Qiu, Yubo Ouyang, Shihan Li, Guorui Zhang, Qi Li, Weirong Chen

Abstract:

This paper aims at exploring the impacts of high dimensional dependencies of wind speed among wind farms on probabilistic optimal power flow. To obtain the reactive power optimization faster and more accurately, a mixture vine Copula structure model combining the K-means clustering, C vine copula and D vine copula is proposed in this paper, through which a more accurate correlation model can be obtained. Moreover, a Modified Backtracking Search Algorithm (MBSA), the three-point estimate method is applied to probabilistic optimal power flow. The validity of the mixture vine copula structure model and the MBSA are respectively tested in IEEE30 node system with measured data of 3 adjacent wind farms in a certain area, and the results indicate effectiveness of these methods.

Keywords: mixture vine copula structure model, three-point estimate method, the probability integral transform, modified backtracking search algorithm, reactive power optimization

Procedia PDF Downloads 243
8687 An Optimization Modelling to Evaluate Flights Scheduling at Tourist Airports

Authors: Dimitrios J. Dimitriou

Abstract:

Airport’s serving a tourist destination are an essential counterpart of the tourist demand supply chain, and their productivity is related to the region’s attractiveness and is enhanced by the air transport business. In this paper, the evaluation framework of the scheduled flights between two tourist airports is taken into consideration. By adopting a systemic approach, the arrivals from an airport that its connectivity heavily depended on the departures of another major airport are reviewed. The methodology framework, based on inventory control theory and the numerical example, promotes the use of the modelling formulation. The results would be essential for comparison and exercising to other similar cases.

Keywords: airport connectivity, inventory control, optimization, optimum allocation

Procedia PDF Downloads 322
8686 Process Optimization and Automation of Information Technology Services in a Heterogenic Digital Environment

Authors: Tasneem Halawani, Yamen Khateeb

Abstract:

With customers’ ever-increasing expectations for fast services provisioning for all their business needs, information technology (IT) organizations, as business partners, have to cope with this demanding environment and deliver their services in the most effective and efficient way. The purpose of this paper is to identify optimization and automation opportunities for the top requested IT services in a heterogenic digital environment and widely spread customer base. In collaboration with systems, processes, and subject matter experts (SMEs), the processes in scope were approached by analyzing four-year related historical data, identifying and surveying stakeholders, modeling the as-is processes, and studying systems integration/automation capabilities. This effort resulted in identifying several pain areas, including standardization, unnecessary customer and IT involvement, manual steps, systems integration, and performance measurement. These pain areas were addressed by standardizing the top five requested IT services, eliminating/automating 43 steps, and utilizing a single platform for end-to-end process execution. In conclusion, the optimization of IT service request processes in a heterogenic digital environment and widely spread customer base is challenging, yet achievable without compromising the service quality and customers’ added value. Further studies can focus on measuring the value of the eliminated/automated process steps to quantify the enhancement impact. Moreover, a similar approach can be utilized to optimize other IT service requests, with a focus on business criticality.

Keywords: automation, customer value, heterogenic, integration, IT services, optimization, processes

Procedia PDF Downloads 100
8685 Meta-Analysis Comparing the Femoral Tunnel Length, Femoral Tunnel Position and Graft Bending Angle of Transtibial, Anteromedial and Outside-In Techniques for Single-Bundle Anterior Cruciate Ligament Reconstruction

Authors: Andrew Tan Hwee Chye, Yeo Zhen Ning

Abstract:

This study aims to meta-analyse clinical studies comparing femoral tunnel position (FTP), femoral tunnel length (FTL) and graft bending angle (GBA) of single-bundle Anterior Cruciate Ligament (ACL) reconstruction using Transtibial (TT), Anteromedial (AM) and Outside-in (OI) techniques. A meta-analysis comparing the FTP, FTL and GBA of single-bundle ACL reconstruction utilising the TT, AM and OI was performed. Prospective Comparative Studies (PCS) and Retrospective Comparative Studies (RCS) from PubMed, Cochrane Library, and Embase were included. A total of 17 studies were included in this study. TT had the longest FTL, when compared to AM (Mean difference = 7.38, 95% CI: 3.76 to 11.00, P < 0.001) and OI (Mean difference = 9.47, 95% CI: 4.89 to 14.05, P < 0.001). In the deep-to-shallow direction, the OI resulted in a significantly deeper femoral tunnel as compared to the TT (Mean difference = 4.36, 95% CI: 1.39 to 7.33, P = 0.004) (Figure 6B). The AM technique also contributed to a significantly lower tunnel position as compared to the OI technique (Mean difference = 2.34, 95% CI: 0.76 to 3.92, P = 0.004). There were no significant differences in the graft bending angle between TT, AM and OI techniques. AM and OI techniques provide a more anatomical position as compared to the TT. Although FTL in the TT is longer than the AM and OI, all three techniques exceed the critical length of 25mm. There are no differences in the GBA between the three techniques.

Keywords: femoral tunnel position, femoral tunnel length, anterior cruciate ligament, transtibial, graft bending angle, anteromedial, outside-in

Procedia PDF Downloads 112
8684 Ant Colony Optimization Control for Multilevel STATCOM

Authors: H. Tédjini, Y. Meslem, B. Guesbaoui, A. Safa

Abstract:

Flexible AC Transmission Systems (FACTS) are potentially becoming more flexible and more economical local controllers in the power system; and because of the high MVA ratings, it would be expensive to provide independent, equal, regulated DC voltage sources to power the multilevel converters which are presently proposed for STATCOMs. DC voltage sources can be derived from the DC link capacitances which are charged by the rectified ac power. In this paper a new stronger control combined of nonlinear control based Lyapunov’s theorem and Ant Colony Algorithm (ACA) to maintain stability of multilevel STATCOM and the utility.

Keywords: Static Compensator (STATCOM), ant colony optimization (ACO), lyapunov control theory, Decoupled power control, neutral point clamped (NPC)

Procedia PDF Downloads 544
8683 Dogs Chest Homogeneous Phantom for Image Optimization

Authors: Maris Eugênia Dela Rosa, Ana Luiza Menegatti Pavan, Marcela De Oliveira, Diana Rodrigues De Pina, Luis Carlos Vulcano

Abstract:

In medical veterinary as well as in human medicine, radiological study is essential for a safe diagnosis in clinical practice. Thus, the quality of radiographic image is crucial. In last year’s there has been an increasing substitution of image acquisition screen-film systems for computed radiology equipment (CR) without technical charts adequacy. Furthermore, to carry out a radiographic examination in veterinary patient is required human assistance for restraint this, which can compromise image quality by generating dose increasing to the animal, for Occupationally Exposed and also the increased cost to the institution. The image optimization procedure and construction of radiographic techniques are performed with the use of homogeneous phantoms. In this study, we sought to develop a homogeneous phantom of canine chest to be applied to the optimization of these images for the CR system. In carrying out the simulator was created a database with retrospectives chest images of computed tomography (CT) of the Veterinary Hospital of the Faculty of Veterinary Medicine and Animal Science - UNESP (FMVZ / Botucatu). Images were divided into four groups according to the animal weight employing classification by sizes proposed by Hoskins & Goldston. The thickness of biological tissues were quantified in a 80 animals, separated in groups of 20 animals according to their weights: (S) Small - equal to or less than 9.0 kg, (M) Medium - between 9.0 and 23.0 kg, (L) Large – between 23.1 and 40.0kg and (G) Giant – over 40.1 kg. Mean weight for group (S) was 6.5±2.0 kg, (M) 15.0±5.0 kg, (L) 32.0±5.5 kg and (G) 50.0 ±12.0 kg. An algorithm was developed in Matlab in order to classify and quantify biological tissues present in CT images and convert them in simulator materials. To classify tissues presents, the membership functions were created from the retrospective CT scans according to the type of tissue (adipose, muscle, bone trabecular or cortical and lung tissue). After conversion of the biologic tissue thickness in equivalent material thicknesses (acrylic simulating soft tissues, bone tissues simulated by aluminum and air to the lung) were obtained four different homogeneous phantoms, with (S) 5 cm of acrylic, 0,14 cm of aluminum and 1,8 cm of air; (M) 8,7 cm of acrylic, 0,2 cm of aluminum and 2,4 cm of air; (L) 10,6 cm of acrylic, 0,27 cm of aluminum and 3,1 cm of air and (G) 14,8 cm of acrylic, 0,33 cm of aluminum and 3,8 cm of air. The developed canine homogeneous phantom is a practical tool, which will be employed in future, works to optimize veterinary X-ray procedures.

Keywords: radiation protection, phantom, veterinary radiology, computed radiography

Procedia PDF Downloads 409
8682 Optimization of Titanium Leaching Process Using Experimental Design

Authors: Arash Rafiei, Carroll Moore

Abstract:

Leaching process as the first stage of hydrometallurgy is a multidisciplinary system including material properties, chemistry, reactor design, mechanics and fluid dynamics. Therefore, doing leaching system optimization by pure scientific methods need lots of times and expenses. In this work, a mixture of two titanium ores and one titanium slag are used for extracting titanium for leaching stage of TiO2 pigment production procedure. Optimum titanium extraction can be obtained from following strategies: i) Maximizing titanium extraction without selective digestion; and ii) Optimizing selective titanium extraction by balancing between maximum titanium extraction and minimum impurity digestion. The main difference between two strategies is due to process optimization framework. For the first strategy, the most important stage of production process is concerned as the main stage and rest of stages would be adopted with respect to the main stage. The second strategy optimizes performance of more than one stage at once. The second strategy has more technical complexity compared to the first one but it brings more economical and technical advantages for the leaching system. Obviously, each strategy has its own optimum operational zone that is not as same as the other one and the best operational zone is chosen due to complexity, economical and practical aspects of the leaching system. Experimental design has been carried out by using Taguchi method. The most important advantages of this methodology are involving different technical aspects of leaching process; minimizing the number of needed experiments as well as time and expense; and concerning the role of parameter interactions due to principles of multifactor-at-time optimization. Leaching tests have been done at batch scale on lab with appropriate control on temperature. The leaching tank geometry has been concerned as an important factor to provide comparable agitation conditions. Data analysis has been done by using reactor design and mass balancing principles. Finally, optimum zone for operational parameters are determined for each leaching strategy and discussed due to their economical and practical aspects.

Keywords: titanium leaching, optimization, experimental design, performance analysis

Procedia PDF Downloads 360
8681 Extraction of Essential Oil From Orange Peels

Authors: Aayush Bhisikar, Neha Rajas, Aditya Bhingare, Samarth Bhandare, Amruta Amrurkar

Abstract:

Orange peels are currently thrown away as garbage in India after orange fruits' edible components are consumed. However, the nation depends on important essential oils for usage in companies that produce goods, including food, beverages, cosmetics, and medicines. This study was conducted to show how to effectively use it. By using various extraction techniques, orange peel is used in the creation of essential oils. Stream distillation, water distillation, and solvent extraction were the techniques taken into consideration in this paper. Due to its relative prevalence among the extraction techniques, Design Expert 7.0 was used to plan an experimental run for solvent extraction. Oil was examined to ascertain its physical and chemical characteristics after extraction. It was determined from the outcomes that the orange peels.

Keywords: orange peels, extraction, essential oil, distillation

Procedia PDF Downloads 65
8680 Extraction of Essential Oil from Orange Peels

Authors: Neha Rajas, Aayush Bhisikar, Samarth Bhandare, Aditya Bhingare, Amruta Amrutkar

Abstract:

Orange peels are currently thrown away as garbage in India after orange fruits' edible components are consumed. However, the nation depends on important essential oils for usage in companies that produce goods, including food, beverages, cosmetics, and medicines. This study was conducted to show how to effectively use it. By using various extraction techniques, orange peel is used in the creation of essential oils. Stream distillation, water distillation, and solvent extraction were the techniques taken into consideration in this paper. Due to its relative prevalence among the extraction techniques, Design Expert 7.0 was used to plan an experimental run for solvent extraction. Oil was examined to ascertain its physical and chemical characteristics after extraction. It was determined from the outcomes that the orange peels.

Keywords: orange peels, extraction, distillation, essential oil

Procedia PDF Downloads 60
8679 Check Red Blood Cells Concentrations of a Blood Sample by Using Photoconductive Antenna

Authors: Ahmed Banda, Alaa Maghrabi, Aiman Fakieh

Abstract:

Terahertz (THz) range lies in the area between 0.1 to 10 THz. The process of generating and detecting THz can be done through different techniques. One of the most familiar techniques is done through a photoconductive antenna (PCA). The process of generating THz radiation at PCA includes applying a laser pump in femtosecond and DC voltage difference. However, photocurrent is generated at PCA, which its value is affected by different parameters (e.g., dielectric properties, DC voltage difference and incident power of laser pump). THz radiation is used for biomedical applications. However, different biomedical fields need new technologies to meet patients’ needs (e.g. blood-related conditions). In this work, a novel method to check the red blood cells (RBCs) concentration of a blood sample using PCA is presented. RBCs constitute 44% of total blood volume. RBCs contain Hemoglobin that transfers oxygen from lungs to body organs. Then it returns to the lungs carrying carbon dioxide, which the body then gets rid of in the process of exhalation. The configuration has been simulated and optimized using COMSOL Multiphysics. The differentiation of RBCs concentration affects its dielectric properties (e.g., the relative permittivity of RBCs in the blood sample). However, the effects of four blood samples (with different concentrations of RBCs) on photocurrent value have been tested. Photocurrent peak value and RBCs concentration are inversely proportional to each other due to the change of dielectric properties of RBCs. It was noticed that photocurrent peak value has dropped from 162.99 nA to 108.66 nA when RBCs concentration has risen from 0% to 100% of a blood sample. The optimization of this method helps to launch new products for diagnosing blood-related conditions (e.g., anemia and leukemia). The resultant electric field from DC components can not be used to count the RBCs of the blood sample.

Keywords: biomedical applications, photoconductive antenna, photocurrent, red blood cells, THz radiation

Procedia PDF Downloads 194
8678 Constraint-Directed Techniques for Transport Scheduling with Capacity Restrictions of Automotive Manufacturing Components

Authors: Martha Ndeley, John Ikome

Abstract:

In this paper, we expand the scope of constraint-directed techniques to deal with the case of transportation schedule with capacity restrictions where the scheduling problem includes alternative activities. That is, not only does the scheduling problem consist of determining when an activity is to be executed, but also determining which set of alternative activities is to be executed at all level of transportation from input to output. Such problems encompass both alternative resource problems and alternative process plan problems. We formulate a constraint-based representation of alternative activities to model problems containing such choices. We then extend existing constraint-directed scheduling heuristic commitment techniques and propagators to reason directly about the fact that an activity does not necessarily have to exist in a final transportation schedule without being completed. Tentative results show that an algorithm using a novel texture-based heuristic commitment technique propagators achieves the best overall performance of the techniques tested.

Keywords: production, transportation, scheduling, integrated

Procedia PDF Downloads 346
8677 Energy Saving Techniques for MIMO Decoders

Authors: Zhuofan Cheng, Qiongda Hu, Mohammed El-Hajjar, Basel Halak

Abstract:

Multiple-input multiple-output (MIMO) systems can allow significantly higher data rates compared to single-antenna-aided systems. They are expected to be a prominent part of the 5G communication standard. However, these decoders suffer from high power consumption. This work presents a design technique in order to improve the energy efficiency of MIMO systems; this facilitates their use in the next generation of battery-operated communication devices such as mobile phones and tablets. The proposed optimization approach consists of the use of low complexity lattice reduction algorithm in combination with an adaptive VLSI implementation. The proposed design has been realized and verified in 65nm technology. The results show that the proposed design is significantly more energy-efficient than conventional K-best MIMO systems.

Keywords: energy, lattice reduction, MIMO, VLSI

Procedia PDF Downloads 320
8676 Cognitive Model of Analogy Based on Operation of the Brain Cells: Glial, Axons and Neurons

Authors: Ozgu Hafizoglu

Abstract:

Analogy is an essential tool of human cognition that enables connecting diffuse and diverse systems with attributional, deep structural, casual relations that are essential to learning, to innovation in artificial worlds, and to discovery in science. Cognitive Model of Analogy (CMA) leads and creates information pattern transfer within and between domains and disciplines in science. This paper demonstrates the Cognitive Model of Analogy (CMA) as an evolutionary approach to scientific research. The model puts forward the challenges of deep uncertainty about the future, emphasizing the need for flexibility of the system in order to enable reasoning methodology to adapt to changing conditions. In this paper, the model of analogical reasoning is created based on brain cells, their fractal, and operational forms within the system itself. Visualization techniques are used to show correspondences. Distinct phases of the problem-solving processes are divided thusly: encoding, mapping, inference, and response. The system is revealed relevant to brain activation considering each of these phases with an emphasis on achieving a better visualization of the brain cells: glial cells, axons, axon terminals, and neurons, relative to matching conditions of analogical reasoning and relational information. It’s found that encoding, mapping, inference, and response processes in four-term analogical reasoning are corresponding with the fractal and operational forms of brain cells: glial, axons, and neurons.

Keywords: analogy, analogical reasoning, cognitive model, brain and glials

Procedia PDF Downloads 174
8675 Multivariate Analysis on Water Quality Attributes Using Master-Slave Neural Network Model

Authors: A. Clementking, C. Jothi Venkateswaran

Abstract:

Mathematical and computational functionalities such as descriptive mining, optimization, and predictions are espoused to resolve natural resource planning. The water quality prediction and its attributes influence determinations are adopted optimization techniques. The water properties are tainted while merging water resource one with another. This work aimed to predict influencing water resource distribution connectivity in accordance to water quality and sediment using an innovative proposed master-slave neural network back-propagation model. The experiment results are arrived through collecting water quality attributes, computation of water quality index, design and development of neural network model to determine water quality and sediment, master–slave back propagation neural network back-propagation model to determine variations on water quality and sediment attributes between the water resources and the recommendation for connectivity. The homogeneous and parallel biochemical reactions are influences water quality and sediment while distributing water from one location to another. Therefore, an innovative master-slave neural network model [M (9:9:2)::S(9:9:2)] designed and developed to predict the attribute variations. The result of training dataset given as an input to master model and its maximum weights are assigned as an input to the slave model to predict the water quality. The developed master-slave model is predicted physicochemical attributes weight variations for 85 % to 90% of water quality as a target values.The sediment level variations also predicated from 0.01 to 0.05% of each water quality percentage. The model produced the significant variations on physiochemical attribute weights. According to the predicated experimental weight variation on training data set, effective recommendations are made to connect different resources.

Keywords: master-slave back propagation neural network model(MSBPNNM), water quality analysis, multivariate analysis, environmental mining

Procedia PDF Downloads 463
8674 Illuminating Human Identity in Theology and Islamic Philosophy

Authors: Khan Shahid, Shahid Zakia

Abstract:

The article demonstrates how Theology and Islamic Philosophy can be illuminated and enhanced through the application of the SOUL framework (Sincere act, Optimization effort, Ultimate goal, Law compliance). The study explores historical development using a phenomenological approach and integrates the SOUL framework to enrich Theology and Islamic Philosophy. The proposed framework highlights the significance of these elements, ultimately leading to a deeper understanding of Theology and Islamic Philosophy.

Keywords: SOUL framework, illuminating human identity, theology, Islamic Philosophy, sincerity act, optimization effort, ultimate goals, law compliance

Procedia PDF Downloads 80
8673 The Optimization of Copper Sulfate and Tincalconite Molar Ratios on the Hydrothermal Synthesis of Copper Borates

Authors: E. Moroydor Derun, N. Tugrul, F. T. Senberber, A. S. Kipcak, S. Piskin

Abstract:

In this research, copper borates are synthesized by the reaction of copper sulfate pentahydrate (CuSO4.5H2O) and tincalconite (Na2O4B7.10H2O). The experimental parameters are selected as 80°C reaction temperature and 60 of reaction time. The effect of mole ratio of CuSO4.5H2O to Na2O4B7.5H2O is studied. For the identification analyses X-Ray Diffraction (XRD) and Fourier Transform Infrared Spectroscopy (FT-IR) techniques are used. At the end of the experiments, synthesized copper borate is matched with the powder diffraction file of “00-001-0472” [Cu(BO2)2] and characteristic vibrations between B and O atoms are seen. The proper crystals are obtained at the mole ratio of 3:1. This study showed that simplified synthesis process is suitable for the production of copper borate minerals.

Keywords: hydrothermal synthesis, copper borates, copper sulfate, tincalconite

Procedia PDF Downloads 369
8672 Hybrid Gravity Gradient Inversion-Ant Colony Optimization Algorithm for Motion Planning of Mobile Robots

Authors: Meng Wu

Abstract:

Motion planning is a common task required to be fulfilled by robots. A strategy combining Ant Colony Optimization (ACO) and gravity gradient inversion algorithm is proposed for motion planning of mobile robots. In this paper, in order to realize optimal motion planning strategy, the cost function in ACO is designed based on gravity gradient inversion algorithm. The obstacles around mobile robot can cause gravity gradient anomalies; the gradiometer is installed on the mobile robot to detect the gravity gradient anomalies. After obtaining the anomalies, gravity gradient inversion algorithm is employed to calculate relative distance and orientation between mobile robot and obstacles. The relative distance and orientation deduced from gravity gradient inversion algorithm is employed as cost function in ACO algorithm to realize motion planning. The proposed strategy is validated by the simulation and experiment results.

Keywords: motion planning, gravity gradient inversion algorithm, ant colony optimization

Procedia PDF Downloads 129
8671 Optimizing the Public Policy Information System under the Environment of E-Government

Authors: Qian Zaijian

Abstract:

E-government is one of the hot issues in the current academic research of public policy and management. As the organic integration of information and communication technology (ICT) and public administration, e-government is one of the most important areas in contemporary information society. Policy information system is a basic subsystem of public policy system, its operation affects the overall effect of the policy process or even exerts a direct impact on the operation of a public policy and its success or failure. The basic principle of its operation is information collection, processing, analysis and release for a specific purpose. The function of E-government for public policy information system lies in the promotion of public access to the policy information resources, information transmission through e-participation, e-consultation in the process of policy analysis and processing of information and electronic services in policy information stored, to promote the optimization of policy information systems. However, due to many factors, the function of e-government to promote policy information system optimization has its practical limits. In the building of E-government in our country, we should take such path as adhering to the principle of freedom of information, eliminating the information divide (gap), expanding e-consultation, breaking down information silos and other major path, so as to promote the optimization of public policy information systems.

Keywords: China, e-consultation, e-democracy, e-government, e-participation, ICTs, public policy information systems

Procedia PDF Downloads 846
8670 A Deep Learning Approach for Optimum Shape Design

Authors: Cahit Perkgöz

Abstract:

Artificial intelligence has brought new approaches to solving problems in almost every research field in recent years. One of these topics is shape design and optimization, which has the possibility of applications in many fields, such as nanotechnology and electronics. A properly constructed cost function can eliminate the need for labeled data required in deep learning and create desired shapes. In this work, the network parameters are optimized differentially, which differs from traditional approaches. The methods are tested for physics-related structures and successful results are obtained. This work is supported by Eskişehir Technical University scientific research project (Project No: 20ADP090)

Keywords: deep learning, shape design, optimization, artificial intelligence

Procedia PDF Downloads 141
8669 Optimization of Soybean Oil by Modified Supercritical Carbon Dioxide

Authors: N. R. Putra, A. H. Abdul Aziz, A. S. Zaini, Z. Idham, F. Idrus, M. Z. Bin Zullyadini, M. A. Che Yunus

Abstract:

The content of omega-3 in soybean oil is important in the development of infants and is an alternative for the omega-3 in fish oils. The investigation of extraction of soybean oil is needed to obtain the bioactive compound in the extract. Supercritical carbon dioxide extraction is modern and green technology to extract herbs and plants to obtain high quality extract due to high diffusivity and solubility of the solvent. The aim of this study was to obtain the optimum condition of soybean oil extraction by modified supercritical carbon dioxide. The soybean oil was extracted by using modified supercritical carbon dioxide (SC-CO2) under the temperatures of 40, 60, 80 °C, pressures of 150, 250, 350 Bar, and constant flow-rate of 10 g/min as the parameters of extraction processes. An experimental design was performed in order to optimize three important parameters of SC-CO2 extraction which are pressure (X1), temperature (X2) to achieve optimum yields of soybean oil. Box Behnken Design was applied for experimental design. From the optimization process, the optimum condition of extraction of soybean oil was obtained at pressure 338 Bar and temperature 80 °C with oil yield of 2.713 g. Effect of pressure is significant on the extraction of soybean oil by modified supercritical carbon dioxide. Increasing of pressure will increase the oil yield of soybean oil.

Keywords: soybean oil, SC-CO₂ extraction, yield, optimization

Procedia PDF Downloads 243
8668 Data Mining and Knowledge Management Application to Enhance Business Operations: An Exploratory Study

Authors: Zeba Mahmood

Abstract:

The modern business organizations are adopting technological advancement to achieve competitive edge and satisfy their consumer. The development in the field of Information technology systems has changed the way of conducting business today. Business operations today rely more on the data they obtained and this data is continuously increasing in volume. The data stored in different locations is difficult to find and use without the effective implementation of Data mining and Knowledge management techniques. Organizations who smartly identify, obtain and then convert data in useful formats for their decision making and operational improvements create additional value for their customers and enhance their operational capabilities. Marketers and Customer relationship departments of firm use Data mining techniques to make relevant decisions, this paper emphasizes on the identification of different data mining and Knowledge management techniques that are applied to different business industries. The challenges and issues of execution of these techniques are also discussed and critically analyzed in this paper.

Keywords: knowledge, knowledge management, knowledge discovery in databases, business, operational, information, data mining

Procedia PDF Downloads 521
8667 Improved Multi-Objective Particle Swarm Optimization Applied to Design Problem

Authors: Kapse Swapnil, K. Shankar

Abstract:

Aiming at optimizing the weight and deflection of cantilever beam subjected to maximum stress and maximum deflection, Multi-objective Particle Swarm Optimization (MOPSO) with Utopia Point based local search is implemented. Utopia point is used to govern the search towards the Pareto Optimal set. The elite candidates obtained during the iterations are stored in an archive according to non-dominated sorting and also the archive is truncated based on least crowding distance. Local search is also performed on elite candidates and the most diverse particle is selected as the global best. This method is implemented on standard test functions and it is observed that the improved algorithm gives better convergence and diversity as compared to NSGA-II in fewer iterations. Implementation on practical structural problem shows that in 5 to 6 iterations, the improved algorithm converges with better diversity as evident by the improvement of cantilever beam on an average of 0.78% and 9.28% in the weight and deflection respectively compared to NSGA-II.

Keywords: Utopia point, multi-objective particle swarm optimization, local search, cantilever beam

Procedia PDF Downloads 508