Search results for: Artificial Bee Colony Algorithm
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 5481

Search results for: Artificial Bee Colony Algorithm

2631 Normal Hematopoietic Stem Cell and the Toxic Effect of Parthenolide

Authors: Alsulami H., Alghamdi N., Alasker A., Almohen N., Shome D.

Abstract:

Most conventional chemotherapeutic agents which are used for the treatment of cancers not only eradicate cancer cells but also affect normal hematopoietic Stem cells (HSCs) that leads to severe pancytopenia during treatment. Therefore, a need exists for novel approaches to treat cancer without or with minimum effect on normal HSCs. Parthenolide (PTL), a herbal product occurring naturally in the plant Feverfew, is a potential new chemotherapeutic agent for the treatment of many cancers such as acute myeloid leukemia (AML) and chronic lymphocytic leukemia (CLL). In this study we investigated the effect of different PTL concentrations on the viability of normal HSCs and also on the ability of these cells to form colonies after they have been treated with PTL in vitro. Methods: In this study, 24 samples of bone marrow and cord blood were collected with consent, and mononuclear cells were separated using density gradient separation. These cells were then exposed to various concentrations of PTL for 24 hours. Cell viability after culture was determined using 7ADD in a flow cytometry test. Additionally, the impact of PTL on hematopoietic stem cells (HSCs) was evaluated using a colony forming unit assay (CFU). Furthermore, the levels of NFҝB expression were assessed by using a PE-labelled anti-pNFκBP65 antibody. Results: this study showed that there was no statistically significant difference in the percentage of cell death between untreated and PTL treated cells with 5 μM PTL (p = 0.7), 10 μM PTL (p = 0.4) and 25 μM (p = 0.09) respectively. However, at higher doses, PTL caused significant increase in the percentage of cell death. These results were significant when compared to untreated control (p < 0.001). The response of cord blood cells (n=4) on the other hand was slightly different from that for bone marrow cells in that the percentage of cell death was significant at 100 μM PTL. Therefore, cord blood cells seemed more resistant than bone marrow cells. Discussion &Conclusion: At concentrations ≤25 μM PTL has a minimum or no effect on HSCs in vitro. Cord blood HSCs are more resistant to PTL compared to bone marrow HSCs. This could be due to the higher percentage of T-lymphocytes, which are resistant to PTL, in CB samples (85% in CB vs. 56% in BM. Additionally, CB samples contained a higher proportion of CD34+ cells, with 14.5% of brightly CD34+ cells compared to only 1% in normal BM. These bright CD34+ cells in CB were mostly negative for early-stage stem cell maturation antigens, making them young and resilient to oxidative stress and high concentrations of PTL.

Keywords: stem cell, parthenolide, NFKB, CLL

Procedia PDF Downloads 37
2630 A Nonlinear Approach for System Identification of a Li-Ion Battery Based on a Non-Linear Autoregressive Exogenous Model

Authors: Meriem Mossaddek, El Mehdi Laadissi, El Mehdi Loualid, Chouaib Ennawaoui, Sohaib Bouzaid, Abdelowahed Hajjaji

Abstract:

An electrochemical system is a subset of mechatronic systems that includes a wide variety of batteries and nickel-cadmium, lead-acid batteries, and lithium-ion. Those structures have several non-linear behaviors and uncertainties in their running range. This paper studies an effective technique for modeling Lithium-Ion (Li-Ion) batteries using a Nonlinear Auto-Regressive model with exogenous input (NARX). The Artificial Neural Network (ANN) is trained to employ the data collected from the battery testing process. The proposed model is implemented on a Li-Ion battery cell. Simulation of this model in MATLAB shows good accuracy of the proposed model.

Keywords: lithium-ion battery, neural network, energy storage, battery model, nonlinear models

Procedia PDF Downloads 102
2629 Multi-Objectives Genetic Algorithm for Optimizing Machining Process Parameters

Authors: Dylan Santos De Pinho, Nabil Ouerhani

Abstract:

Energy consumption of machine-tools is becoming critical for machine-tool builders and end-users because of economic, ecological and legislation-related reasons. Many machine-tool builders are seeking for solutions that allow the reduction of energy consumption of machine-tools while preserving the same productivity rate and the same quality of machined parts. In this paper, we present the first results of a project conducted jointly by academic and industrial partners to reduce the energy consumption of a Swiss-Type lathe. We employ genetic algorithms to find optimal machining parameters – the set of parameters that lead to the best trade-off between energy consumption, part quality and tool lifetime. Three main machining process parameters are considered in our optimization technique, namely depth of cut, spindle rotation speed and material feed rate. These machining process parameters have been identified as the most influential ones in the configuration of the Swiss-type machining process. A state-of-the-art multi-objective genetic algorithm has been used. The algorithm combines three fitness functions, which are objective functions that permit to evaluate a set of parameters against the three objectives: energy consumption, quality of the machined parts, and tool lifetime. In this paper, we focus on the investigation of the fitness function related to energy consumption. Four different energy consumption related fitness functions have been investigated and compared. The first fitness function refers to the Kienzle cutting force model. The second fitness function uses the Material Removal Rate (RMM) as an indicator of energy consumption. The two other fitness functions are non-deterministic, learning-based functions. One fitness function uses a simple Neural Network to learn the relation between the process parameters and the energy consumption from experimental data. Another fitness function uses Lasso regression to determine the same relation. The goal is, then, to find out which fitness functions predict best the energy consumption of a Swiss-Type machining process for the given set of machining process parameters. Once determined, these functions may be used for optimization purposes – determine the optimal machining process parameters leading to minimum energy consumption. The performance of the four fitness functions has been evaluated. The Tornos DT13 Swiss-Type Lathe has been used to carry out the experiments. A mechanical part including various Swiss-Type machining operations has been selected for the experiments. The evaluation process starts with generating a set of CNC (Computer Numerical Control) programs for machining the part at hand. Each CNC program considers a different set of machining process parameters. During the machining process, the power consumption of the spindle is measured. All collected data are assigned to the appropriate CNC program and thus to the set of machining process parameters. The evaluation approach consists in calculating the correlation between the normalized measured power consumption and the normalized power consumption prediction for each of the four fitness functions. The evaluation shows that the Lasso and Neural Network fitness functions have the highest correlation coefficient with 97%. The fitness function “Material Removal Rate” (MRR) has a correlation coefficient of 90%, whereas the Kienzle-based fitness function has a correlation coefficient of 80%.

Keywords: adaptive machining, genetic algorithms, smart manufacturing, parameters optimization

Procedia PDF Downloads 142
2628 Crushing Analysis of Foam-Filled Thin-Walled Aluminum Profiles Subjected to Axial Loading

Authors: Michał Rogala, Jakub Gajewski

Abstract:

As the automotive industry develops, passive safety is becoming an increasingly important aspect when designing motor vehicles. A commonly used solution is energy absorption by thin-walled construction. One such structure is a closed thin-walled profile fixed to the vehicle stringers. The article presents numerical tests of conical thin-walled profiles filled with aluminum foam. The columns were loaded axially with constant energy. On the basis of the results obtained, efficiency indicators were calculated. The efficiency of the foam filling was evaluated. Artificial neural networks were used for data analysis. The application of regression analysis was used as a tool to study the relationship between the quantities characteristic of the dynamic crush.

Keywords: aluminium foam, crashworthiness, neural networks, thin-walled structure

Procedia PDF Downloads 142
2627 Non-Canonical Beclin-1-Independent Autophagy and Apoptosis in Cell Death Induced by Rhus coriaria in Human Colon HT-29 Cancer Cells

Authors: Rabah Iratni, Husain El Hasasna, Khawlah Athamneh, Halima Al Sameri, Nehla Benhalilou, Asma Al Rashedi

Abstract:

Background: Cancer therapies have witnessed great advances in the recent past, however, cancer continues to be a leading cause of death, with colorectal cancer being the fourth cause of cancer-related deaths. Colorectal cancer affects both sexes equally with poor survival rate once it metastasizes. Phytochemicals, which are plant derived compounds, have been on a steady rise as anti-cancer drugs due to the accumulation of evidences that support their potential. Here, we investigated the anticancer effect of Rhus coriaria on colon cancer cells. Material and Method: Human colon cancer HT-29 cell line was used. Protein expression and protein phosphorylation were examined using Western blotting. Transcription activity was measure using Quantitative RT-PCR. Human tumoral clonogenic assay was used to assess cell survival. Senescence was assessed by the senescence-associated beta-galactosidase assay. Results: Rhus coriaria extract (RCE) was found to significantly inhibit the viability and colony growth of human HT-29 colon cancer cells. RCE induced senescence and cell cycle arrest at G1 phase. These changes were concomitant with upregulation of p21, p16, downregulation of cyclin D1, p27, c-myc and expression of Senescence-associated-β-Galactosidase activity. Moreover, RCE induced non-canonical beclin-1independent autophagy and subsequent apoptotic cell death through activation of activation caspase 8 and caspase 7. The blocking of autophagy by 3-methyladenine (3-MA) or chloroquine (CQ) reduced RCE-induced cell death. Further, RCE induced DNA damage, reduced mutant p53 protein level and downregulated phospho-AKT and phospho-mTOR, events that preceded autophagy. Mechanistically, we found that RCE inhibited the AKT and mTOR pathway, a regulator of autophagy, by promoting the proteasome-dependent degradation of both AKT and mTOR proteins. Conclusion: Our findings provide strong evidence that Rhus coriaria possesses strong anti-colon cancer activity through induction of senescence and autophagic cell death, making it a promising alternative or adjunct therapeutic candidate against colon cancer.

Keywords: autophagy, proteasome degradation, senescence, mTOR, apoptosis, Beclin-1

Procedia PDF Downloads 253
2626 Performance Evaluation of Contemporary Classifiers for Automatic Detection of Epileptic EEG

Authors: K. E. Ch. Vidyasagar, M. Moghavvemi, T. S. S. T. Prabhat

Abstract:

Epilepsy is a global problem, and with seizures eluding even the smartest of diagnoses a requirement for automatic detection of the same using electroencephalogram (EEG) would have a huge impact in diagnosis of the disorder. Among a multitude of methods for automatic epilepsy detection, one should find the best method out, based on accuracy, for classification. This paper reasons out, and rationalizes, the best methods for classification. Accuracy is based on the classifier, and thus this paper discusses classifiers like quadratic discriminant analysis (QDA), classification and regression tree (CART), support vector machine (SVM), naive Bayes classifier (NBC), linear discriminant analysis (LDA), K-nearest neighbor (KNN) and artificial neural networks (ANN). Results show that ANN is the most accurate of all the above stated classifiers with 97.7% accuracy, 97.25% specificity and 98.28% sensitivity in its merit. This is followed closely by SVM with 1% variation in result. These results would certainly help researchers choose the best classifier for detection of epilepsy.

Keywords: classification, seizure, KNN, SVM, LDA, ANN, epilepsy

Procedia PDF Downloads 509
2625 The Impact of Artificial Intelligence on Human Rights Development and Obligations

Authors: Bola George Asaad Bekledas

Abstract:

Relationship between development and human rights has long been the subject of academic debate. To understand the dynamics between these two concepts, various principles are adopted, from the right to development to development-based human rights. Despite the initiatives taken, the relationship between development and human rights remains unclear. However, the compatibility between these two concepts and the idea that these efforts should be made to respect human rights guarantees have gained momentum in recent years. It is then evaluated whether the right to sustainable development is acceptable or not. The article concludes that the principles of sustainable development are directly or indirectly recognized in various human rights instruments, which is a good answer to the question posed above. This study, therefore, cites regional and international human rights agreements such as, as well as the jurisprudence and interpretative guidelines of human rights institutions, to prove this hypothesis.

Keywords: sustainable development, human rights, the right to development, the human rights-based approach to development, environmental rights, economic development, social sustainability human rights protection, human rights violations, workers’ rights, justice, security.

Procedia PDF Downloads 43
2624 Reconstruction of Binary Matrices Satisfying Neighborhood Constraints by Simulated Annealing

Authors: Divyesh Patel, Tanuja Srivastava

Abstract:

This paper considers the NP-hard problem of reconstructing binary matrices satisfying exactly-1-4-adjacency constraint from its row and column projections. This problem is formulated into a maximization problem. The objective function gives a measure of adjacency constraint for the binary matrices. The maximization problem is solved by the simulated annealing algorithm and experimental results are presented.

Keywords: discrete tomography, exactly-1-4-adjacency, simulated annealing, binary matrices

Procedia PDF Downloads 401
2623 Rapid Parallel Algorithm for GPS Signal Acquisition

Authors: Fabricio Costa Silva, Samuel Xavier de Souza

Abstract:

A Global Positioning System (GPS) receiver is responsible to determine position, velocity and timing information by using satellite information. To get this information's are necessary to combine an incoming and a locally generated signal. The procedure called acquisition need to found two information, the frequency and phase of the incoming signal. This is very time consuming, so there are several techniques to reduces the computational complexity, but each of then put projects issues in conflict. I this papers we present a method that can reduce the computational complexity by reducing the search space and paralleling the search.

Keywords: GPS, acquisition, low complexity, parallelism

Procedia PDF Downloads 493
2622 Artificial Neural Network in FIRST Robotics Team-Based Prediction System

Authors: Cedric Leong, Parth Desai, Parth Patel

Abstract:

The purpose of this project was to develop a neural network based on qualitative team data to predict alliance scores to determine winners of matches in the FIRST Robotics Competition (FRC). The game for the competition changes every year with different objectives and game objects, however the idea was to create a prediction system which can be reused year by year using some of the statistics that are constant through different games, making our system adaptable to future games as well. Aerial Assist is the FRC game for 2014, and is played in alliances of 3 teams going against one another, namely the Red and Blue alliances. This application takes any 6 teams paired into 2 alliances of 3 teams and generates the prediction for the final score between them.

Keywords: artifical neural network, prediction system, qualitative team data, FIRST Robotics Competition (FRC)

Procedia PDF Downloads 507
2621 An Evaluation of the Artificial Neural Network and Adaptive Neuro Fuzzy Inference System Predictive Models for the Remediation of Crude Oil-Contaminated Soil Using Vermicompost

Authors: Precious Ehiomogue, Ifechukwude Israel Ahuchaogu, Isiguzo Edwin Ahaneku

Abstract:

Vermicompost is the product of the decomposition process using various species of worms, to create a mixture of decomposing vegetable or food waste, bedding materials, and vemicast. This process is called vermicomposting, while the rearing of worms for this purpose is called vermiculture. Several works have verified the adsorption of toxic metals using vermicompost but the application is still scarce for the retention of organic compounds. This research brings to knowledge the effectiveness of earthworm waste (vermicompost) for the remediation of crude oil contaminated soils. The remediation methods adopted in this study were two soil washing methods namely, batch and column process which represent laboratory and in-situ remediation. Characterization of the vermicompost and crude oil contaminated soil were performed before and after the soil washing using Fourier transform infrared (FTIR), scanning electron microscopy (SEM), X-ray fluorescence (XRF), X-ray diffraction (XRD) and Atomic adsorption spectrometry (AAS). The optimization of washing parameters, using response surface methodology (RSM) based on Box-Behnken Design was performed on the response from the laboratory experimental results. This study also investigated the application of machine learning models [Artificial neural network (ANN), Adaptive neuro fuzzy inference system (ANFIS). ANN and ANFIS were evaluated using the coefficient of determination (R²) and mean square error (MSE)]. Removal efficiency obtained from the Box-Behnken design experiment ranged from 29% to 98.9% for batch process remediation. Optimization of the experimental factors carried out using numerical optimization techniques by applying desirability function method of the response surface methodology (RSM) produce the highest removal efficiency of 98.9% at absorbent dosage of 34.53 grams, adsorbate concentration of 69.11 (g/ml), contact time of 25.96 (min), and pH value of 7.71, respectively. Removal efficiency obtained from the multilevel general factorial design experiment ranged from 56% to 92% for column process remediation. The coefficient of determination (R²) for ANN was (0.9974) and (0.9852) for batch and column process, respectively, showing the agreement between experimental and predicted results. For batch and column precess, respectively, the coefficient of determination (R²) for RSM was (0.9712) and (0.9614), which also demonstrates agreement between experimental and projected findings. For the batch and column processes, the ANFIS coefficient of determination was (0.7115) and (0.9978), respectively. It can be concluded that machine learning models can predict the removal of crude oil from polluted soil using vermicompost. Therefore, it is recommended to use machines learning models to predict the removal of crude oil from contaminated soil using vermicompost.

Keywords: ANFIS, ANN, crude-oil, contaminated soil, remediation and vermicompost

Procedia PDF Downloads 101
2620 Screening Maize for Compatibility with F. Oxysporum to Enhance Striga asiatica (L.) Kuntze Resistance

Authors: Admire Isaac Tichafa Shayanowako, Mark Laing, Hussein Shimelis

Abstract:

Striga asiatica is among the leading abiotic constraints to maize production under small-holder farming communities in southern African. However, confirmed sources of resistance to the parasitic weed are still limited. Conventional breeding programmes have been progressing slowly due to the complex nature of the inheritance of Striga resistance, hence there is a need for more innovative approaches. This study aimed to achieve partial resistance as well as to breed for compatibility with Fusarium oxysporum fsp strigae, a soil fungus that is highly specific in its pathogenicity. The agar gel and paper roll assays in conjunction with a glass house pot trial were done to select genotypes based on their potential to stimulate germination of Striga and to test the efficacy of Fusarium oxysporum as a biocontrol agent. Results from agar gel assays showed a moderate to high potential in the release of Strigalactones among the 33 OPVs. Maximum Striga germination distances from the host root of 1.38 cm and up to 46% germination were observed in most of the populations. Considerable resistance was observed in a landrace ‘8lines’ which had the least Striga germination percentage (19%) with a maximum distance of 0.93 cm compared to the resistant check Z-DPLO-DTC1 that had 23% germination at a distance of 1.4cm. The number of fusarium colony forming units significantly deferred (P < 0.05) amongst the genotypes growing between germination papers. The number of crown roots, length of primary root and fresh weight of shoot and roots were highly correlated with concentration of fusarium macrospore counts. Pot trials showed significant differences between the fusarium coated and the uncoated treatments in terms of plant height, leaf counts, anthesis-silks intervals, Striga counts, Striga damage rating and Striga vigour. Striga emergence counts and Striga flowers were low in fusarium treated pots. Plants in fusarium treated pots had non-significant differences in height with the control treatment. This suggests that foxy 2 reduces the impact of Striga damage severity. Variability within fusarium treated genotypes with respect to traits under evaluation indicates the varying degree of compatibility with the biocontrol.

Keywords: maize, Striga asiaitca, resistance, compatibility, F. oxysporum

Procedia PDF Downloads 241
2619 Neural Network and Support Vector Machine for Prediction of Foot Disorders Based on Foot Analysis

Authors: Monireh Ahmadi Bani, Adel Khorramrouz, Lalenoor Morvarid, Bagheri Mahtab

Abstract:

Background:- Foot disorders are common in musculoskeletal problems. Plantar pressure distribution measurement is one the most important part of foot disorders diagnosis for quantitative analysis. However, the association of plantar pressure and foot disorders is not clear. With the growth of dataset and machine learning methods, the relationship between foot disorders and plantar pressures can be detected. Significance of the study:- The purpose of this study was to predict the probability of common foot disorders based on peak plantar pressure distribution and center of pressure during walking. Methodologies:- 2323 participants were assessed in a foot therapy clinic between 2015 and 2021. Foot disorders were diagnosed by an experienced physician and then they were asked to walk on a force plate scanner. After the data preprocessing, due to the difference in walking time and foot size, we normalized the samples based on time and foot size. Some of force plate variables were selected as input to a deep neural network (DNN), and the probability of any each foot disorder was measured. In next step, we used support vector machine (SVM) and run dataset for each foot disorder (classification of yes or no). We compared DNN and SVM for foot disorders prediction based on plantar pressure distributions and center of pressure. Findings:- The results demonstrated that the accuracy of deep learning architecture is sufficient for most clinical and research applications in the study population. In addition, the SVM approach has more accuracy for predictions, enabling applications for foot disorders diagnosis. The detection accuracy was 71% by the deep learning algorithm and 78% by the SVM algorithm. Moreover, when we worked with peak plantar pressure distribution, it was more accurate than center of pressure dataset. Conclusion:- Both algorithms- deep learning and SVM will help therapist and patients to improve the data pool and enhance foot disorders prediction with less expense and error after removing some restrictions properly.

Keywords: deep neural network, foot disorder, plantar pressure, support vector machine

Procedia PDF Downloads 336
2618 Robust Control of a Single-Phase Inverter Using Linear Matrix Inequality Approach

Authors: Chivon Choeung, Heng Tang, Panha Soth, Vichet Huy

Abstract:

This paper presents a robust control strategy for a single-phase DC-AC inverter with an output LC-filter. An all-pass filter is utilized to create an artificial β-signal so that the proposed controller can be simply used in dq-synchronous frame. The proposed robust controller utilizes a state feedback control with integral action in the dq-synchronous frame. A linear matrix inequality-based optimization scheme is used to determine stabilizing gains of the controllers to maximize the convergence rate to steady state in the presence of uncertainties. The uncertainties of the system are described as the potential variation range of the inductance and resistance in the LC-filter.

Keywords: single-phase inverter, linear matrix inequality, robust control, all-pass filter

Procedia PDF Downloads 132
2617 Design and Fabrication of a Scaffold with Appropriate Features for Cartilage Tissue Engineering

Authors: S. S. Salehi, A. Shamloo

Abstract:

Poor ability of cartilage tissue when experiencing a damage leads scientists to use tissue engineering as a reliable and effective method for regenerating or replacing damaged tissues. An artificial tissue should have some features such as biocompatibility, biodegradation and, enough mechanical properties like the original tissue. In this work, a composite hydrogel is prepared by using natural and synthetic materials that has high porosity. Mechanical properties of different combinations of polymers such as modulus of elasticity were tested, and a hydrogel with good mechanical properties was selected. Bone marrow derived mesenchymal stem cells were also seeded into the pores of the sponge, and the results showed the adhesion and proliferation of cells within the hydrogel after one month. In comparison with previous works, this study offers a new and efficient procedure for the fabrication of cartilage like tissue and further cartilage repair.

Keywords: cartilage tissue engineering, hydrogel, mechanical strength, mesenchymal stem cell

Procedia PDF Downloads 293
2616 Nature, Elixir of Architecture: A Contemplation on Human, Nature and Architecture in Islam

Authors: A. Kabiri-Samani, M. J. Seddighi

Abstract:

There is no doubt that a key factor in the manifestation of architecture is the interaction of human and nature. Explaining the type of relationship defined by “the architect” between architecture and nature opens a window towards understanding the theoretical conceptions of the architect as the creator of “architecture”. Now, if these theoretical foundations are put under scrutiny from the viewpoint of Islam, and an architect considers the relationship of human and nature within the context of Islam, he would let nature to manifest itself in architecture. The reasons for such a relationship is explicable in terms of the degree and nature of knowledge of the architect about nature; while the way it comes to existence is explained by defining the force of nature – ruling the entire nature – and its acts. It is by the scientific command of the architect and his mastery in the hermetic force of nature that the material bodies of buildings evolve from artificial to natural. Additionally, the presence of nature creates hermetic architectural spaces for the spiritual development of humans while serving for living at different levels.

Keywords: nature, Islam, cognition, science, presence, elixir

Procedia PDF Downloads 481
2615 Maximum Power Point Tracking Using FLC Tuned with GA

Authors: Mohamed Amine Haraoubia, Abdelaziz Hamzaoui, Najib Essounbouli

Abstract:

The pursuit of the MPPT has led to the development of many kinds of controllers, one of which is the Fuzzy Logic Controller, which has proven its worth. To further tune this controller this paper will discuss and analyze the use of Genetic Algorithms to tune the Fuzzy Logic Controller. It will provide an introduction to both systems, and test their compatibility and performance.

Keywords: fuzzy logic controller, fuzzy logic, genetic algorithm, maximum power point, maximum power point tracking

Procedia PDF Downloads 369
2614 The Impact of Artificial Intelligence on Marketing Principles and Targets

Authors: Felib Ayman Shawky Salem

Abstract:

Experiential marketing means an unforgettable experience that remains deeply anchored in the customer's memory. Furthermore, customer satisfaction is defined as the emotional response to the experiences provided that relate to specific products or services purchased. Therefore, experiential marketing activities can influence the level of customer satisfaction and loyalty. In this context, the study aims to examine the relationship between experiential marketing, customer satisfaction and loyalty of beauty products in Konya. The results of this study showed that experiential marketing is an important indicator of customer satisfaction and loyalty and that experiential marketing has a significant positive impact on customer satisfaction and loyalty.

Keywords: sponsorship, marketing communication theories, marketing communication tools internet, marketing, tourism, tourism management corporate responsibility, employee organizational performance, internal marketing, internal customer experiential marketing, customer satisfaction, customer loyalty, social sciences.

Procedia PDF Downloads 60
2613 Comparison of Machine Learning Models for the Prediction of System Marginal Price of Greek Energy Market

Authors: Ioannis P. Panapakidis, Marios N. Moschakis

Abstract:

The Greek Energy Market is structured as a mandatory pool where the producers make their bid offers in day-ahead basis. The System Operator solves an optimization routine aiming at the minimization of the cost of produced electricity. The solution of the optimization problem leads to the calculation of the System Marginal Price (SMP). Accurate forecasts of the SMP can lead to increased profits and more efficient portfolio management from the producer`s perspective. Aim of this study is to provide a comparative analysis of various machine learning models such as artificial neural networks and neuro-fuzzy models for the prediction of the SMP of the Greek market. Machine learning algorithms are favored in predictions problems since they can capture and simulate the volatilities of complex time series.

Keywords: deregulated energy market, forecasting, machine learning, system marginal price

Procedia PDF Downloads 207
2612 Exploiting Kinetic and Kinematic Data to Plot Cyclograms for Managing the Rehabilitation Process of BKAs by Applying Neural Networks

Authors: L. Parisi

Abstract:

Kinematic data wisely correlate vector quantities in space to scalar parameters in time to assess the degree of symmetry between the intact limb and the amputated limb with respect to a normal model derived from the gait of control group participants. Furthermore, these particular data allow a doctor to preliminarily evaluate the usefulness of a certain rehabilitation therapy. Kinetic curves allow the analysis of ground reaction forces (GRFs) to assess the appropriateness of human motion. Electromyography (EMG) allows the analysis of the fundamental lower limb force contributions to quantify the level of gait asymmetry. However, the use of this technological tool is expensive and requires patient’s hospitalization. This research work suggests overcoming the above limitations by applying artificial neural networks.

Keywords: kinetics, kinematics, cyclograms, neural networks, transtibial amputation

Procedia PDF Downloads 438
2611 Optical Flow Technique for Supersonic Jet Measurements

Authors: Haoxiang Desmond Lim, Jie Wu, Tze How Daniel New, Shengxian Shi

Abstract:

This paper outlines the development of a novel experimental technique in quantifying supersonic jet flows, in an attempt to avoid seeding particle problems frequently associated with particle-image velocimetry (PIV) techniques at high Mach numbers. Based on optical flow algorithms, the idea behind the technique involves using high speed cameras to capture Schlieren images of the supersonic jet shear layers, before they are subjected to an adapted optical flow algorithm based on the Horn-Schnuck method to determine the associated flow fields. The proposed method is capable of offering full-field unsteady flow information with potentially higher accuracy and resolution than existing point-measurements or PIV techniques. Preliminary study via numerical simulations of a circular de Laval jet nozzle successfully reveals flow and shock structures typically associated with supersonic jet flows, which serve as useful data for subsequent validation of the optical flow based experimental results. For experimental technique, a Z-type Schlieren setup is proposed with supersonic jet operated in cold mode, stagnation pressure of 8.2 bar and exit velocity of Mach 1.5. High-speed single-frame or double-frame cameras are used to capture successive Schlieren images. As implementation of optical flow technique to supersonic flows remains rare, the current focus revolves around methodology validation through synthetic images. The results of validation test offers valuable insight into how the optical flow algorithm can be further improved to improve robustness and accuracy. Details of the methodology employed and challenges faced will be further elaborated in the final conference paper should the abstract be accepted. Despite these challenges however, this novel supersonic flow measurement technique may potentially offer a simpler way to identify and quantify the fine spatial structures within the shock shear layer.

Keywords: Schlieren, optical flow, supersonic jets, shock shear layer

Procedia PDF Downloads 310
2610 Freight Time and Cost Optimization in Complex Logistics Networks, Using a Dimensional Reduction Method and K-Means Algorithm

Authors: Egemen Sert, Leila Hedayatifar, Rachel A. Rigg, Amir Akhavan, Olha Buchel, Dominic Elias Saadi, Aabir Abubaker Kar, Alfredo J. Morales, Yaneer Bar-Yam

Abstract:

The complexity of providing timely and cost-effective distribution of finished goods from industrial facilities to customers makes effective operational coordination difficult, yet effectiveness is crucial for maintaining customer service levels and sustaining a business. Logistics planning becomes increasingly complex with growing numbers of customers, varied geographical locations, the uncertainty of future orders, and sometimes extreme competitive pressure to reduce inventory costs. Linear optimization methods become cumbersome or intractable due to a large number of variables and nonlinear dependencies involved. Here we develop a complex systems approach to optimizing logistics networks based upon dimensional reduction methods and apply our approach to a case study of a manufacturing company. In order to characterize the complexity in customer behavior, we define a “customer space” in which individual customer behavior is described by only the two most relevant dimensions: the distance to production facilities over current transportation routes and the customer's demand frequency. These dimensions provide essential insight into the domain of effective strategies for customers; direct and indirect strategies. In the direct strategy, goods are sent to the customer directly from a production facility using box or bulk trucks. In the indirect strategy, in advance of an order by the customer, goods are shipped to an external warehouse near a customer using trains and then "last-mile" shipped by trucks when orders are placed. Each strategy applies to an area of the customer space with an indeterminate boundary between them. Specific company policies determine the location of the boundary generally. We then identify the optimal delivery strategy for each customer by constructing a detailed model of costs of transportation and temporary storage in a set of specified external warehouses. Customer spaces help give an aggregate view of customer behaviors and characteristics. They allow policymakers to compare customers and develop strategies based on the aggregate behavior of the system as a whole. In addition to optimization over existing facilities, using customer logistics and the k-means algorithm, we propose additional warehouse locations. We apply these methods to a medium-sized American manufacturing company with a particular logistics network, consisting of multiple production facilities, external warehouses, and customers along with three types of shipment methods (box truck, bulk truck and train). For the case study, our method forecasts 10.5% savings on yearly transportation costs and an additional 4.6% savings with three new warehouses.

Keywords: logistics network optimization, direct and indirect strategies, K-means algorithm, dimensional reduction

Procedia PDF Downloads 134
2609 DHL CSI Solution Design Project

Authors: Mohammed Al-Yamani, Yaser Miaji

Abstract:

DHL Customer Solutions and Innovation Department (CSI) have been experiencing difficulties while comparing quotes for different customers in different years. Currently, the employees are processing data by opening several loaded Excel files where the quotes are and manually copying values to another Excel Workbook where the comparison is made. This project consists of developing a new and effective database for DHL CSI department so that information is stored altogether on the same catalog. That being said, we have been assigned to find an efficient algorithm that can deal with the different formats of the Excel Workbooks to copy and store the express customer rates for core products (DOX, WPX, IMP) for comparisons purposes.

Keywords: DHL, solution design, ORACLE, EXCEL

Procedia PDF Downloads 402
2608 ANN Modeling for Cadmium Biosorption from Potable Water Using a Packed-Bed Column Process

Authors: Dariush Jafari, Seyed Ali Jafari

Abstract:

The recommended limit for cadmium concentration in potable water is less than 0.005 mg/L. A continuous biosorption process using indigenous red seaweed, Gracilaria corticata, was performed to remove cadmium from the potable water. The process was conducted under fixed conditions and the breakthrough curves were achieved for three consecutive sorption-desorption cycles. A modeling based on Artificial Neural Network (ANN) was employed to fit the experimental breakthrough data. In addition, a simplified semi empirical model, Thomas, was employed for this purpose. It was found that ANN well described the experimental data (R2>0.99) while the Thomas prediction were a bit less successful with R2>0.97. The adjusted design parameters using the nonlinear form of Thomas model was in a good agreement with the experimentally obtained ones. The results approve the capability of ANN to predict the cadmium concentration in potable water.

Keywords: ANN, biosorption, cadmium, packed-bed, potable water

Procedia PDF Downloads 420
2607 Wireless FPGA-Based Motion Controller Design by Implementing 3-Axis Linear Trajectory

Authors: Kiana Zeighami, Morteza Ozlati Moghadam

Abstract:

Designing a high accuracy and high precision motion controller is one of the important issues in today’s industry. There are effective solutions available in the industry but the real-time performance, smoothness and accuracy of the movement can be further improved. This paper discusses a complete solution to carry out the movement of three stepper motors in three dimensions. The objective is to provide a method to design a fully integrated System-on-Chip (SOC)-based motion controller to reduce the cost and complexity of production by incorporating Field Programmable Gate Array (FPGA) into the design. In the proposed method the FPGA receives its commands from a host computer via wireless internet communication and calculates the motion trajectory for three axes. A profile generator module is designed to realize the interpolation algorithm by translating the position data to the real-time pulses. This paper discusses an approach to implement the linear interpolation algorithm, since it is one of the fundamentals of robots’ movements and it is highly applicable in motion control industries. Along with full profile trajectory, the triangular drive is implemented to eliminate the existence of error at small distances. To integrate the parallelism and real-time performance of FPGA with the power of Central Processing Unit (CPU) in executing complex and sequential algorithms, the NIOS II soft-core processor was added into the design. This paper presents different operating modes such as absolute, relative positioning, reset and velocity modes to fulfill the user requirements. The proposed approach was evaluated by designing a custom-made FPGA board along with a mechanical structure. As a result, a precise and smooth movement of stepper motors was observed which proved the effectiveness of this approach.

Keywords: 3-axis linear interpolation, FPGA, motion controller, micro-stepping

Procedia PDF Downloads 203
2606 Sparse Representation Based Spatiotemporal Fusion Employing Additional Image Pairs to Improve Dictionary Training

Authors: Dacheng Li, Bo Huang, Qinjin Han, Ming Li

Abstract:

Remotely sensed imagery with the high spatial and temporal characteristics, which it is hard to acquire under the current land observation satellites, has been considered as a key factor for monitoring environmental changes over both global and local scales. On a basis of the limited high spatial-resolution observations, challenged studies called spatiotemporal fusion have been developed for generating high spatiotemporal images through employing other auxiliary low spatial-resolution data while with high-frequency observations. However, a majority of spatiotemporal fusion approaches yield to satisfactory assumption, empirical but unstable parameters, low accuracy or inefficient performance. Although the spatiotemporal fusion methodology via sparse representation theory has advantage in capturing reflectance changes, stability and execution efficiency (even more efficient when overcomplete dictionaries have been pre-trained), the retrieval of high-accuracy dictionary and its response to fusion results are still pending issues. In this paper, we employ additional image pairs (here each image-pair includes a Landsat Operational Land Imager and a Moderate Resolution Imaging Spectroradiometer acquisitions covering the partial area of Baotou, China) only into the coupled dictionary training process based on K-SVD (K-means Singular Value Decomposition) algorithm, and attempt to improve the fusion results of two existing sparse representation based fusion models (respectively utilizing one and two available image-pair). The results show that more eligible image pairs are probably related to a more accurate overcomplete dictionary, which generally indicates a better image representation, and is then contribute to an effective fusion performance in case that the added image-pair has similar seasonal aspects and image spatial structure features to the original image-pair. It is, therefore, reasonable to construct multi-dictionary training pattern for generating a series of high spatial resolution images based on limited acquisitions.

Keywords: spatiotemporal fusion, sparse representation, K-SVD algorithm, dictionary learning

Procedia PDF Downloads 254
2605 Comparing Performance of Neural Network and Decision Tree in Prediction of Myocardial Infarction

Authors: Reza Safdari, Goli Arji, Robab Abdolkhani Maryam zahmatkeshan

Abstract:

Background and purpose: Cardiovascular diseases are among the most common diseases in all societies. The most important step in minimizing myocardial infarction and its complications is to minimize its risk factors. The amount of medical data is increasingly growing. Medical data mining has a great potential for transforming these data into information. Using data mining techniques to generate predictive models for identifying those at risk for reducing the effects of the disease is very helpful. The present study aimed to collect data related to risk factors of heart infarction from patients’ medical record and developed predicting models using data mining algorithm. Methods: The present work was an analytical study conducted on a database containing 350 records. Data were related to patients admitted to Shahid Rajaei specialized cardiovascular hospital, Iran, in 2011. Data were collected using a four-sectioned data collection form. Data analysis was performed using SPSS and Clementine version 12. Seven predictive algorithms and one algorithm-based model for predicting association rules were applied to the data. Accuracy, precision, sensitivity, specificity, as well as positive and negative predictive values were determined and the final model was obtained. Results: five parameters, including hypertension, DLP, tobacco smoking, diabetes, and A+ blood group, were the most critical risk factors of myocardial infarction. Among the models, the neural network model was found to have the highest sensitivity, indicating its ability to successfully diagnose the disease. Conclusion: Risk prediction models have great potentials in facilitating the management of a patient with a specific disease. Therefore, health interventions or change in their life style can be conducted based on these models for improving the health conditions of the individuals at risk.

Keywords: decision trees, neural network, myocardial infarction, Data Mining

Procedia PDF Downloads 426
2604 Frequency Decomposition Approach for Sub-Band Common Spatial Pattern Methods for Motor Imagery Based Brain-Computer Interface

Authors: Vitor M. Vilas Boas, Cleison D. Silva, Gustavo S. Mafra, Alexandre Trofino Neto

Abstract:

Motor imagery (MI) based brain-computer interfaces (BCI) uses event-related (de)synchronization (ERS/ ERD), typically recorded using electroencephalography (EEG), to translate brain electrical activity into control commands. To mitigate undesirable artifacts and noise measurements on EEG signals, methods based on band-pass filters defined by a specific frequency band (i.e., 8 – 30Hz), such as the Infinity Impulse Response (IIR) filters, are typically used. Spatial techniques, such as Common Spatial Patterns (CSP), are also used to estimate the variations of the filtered signal and extract features that define the imagined motion. The CSP effectiveness depends on the subject's discriminative frequency, and approaches based on the decomposition of the band of interest into sub-bands with smaller frequency ranges (SBCSP) have been suggested to EEG signals classification. However, despite providing good results, the SBCSP approach generally increases the computational cost of the filtering step in IM-based BCI systems. This paper proposes the use of the Fast Fourier Transform (FFT) algorithm in the IM-based BCI filtering stage that implements SBCSP. The goal is to apply the FFT algorithm to reduce the computational cost of the processing step of these systems and to make them more efficient without compromising classification accuracy. The proposal is based on the representation of EEG signals in a matrix of coefficients resulting from the frequency decomposition performed by the FFT, which is then submitted to the SBCSP process. The structure of the SBCSP contemplates dividing the band of interest, initially defined between 0 and 40Hz, into a set of 33 sub-bands spanning specific frequency bands which are processed in parallel each by a CSP filter and an LDA classifier. A Bayesian meta-classifier is then used to represent the LDA outputs of each sub-band as scores and organize them into a single vector, and then used as a training vector of an SVM global classifier. Initially, the public EEG data set IIa of the BCI Competition IV is used to validate the approach. The first contribution of the proposed method is that, in addition to being more compact, because it has a 68% smaller dimension than the original signal, the resulting FFT matrix maintains the signal information relevant to class discrimination. In addition, the results showed an average reduction of 31.6% in the computational cost in relation to the application of filtering methods based on IIR filters, suggesting FFT efficiency when applied in the filtering step. Finally, the frequency decomposition approach improves the overall system classification rate significantly compared to the commonly used filtering, going from 73.7% using IIR to 84.2% using FFT. The accuracy improvement above 10% and the computational cost reduction denote the potential of FFT in EEG signal filtering applied to the context of IM-based BCI implementing SBCSP. Tests with other data sets are currently being performed to reinforce such conclusions.

Keywords: brain-computer interfaces, fast Fourier transform algorithm, motor imagery, sub-band common spatial patterns

Procedia PDF Downloads 126
2603 Methods for Enhancing Ensemble Learning or Improving Classifiers of This Technique in the Analysis and Classification of Brain Signals

Authors: Seyed Mehdi Ghezi, Hesam Hasanpoor

Abstract:

This scientific article explores enhancement methods for ensemble learning with the aim of improving the performance of classifiers in the analysis and classification of brain signals. The research approach in this field consists of two main parts, each with its own strengths and weaknesses. The choice of approach depends on the specific research question and available resources. By combining these approaches and leveraging their respective strengths, researchers can enhance the accuracy and reliability of classification results, consequently advancing our understanding of the brain and its functions. The first approach focuses on utilizing machine learning methods to identify the best features among the vast array of features present in brain signals. The selection of features varies depending on the research objective, and different techniques have been employed for this purpose. For instance, the genetic algorithm has been used in some studies to identify the best features, while optimization methods have been utilized in others to identify the most influential features. Additionally, machine learning techniques have been applied to determine the influential electrodes in classification. Ensemble learning plays a crucial role in identifying the best features that contribute to learning, thereby improving the overall results. The second approach concentrates on designing and implementing methods for selecting the best classifier or utilizing meta-classifiers to enhance the final results in ensemble learning. In a different section of the research, a single classifier is used instead of multiple classifiers, employing different sets of features to improve the results. The article provides an in-depth examination of each technique, highlighting their advantages and limitations. By integrating these techniques, researchers can enhance the performance of classifiers in the analysis and classification of brain signals. This advancement in ensemble learning methodologies contributes to a better understanding of the brain and its functions, ultimately leading to improved accuracy and reliability in brain signal analysis and classification.

Keywords: ensemble learning, brain signals, classification, feature selection, machine learning, genetic algorithm, optimization methods, influential features, influential electrodes, meta-classifiers

Procedia PDF Downloads 70
2602 Contribution to the Study of the Rill Density Effects on Soil Erosion: Laboratory Experiments

Authors: L. Mouzai, M. Bouhadef

Abstract:

Rills begin to be generated once overland flow shear capacity overcomes the soil surface resistance. This resistance depends on soil texture, the arrangement of soil particles and on chemical and physical properties. The rill density could affect soil erosion, especially when the distance between the rills (interrill) contributes to the variation of the rill characteristics, and consequently on sediment concentration. To investigate this point, agricultural sandy soil, a soil tray of 0.2x1x3m³ and a piece of hardwood rectangular in shape to build up rills were the base of this work. The results have shown that small lines have been developed between the rills and the flow acceleration increased in comparison to the flow on the flat surface (interrill). Sediment concentration increased with increasing rill number (density).

Keywords: artificial rainfall, experiments, rills, soil erosion, transport capacity

Procedia PDF Downloads 154