Search results for: input output linearization
1240 Performance of BLDC Motor under Kalman Filter Sensorless Drive
Authors: Yuri Boiko, Ci Lin, Iluju Kiringa, Tet Yeap
Abstract:
The performance of a BLDC motor controlled by the Kalman filter-based position-sensorless drive is studied in terms of its dependence on the system’s parameters' variations. The effects of system’s parameters changes on the dynamic behavior of state variables are verified. Simulated is a closed-loop control scheme with a Kalman filter in the feedback line. Distinguished are two separate data sampling modes in analyzing feedback output from the BLDC motor: (1) equal angular separation and (2) equal time intervals. In case (1), the data are collected via equal intervals Δθ of rotor’s angular position θᵢ, i.e., keeping Δθ=const. In case (2), the data collection time points tᵢ are separated by equal sampling time intervals Δt=const. Demonstrated are the effects of the parameters changes on the sensorless control flow, in particular, reduction of the torque ripples, switching spikes, torque load balancing. It is specifically shown that an efficient suppression of commutation induced torque ripples is achievable selection of the sampling rate in the Kalman filter settings above certain critical value. The computational cost of such suppression is shown to be higher for the motors with lower induction values of the windings.Keywords: BLDC motor, Kalman filter, sensorless drive, state variables, torque ripples reduction, sampling rate
Procedia PDF Downloads 1481239 Using Machine Learning to Classify Human Fetal Health and Analyze Feature Importance
Authors: Yash Bingi, Yiqiao Yin
Abstract:
Reduction of child mortality is an ongoing struggle and a commonly used factor in determining progress in the medical field. The under-5 mortality number is around 5 million around the world, with many of the deaths being preventable. In light of this issue, Cardiotocograms (CTGs) have emerged as a leading tool to determine fetal health. By using ultrasound pulses and reading the responses, CTGs help healthcare professionals assess the overall health of the fetus to determine the risk of child mortality. However, interpreting the results of the CTGs is time-consuming and inefficient, especially in underdeveloped areas where an expert obstetrician is hard to come by. Using a support vector machine (SVM) and oversampling, this paper proposed a model that classifies fetal health with an accuracy of 99.59%. To further explain the CTG measurements, an algorithm based on Randomized Input Sampling for Explanation ((RISE) of Black-box Models was created, called Feature Alteration for explanation of Black Box Models (FAB), and compared the findings to Shapley Additive Explanations (SHAP) and Local Interpretable Model Agnostic Explanations (LIME). This allows doctors and medical professionals to classify fetal health with high accuracy and determine which features were most influential in the process.Keywords: machine learning, fetal health, gradient boosting, support vector machine, Shapley values, local interpretable model agnostic explanations
Procedia PDF Downloads 1441238 Spatiotemporal Analysis of Visual Evoked Responses Using Dense EEG
Authors: Rima Hleiss, Elie Bitar, Mahmoud Hassan, Mohamad Khalil
Abstract:
A comprehensive study of object recognition in the human brain requires combining both spatial and temporal analysis of brain activity. Here, we are mainly interested in three issues: the time perception of visual objects, the ability of discrimination between two particular categories (objects vs. animals), and the possibility to identify a particular spatial representation of visual objects. Our experiment consisted of acquiring dense electroencephalographic (EEG) signals during a picture-naming task comprising a set of objects and animals’ images. These EEG responses were recorded from nine participants. In order to determine the time perception of the presented visual stimulus, we analyzed the Event Related Potentials (ERPs) derived from the recorded EEG signals. The analysis of these signals showed that the brain perceives animals and objects with different time instants. Concerning the discrimination of the two categories, the support vector machine (SVM) was applied on the instantaneous EEG (excellent temporal resolution: on the order of millisecond) to categorize the visual stimuli into two different classes. The spatial differences between the evoked responses of the two categories were also investigated. The results showed a variation of the neural activity with the properties of the visual input. Results showed also the existence of a spatial pattern of electrodes over particular regions of the scalp in correspondence to their responses to the visual inputs.Keywords: brain activity, categorization, dense EEG, evoked responses, spatio-temporal analysis, SVM, time perception
Procedia PDF Downloads 4221237 Time Series Forecasting (TSF) Using Various Deep Learning Models
Authors: Jimeng Shi, Mahek Jain, Giri Narasimhan
Abstract:
Time Series Forecasting (TSF) is used to predict the target variables at a future time point based on the learning from previous time points. To keep the problem tractable, learning methods use data from a fixed-length window in the past as an explicit input. In this paper, we study how the performance of predictive models changes as a function of different look-back window sizes and different amounts of time to predict the future. We also consider the performance of the recent attention-based Transformer models, which have had good success in the image processing and natural language processing domains. In all, we compare four different deep learning methods (RNN, LSTM, GRU, and Transformer) along with a baseline method. The dataset (hourly) we used is the Beijing Air Quality Dataset from the UCI website, which includes a multivariate time series of many factors measured on an hourly basis for a period of 5 years (2010-14). For each model, we also report on the relationship between the performance and the look-back window sizes and the number of predicted time points into the future. Our experiments suggest that Transformer models have the best performance with the lowest Mean Average Errors (MAE = 14.599, 23.273) and Root Mean Square Errors (RSME = 23.573, 38.131) for most of our single-step and multi-steps predictions. The best size for the look-back window to predict 1 hour into the future appears to be one day, while 2 or 4 days perform the best to predict 3 hours into the future.Keywords: air quality prediction, deep learning algorithms, time series forecasting, look-back window
Procedia PDF Downloads 1541236 Characterization of Aluminosilicates and Verification of Their Impact on Quality of Ceramic Proppants Intended for Shale Gas Output
Authors: Joanna Szymanska, Paulina Wawulska-Marek, Jaroslaw Mizera
Abstract:
Nowadays, the rapid growth of global energy consumption and uncontrolled depletion of natural resources become a serious problem. Shale rocks are the largest and potential global basins containing hydrocarbons, trapped in closed pores of the shale matrix. Regardless of the shales origin, mining conditions are extremely unfavourable due to high reservoir pressure, great depths, increased clay minerals content and limited permeability (nanoDarcy) of the rocks. Taking into consideration such geomechanical barriers, effective extraction of natural gas from shales with plastic zones demands effective operations. Actually, hydraulic fracturing is the most developed technique based on the injection of pressurized fluid into a wellbore, to initiate fractures propagation. However, a rapid drop of pressure after fluid suction to the ground induces a fracture closure and conductivity reduction. In order to minimize this risk, proppants should be applied. They are solid granules transported with hydraulic fluids to locate inside the rock. Proppants act as a prop for the closing fracture, thus gas migration to a borehole is effective. Quartz sands are commonly applied proppants only at shallow deposits (USA). Whereas, ceramic proppants are designed to meet rigorous downhole conditions to intensify output. Ceramic granules predominate with higher mechanical strength, stability in strong acidic environment, spherical shape and homogeneity as well. Quality of ceramic proppants is conditioned by raw materials selection. Aim of this study was to obtain the proppants from aluminosilicates (the kaolinite subgroup) and mix of minerals with a high alumina content. These loamy minerals contain a tubular and platy morphology that improves mechanical properties and reduces their specific weight. Moreover, they are distinguished by well-developed surface area, high porosity, fine particle size, superb dispersion and nontoxic properties - very crucial for particles consolidation into spherical and crush-resistant granules in mechanical granulation process. The aluminosilicates were mixed with water and natural organic binder to improve liquid-bridges and pores formation between particles. Afterward, the green proppants were subjected to sintering at high temperatures. Evaluation of the minerals utility was based on their particle size distribution (laser diffraction study) and thermal stability (thermogravimetry). Scanning Electron Microscopy was useful for morphology and shape identification combined with specific surface area measurement (BET). Chemical composition was verified by Energy Dispersive Spectroscopy and X-ray Fluorescence. Moreover, bulk density and specific weight were measured. Such comprehensive characterization of loamy materials confirmed their favourable impact on the proppants granulation. The sintered granules were analyzed by SEM to verify the surface topography and phase transitions after sintering. Pores distribution was identified by X-Ray Tomography. This method enabled also the simulation of proppants settlement in a fracture, while measurement of bulk density was essential to predict their amount to fill a well. Roundness coefficient was also evaluated, whereas impact on mining environment was identified by turbidity and solubility in acid - to indicate risk of the material decay in a well. The obtained outcomes confirmed a positive influence of the loamy minerals on ceramic proppants properties with respect to the strict norms. This research is perspective for higher quality proppants production with costs reduction.Keywords: aluminosilicates, ceramic proppants, mechanical granulation, shale gas
Procedia PDF Downloads 1631235 Static Analysis of Security Issues of the Python Packages Ecosystem
Authors: Adam Gorine, Faten Spondon
Abstract:
Python is considered the most popular programming language and offers its own ecosystem for archiving and maintaining open-source software packages. This system is called the python package index (PyPI), the repository of this programming language. Unfortunately, one-third of these software packages have vulnerabilities that allow attackers to execute code automatically when a vulnerable or malicious package is installed. This paper contributes to large-scale empirical studies investigating security issues in the python ecosystem by evaluating package vulnerabilities. These provide a series of implications that can help the security of software ecosystems by improving the process of discovering, fixing, and managing package vulnerabilities. The vulnerable dataset is generated using the NVD, the national vulnerability database, and the Snyk vulnerability dataset. In addition, we evaluated 807 vulnerability reports in the NVD and 3900 publicly known security vulnerabilities in Python Package Manager (pip) from the Snyk database from 2002 to 2022. As a result, many Python vulnerabilities appear in high severity, followed by medium severity. The most problematic areas have been improper input validation and denial of service attacks. A hybrid scanning tool that combines the three scanners bandit, snyk and dlint, which provide a clear report of the code vulnerability, is also described.Keywords: Python vulnerabilities, bandit, Snyk, Dlint, Python package index, ecosystem, static analysis, malicious attacks
Procedia PDF Downloads 1401234 Test Suite Optimization Using an Effective Meta-Heuristic BAT Algorithm
Authors: Anuradha Chug, Sunali Gandhi
Abstract:
Regression Testing is a very expensive and time-consuming process carried out to ensure the validity of modified software. Due to the availability of insufficient resources to re-execute all the test cases in time constrained environment, efforts are going on to generate test data automatically without human efforts. Many search based techniques have been proposed to generate efficient, effective as well as optimized test data, so that the overall cost of the software testing can be minimized. The generated test data should be able to uncover all potential lapses that exist in the software or product. Inspired from the natural behavior of bat for searching her food sources, current study employed a meta-heuristic, search-based bat algorithm for optimizing the test data on the basis certain parameters without compromising their effectiveness. Mathematical functions are also applied that can effectively filter out the redundant test data. As many as 50 Java programs are used to check the effectiveness of proposed test data generation and it has been found that 86% saving in testing efforts can be achieved using bat algorithm while covering 100% of the software code for testing. Bat algorithm was found to be more efficient in terms of simplicity and flexibility when the results were compared with another nature inspired algorithms such as Firefly Algorithm (FA), Hill Climbing Algorithm (HC) and Ant Colony Optimization (ACO). The output of this study would be useful to testers as they can achieve 100% path coverage for testing with minimum number of test cases.Keywords: regression testing, test case selection, test case prioritization, genetic algorithm, bat algorithm
Procedia PDF Downloads 3811233 Communication Styles of Business Students: A Comparison of Four National Cultures
Authors: Tiina Brandt, Isaac Wanasika
Abstract:
Culturally diverse global companies need to understand cultural differences between leaders and employees from different backgrounds. Communication is culturally contingent and has a significant impact on effective execution of leadership goals. The awareness of cultural variations related to communication and interactions will help leaders modify their own behavior, and consequently improve the execution of goals and avoid unnecessary faux pas. Our focus is on young adults that have experienced cultural integration, culturally diverse surroundings in schools and universities, and cultural travels. Our central research problem is to understand the impact of different national cultures on communication. We focus on four countries with distinct national cultures and spatial distribution. The countries are Finland, Indonesia, Russia and USA. Our sample is based on business students (n = 225) from various backgrounds in the four countries. Their responses of communication and leadership styles were analyzed using ANOVA and post-hoc test. Results indicate that culture impacts on communication behavior. Even young culturally-exposed adults with cultural awareness and experience demonstrate cultural differences in their behavior. Apparently, culture is a deeply seated trait that cannot be completely neutralized by environmental variables. Our study offers valuable input for leadership training programs and for expatriates when recognizing specific differences on leaders’ behavior due to culture.Keywords: communication, culture, interaction, leadership
Procedia PDF Downloads 1131232 Windstorm Risk Assessment for Offshore Wind Farms in the North Sea
Authors: Paul Buchana, Patrick E. Mc Sharry
Abstract:
In 2017 there will be about 38 wind farms in the North Sea belonging to 5 different countries. The North Sea is ideal for offshore wind power generation and is thus attractive to offshore wind energy developers and investors. With concerns about the potential for offshore wind turbines to sustain substantial damage as a result of extreme weather conditions, particularly windstorms, this poses a unique challenge to insurers and reinsurers as to adequately quantify the risk and offer appropriate insurance cover for these assets. The need to manage this risk also concerns regulators, who provide the oversight needed to ensure that if a windstorm or a series of storms occur in this area over a one-year time frame, the insurers of these assets in the EU remain solvent even after meeting consequent damage costs. In this paper, using available European windstorm data for the past 33 years and actual wind farm locations together with information pertaining to each of the wind farms (number of turbines, total capacity and financial value), we present a Monte Carlo simulation approach to assess the number of turbines that would be buckled in each of the wind farms using maximum wind speeds reaching each of them. These wind speeds are drawn from historical windstorm data. From the number of turbines buckled, associated financial loss and output capacity can be deduced. The results presented in this paper are targeted towards offshore wind energy developers, insurance and reinsurance companies and regulators.Keywords: catastrophe modeling, North Sea wind farms, offshore wind power, risk analysis
Procedia PDF Downloads 2991231 Reduction Conditions of Briquetted Solid Wastes Generated by the Integrated Iron and Steel Plant
Authors: Gökhan Polat, Dicle Kocaoğlu Yılmazer, Muhlis Nezihi Sarıdede
Abstract:
Iron oxides are the main input to produce iron in integrated iron and steel plants. During production of iron from iron oxides, some wastes with high iron content occur. These main wastes can be classified as basic oxygen furnace (BOF) sludge, flue dust and rolling scale. Recycling of these wastes has a great importance for both environmental effects and reduction of production costs. In this study, recycling experiments were performed on basic oxygen furnace sludge, flue dust and rolling scale which contain 53.8%, 54.3% and 70.2% iron respectively. These wastes were mixed together with coke as reducer and these mixtures are pressed to obtain cylindrical briquettes. These briquettes were pressed under various compacting forces from 1 ton to 6 tons. Also, both stoichiometric and twice the stoichiometric cokes were added to investigate effect of coke amount on reduction properties of the waste mixtures. Then, these briquettes were reduced at 1000°C and 1100°C during 30, 60, 90, 120 and 150 min in a muffle furnace. According to the results of reduction experiments, the effect of compacting force, temperature and time on reduction ratio of the wastes were determined. It is found that 1 ton compacting force, 150 min reduction time and 1100°C are the optimum conditions to obtain reduction ratio higher than 75%.Keywords: Coke, iron oxide wastes, recycling, reduction
Procedia PDF Downloads 3401230 Real-Time Multi-Vehicle Tracking Application at Intersections Based on Feature Selection in Combination with Color Attribution
Authors: Qiang Zhang, Xiaojian Hu
Abstract:
In multi-vehicle tracking, based on feature selection, the tracking system efficiently tracks vehicles in a video with minimal error in combination with color attribution, which focuses on presenting a simple and fast, yet accurate and robust solution to the problem such as inaccurately and untimely responses of statistics-based adaptive traffic control system in the intersection scenario. In this study, a real-time tracking system is proposed for multi-vehicle tracking in the intersection scene. Considering the complexity and application feasibility of the algorithm, in the object detection step, the detection result provided by virtual loops were post-processed and then used as the input for the tracker. For the tracker, lightweight methods were designed to extract and select features and incorporate them into the adaptive color tracking (ACT) framework. And the approbatory online feature selection algorithms are integrated on the mature ACT system with good compatibility. The proposed feature selection methods and multi-vehicle tracking method are evaluated on KITTI datasets and show efficient vehicle tracking performance when compared to the other state-of-the-art approaches in the same category. And the system performs excellently on the video sequences recorded at the intersection. Furthermore, the presented vehicle tracking system is suitable for surveillance applications.Keywords: real-time, multi-vehicle tracking, feature selection, color attribution
Procedia PDF Downloads 1631229 Ground Motion Modelling in Bangladesh Using Stochastic Method
Authors: Mizan Ahmed, Srikanth Venkatesan
Abstract:
Geological and tectonic framework indicates that Bangladesh is one of the most seismically active regions in the world. The Bengal Basin is at the junction of three major interacting plates: the Indian, Eurasian, and Burma Plates. Besides there are many active faults within the region, e.g. the large Dauki fault in the north. The country has experienced a number of destructive earthquakes due to the movement of these active faults. Current seismic provisions of Bangladesh are mostly based on earthquake data prior to the 1990. Given the record of earthquakes post 1990, there is a need to revisit the design provisions of the code. This paper compares the base shear demand of three major cities in Bangladesh: Dhaka (the capital city), Sylhet, and Chittagong for earthquake scenarios of magnitudes 7.0MW, 7.5MW, 8.0MW and 8.5MW using a stochastic model. In particular, the stochastic model allows the flexibility to input region specific parameters such as shear wave velocity profile (that were developed from Global Crustal Model CRUST2.0) and include the effects of attenuation as individual components. Effects of soil amplification were analysed using the Extended Component Attenuation Model (ECAM). Results show that the estimated base shear demand is higher in comparison with code provisions leading to the suggestion of additional seismic design consideration in the study regions.Keywords: attenuation, earthquake, ground motion, Stochastic, seismic hazard
Procedia PDF Downloads 2491228 Efficiency of Membrane Distillation to Produce Fresh Water
Authors: Sabri Mrayed, David Maccioni, Greg Leslie
Abstract:
Seawater desalination has been accepted as one of the most effective solutions to the growing problem of a diminishing clean drinking water supply. Currently, two desalination technologies dominate the market – the thermally driven multi-stage flash distillation (MSF) and the membrane based reverse osmosis (RO). However, in recent years membrane distillation (MD) has emerged as a potential alternative to the established means of desalination. This research project intended to determine the viability of MD as an alternative process to MSF and RO for seawater desalination. Specifically the project involves conducting a thermodynamic analysis of the process based on the second law of thermodynamics to determine the efficiency of the MD. Data was obtained from experiments carried out on a laboratory rig. In order to determine exergy values required for the exergy analysis, two separate models were built in Engineering Equation Solver – the ’Minimum Separation Work Model’ and the ‘Stream Exergy Model’. The efficiency of MD process was found to be 17.3 %, and the energy consumption was determined to be 4.5 kWh to produce one cubic meter of fresh water. The results indicate MD has potential as a technique for seawater desalination compared to RO and MSF. However, it was shown that this was only the case if an alternate energy source such as green or waste energy was available to provide the thermal energy input to the process. If the process was required to power itself, it was shown to be highly inefficient and in no way thermodynamically viable as a commercial desalination process.Keywords: desalination, exergy, membrane distillation, second law efficiency
Procedia PDF Downloads 3641227 Computational Approach for Grp78–Nf-ΚB Binding Interactions in the Context of Neuroprotective Pathway in Brain Injuries
Authors: Janneth Gonzalez, Marco Avila, George Barreto
Abstract:
GRP78 participates in multiple functions in the cell during normal and pathological conditions, controlling calcium homeostasis, protein folding and unfolded protein response. GRP78 is located in the endoplasmic reticulum, but it can change its location under stress, hypoxic and apoptotic conditions. NF-κB represents the keystone of the inflammatory process and regulates the transcription of several genes related with apoptosis, differentiation, and cell growth. The possible relationship between GRP78-NF-κB could support and explain several mechanisms that may regulate a variety of cell functions, especially following brain injuries. Although several reports show interactions between NF-κB and heat shock proteins family members, there is a lack of information on how GRP78 may be interacting with NF-κB, and possibly regulating its downstream activation. Therefore, we assessed the computational predictions of the GRP78 (Chain A) and NF-κB complex (IkB alpha and p65) protein-protein interactions. The interaction interface of the docking model showed that the amino acids ASN 47, GLU 215, GLY 403 of GRP78 and THR 54, ASN 182 and HIS 184 of NF-κB are key residues involved in the docking. The electrostatic field between GRP78-NF-κB interfaces and molecular dynamic simulations support the possible interaction between the proteins. In conclusion, this work shed some light in the possible GRP78-NF-κB complex indicating key residues in this crosstalk, which may be used as an input for better drug design strategy targeting NF-κB downstream signaling as a new therapeutic approach following brain injuries.Keywords: computational biology, protein interactions, Grp78, bioinformatics, molecular dynamics
Procedia PDF Downloads 3421226 X-Ray Dynamical Diffraction Rocking Curves in Case of Third Order Nonlinear Renninger Effect
Authors: Minas Balyan
Abstract:
In the third-order nonlinear Takagi’s equations for monochromatic waves and in the third-order nonlinear time-dependent dynamical diffraction equations for X-ray pulses for forbidden reflections the Fourier-coefficients of the linear and the third order nonlinear susceptibilities are zero. The dynamical diffraction in the nonlinear case is related to the presence in the nonlinear equations the terms proportional to the zero order and the second order nonzero Fourier coefficients of the third order nonlinear susceptibility. Thus in the third order nonlinear Bragg diffraction case a nonlinear analogue of the well known Renninger effect takes place. In this work, the ‘third order nonlinear Renninger effect’ is considered theoretically and numerically. If the reflection exactly is forbidden the diffracted wave’s amplitude is zero both in Laue and Bragg cases since the boundary conditions and dynamical diffraction equations are compatible with zero solution. But in real crystals due to some percent of dislocations and other localized defects, the atoms are displaced with respect to their equilibrium positions. Thus in real crystals susceptibilities of forbidden reflection are by some order small than for usual not forbidden reflections but are not exactly equal to zero. The numerical calculations for susceptibilities two order less than for not forbidden reflection show that in Bragg geometry case the nonlinear reflection curve’s behavior is the same as for not forbidden reflection, but for forbidden reflection the rocking curves’ width, center and boundaries are two order sensitive on the input intensity value. This gives an opportunity to investigate third order nonlinear X-ray dynamical diffraction for not intense beams – 0.001 in the units of critical intensity.Keywords: third order nonlinearity, Bragg diffraction, nonlinear Renninger effect, rocking curves
Procedia PDF Downloads 4071225 Principle Component Analysis on Colon Cancer Detection
Authors: N. K. Caecar Pratiwi, Yunendah Nur Fuadah, Rita Magdalena, R. D. Atmaja, Sofia Saidah, Ocky Tiaramukti
Abstract:
Colon cancer or colorectal cancer is a type of cancer that attacks the last part of the human digestive system. Lymphoma and carcinoma are types of cancer that attack human’s colon. Colon cancer causes deaths about half a million people every year. In Indonesia, colon cancer is the third largest cancer case for women and second in men. Unhealthy lifestyles such as minimum consumption of fiber, rarely exercising and lack of awareness for early detection are factors that cause high cases of colon cancer. The aim of this project is to produce a system that can detect and classify images into type of colon cancer lymphoma, carcinoma, or normal. The designed system used 198 data colon cancer tissue pathology, consist of 66 images for Lymphoma cancer, 66 images for carcinoma cancer and 66 for normal / healthy colon condition. This system will classify colon cancer starting from image preprocessing, feature extraction using Principal Component Analysis (PCA) and classification using K-Nearest Neighbor (K-NN) method. Several stages in preprocessing are resize, convert RGB image to grayscale, edge detection and last, histogram equalization. Tests will be done by trying some K-NN input parameter setting. The result of this project is an image processing system that can detect and classify the type of colon cancer with high accuracy and low computation time.Keywords: carcinoma, colorectal cancer, k-nearest neighbor, lymphoma, principle component analysis
Procedia PDF Downloads 2051224 Pricing, Production and Inventory Policies Manufacturing under Stochastic Demand and Continuous Prices
Authors: Masoud Rabbani, Majede Smizadeh, Hamed Farrokhi-Asl
Abstract:
We study jointly determining prices and production in a multiple period horizon under a general non-stationary stochastic demand with continuous prices. In some periods we need to increase capacity of production to satisfy demand. This paper presents a model to aid multi-period production capacity planning by quantifying the trade-off between product quality and production cost. The product quality is estimated as the statistical variation from the target performances obtained from the output tolerances of the production machines that manufacture the components. We consider different tolerance for different machines that use to increase capacity. The production cost is estimated as the total cost of owning and operating a production facility during the planning horizon.so capacity planning has cost that impact on price. Pricing products often turns out to be difficult to measure them because customers have a reservation price to pay that impact on price and demand. We decide to determine prices and production for periods after enhance capacity and consider reservation price to determine price. First we use an algorithm base on fuzzy set of the optimal objective function values to determine capacity planning by determine maximize interval from upper bound in minimum objectives and define weight for objectives. Then we try to determine inventory and pricing policies. We can use a lemma to solve a problem in MATLAB and find exact answer.Keywords: price policy, inventory policy, capacity planning, product quality, epsilon -constraint
Procedia PDF Downloads 5691223 Increasing the Efficiency of the Biomass Gasification Technology with Using the Organic Rankin Cycle
Authors: Jaroslav Frantík, Jan Najser
Abstract:
This article deals with increasing the energy efficiency of a plant in terms of optimizing the process. The European Union is striving to achieve the climate-energy package in the area increasing of energy efficiency. The goal of energy efficiency is to reduce primary energy consumption by 20% within the EU until 2020. The objective of saving energy consumption in the Czech Republic was set at 47.84 PJ (13.29 TWh). For reducing electricity consumption, it is possible to choose: a) mandatory increasing of energy efficiency, b) alternative scheme, c) combination of both actions. The Czech Republic has chosen for reducing electricity consumption using-alternative scheme. The presentation is focused on the proposal of a technological unit dealing with the gasification process of processing of biomass with an increase of power in the output. The synthesis gas after gasification of biomass is used as fuel in a cogeneration process of reciprocating internal combustion engine with the classic production of heat and electricity. Subsequently, there is an explanation of the ORC system dealing with the conversion of waste heat to electricity with the using closed cycle of the steam process with organic medium. The arising electricity is distributed to the power grid as a further energy source, or it is used for needs of the partial coverage of the technological unit. Furthermore, there is a presented schematic description of the technology with the identification of energy flows starting from the biomass treatment by drying, through its conversion to gaseous fuel, producing of electricity and utilize of thermal energy with minimizing losses. It has been found that using of ORC system increased the efficiency of the produced electricity by 7.5%.Keywords: biomass, efficiency, gasification, ORC system
Procedia PDF Downloads 2171222 First and Second Analysis on the Reheat Organic Rankine Cycle
Authors: E. Moradimaram, H. Sayehvand
Abstract:
In recent years the increasing use of fossil fuels has led to various environmental problems including urban pollution, ozone layer depletion and acid rains. Moreover, with the increased number of industrial centers and higher consumption of these fuels, the end point of the fossil energy reserves has become more evident. Considering the environmental pollution caused by fossil fuels and their limited availability, renewable sources can be considered as the main substitute for non-renewable resources. One of these resources is the Organic Rankine Cycles (ORCs). These cycles while having high safety, have low maintenance requirements. Combining the ORCs with other systems, such as ejector and reheater will increase overall cycle efficiency. In this study, ejector and reheater are used to improve the thermal efficiency (ηth), exergy efficiency (η_ex) and net output power (w_net); therefore, the ORCs with reheater (RORCs) are proposed. A computational program has been developed to calculate the thermodynamic parameters required in Engineering Equations Solver (EES). In this program, the analysis of the first and second law in RORC is conducted, and a comparison is made between them and the ORCs with Ejector (EORC). R245fa is selected as the working fluid and water is chosen as low temperature heat source with a temperature of 95 °C and a mass transfer rate of 1 kg/s. The pressures of the second evaporator and reheater are optimized in terms of maximum exergy efficiency. The environment is at 298.15 k and at 101.325 kpa. The results indicate that the thermodynamic parameters in the RORC have improved compared to EORC.Keywords: Organic Rankine Cycle (ORC), Organic Rankine Cycle with Reheater (RORC), Organic Rankine Cycle with Ejector (EORC), exergy efficiency
Procedia PDF Downloads 1631221 Fuzzy Inference-Assisted Saliency-Aware Convolution Neural Networks for Multi-View Summarization
Authors: Tanveer Hussain, Khan Muhammad, Amin Ullah, Mi Young Lee, Sung Wook Baik
Abstract:
The Big Data generated from distributed vision sensors installed on large scale in smart cities create hurdles in its efficient and beneficial exploration for browsing, retrieval, and indexing. This paper presents a three-folded framework for effective video summarization of such data and provide a compact and representative format of Big Video Data. In the first fold, the paper acquires input video data from the installed cameras and collect clues such as type and count of objects and clarity of the view from a chunk of pre-defined number of frames of each view. The decision of representative view selection for a particular interval is based on fuzzy inference system, acquiring a precise and human resembling decision, reinforced by the known clues as a part of the second fold. In the third fold, the paper forwards the selected view frames to the summary generation mechanism that is supported by a saliency-aware convolution neural network (CNN) model. The new trend of fuzzy rules for view selection followed by CNN architecture for saliency computation makes the multi-view video summarization (MVS) framework a suitable candidate for real-world practice in smart cities.Keywords: big video data analysis, fuzzy logic, multi-view video summarization, saliency detection
Procedia PDF Downloads 1881220 Single Chip Controller Design for Piezoelectric Actuators with Mixed Signal FPGA
Authors: Han-Bin Park, Taesam Kang, SunKi Hong, Jeong Hoi Gu
Abstract:
The piezoelectric material is being used widely for actuators due to its large power density with simple structure. It can generate a larger force than the conventional actuators with the same size. Furthermore, the response time of piezoelectric actuators is very short, and thus, it can be used for very fast system applications with compact size. To control the piezoelectric actuator, we need analog signal conditioning circuits as well as digital microcontrollers. Conventional microcontrollers are not equipped with analog parts and thus the control system becomes bulky compared with the small size of the piezoelectric devices. To overcome these weaknesses, we are developing one-chip micro controller that can handle analog and digital signals simultaneously using mixed signal FPGA technology. We used the SmartFusion™ FPGA device that integrates ARM®Cortex-M3, analog interface and FPGA fabric in a single chip and offering full customization. It gives more flexibility than traditional fixed-function microcontrollers with the excessive cost of soft processor cores on traditional FPGAs. In this paper we introduce the design of single chip controller using mixed signal FPGA, SmartFusion™[1] device. To demonstrate its performance, we implemented a PI controller for power driving circuit and a 5th order H-infinity controller for the system with piezoelectric actuator in the FPGA fabric. We also demonstrated the regulation of a power output and the operation speed of a 5th order H-infinity controller.Keywords: mixed signal FPGA, PI control, piezoelectric actuator, SmartFusion™
Procedia PDF Downloads 5201219 Bias Prevention in Automated Diagnosis of Melanoma: Augmentation of a Convolutional Neural Network Classifier
Authors: Kemka Ihemelandu, Chukwuemeka Ihemelandu
Abstract:
Melanoma remains a public health crisis, with incidence rates increasing rapidly in the past decades. Improving diagnostic accuracy to decrease misdiagnosis using Artificial intelligence (AI) continues to be documented. Unfortunately, unintended racially biased outcomes, a product of lack of diversity in the dataset used, with a noted class imbalance favoring lighter vs. darker skin tone, have increasingly been recognized as a problem.Resulting in noted limitations of the accuracy of the Convolutional neural network (CNN)models. CNN models are prone to biased output due to biases in the dataset used to train them. Our aim in this study was the optimization of convolutional neural network algorithms to mitigate bias in the automated diagnosis of melanoma. We hypothesized that our proposed training algorithms based on a data augmentation method to optimize the diagnostic accuracy of a CNN classifier by generating new training samples from the original ones will reduce bias in the automated diagnosis of melanoma. We applied geometric transformation, including; rotations, translations, scale change, flipping, and shearing. Resulting in a CNN model that provided a modifiedinput data making for a model that could learn subtle racial features. Optimal selection of the momentum and batch hyperparameter increased our model accuracy. We show that our augmented model reduces bias while maintaining accuracy in the automated diagnosis of melanoma.Keywords: bias, augmentation, melanoma, convolutional neural network
Procedia PDF Downloads 2111218 Impact of Climate on Productivity of Major Cereal Crops in Sokoto State, Nigeria
Authors: M. B. Sokoto, L. Tanko, Y. M. Abdullahi
Abstract:
The study aimed at examining the impact of climatic factors (rainfall, minimum and maximum temperature) on the productivity of major cereals in Sokoto state, Nigeria. Secondary data from 1997-2008 were used in respect of annual yield of Major cereals crops (maize, millet, rice, and sorghum (t ha-1). Data in respect of climate was collected from Sokoto Energy Research Centre (SERC) for the period under review. Data collected was analyzed using descriptive statistics, correlation and regression analysis. The result of the research reveals that there is variation in the trend of the climatic factors and also variation in cereals output. The effect of average temperature on yields has a negative effect on crop yields. Similarly, rainfall is not significant in explaining the effect of climate on cereal crops production. The study has revealed to some extend the effect of climatic variables, such as rainfall, relative humidity, maximum and minimum temperature on major cereals production in Sokoto State. This will assist in planning ahead in cereals production in the area. Other factors such as soil fertility, correct timing of planting and good cultural practices (such as spacing of strands), protection of crops from weeds, pests and diseases and planting of high yielding varieties should also be taken into consideration for increase yield of cereals.Keywords: cereals, climate, impact, major, productivity
Procedia PDF Downloads 3901217 Comparison between Deterministic and Probabilistic Stability Analysis, Featuring Consequent Risk Assessment
Authors: Isabela Moreira Queiroz
Abstract:
Slope stability analyses are largely carried out by deterministic methods and evaluated through a single security factor. Although it is known that the geotechnical parameters can present great dispersal, such analyses are considered fixed and known. The probabilistic methods, in turn, incorporate the variability of input key parameters (random variables), resulting in a range of values of safety factors, thus enabling the determination of the probability of failure, which is an essential parameter in the calculation of the risk (probability multiplied by the consequence of the event). Among the probabilistic methods, there are three frequently used methods in geotechnical society: FOSM (First-Order, Second-Moment), Rosenblueth (Point Estimates) and Monte Carlo. This paper presents a comparison between the results from deterministic and probabilistic analyses (FOSM method, Monte Carlo and Rosenblueth) applied to a hypothetical slope. The end was held to evaluate the behavior of the slope and consequent risk analysis, which is used to calculate the risk and analyze their mitigation and control solutions. It can be observed that the results obtained by the three probabilistic methods were quite close. It should be noticed that the calculation of the risk makes it possible to list the priority to the implementation of mitigation measures. Therefore, it is recommended to do a good assessment of the geological-geotechnical model incorporating the uncertainty in viability, design, construction, operation and closure by means of risk management.Keywords: probabilistic methods, risk assessment, risk management, slope stability
Procedia PDF Downloads 3921216 Optimization Techniques of Doubly-Fed Induction Generator Controller Design for Reliability Enhancement of Wind Energy Conversion Systems
Authors: Om Prakash Bharti, Aanchal Verma, R. K. Saket
Abstract:
The Doubly-Fed Induction Generator (DFIG) is suggested for Wind Energy Conversion System (WECS) to extract wind power. DFIG is preferably employed due to its robustness towards variable wind and rotor speed. DFIG has the adaptable property because the system parameters are smoothly dealt with, including real power, reactive power, DC-link voltage, and the transient and dynamic responses, which are needed to analyze constantly. The analysis becomes more prominent during any unusual condition in the electrical power system. Hence, the study and improvement in the system parameters and transient response performance of DFIG are required to be accomplished using some controlling techniques. For fulfilling the task, the present work implements and compares the optimization methods for the design of the DFIG controller for WECS. The bio-inspired optimization techniques are applied to get the optimal controller design parameters for DFIG-based WECS. The optimized DFIG controllers are then used to retrieve the transient response performance of the six-order DFIG model with a step input. The results using MATLAB/Simulink show the betterment of the Firefly algorithm (FFA) over other control techniques when compared with the other controller design methods.Keywords: doubly-fed induction generator, wind turbine, wind energy conversion system, induction generator, transfer function, proportional, integral, derivatives
Procedia PDF Downloads 931215 BIM-Based Tool for Sustainability Assessment and Certification Documents Provision
Authors: Taki Eddine Seghier, Mohd Hamdan Ahmad, Yaik-Wah Lim, Samuel Opeyemi Williams
Abstract:
The assessment of building sustainability to achieve a specific green benchmark and the preparation of the required documents in order to receive a green building certification, both are considered as major challenging tasks for green building design team. However, this labor and time-consuming process can take advantage of the available Building Information Modeling (BIM) features such as material take-off and scheduling. Furthermore, the workflow can be automated in order to track potentially achievable credit points and provide rating feedback for several design options by using integrated Visual Programing (VP) to handle the stored parameters within the BIM model. Hence, this study proposes a BIM-based tool that uses Green Building Index (GBI) rating system requirements as a unique input case to evaluate the building sustainability in the design stage of the building project life cycle. The tool covers two key models for data extraction, firstly, a model for data extraction, calculation and the classification of achievable credit points in a green template, secondly, a model for the generation of the required documents for green building certification. The tool was validated on a BIM model of residential building and it serves as proof of concept that building sustainability assessment of GBI certification can be automatically evaluated and documented through BIM.Keywords: green building rating system, GBRS, building information modeling, BIM, visual programming, VP, sustainability assessment
Procedia PDF Downloads 3261214 Efficient Estimation for the Cox Proportional Hazards Cure Model
Authors: Khandoker Akib Mohammad
Abstract:
While analyzing time-to-event data, it is possible that a certain fraction of subjects will never experience the event of interest, and they are said to be cured. When this feature of survival models is taken into account, the models are commonly referred to as cure models. In the presence of covariates, the conditional survival function of the population can be modelled by using the cure model, which depends on the probability of being uncured (incidence) and the conditional survival function of the uncured subjects (latency), and a combination of logistic regression and Cox proportional hazards (PH) regression is used to model the incidence and latency respectively. In this paper, we have shown the asymptotic normality of the profile likelihood estimator via asymptotic expansion of the profile likelihood and obtain the explicit form of the variance estimator with an implicit function in the profile likelihood. We have also shown the efficient score function based on projection theory and the profile likelihood score function are equal. Our contribution in this paper is that we have expressed the efficient information matrix as the variance of the profile likelihood score function. A simulation study suggests that the estimated standard errors from bootstrap samples (SMCURE package) and the profile likelihood score function (our approach) are providing similar and comparable results. The numerical result of our proposed method is also shown by using the melanoma data from SMCURE R-package, and we compare the results with the output obtained from the SMCURE package.Keywords: Cox PH model, cure model, efficient score function, EM algorithm, implicit function, profile likelihood
Procedia PDF Downloads 1441213 Modeling and Simulation of Pad Surface Topography by Diamond Dressing in Chemical-Mechanical Polishing Process
Authors: A.Chen Chao-Chang, Phong Pham-Quoc
Abstract:
Chemical-mechanical polishing (CMP) process has been widely applied on fabricating integrated circuits (IC) with a soft polishing pad combined with slurry composed of micron or nano-scaled abrasives for generating chemical reaction to remove substrate or film materials from wafer. During CMP process, pad uniformity usually works as a datum surface of wafer planarization and pad asperities can dominate the microscopic pad-slurry-wafer interaction. However, pad topography can be changed by related mechanism factors of CMP and it needs to be re-conditioned or dressed by a diamond dresser of well-distributed diamond grits on a disc surface. It is still very complicated to analyze and understand kinematic of diamond dressing process under the effects of input variables including oscillatory of diamond dresser and rotation speed ratio between the pad and the diamond dresser. This paper has developed a generic geometric model to clarify the kinematic modeling of diamond dressing processes such as dresser/pad motion, pad cutting locus, the relative velocity of the diamond abrasive grits on pad surface, and overlap of cutting for prediction of pad surface topography. Simulation results focus on comparing and analysis kinematics of the diamond dressing on certain CMP tools. Results have shown the significant parameters for diamond dressing process and also discussed. Future study can apply on diamond dresser design and experimental verification of pad dressing process.Keywords: kinematic modeling, diamond dresser, pad cutting locus, CMP
Procedia PDF Downloads 2551212 Excitation of Guided Waves in Finite Width Plates Using a Numerical Approach
Authors: Wenbo Duan, Hossein Habibi, Vassilios Kappatos, Cem Selcuk, Tat-Hean Gan
Abstract:
Ultrasonic guided waves are often used to remove ice or fouling in different structures, such as ship hulls, wind turbine blades and so on. To achieve maximum sound power output, it is important that multiple transducers are arranged in a particular way so that a desired mode can be excited. The objective of this paper is thus to provide a theoretical basis for generating a particular mode in a finite width rectangular plate which can be used for removing potential ice or fouling on the plate. The number of transducers and their locations with respect to a particular mode will be investigated, and the link between dispersion curves and practical applications will be explored. To achieve this, a semi-analytical finite element (SAFE) method is used to study the dispersion characteristics of all the modes in the ultrasonic frequency range. The detailed modal shapes will be revealed, and from the modal analysis, the particular mode with the strongest yet continuous transverse and axial displacements on the surfaces of the plate will be chosen for the purpose of removing potential ice or fouling on the plate. The modal analysis is followed by providing information on the number, location and amplitude of transducers needed to excite this particular mode. Modal excitation is then implemented in a standard finite element commercial package, namely COMSOL Multiphysics. Wave motion is visualized in COMSOL, and the mode shapes generated in SAFE is found to be consistent with the mode shapes generated in COMSOL.Keywords: dispersion analysis, finite width plate, guided wave, modal excitation
Procedia PDF Downloads 4741211 Control System Design for a Simulated Microbial Electrolysis Cell
Authors: Pujari Muruga, T. K. Radhakrishnan, N. Samsudeen
Abstract:
Hydrogen is considered as the most important energy carrier and fuel of the future because of its high energy density and zero emission properties. Microbial Electrolysis Cell (MEC) is a new and promising approach for hydrogen production from organic matter, including wastewater and other renewable resources. By utilizing anode microorganism activity, MEC can produce hydrogen gas with smaller voltages (as low as 0.2 V) than those required for electrolytic hydrogen production ( ≥ 1.23 V). The hydrogen production processes of the MEC reactor are very nonlinear and highly complex because of the presence of microbial interactions and highly complex phenomena in the system. Increasing the hydrogen production rate and lowering the energy input are two important challenges of MEC technology. The mathematical model of the MEC is based on material balance with the integration of bioelectrochemical reactions. The main objective of the research is to produce biohydrogen by selecting the optimum current and controlling applied voltage to the MEC. Precise control is required for the MEC reactor, so that the amount of current required to produce hydrogen gas can be controlled according to the composition of the substrate in the reactor. Various simulation tests involving multiple set-point changes disturbance and noise rejection were performed to evaluate the performance using PID controller tuned with Ziegler Nichols settings. Simulation results shows that other good controller can provide better control effect on the MEC system, so that higher hydrogen production can be obtained.Keywords: microbial electrolysis cell, hydrogen production, applied voltage, PID controller
Procedia PDF Downloads 247