Search results for: sequential extraction
541 Adaptive Gait Pattern Generation of Biped Robot based on Human's Gait Pattern Analysis
Authors: Seungsuk Ha, Youngjoon Han, Hernsoo Hahn
Abstract:
This paper proposes a method of adaptively generating a gait pattern of biped robot. The gait synthesis is based on human's gait pattern analysis. The proposed method can easily be applied to generate the natural and stable gait pattern of any biped robot. To analyze the human's gait pattern, sequential images of the human's gait on the sagittal plane are acquired from which the gait control values are extracted. The gait pattern of biped robot on the sagittal plane is adaptively generated by a genetic algorithm using the human's gait control values. However, gait trajectories of the biped robot on the sagittal plane are not enough to construct the complete gait pattern because the biped robot moves on 3-dimension space. Therefore, the gait pattern on the frontal plane, generated from Zero Moment Point (ZMP), is added to the gait one acquired on the sagittal plane. Consequently, the natural and stable walking pattern for the biped robot is obtained.
Keywords: Biped robot, gait pattern, genetic algorithm.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2272540 Deep Reinforcement Learning for Optimal Decision-making in Supply Chains
Authors: Nitin Singh, Meng Ling, Talha Ahmed, Tianxia Zhao, Reinier van de Pol
Abstract:
We propose the use of Reinforcement Learning (RL) as a viable alternative for optimizing supply chain management, particularly in scenarios with stochasticity in product demands. RL’s adaptability to changing conditions and its demonstrated success in diverse fields of sequential decision-making make it a promising candidate for addressing supply chain problems. We investigate the impact of demand fluctuations in a multi-product supply chain system and develop RL agents with learned generalizable policies. We provide experimentation details for training RL agents and a statistical analysis of the results. We study generalization ability of RL agents for different demand uncertainty scenarios and observe superior performance compared to the agents trained with fixed demand curves. The proposed methodology has the potential to lead to cost reduction and increased profit for companies dealing with frequent inventory movement between supply and demand nodes.
Keywords: Inventory Management, Reinforcement Learning, Supply Chain Optimization, Uncertainty.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 390539 Sparse Networks-Based Speedup Technique for Proteins Betweenness Centrality Computation
Authors: Razvan Bocu, Dr Sabin Tabirca
Abstract:
The study of proteomics reached unexpected levels of interest, as a direct consequence of its discovered influence over some complex biological phenomena, such as problematic diseases like cancer. This paper presents the latest authors- achievements regarding the analysis of the networks of proteins (interactome networks), by computing more efficiently the betweenness centrality measure. The paper introduces the concept of betweenness centrality, and then describes how betweenness computation can help the interactome net- work analysis. Current sequential implementations for the between- ness computation do not perform satisfactory in terms of execution times. The paper-s main contribution is centered towards introducing a speedup technique for the betweenness computation, based on modified shortest path algorithms for sparse graphs. Three optimized generic algorithms for betweenness computation are described and implemented, and their performance tested against real biological data, which is part of the IntAct dataset.Keywords: Betweenness centrality, interactome networks, protein-protein interactions, sub-communities, sparse networks, speedup tech-nique, IntAct.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1509538 Single Event Transient Tolerance Analysis in 8051 Microprocessor Using Scan Chain
Authors: Jun Sung Go, Jong Kang Park, Jong Tae Kim
Abstract:
As semi-conductor manufacturing technology evolves; the single event transient problem becomes more significant issue. Single event transient has a critical impact on both combinational and sequential logic circuits, so it is important to evaluate the soft error tolerance of the circuits at the design stage. In this paper, we present a soft error detecting simulation using scan chain. The simulation model generates a single event transient randomly in the circuit, and detects the soft error during the execution of the test patterns. We verified this model by inserting a scan chain in an 8051 microprocessor using 65 nm CMOS technology. While the test patterns generated by ATPG program are passing through the scan chain, we insert a single event transient and detect the number of soft errors per sub-module. The experiments show that the soft error rates per cell area of the SFR module is 277% larger than other modules.Keywords: Scan chain, single event transient, soft error, 8051 processor.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1492537 Lung Cancer Detection and Multi Level Classification Using Discrete Wavelet Transform Approach
Authors: V. Veeraprathap, G. S. Harish, G. Narendra Kumar
Abstract:
Uncontrolled growth of abnormal cells in the lung in the form of tumor can be either benign (non-cancerous) or malignant (cancerous). Patients with Lung Cancer (LC) have an average of five years life span expectancy provided diagnosis, detection and prediction, which reduces many treatment options to risk of invasive surgery increasing survival rate. Computed Tomography (CT), Positron Emission Tomography (PET), and Magnetic Resonance Imaging (MRI) for earlier detection of cancer are common. Gaussian filter along with median filter used for smoothing and noise removal, Histogram Equalization (HE) for image enhancement gives the best results without inviting further opinions. Lung cavities are extracted and the background portion other than two lung cavities is completely removed with right and left lungs segmented separately. Region properties measurements area, perimeter, diameter, centroid and eccentricity measured for the tumor segmented image, while texture is characterized by Gray-Level Co-occurrence Matrix (GLCM) functions, feature extraction provides Region of Interest (ROI) given as input to classifier. Two levels of classifications, K-Nearest Neighbor (KNN) is used for determining patient condition as normal or abnormal, while Artificial Neural Networks (ANN) is used for identifying the cancer stage is employed. Discrete Wavelet Transform (DWT) algorithm is used for the main feature extraction leading to best efficiency. The developed technology finds encouraging results for real time information and on line detection for future research.
Keywords: ANN, DWT, GLCM, KNN, ROI, artificial neural networks, discrete wavelet transform, gray-level co-occurrence matrix, k-nearest neighbor, region of interest.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 961536 Process Analysis through Length Consistency
Authors: James E. Ponder
Abstract:
The requirement for consistency in physics can sometimes offer a common ground between disciplines such that their fundamental equations share a common parameter set and mathematical method for equation extraction. The parameter set shared by Relativity and Quantum Wave Mechanics enables an analysis which will be seen to be very straightforward, primarily classical in nature using linear algebra concepts, yet deriving a theoretical estimate of the value of the Gravitational Constant along with dependencies never before known.
Keywords: Gravitational Constant, Physical Consistency, Quantum Mechanics, Relativity.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1543535 Using Data Fusion for Biometric Verification
Authors: Richard A. Wasniowski
Abstract:
A wide spectrum of systems require reliable personal recognition schemes to either confirm or determine the identity of an individual person. This paper considers multimodal biometric system and their applicability to access control, authentication and security applications. Strategies for feature extraction and sensor fusion are considered and contrasted. Issues related to performance assessment, deployment and standardization are discussed. Finally future directions of biometric systems development are discussed.Keywords: Multimodal, biometric, recognition, fusion.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1772534 Barriers to the Use of Factoring Accounts Receivables: The Ghanaian Contractor’s Perception
Authors: E. Kissi, V. K. Acheamfour, J. J. Gyimah, T. Adjei-Kumi
Abstract:
Factoring accounts receivable is widely accepted as an alternative financing source and utilized in almost every industry that sells business-to-business or business-to-government. However, its patronage in the construction industry is very limited as some barriers hinder its application in the construction industry. This study aims at assessing the barriers to the use of factoring accounts receivables in the Ghanaian construction industry. The study adopted the sequential exploratory research method where structured and unstructured questionnaires were conveniently distributed to D1K1 and D2K2 construction firms in Ghana. Using the one-sample t-test and Kendall’s Coefficient of concordance data were analyzed. The most severe challenge concluded is the high cost of factoring patronage. Other critical challenges identified were low knowledge on factoring processes, inadequate access to information on factoring, and high risks involved in factoring. Hence, it is recommended that contractors should be made aware of the prospects of factoring of accounts receivables in the construction industry. This study serves as basis for further rigorous research into factoring of accounts receivables in the industry.
Keywords: Barriers, contractors, factoring accounts receivables, Ghanaian, perception.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 552533 Multidimensional Compromise Programming Evaluation of Digital Commerce Websites
Authors: C. Ardil
Abstract:
Multidimensional compromise programming evaluation of digital commerce websites is essential not only to have recommendations for improvement, but also to make comparisons with global business competitors. This research provides a multidimensional decision making model that prioritizes the objective criteria weights of various commerce websites using multidimensional compromise solution. Evaluation of digital commerce website quality can be considered as a complex information system structure including qualitative and quantitative factors for a multicriteria decision making problem. The proposed multicriteria decision making approach mainly consists of three sequential steps for the selection problem. In the first step, three major different evaluation criteria are characterized for website ranking problem. In the second step, identified critical criteria are weighted using the standard deviation procedure. In the third step, the multidimensional compromise programming is applied to rank the digital commerce websites.
Keywords: Standard deviation, commerce website, website evaluation, multicriteria decision making, multicriteria compromise programming, website quality, multidimensional decision analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 817532 Translator Design to Model Cpp Files
Authors: Er. Satwinder Singh, Dr. K.S. Kahlon, Rakesh Kumar, Er. Gurjeet Singh
Abstract:
The most reliable and accurate description of the actual behavior of a software system is its source code. However, not all questions about the system can be answered directly by resorting to this repository of information. What the reverse engineering methodology aims at is the extraction of abstract, goal-oriented “views" of the system, able to summarize relevant properties of the computation performed by the program. While concentrating on reverse engineering we had modeled the C++ files by designing the translator.
Keywords: Translator, Modeling, UML, DYNO, ISVis, TED.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1535531 Understanding Physical Activity Behavior of Type 2 Diabetics Using the Theory of Planned Behavior and Structural Equation Modeling
Authors: D. O. Omondi, M. K. Walingo, G. M. Mbagaya, L. O. A. Othuon
Abstract:
Understanding patient factors related to physical activity behavior is important in the management of Type 2 Diabetes. This study applied the Theory of Planned Behavior model to understand physical activity behavior among sampled Type 2 diabetics in Kenya. The study was conducted within the diabetic clinic at Kisii Level 5 Hospital and adopted sequential mixed methods design beginning with qualitative phase and ending with quantitative phase. Qualitative data was analyzed using grounded theory analysis method. Structural equation modeling using maximum likelihood was used to analyze quantitative data. The common fit indices revealed that the theory of planned behavior fitted the data acceptably well among the Type 2 diabetes and within physical activity behavior {¤ç2 = 213, df = 84, n=230, p = .061, ¤ç2/df = 2.53; TLI = .97; CFI =.96; RMSEA (90CI) = .073(.029, .08)}. This theory proved to be useful in understanding physical activity behavior among Type 2 diabetics.
Keywords: Physical activity, Theory of Planned Behavior, Type2 diabetes, Kenya.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1984530 Developing New Processes and Optimizing Performance Using Response Surface Methodology
Authors: S. Raissi
Abstract:
Response surface methodology (RSM) is a very efficient tool to provide a good practical insight into developing new process and optimizing them. This methodology could help engineers to raise a mathematical model to represent the behavior of system as a convincing function of process parameters. Through this paper the sequential nature of the RSM surveyed for process engineers and its relationship to design of experiments (DOE), regression analysis and robust design reviewed. The proposed four-step procedure in two different phases could help system analyst to resolve the parameter design problem involving responses. In order to check accuracy of the designed model, residual analysis and prediction error sum of squares (PRESS) described. It is believed that the proposed procedure in this study can resolve a complex parameter design problem with one or more responses. It can be applied to those areas where there are large data sets and a number of responses are to be optimized simultaneously. In addition, the proposed procedure is relatively simple and can be implemented easily by using ready-made standard statistical packages.Keywords: Response Surface Methodology (RSM), Design of Experiments (DOE), Process modeling, Process setting, Process optimization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1838529 Design of an Efficient Retimed CIC Compensation Filter
Authors: Vishal Awasthi, Krishna Raj
Abstract:
Unwanted side effects because of spectral aliasing and spectral imaging during signal processing would be the major concern over the sampling rate alteration. Multirate-multistage implementation of digital filter could come about a large computational saving than single rate filter suitable for sample rate conversion. This implementation can further improve through high-level architectural transformation in circuit level. Reallocating registers and relocating flip-flops across logic gates through retiming certainly a prominent sequential transformation technology, that optimize hardware circuits to achieve faster clocking speed without affecting the functionality. In this paper, we proposed an efficient compensated cascade Integrator comb (CIC) decimation filter structure that analyze the consequence of filter order variation which has a retimed FIR filter being compensator while using the cutset retiming technique and achieved an improvement in the passband droop by 14% to 39%, in computation time by 38.04%, 25.78%, 12.21%, 6.69% and 4.44% and reduction in path delay by 62.27%, 72%, 86.63%, 91.56% and 94.42% of 3, 6, 8, 12 and 24 order filter respectively than the non-retimed CIC compensation filter.
Keywords: Multirate Filtering, CIC decimation filter, Compensation theory, Retiming, Retiming algorithm, Filter order, Synchronous dataflow graph.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3706528 MABENA Strategic Management Model for Local Companies
Authors: Kaveh Mohammad Cyrus, Shadi Sanagoo
Abstract:
MABENA model is a complementary model in comparison with traditional models such as HCMS, CMS and etc. New factors, which have effects on preparation of strategic plans and their sequential order in MABENA model is the platform of presented road map in this paper.Study review shows, factors such as emerging new critical success factors for strategic planning, improvement of international strategic models, increasing the maturity of companies and emerging new needs leading to design a new model which can be responsible for new critical factors and solve the limitations of previous strategic management models. Preparation of strategic planning need more factors than introduced in traditional models. The needed factors includes determining future Critical Success Factors and competencies, defining key processes, determining the maturity of the processes, considering all aspects of the external environment etc. Description of aforementioned requirements, the outcomes and their order is developing and presenting the MABENA model-s road map in this paper. This study presents a road map for strategic planning of the Iranian organizations.Keywords: Competitive Advantage, Process Maturity, StrategicPlanning, Strategic potential
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2161527 Efficient Large Numbers Karatsuba-Ofman Multiplier Designs for Embedded Systems
Authors: M.Machhout, M.Zeghid, W.El hadj youssef, B.Bouallegue, A.Baganne, R.Tourki
Abstract:
Long number multiplications (n ≥ 128-bit) are a primitive in most cryptosystems. They can be performed better by using Karatsuba-Ofman technique. This algorithm is easy to parallelize on workstation network and on distributed memory, and it-s known as the practical method of choice. Multiplying long numbers using Karatsuba-Ofman algorithm is fast but is highly recursive. In this paper, we propose different designs of implementing Karatsuba-Ofman multiplier. A mixture of sequential and combinational system design techniques involving pipelining is applied to our proposed designs. Multiplying large numbers can be adapted flexibly to time, area and power criteria. Computationally and occupation constrained in embedded systems such as: smart cards, mobile phones..., multiplication of finite field elements can be achieved more efficiently. The proposed designs are compared to other existing techniques. Mathematical models (Area (n), Delay (n)) of our proposed designs are also elaborated and evaluated on different FPGAs devices.Keywords: finite field, Karatsuba-Ofman, long numbers, multiplication, mathematical model, recursivity.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2532526 Javanese Character Recognition Using Hidden Markov Model
Authors: Anastasia Rita Widiarti, Phalita Nari Wastu
Abstract:
Hidden Markov Model (HMM) is a stochastic method which has been used in various signal processing and character recognition. This study proposes to use HMM to recognize Javanese characters from a number of different handwritings, whereby HMM is used to optimize the number of state and feature extraction. An 85.7 % accuracy is obtained as the best result in 16-stated vertical model using pure HMM. This initial result is satisfactory for prompting further research.Keywords: Character recognition, off-line handwritingrecognition, Hidden Markov Model.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1989525 Web Data Scraping Technology Using Term Frequency Inverse Document Frequency to Enhance the Big Data Quality on Sentiment Analysis
Authors: Sangita Pokhrel, Nalinda Somasiri, Rebecca Jeyavadhanam, Swathi Ganesan
Abstract:
Tourism is a booming industry with huge future potential for global wealth and employment. There are countless data generated over social media sites every day, creating numerous opportunities to bring more insights to decision-makers. The integration of big data technology into the tourism industry will allow companies to conclude where their customers have been and what they like. This information can then be used by businesses, such as those in charge of managing visitor centres or hotels, etc., and the tourist can get a clear idea of places before visiting. The technical perspective of natural language is processed by analysing the sentiment features of online reviews from tourists, and we then supply an enhanced long short-term memory (LSTM) framework for sentiment feature extraction of travel reviews. We have constructed a web review database using a crawler and web scraping technique for experimental validation to evaluate the effectiveness of our methodology. The text form of sentences was first classified through VADER and RoBERTa model to get the polarity of the reviews. In this paper, we have conducted study methods for feature extraction, such as Count Vectorization and Term Frequency – Inverse Document Frequency (TFIDF) Vectorization and implemented Convolutional Neural Network (CNN) classifier algorithm for the sentiment analysis to decide if the tourist’s attitude towards the destinations is positive, negative, or simply neutral based on the review text that they posted online. The results demonstrated that from the CNN algorithm, after pre-processing and cleaning the dataset, we received an accuracy of 96.12% for the positive and negative sentiment analysis.
Keywords: Counter vectorization, Convolutional Neural Network, Crawler, data technology, Long Short-Term Memory, LSTM, Web Scraping, sentiment analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 182524 A Review on Design and Fabrication of Fuel Fired Crucible Furnace
Authors: Oluwaseyi O. Taiwo, Adeolu A. Adediran, Abayomi A. Akinwande, Frank C. Okoyeh
Abstract:
The use of fuel fired crucible furnace is essential in the foundries of developing countries owing to the luxury of electricity. Fuel fired crucible furnace are commonly used in recycling, casting, research and training activities in tertiary institutions, therefore, several attempts are being made to improve the performance and service life of fuel fired crucible. The current study reviews the sequential stages involved in the designs and fabrication of fuel fired crucible furnace which include; design, material selection, modelling and simulation as well as performance evaluation. The study shows that selecting appropriate materials for the different units in the fabrication process is important to the efficiency and service life of fuel fired crucible furnaces. Also, efficiency and performance of fuel fired furnaces are independent of cost of fabrication and their capacity. The importance of modelling and simulation tools in the fabrication process are identified while their non-frequent usage in several works is observed. The need to widen performance evaluations in further studies beyond efficiency determination to give a more detailed assessment of fuel fired crucible furnaces is also observed.
Keywords: Crucible furnace, furnace design, fabrication, fuel.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 480523 Processing and Economic Analysis of Rain Tree (Samanea saman) Pods for Village Level Hydrous Bioethanol Production
Authors: Dharell B. Siano, Wendy C. Mateo, Victorino T. Taylan, Francisco D. Cuaresma
Abstract:
Biofuel is one of the renewable energy sources adapted by the Philippine government in order to lessen the dependency on foreign fuel and to reduce carbon dioxide emissions. Rain tree pods were seen to be a promising source of bioethanol since it contains significant amount of fermentable sugars. The study was conducted to establish the complete procedure in processing rain tree pods for village level hydrous bioethanol production. Production processes were done for village level hydrous bioethanol production from collection, drying, storage, shredding, dilution, extraction, fermentation, and distillation. The feedstock was sundried, and moisture content was determined at a range of 20% to 26% prior to storage. Dilution ratio was 1:1.25 (1 kg of pods = 1.25 L of water) and after extraction process yielded a sugar concentration of 22 0Bx to 24 0Bx. The dilution period was three hours. After three hours of diluting the samples, the juice was extracted using extractor with a capacity of 64.10 L/hour. 150 L of rain tree pods juice was extracted and subjected to fermentation process using a village level anaerobic bioreactor. Fermentation with yeast (Saccharomyces cerevisiae) can fasten up the process, thus producing more ethanol at a shorter period of time; however, without yeast fermentation, it also produces ethanol at lower volume with slower fermentation process. Distillation of 150 L of fermented broth was done for six hours at 85 °C to 95 °C temperature (feedstock) and 74 °C to 95 °C temperature of the column head (vapor state of ethanol). The highest volume of ethanol recovered was established at with yeast fermentation at five-day duration with a value of 14.89 L and lowest actual ethanol content was found at without yeast fermentation at three-day duration having a value of 11.63 L. In general, the results suggested that rain tree pods had a very good potential as feedstock for bioethanol production. Fermentation of rain tree pods juice can be done with yeast and without yeast.
Keywords: Fermentation, hydrous bioethanol, rain tree pods, village level.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1566522 A Comparison of Signal Processing Techniques for the Extraction of Breathing Rate from the Photoplethysmogram
Authors: Susannah G. Fleming Lionel Tarassenko
Abstract:
The photoplethysmogram (PPG) is the pulsatile waveform produced by the pulse oximeter, which is widely used for monitoring arterial oxygen saturation in patients. Various methods for extracting the breathing rate from the PPG waveform have been compared using a consistent data set, and a novel technique using autoregressive modelling is presented. This novel technique is shown to outperform the existing techniques, with a mean error in breathing rate of 0.04 breaths per minute.Keywords: Autoregressive modelling, breathing rate, photoplethysmogram, pulse oximetry.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3323521 General Purpose Graphic Processing Units Based Real Time Video Tracking System
Authors: Mallikarjuna Rao Gundavarapu, Ch. Mallikarjuna Rao, K. Anuradha Bai
Abstract:
Real Time Video Tracking is a challenging task for computing professionals. The performance of video tracking techniques is greatly affected by background detection and elimination process. Local regions of the image frame contain vital information of background and foreground. However, pixel-level processing of local regions consumes a good amount of computational time and memory space by traditional approaches. In our approach we have explored the concurrent computational ability of General Purpose Graphic Processing Units (GPGPU) to address this problem. The Gaussian Mixture Model (GMM) with adaptive weighted kernels is used for detecting the background. The weights of the kernel are influenced by local regions and are updated by inter-frame variations of these corresponding regions. The proposed system has been tested with GPU devices such as GeForce GTX 280, GeForce GTX 280 and Quadro K2000. The results are encouraging with maximum speed up 10X compared to sequential approach.
Keywords: Connected components, Embrace threads, Local weighted kernel, Structuring element.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1174520 Ride Control of Passenger Cars with Semi-active Suspension System Using a Linear Quadratic Regulator and Hybrid Optimization Algorithm
Authors: Ali Fellah Jahromi, Wen Fang Xie, Rama B. Bhat
Abstract:
A semi-active control strategy for suspension systems of passenger cars is presented employing Magnetorheological (MR) dampers. The vehicle is modeled with seven DOFs including the, roll pitch and bounce of car body, and the vertical motion of the four tires. In order to design an optimal controller based on the actuator constraints, a Linear-Quadratic Regulator (LQR) is designed. The design procedure of the LQR consists of selecting two weighting matrices to minimize the energy of the control system. This paper presents a hybrid optimization procedure which is a combination of gradient-based and evolutionary algorithms to choose the weighting matrices with regards to the actuator constraint. The optimization algorithm is defined based on maximum comfort and actuator constraints. It is noted that utilizing the present control algorithm may significantly reduce the vibration response of the passenger car, thus, providing a comfortable ride.Keywords: Full car model, Linear Quadratic Regulator, Sequential Quadratic Programming, Genetic Algorithm
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2944519 Thermodynamic Evaluation of Coupling APR1400 with a Thermal Desalination Plant
Authors: M. Gomaa Abdoelatef, Robert M. Field, Lee, Yong-Kwan
Abstract:
Growing human population has placed increased demands on water supplies and spurred a heightened interest in desalination infrastructure. Key elements of the economics of desalination projects are thermal and electrical inputs. With growing concerns over use of fossil fuels to (indirectly) supply these inputs, coupling of desalination with nuclear power production represents a significant opportunity. Individually, nuclear and desalination technologies have a long history and are relatively mature. For desalination, Reverse Osmosis (RO) has the lowest energy inputs. However, the economically driven output quality of the water produced using RO, which uses only electrical inputs, is lower than the output water quality from thermal desalination plants. Therefore, modern desalination projects consider that RO should be coupled with thermal desalination technologies (MSF, MED, or MED-TVC) with attendant steam inputs to permit blending to produce various qualities of water. A large nuclear facility is well positioned to dispatch large quantities of both electrical and thermal power. This paper considers the supply of thermal energy to a large desalination facility to examine heat balance impact on the nuclear steam cycle. The APR1400 nuclear plant is selected as prototypical from both a capacity and turbine cycle heat balance perspective to examine steam supply and the impact on electrical output. Extraction points and quantities of steam are considered parametrically along with various types of thermal desalination technologies to form the basis for further evaluations of economically optimal approaches to the interface of nuclear power production with desalination projects. In our study, the thermodynamic evaluation will be executed by DE-TOP, an IAEA sponsored program. DE-TOP has capabilities to analyze power generation systems coupled to desalination plants through various steam extraction positions, taking into consideration the isolation loop between the nuclear and the thermal desalination facilities (i.e., for radiological isolation).Keywords: APR1400, Cogeneration, Desalination, DE-TOP, IAEA, MED, MED-TVC, MSF, RO.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2837518 Some Improvements on Kumlander-s Maximum Weight Clique Extraction Algorithm
Authors: Satoshi Shimizu, Kazuaki Yamaguchi, Toshiki Saitoh, Sumio Masuda
Abstract:
Some fast exact algorithms for the maximum weight clique problem have been proposed. Östergard’s algorithm is one of them. Kumlander says his algorithm is faster than it. But we confirmed that the straightforwardly implemented Kumlander’s algorithm is slower than O¨ sterga˚rd’s algorithm. We propose some improvements on Kumlander’s algorithm.
Keywords: Maximum weight clique, exact algorithm, branch-andbound, NP-hard.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1858517 Cellulose Extraction from Pomelo Peel: Synthesis of Carboxymethyl Cellulose
Authors: J. Chumee, D. Seeburin
Abstract:
The cellulose was extracted from pomelo peel and an etherification reaction used for converting cellulose to carboxymethyl cellulose (CMC). The pomelo peel was refluxed with 0.5 M HCl and 1 M NaOH solution at 90°C for 1 h and 2 h, respectively. The cellulose was bleached with calcium hypochlorite and used as precursor. The precursor was soaked in mixed solution between isopropyl alcohol and 40%w/v NaOH for 12 h. After that, chloroacetic acid was added and reacted at 55°C for 6 h. The optimum condition was 5 g of cellulose: 0.25 mole of NaOH : 0.07 mole of ClCH2COOH with 78.00% of yield. Moreover, the product had 0.54 of degree of substitution (DS).
Keywords: Pomelo peel, Carboxymethyl cellulose, Cellulose.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4310516 Dynamic Power Reduction in Sequential Circuits Using Look Ahead Clock Gating Technique
Authors: R. Manjith, C. Muthukumari
Abstract:
In this paper, a novel Linear Feedback Shift Register (LFSR) with Look Ahead Clock Gating (LACG) technique is presented to reduce the power consumption in modern processors and System-on-Chip. Clock gating is a predominant technique used to reduce unwanted switching of clock signals. Several clock gating techniques to reduce the dynamic power have been developed, of which LACG is predominant. LACG computes the clock enabling signals of each flip-flop (FF) one cycle ahead of time, based on the present cycle data of the flip-flops on which it depends. It overcomes the timing problems in the existing clock gating methods like datadriven clock gating and Auto-Gated flip-flops (AGFF) by allotting a full clock cycle for the determination of the clock enabling signals. Further to reduce the power consumption in LACG technique, FFs can be grouped so that they share a common clock enabling signal. Simulation results show that the novel grouped LFSR with LACG achieves 15.03% power savings than conventional LFSR with LACG and 44.87% than data-driven clock gating.Keywords: AGFF, data-driven, LACG, LFSR.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1746515 Optimization of Shale Gas Production by Advanced Hydraulic Fracturing
Authors: Fazl Ullah, Rahmat Ullah
Abstract:
This paper shows a comprehensive learning focused on the optimization of gas production in shale gas reservoirs through hydraulic fracturing. Shale gas has emerged as an important unconventional vigor resource, necessitating innovative techniques to enhance its extraction. The key objective of this study is to examine the influence of fracture parameters on reservoir productivity and formulate strategies for production optimization. A sophisticated model integrating gas flow dynamics and real stress considerations is developed for hydraulic fracturing in multi-stage shale gas reservoirs. This model encompasses distinct zones: a single-porosity medium region, a dual-porosity average region, and a hydraulic fracture region. The apparent permeability of the matrix and fracture system is modeled using principles like effective stress mechanics, porous elastic medium theory, fractal dimension evolution, and fluid transport apparatuses. The developed model is then validated using field data from the Barnett and Marcellus formations, enhancing its reliability and accuracy. By solving the partial differential equation by means of COMSOL software, the research yields valuable insights into optimal fracture parameters. The findings reveal the influence of fracture length, diversion capacity, and width on gas production. For reservoirs with higher permeability, extending hydraulic fracture lengths proves beneficial, while complex fracture geometries offer potential for low-permeability reservoirs. Overall, this study contributes to a deeper understanding of hydraulic cracking dynamics in shale gas reservoirs and provides essential guidance for optimizing gas production. The research findings are instrumental for energy industry professionals, researchers, and policymakers alike, shaping the future of sustainable energy extraction from unconventional resources.
Keywords: Fluid-solid coupling, apparent permeability, shale gas reservoir, fracture property, numerical simulation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 173514 2D Graphical Analysis of Wastewater Influent Capacity Time Series
Authors: Monika Chuchro, Maciej Dwornik
Abstract:
The extraction of meaningful information from image could be an alternative method for time series analysis. In this paper, we propose a graphical analysis of time series grouped into table with adjusted colour scale for numerical values. The advantages of this method are also discussed. The proposed method is easy to understand and is flexible to implement the standard methods of pattern recognition and verification, especially for noisy environmental data.Keywords: graphical analysis, time series, seasonality, noisy environmental data
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1451513 Influence of Fiber Packing on Transverse Plastic Properties of Metal Matrix Composites
Authors: Mohammad Tahaye Abadi
Abstract:
The present paper concerns with the influence of fiber packing on the transverse plastic properties of metal matrix composites. A micromechanical modeling procedure is used to predict the effective mechanical properties of composite materials at large tensile and compressive deformations. Microstructure is represented by a repeating unit cell (RUC). Two fiber arrays are considered including ideal square fiber packing and random fiber packing defined by random sequential algorithm. The micromechanical modeling procedure is implemented for graphite/aluminum metal matrix composite in which the reinforcement behaves as elastic, isotropic solids and the matrix is modeled as an isotropic elastic-plastic solid following the von Mises criterion with isotropic hardening and the Ramberg-Osgood relationship between equivalent true stress and logarithmic strain. The deformation is increased to a considerable value to evaluate both elastic and plastic behaviors of metal matrix composites. The yields strength and true elastic-plastic stress are determined for graphite/aluminum composites.Keywords: Fiber packing, metal matrix composites, micromechanics, plastic deformation, random
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1646512 Optimization of the Headspace Solid-Phase Microextraction Gas Chromatography for Volatile Compounds Determination in Phytophthora Cinnamomi Rands
Authors: Rui Qiu, Giles Hardy, Dong Qu, Robert Trengove, Manjree Agarwal, YongLin Ren
Abstract:
Phytophthora cinnamomi (P. c) is a plant pathogenic oomycete that is capable of damaging plants in commercial production systems and natural ecosystems worldwide. The most common methods for the detection and diagnosis of P. c infection are expensive, elaborate and time consuming. This study was carried out to examine whether species specific and life cycle specific volatile organic compounds (VOCs) can be absorbed by solid-phase microextraction fibers and detected by gas chromatography that are produced by P. c and another oomycete Pythium dissotocum. A headspace solid-phase microextraction (HS-SPME) together with gas chromatography (GC) method was developed and optimized for the identification of the VOCs released by P. c. The optimized parameters included type of fiber, exposure time, desorption temperature and desorption time. Optimization was achieved with the analytes of P. c+V8A and V8A alone. To perform the HS-SPME, six types of fiber were assayed and compared: 7μm Polydimethylsiloxane (PDMS), 100μm Polydimethylsiloxane (PDMS), 50/30μm Divinylbenzene/CarboxenTM/Polydimethylsiloxane DVB/CAR/PDMS), 65μm Polydimethylsiloxane/Divinylbenzene (PDMS/DVB), 85μm Polyacrylate (PA) fibre and 85μm CarboxenTM/ Polydimethylsiloxane (Carboxen™/PDMS). In a comparison of the efficacy of the fibers, the bipolar fiber DVB/CAR/PDMS had a higher extraction efficiency than the other fibers. An exposure time of 16h with DVB/CAR/PDMS fiber in the sample headspace was enough to reach the maximum extraction efficiency. A desorption time of 3min in the GC injector with the desorption temperature of 250°C was enough for the fiber to desorb the compounds of interest. The chromatograms and morphology study confirmed that the VOCs from P. c+V8A had distinct differences from V8A alone, as did different life cycle stages of P. c and different taxa such as Pythium dissotocum. The study proved that P. c has species and life cycle specific VOCs, which in turn demonstrated the feasibility of this method as means ofKeywords: Gas chromatography, headspace solid-phase microextraction, optimization, volatile compounds.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1885