Search results for: nonlinear boundary problem
575 Retrospective Synthetic Focusing with Correlation Weighting for Very High Frame Rate Ultrasound
Authors: Chang-Lin Hu, Yao-You Cheng, Meng-Lin Li
Abstract:
The need of high frame-rate imaging has been triggered by the new applications of ultrasound imaging to transient elastography and real-time 3D ultrasound. Using plane wave excitation (PWE) is one of the methods to achieve very high frame-rate imaging since an image can be formed with a single insonification. However, due to the lack of transmit focusing, the image quality with PWE is lower compared with those using conventional focused transmission. To solve this problem, we propose a filter-retrieved transmit focusing (FRF) technique combined with cross-correlation weighting (FRF+CC weighting) for high frame-rate imaging with PWE. A restrospective focusing filter is designed to simultaneously minimize the predefined sidelobe energy associated with single PWE and the filter energy related to the signal-to-noise-ratio (SNR). This filter attempts to maintain the mainlobe signals and to reduce the sidelobe ones, which gives similar mainlobe signals and different sidelobes between the original PWE and the FRF baseband data. Normalized cross-correlation coefficient at zero lag is calculated to quantify the degree of similarity at each imaging point and used as a weighting matrix to the FRF baseband data to further suppress sidelobes, thus improving the filter-retrieved focusing quality.
Keywords: retrospective synthetic focusing, high frame rate, correlation weighting.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1852574 Improved Body Mass Index Classification for Football Code Masters Athletes, A Comparison to the Australian National Population
Authors: Joe Walsh, Mike Climstein, Ian Timothy Heazlewood, Stephen Burke, Jyrki Kettunen, Kent Adams, Mark DeBeliso
Abstract:
Thousands of masters athletes participate quadrennially in the World Masters Games (WMG), yet this cohort of athletes remains proportionately under-investigated. Due to a growing global obesity pandemic in context of benefits of physical activity across the lifespan, the prevalence of obesity in this unique population was of particular interest. Data gathered on a sub-sample of 535 football code athletes, aged 31-72 yrs ( =47.4, s =±7.1), competing at the Sydney World Masters Games (2009) demonstrated a significantly (p<0.001), reduced classification of obesity using Body Mass Index (BMI) when compared to data on the Australian national population. This evidence of improved classification in one index of health (BMI<30) implies there are either improved levels of this index of health due to adherence to sport or possibly the reduced BMI is advantageous and contributes to this cohort adhering (or being attracted) to masters sport. Given the worldwide focus on the obesity epidemic and the need for a multi-faceted solution to this problem, demonstration of these middle to older aged adults having improved BMI over the general population is of particular interest.Keywords: BMI, masters athlete, rugby union, soccer, touch football
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1678573 Attribute Analysis of Quick Response Code Payment Users Using Discriminant Non-negative Matrix Factorization
Authors: Hironori Karachi, Haruka Yamashita
Abstract:
Recently, the system of quick response (QR) code is getting popular. Many companies introduce new QR code payment services and the services are competing with each other to increase the number of users. For increasing the number of users, we should grasp the difference of feature of the demographic information, usage information, and value of users between services. In this study, we conduct an analysis of real-world data provided by Nomura Research Institute including the demographic data of users and information of users’ usages of two services; LINE Pay, and PayPay. For analyzing such data and interpret the feature of them, Nonnegative Matrix Factorization (NMF) is widely used; however, in case of the target data, there is a problem of the missing data. EM-algorithm NMF (EMNMF) to complete unknown values for understanding the feature of the given data presented by matrix shape. Moreover, for comparing the result of the NMF analysis of two matrices, there is Discriminant NMF (DNMF) shows the difference of users features between two matrices. In this study, we combine EMNMF and DNMF and also analyze the target data. As the interpretation, we show the difference of the features of users between LINE Pay and Paypay.
Keywords: Data science, non-negative matrix factorization, missing data, quality of services.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 453572 A Robust and Adaptive Unscented Kalman Filter for the Air Fine Alignment of the Strapdown Inertial Navigation System/GPS
Authors: Jian Shi, Baoguo Yu, Haonan Jia, Meng Liu, Ping Huang
Abstract:
Adapting to the flexibility of war, a large number of guided weapons launch from aircraft. Therefore, the inertial navigation system loaded in the weapon needs to undergo an alignment process in the air. This article proposes the following methods to the problem of inaccurate modeling of the system under large misalignment angles, the accuracy reduction of filtering caused by outliers, and the noise changes in GPS signals: first, considering the large misalignment errors of Strapdown Inertial Navigation System (SINS)/GPS, a more accurate model is made rather than to make a small-angle approximation, and the Unscented Kalman Filter (UKF) algorithms are used to estimate the state; then, taking into account the impact of GPS noise changes on the fine alignment algorithm, the innovation adaptive filtering algorithm is introduced to estimate the GPS’s noise in real-time; at the same time, in order to improve the anti-interference ability of the air fine alignment algorithm, a robust filtering algorithm based on outlier detection is combined with the air fine alignment algorithm to improve the robustness of the algorithm. The algorithm can improve the alignment accuracy and robustness under interference conditions, which is verified by simulation.
Keywords: Air alignment, fine alignment, inertial navigation system, integrated navigation system, UKF.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 541571 An Intelligent Approach for Management of Hybrid DG System
Authors: Ali Vaseghi Ardekani, Hamid Reza Forutan, Amir Habibi, Ali Reza Rajabi, Hasan Adloo
Abstract:
Distributed generation units (DGs) are grid-connected or stand-alone electric generation units located within the electric distribution system at or near the end user. It is generally accepted that centralized electric power plants will remain the major source of the electric power supply for the near future. DGs, however, can complement central power by providing incremental capacity to the utility grid or to an end user. This paper presents an efficient power dispatching model for hybrid wind-Solar power generation system, to satisfy some essential requirements, such as dispersed electric power demand, electric power quality and reducing generation cost and so on. In this paper, presented some elements of the main parts in the hybrid system; and then made fundamental dispatching strategies according to different situations; then pointed out four improving measures to improve genetic algorithm, such as: modify the producing way of selection probability, improve the way of crossover, protect excellent chromosomes, and change mutation range and so on. Finally, propose a technique for solving the unit's commitment for dispatching problem based on an improved genetic algorithm.
Keywords: Power Quality, Wind-Solar System, Genetic Algorithm, Hybrid System.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1645570 Face Authentication for Access Control based on SVM using Class Characteristics
Authors: SeHun Lim, Sanghoon Kim, Sun-Tae Chung, Seongwon Cho
Abstract:
Face authentication for access control is a face membership authentication which passes the person of the incoming face if he turns out to be one of an enrolled person based on face recognition or rejects if not. Face membership authentication belongs to the two class classification problem where SVM(Support Vector Machine) has been successfully applied and shows better performance compared to the conventional threshold-based classification. However, most of previous SVMs have been trained using image feature vectors extracted from face images of each class member(enrolled class/unenrolled class) so that they are not robust to variations in illuminations, poses, and facial expressions and much affected by changes in member configuration of the enrolled class In this paper, we propose an effective face membership authentication method based on SVM using class discriminating features which represent an incoming face image-s associability with each class distinctively. These class discriminating features are weakly related with image features so that they are less affected by variations in illuminations, poses and facial expression. Through experiments, it is shown that the proposed face membership authentication method performs better than the threshold rule-based or the conventional SVM-based authentication methods and is relatively less affected by changes in member size and membership.Keywords: Face Authentication, Access control, member ship authentication, SVM.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1508569 The Robust Clustering with Reduction Dimension
Authors: Dyah E. Herwindiati
Abstract:
A clustering is process to identify a homogeneous groups of object called as cluster. Clustering is one interesting topic on data mining. A group or class behaves similarly characteristics. This paper discusses a robust clustering process for data images with two reduction dimension approaches; i.e. the two dimensional principal component analysis (2DPCA) and principal component analysis (PCA). A standard approach to overcome this problem is dimension reduction, which transforms a high-dimensional data into a lower-dimensional space with limited loss of information. One of the most common forms of dimensionality reduction is the principal components analysis (PCA). The 2DPCA is often called a variant of principal component (PCA), the image matrices were directly treated as 2D matrices; they do not need to be transformed into a vector so that the covariance matrix of image can be constructed directly using the original image matrices. The decomposed classical covariance matrix is very sensitive to outlying observations. The objective of paper is to compare the performance of robust minimizing vector variance (MVV) in the two dimensional projection PCA (2DPCA) and the PCA for clustering on an arbitrary data image when outliers are hiden in the data set. The simulation aspects of robustness and the illustration of clustering images are discussed in the end of paperKeywords: Breakdown point, Consistency, 2DPCA, PCA, Outlier, Vector Variance
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1697568 Studies on the Mechanical Behavior of Bottom Ash for a Sustainable Environment
Authors: B. A. Mir, Asim Malik
Abstract:
Bottom ash is a by-product of the combustion process of coal in furnaces in the production of electricity in thermal power plants. In India, about 75% of total power is produced by using pulverized coal. The coal of India has a high ash content which leads to the generation of a huge quantity of bottom ash per year posing the dual problem of environmental pollution and difficulty in disposal. This calls for establishing strategies to use this industry by-product effectively and efficiently. However, its large-scale utilization is possible only in geotechnical applications, either alone or with soil. In the present investigation, bottom ash was collected from National Capital Power Station Dadri, Uttar Pradesh, India. Test samples of bottom ash admixed with 20% clayey soil were prepared and treated with different cement content by weight and subjected to various laboratory tests for assessing its suitability as an engineered construction material. This study has shown that use of 10% cement content is a viable chemical additive to enhance the mechanical properties of bottom ash, which can be used effectively as an engineered construction material in various geotechnical applications. More importantly, it offers an interesting potential for making use of an industrial waste to overcome challenges posed by bottom ash for a sustainable environment.
Keywords: Bottom ash, environmental pollution, solid waste, sustainable environment, waste utilization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1720567 Using the Combined Model of PROMETHEE and Fuzzy Analytic Network Process for Determining Question Weights in Scientific Exams through Data Mining Approach
Authors: Hassan Haleh, Amin Ghaffari, Parisa Farahpour
Abstract:
Need for an appropriate system of evaluating students- educational developments is a key problem to achieve the predefined educational goals. Intensity of the related papers in the last years; that tries to proof or disproof the necessity and adequacy of the students assessment; is the corroborator of this matter. Some of these studies tried to increase the precision of determining question weights in scientific examinations. But in all of them there has been an attempt to adjust the initial question weights while the accuracy and precision of those initial question weights are still under question. Thus In order to increase the precision of the assessment process of students- educational development, the present study tries to propose a new method for determining the initial question weights by considering the factors of questions like: difficulty, importance and complexity; and implementing a combined method of PROMETHEE and fuzzy analytic network process using a data mining approach to improve the model-s inputs. The result of the implemented case study proves the development of performance and precision of the proposed model.Keywords: Assessing students, Analytic network process, Clustering, Data mining, Fuzzy sets, Multi-criteria decision making, and Preference function.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1581566 Application of Machine Learning Methods to Online Test Error Detection in Semiconductor Test
Authors: Matthias Kirmse, Uwe Petersohn, Elief Paffrath
Abstract:
As in today's semiconductor industries test costs can make up to 50 percent of the total production costs, an efficient test error detection becomes more and more important. In this paper, we present a new machine learning approach to test error detection that should provide a faster recognition of test system faults as well as an improved test error recall. The key idea is to learn a classifier ensemble, detecting typical test error patterns in wafer test results immediately after finishing these tests. Since test error detection has not yet been discussed in the machine learning community, we define central problem-relevant terms and provide an analysis of important domain properties. Finally, we present comparative studies reflecting the failure detection performance of three individual classifiers and three ensemble methods based upon them. As base classifiers we chose a decision tree learner, a support vector machine and a Bayesian network, while the compared ensemble methods were simple and weighted majority vote as well as stacking. For the evaluation, we used cross validation and a specially designed practical simulation. By implementing our approach in a semiconductor test department for the observation of two products, we proofed its practical applicability.
Keywords: Ensemble methods, fault detection, machine learning, semiconductor test.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2274565 SLM Using Riemann Sequence Combined with DCT Transform for PAPR Reduction in OFDM Communication Systems
Authors: Pepin Magnangana Zoko Goyoro, Ibrahim James Moumouni, Sroy Abouty
Abstract:
Orthogonal Frequency Division Multiplexing (OFDM) is an efficient method of data transmission for high speed communication systems. However, the main drawback of OFDM systems is that, it suffers from the problem of high Peak-to-Average Power Ratio (PAPR) which causes inefficient use of the High Power Amplifier and could limit transmission efficiency. OFDM consist of large number of independent subcarriers, as a result of which the amplitude of such a signal can have high peak values. In this paper, we propose an effective reduction scheme that combines DCT and SLM techniques. The scheme is composed of the DCT followed by the SLM using the Riemann matrix to obtain phase sequences for the SLM technique. The simulation results show PAPR can be greatly reduced by applying the proposed scheme. In comparison with OFDM, while OFDM had high values of PAPR –about 10.4dB our proposed method achieved about 4.7dB reduction of the PAPR with low complexities computation. This approach also avoids randomness in phase sequence selection, which makes it simpler to decode at the receiver. As an added benefit, the matrices can be generated at the receiver end to obtain the data signal and hence it is not required to transmit side information (SI).Keywords: DCT transform, OFDM, PAPR, Riemann matrix, SLM.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2639564 Representing Shared Join Points with State Charts: A High Level Design Approach
Authors: Muhammad Naveed, Muhammad Khalid Abdullah, Khalid Rashid, Hafiz Farooq Ahmad
Abstract:
Aspect Oriented Programming promises many advantages at programming level by incorporating the cross cutting concerns into separate units, called aspects. Join Points are distinguishing features of Aspect Oriented Programming as they define the points where core requirements and crosscutting concerns are (inter)connected. Currently, there is a problem of multiple aspects- composition at the same join point, which introduces the issues like ordering and controlling of these superimposed aspects. Dynamic strategies are required to handle these issues as early as possible. State chart is an effective modeling tool to capture dynamic behavior at high level design. This paper provides methodology to formulate the strategies for multiple aspect composition at high level, which helps to better implement these strategies at coding level. It also highlights the need of designing shared join point at high level, by providing the solutions of these issues using state chart diagrams in UML 2.0. High level design representation of shared join points also helps to implement the designed strategy in systematic way.Keywords: Aspect Oriented Software Development, Shared Join Points.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1717563 Intelligent Neural Network Based STLF
Authors: H. Shayeghi, H. A. Shayanfar, G. Azimi
Abstract:
Short-Term Load Forecasting (STLF) plays an important role for the economic and secure operation of power systems. In this paper, Continuous Genetic Algorithm (CGA) is employed to evolve the optimum large neural networks structure and connecting weights for one-day ahead electric load forecasting problem. This study describes the process of developing three layer feed-forward large neural networks for load forecasting and then presents a heuristic search algorithm for performing an important task of this process, i.e. optimal networks structure design. The proposed method is applied to STLF of the local utility. Data are clustered due to the differences in their characteristics. Special days are extracted from the normal training sets and handled separately. In this way, a solution is provided for all load types, including working days and weekends and special days. We find good performance for the large neural networks. The proposed methodology gives lower percent errors all the time. Thus, it can be applied to automatically design an optimal load forecaster based on historical data.
Keywords: Feed-forward Large Neural Network, Short-TermLoad Forecasting, Continuous Genetic Algorithm.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1830562 Regulation of Transfer of 137cs by Polymeric Sorbents for Grow Ecologically Sound Biomass
Authors: A. H. Tadevosyan, S. K. Mayrapetyan, N. B. Tavakalyan, K. I. Pyuskyulyan, A. H. Hovsepyan, S. N. Sergeeva
Abstract:
Soil contamination with radiocesium has a long-term radiological impact due to its long physical half-life (30.1 years for 137Cs and 2 years for 134Cs) and its high biological availability. 137Cs causes the largest concerns because of its deleterious effect on agriculture and stock farming, and, thus, human life for decades. One of the important aspects of the problem of contaminated soils remediation is understand of protective actions aimed at the reduction of biological migration of radionuclides in soil-plant system. The most effective way to bind radionuclides is the use of selective sorbents. The proposed research mainly aims to achieve control on transfer of 137Cs in a system growing media – plant due to counter ions variation in the polymeric sorbents. As research object Japanese basil - Perilla frutescens was chosen. Productivity of plants depending on the presence (control-without presence of polymer) and type of polymer material, as well as content of 137Cs in plant material has been determined. The character of different polymers influences on the 137Cs migration in growing media – plant system as well as accumulation in the plants has been cleared up.
Keywords: Radioceaseum, Japanese basil, polymer, soil-plant system.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1842561 MGAUM—Towards a Mobile Government Adoption and Utilization Model: The Case of Saudi Arabia
Authors: Mohammed Alonazi, Natalia Beloff, Martin White
Abstract:
This paper presents a proposal for a mobile government adoption and utilization model (MGAUM), which is a framework designed to increase the adoption rate of m-government services in Saudi Arabia. Recent advances in mobile technologies such are Mobile compatibilities, The development of wireless communication, mobile applications and devices are enabling governments to deliver services in new ways to citizens more efficiently and economically. In the last decade, many governments around the globe are utilizing these advances effectively to develop their next generation of e-government services. However, a low adoption rate of m-government services by citizens is a common problem in Arabian countries, including Saudi Arabia. Yet, to our knowledge, very little research has been conducted focused on understanding the factors that influence citizen adoption of these m-government services in this part of the world. A set of social, cultural and technological factors have been identified in the literature, which has led to the formulation of associated research questions and hypotheses. These hypotheses will be tested on Saudi citizens using questionnaires and interview methods based around the technology acceptance model. A key objective of the MGAUM framework is to investigate and understand Saudi citizens perception towards adoption and utilization of m-government services.Keywords: E-Government, M-Government, citizen services quality, technology acceptance model, Saudi Arabia, adoption framework.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1644560 Fault and Theft Recognition Using Toro Dial Sensor in Programmable Current Relay for Feeder Security
Authors: R. Kamalakannan, N. Ravi Kumar
Abstract:
Feeder protection is important in transmission and distribution side because if any fault occurs in any feeder or transformer, man power is needed to identify the problem and it will take more time. In the existing system, directional overcurrent elements with load further secured by a load encroachment function can be used to provide necessary security and sensitivity for faults on remote points in a circuit. It is validated only in renewable plant collector circuit protection applications over a wide range of operating conditions. In this method, the directional overcurrent feeder protection is developed by using monitoring of feeder section through internet. In this web based monitoring, the fault and power theft are identified by using Toro dial sensor and its information is received by SCADA (Supervisory Control and Data Acquisition) and controlled by ARM microcontroller. This web based monitoring is also used to monitor the feeder management, directional current detection, demand side management, overload fault. This monitoring system is capable of monitoring the distribution feeder over a large area depending upon the cost. It is also used to reduce the power theft, time and man power. The simulation is done by MATLAB software.
Keywords: Current sensor, distribution feeder protection, directional overcurrent, power theft, protective relay.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 793559 An Investigation of the Effect of the Different Mix Constituents on Concrete Electric Resistivity
Authors: H. M. Ghasemzadeh, Y. Mohammadi, Gh. Nouri, S. E. Nabavi
Abstract:
Steel corrosion in concrete is considered as a main engineering problems for many countries and lots of expenses has been paid for their repair and maintenance annually. This problem may occur in all engineering structures whether in coastal and offshore or other areas. Hence, concrete structures should be able to withstand corrosion factors existing in water or soil. Reinforcing steel corrosion enhancement can be measured by use of concrete electrical resistance; and maintaining high electric resistivity in concrete is necessary for steel corrosion prevention. Lots of studies devoted to different aspects of the subjects worldwide. In this paper, an evaluation of the effects of W/C ratio, cementitious materials, and percent increase in silica fume were investigated on electric resistivity of high strength concrete. To do that, sixteen mix design with one aggregate grading was planned. Five of them had varying amount of W/C ratio and other eleven mixes was prepared with constant W/C ratio but different amount of cementitious materials. Silica fume and super plasticizer were used with different proportions in all specimens. Specimens were tested after moist curing for 28 days. A total of 80 cube specimens (50 mm) were tested for concrete electrical resistance. Results show that concrete electric resistivity can be increased with increasing amount of cementitious materials and silica fume.Keywords: Corrosion, Electric resistivity, Mix design, Silica fume
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1559558 Topographic Arrangement of 3D Design Components on 2D Maps by Unsupervised Feature Extraction
Authors: Stefan Menzel
Abstract:
As a result of the daily workflow in the design development departments of companies, databases containing huge numbers of 3D geometric models are generated. According to the given problem engineers create CAD drawings based on their design ideas and evaluate the performance of the resulting design, e.g. by computational simulations. Usually, new geometries are built either by utilizing and modifying sets of existing components or by adding single newly designed parts to a more complex design. The present paper addresses the two facets of acquiring components from large design databases automatically and providing a reasonable overview of the parts to the engineer. A unified framework based on the topographic non-negative matrix factorization (TNMF) is proposed which solves both aspects simultaneously. First, on a given database meaningful components are extracted into a parts-based representation in an unsupervised manner. Second, the extracted components are organized and visualized on square-lattice 2D maps. It is shown on the example of turbine-like geometries that these maps efficiently provide a wellstructured overview on the database content and, at the same time, define a measure for spatial similarity allowing an easy access and reuse of components in the process of design development.Keywords: Design decomposition, topographic non-negative matrix factorization, parts-based representation, self-organization, unsupervised feature extraction.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1379557 Contamination in Industrial Areas and Environmental Management in Latvia
Authors: Juris Burlakovs, Maris Klavins, Raimonds Ernsteins, Armands Ruskulis
Abstract:
Environmental contamination is a common problem in ex-industrial and industrial sites. This article gives a brief description of general applied environmental investigation methodologies and possible remediation applications in Latvia. Most of contaminated areas are situated in former and active industrial, military areas and ports. Industrial and logistic activities very often have been with great impact for more than hundred years thus the contamination level with heavy metals, hydrocarbons, pesticides, persistent organic pollutants is high and is threatening health and environment in general. 242 territories now are numbered as contaminated and fixed in the National Register of contaminated territories in Latvia. Research and remediation of contamination in densely populated areas are of important environmental policy domain. Four different investigation case studies of contaminated areas are given describing the history of use, environmental quality assessment as well as planned environmental management actions. All four case study locations are situated in Riga - the capital of the Republic of Latvia. The aim of this paper is to analyze the situation and problems with management of contaminated areas in Latvia, give description of field research methods and recommendations for remediation industry based on scientific data and innovations.
Keywords: Remediation technology, environmental quality assessment, heavy metals, hydrocarbon contamination, environmental management.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2301556 A Finite Volume Procedure on Unstructured Meshes for Fluid-Structure Interaction Problems
Authors: P I Jagad, B P Puranik, A W Date
Abstract:
Flow through micro and mini channels requires relatively high driving pressure due to the large fluid pressure drop through these channels. Consequently the forces acting on the walls of the channel due to the fluid pressure are also large. Due to these forces there are displacement fields set up in the solid substrate containing the channels. If the movement of the substrate is constrained at some points, then stress fields are established in the substrate. On the other hand, if the deformation of the channel shape is sufficiently large then its effect on the fluid flow is important to be calculated. Such coupled fluid-solid systems form a class of problems known as fluidstructure interactions. In the present work a co-located finite volume discretization procedure on unstructured meshes is described for solving fluid-structure interaction type of problems. A linear elastic solid is assumed for which the effect of the channel deformation on the flow is neglected. Thus the governing equations for the fluid and the solid are decoupled and are solved separately. The procedure is validated by solving two benchmark problems, one from fluid mechanics and another from solid mechanics. A fluid-structure interaction problem of flow through a U-shaped channel embedded in a plate is solved.Keywords: Finite volume method, flow induced stresses, fluidstructureinteraction, unstructured meshes.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1890555 Forecast of Polyethylene Properties in the Gas Phase Polymerization Aided by Neural Network
Authors: Nasrin Bakhshizadeh, Ashkan Forootan
Abstract:
A major problem that affects the quality control of polymer in the industrial polymerization is the lack of suitable on-line measurement tools to evaluate the properties of the polymer such as melt and density indices. Controlling the polymerization in ordinary method is performed manually by taking samples, measuring the quality of polymer in the lab and registry of results. This method is highly time consuming and leads to producing large number of incompatible products. An online application for estimating melt index and density proposed in this study is a neural network based on the input-output data of the polyethylene production plant. Temperature, the level of reactors' bed, the intensity of ethylene mass flow, hydrogen and butene-1, the molar concentration of ethylene, hydrogen and butene-1 are used for the process to establish the neural model. The neural network is taught based on the actual operational data and back-propagation and Levenberg-Marquart techniques. The simulated results indicate that the neural network process model established with three layers (one hidden layer) for forecasting the density and the four layers for the melt index is able to successfully predict those quality properties.
Keywords: Polyethylene, polymerization, density, melt index, neural network.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 685554 Optimal Portfolio Selection in a DC Pension with Multiple Contributors and the Impact of Stochastic Additional Voluntary Contribution on the Optimal Investment Strategy
Authors: Edikan E. Akpanibah, Okwigbedi Oghen’Oro
Abstract:
In this paper, we studied the optimal portfolio selection in a defined contribution (DC) pension scheme with multiple contributors under constant elasticity of variance (CEV) model and the impact of stochastic additional voluntary contribution on the investment strategies. We assume that the voluntary contributions are stochastic and also consider investments in a risk free asset and a risky asset to increase the expected returns of the contributing members. We derived a stochastic differential equation which consists of the members’ monthly contributions and the invested fund and obtained an optimized problem with the help of Hamilton Jacobi Bellman equation. Furthermore, we find an explicit solution for the optimal investment strategy with stochastic voluntary contribution using power transformation and change of variables method and the corresponding optimal fund size was obtained. We discussed the impact of the voluntary contribution on the optimal investment strategy with numerical simulations and observed that the voluntary contribution reduces the optimal investment strategy of the risky asset.
Keywords: DC pension fund, Hamilton-Jacobi-Bellman, optimal investment strategies, power transformation method, stochastic, voluntary contribution.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 832553 Reduced Dynamic Time Warping for Handwriting Recognition Based on Multidimensional Time Series of a Novel Pen Device
Authors: Muzaffar Bashir, Jürgen Kempf
Abstract:
The purpose of this paper is to present a Dynamic Time Warping technique which reduces significantly the data processing time and memory size of multi-dimensional time series sampled by the biometric smart pen device BiSP. The acquisition device is a novel ballpoint pen equipped with a diversity of sensors for monitoring the kinematics and dynamics of handwriting movement. The DTW algorithm has been applied for time series analysis of five different sensor channels providing pressure, acceleration and tilt data of the pen generated during handwriting on a paper pad. But the standard DTW has processing time and memory space problems which limit its practical use for online handwriting recognition. To face with this problem the DTW has been applied to the sum of the five sensor signals after an adequate down-sampling of the data. Preliminary results have shown that processing time and memory size could significantly be reduced without deterioration of performance in single character and word recognition. Further excellent accuracy in recognition was achieved which is mainly due to the reduced dynamic time warping RDTW technique and a novel pen device BiSP.Keywords: Biometric character recognition, biometric person authentication, biometric smart pen BiSP, dynamic time warping DTW, online-handwriting recognition, multidimensional time series.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2406552 A Modified Run Length Coding Technique for Test Data Compression Based on Multi-Level Selective Huffman Coding
Authors: C. Kalamani, K. Paramasivam
Abstract:
Test data compression is an efficient method for reducing the test application cost. The problem of reducing test data has been addressed by researchers in three different aspects: Test Data Compression, Built-in-Self-Test (BIST) and Test set compaction. The latter two methods are capable of enhancing fault coverage with cost of hardware overhead. The drawback of the conventional methods is that they are capable of reducing the test storage and test power but when test data have redundant length of runs, no additional compression method is followed. This paper presents a modified Run Length Coding (RLC) technique with Multilevel Selective Huffman Coding (MLSHC) technique to reduce test data volume, test pattern delivery time and power dissipation in scan test applications where redundant length of runs is encountered then the preceding run symbol is replaced with tiny codeword. Experimental results show that the presented method not only improves the test data compression but also reduces the overall test data volume compared to recent schemes. Experiments for the six largest ISCAS-98 benchmarks show that our method outperforms most known techniques.
Keywords: Modified run length coding, multilevel selective Huffman coding, built-in-self-test modified selective Huffman coding, automatic test equipment.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1274551 Are Adolescent Girls More Depressive than Adolescent Boys in Turkey?
Authors: Hatice Odacı
Abstract:
Depression is a serious mental health problem that affects people of all ages, including children and adolescents. Studies showed that female gender is one of the risk factors may influence the development of depression in adolescents. However, some of the studies from Turkey suggested that gender does not lead to any significant difference in the youth depression level. Therefore, the presented study investigated whether girls differ from boys in respect of depression. The association between genders and test scores for the adolescents in a population of primary and secondary school students was also evaluated. The study was consisting of 254 adolescents (122 boys and 132 girls) with a mean age of 13.86±1.43 (Mean±SD) ranging from 12-16 years. Psychological assessment was performed using Children-s Depression Inventory (CDI). Chi-square and Student-s t-test statistics were employed to analyze the data. The mean of the CDI scores of the girls were higher than boys- CDI scores (t = -4.580, p = 0.001). Higher ratio appeared for the girls when they compared with boy group-s depression levels using a CDI cut-off point of 19 (p = 0.001, Odds Ratio = 2,603). The findings of the present study suggested that adolescent girls have high level of depression than adolescent boys aged between 12-16 years in Turkey. Although some studies reported that there is no any differences depression level between adolescent boys and girls in Turkey, result of the present study showed that adolescent girls have high level of depression than adolescent boys in Turkey.Keywords: Depression, Adolescent, Turkey, Female Gender, Male Gender.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1937550 On the Exact Solution of Non-Uniform Torsion for Beams with Asymmetric Cross-Section
Authors: A.Campanile, M. Mandarino, V. Piscopo
Abstract:
This paper deals with the problem of non-uniform torsion in thin-walled elastic beams with asymmetric cross-section, removing the basic concept of a fixed center of twist, necessary in the Vlasov-s and Benscoter-s theories to obtain a warping stress field equivalent to zero. In this new torsion/flexure theory, despite of the classical ones, the warping function will punctually satisfy the first indefinite equilibrium equation along the beam axis and it wont- be necessary to introduce the classical congruence condition, to take into account the effect of the beam restraints. The solution, based on the Fourier development of the displacement field, is obtained assuming that the applied external torque is constant along the beam axis and on both beam ends the unit twist angle and the warping axial displacement functions are totally restrained. Finally, in order to verify the feasibility of the proposed method and to compare it with the classical theories, two applications are carried out. The first one, relative to an open profile, is necessary to test the numerical method adopted to find the solution; the second one, instead, is relative to a simplified containership section, considered as full restrained in correspondence of two adjacent transverse bulkheads.Keywords: Non-uniform torsion, Asymmetric cross-section, Fourier series, Helmholtz equation, FE method.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1943549 Domain Driven Design vs Soft Domain Driven Design Frameworks
Authors: Mohammed Salahat, Steve Wade
Abstract:
This paper presents and compares the SSDDD “Systematic Soft Domain Driven Design Framework” to DDD “Domain Driven Design Framework” as a soft system approach of information systems development. The framework use SSM as a guiding methodology within which we have embedded a sequence of design tasks based on the UML leading to the implementation of a software system using the Naked Objects framework. This framework has been used in action research projects that have involved the investigation and modelling of business processes using object-oriented domain models and the implementation of software systems based on those domain models. Within this framework, Soft Systems Methodology (SSM) is used as a guiding methodology to explore the problem situation and to develop the domain model using UML for the given business domain. The framework is proposed and evaluated in our previous works, a comparison between SSDDD and DDD is presented in this paper, to show how SSDDD improved DDD as an approach to modelling and implementing business domain perspectives for Information Systems Development. The comparison process, the results, and the improvements are presented in the following sections of this paper.Keywords: SSM, UML, domain-driven design, soft domain-driven design, naked objects, soft language, information retrieval, multimethodology.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1978548 Study on the Optimization of Completely Batch Water-using Network with Multiple Contaminants Considering Flow Change
Authors: Jian Du, Shui Hong Hong, Lu Meng, Qing Wei Meng
Abstract:
This work addresses the problem of optimizing completely batch water-using network with multiple contaminants where the flow change caused by mass transfer is taken into consideration for the first time. A mathematical technique for optimizing water-using network is proposed based on source-tank-sink superstructure. The task is to obtain the freshwater usage, recycle assignments among water-using units, wastewater discharge and a steady water-using network configuration by following steps. Firstly, operating sequences of water-using units are determined by time constraints. Next, superstructure is simplified by eliminating the reuse and recycle from water-using units with maximum concentration of key contaminants. Then, the non-linear programming model is solved by GAMS (General Algebra Model System) for minimum freshwater usage, maximum water recycle and minimum wastewater discharge. Finally, numbers of operating periods are calculated to acquire the steady network configuration. A case study is solved to illustrate the applicability of the proposed approach.Keywords: Completely batch process, flow change, multiple contaminants, water-using network.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1451547 Efficacy of Polyfluoroalkyl Substances Filtration with Low-Cost Organic Fiber Filter
Authors: Gautham Das, Edward Morrone, Erik Treble, Clinton Binder
Abstract:
The purpose of this study was to evaluate the efficacy of a low-cost filter regarding per- and polyfluoroalkyl substances (PFAS). PFAS is a commonly used man-made chemical that can be found in a variety of household and industrial products with deleterious effects on humans. The filter consists of a combination of low-cost materials which could be locally procured. Water testing results for 4 different PFAS contaminants indicated that for Perfluorooctane sulfonic acid (PFOS), the Agency for Toxic Substances and Disease Registry (ATSDR) regulation is 7 ppt, the initial concentration was 15 ppt, and the final concentration was 3.9 ppt. For Perfluorononanoic acid (PFNA), the ATSDR regulation is 10.5 ppt, the initial concentration was 15 ppt, and the final concentration was 3.9 ppt. For Perfluorooctanoic acid (PFOA), the ATSDR regulation is 11 ppt, the initial concentration was 15 ppt, and the final concentration was 3.9 ppt. For Perfluorohexane sulfonic acid (PFHxS), the ATSDR regulation is 70 ppt, the initial concentration was 15 ppt, and the final concentration was 3.9 ppt. The results indicated a 74% reduction in PFAS concentration in filtered samples. Statistical data through regression analysis showed 0.9 validity of the sample data. Initial tests show the efficiency of the proposed filter described could be far greater if tested at a greater scale. It is highly recommended further testing to be conducted to validate the data for an innovative solution to a ubiquitous problem.
Keywords: PFAS, PFOS, PFOA, PFHxS, low-cost filter.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 653546 The Patterns of Unemployment and the Geography of Social Housing
Authors: Sónia Alves
Abstract:
During the last few decades in the academic field, the debate has increased on the effects of social geography on the opportunities of socioeconomic integration. On one hand, it has been discussed how the contents of the urban structure and social geography affect not only the way people interact, but also their chances of social and economic integration. On the other hand, it has also been discussed how the urban structure is also constrained and transformed by the action of social actors. Without questioning the powerful influence of structural factors, related to the logic of the production system, labor markets, education and training, the research has shown the role played by place of residence in shaping individual outcomes such as unemployment. In the context of this debate the importance of territory of residence with respect to the problem of unemployment has been highlighted. Although statistics of unemployment have already demonstrated the unequal incidence of the phenomenon in social groups, the issue of uneven territorial impact on the phenomenon at intra-urban level remains relatively unknown. The purpose of this article is to show and to interpret the spatial patterns of unemployment in the city of Porto using GIS (Geographic Information System - GIS) technology. Under this analysis the overlap of the spatial patterns of unemployment with the spatial distribution of social housing, allows the discussion of the relationship that occurs between these patterns and the reasons that might explain the relative immutability of socioeconomic problems in some neighborhoods.Keywords: Unemployment, area effects, urban planning, Porto.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2173