Search results for: Software Component
2447 Neuromarketing: Discovering the Somathyc Marker in the Consumer´s Brain
Authors: Mikel Alonso López, María Francisca Blasco López, Víctor Molero Ayala
Abstract:
The present study explains the somatic marker theory of Antonio Damasio, which indicates that when making a decision, the stored or possible future scenarios (future memory) images allow people to feel for a moment what would happen when they make a choice, and how this is emotionally marked. This process can be conscious or unconscious. The development of new Neuromarketing techniques such as functional magnetic resonance imaging (fMRI), carries a greater understanding of how the brain functions and consumer behavior. In the results observed in different studies using fMRI, the evidence suggests that the somatic marker and future memories influence the decision-making process, adding a positive or negative emotional component to the options. This would mean that all decisions would involve a present emotional component, with a rational cost-benefit analysis that can be performed later.
Keywords: Emotions, decision making, somatic marker, consumer´s brain.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21362446 Modeling the Transport of Charge Carriers in the Active Devices MESFET, Based of GaInP by the Monte Carlo Method
Authors: N. Massoum, A. Guen. Bouazza, B. Bouazza, A. El Ouchdi
Abstract:
The progress of industry integrated circuits in recent years has been pushed by continuous miniaturization of transistors. With the reduction of dimensions of components at 0.1 micron and below, new physical effects come into play as the standard simulators of two dimensions (2D) do not consider. In fact the third dimension comes into play because the transverse and longitudinal dimensions of the components are of the same order of magnitude. To describe the operation of such components with greater fidelity, we must refine simulation tools and adapted to take into account these phenomena. After an analytical study of the static characteristics of the component, according to the different operating modes, a numerical simulation is performed of field-effect transistor with submicron gate MESFET GaInP. The influence of the dimensions of the gate length is studied. The results are used to determine the optimal geometric and physical parameters of the component for their specific applications and uses.
Keywords: Monte Carlo simulation, transient electron transport, MESFET device.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16652445 Incremental Learning of Independent Topic Analysis
Authors: Takahiro Nishigaki, Katsumi Nitta, Takashi Onoda
Abstract:
In this paper, we present a method of applying Independent Topic Analysis (ITA) to increasing the number of document data. The number of document data has been increasing since the spread of the Internet. ITA was presented as one method to analyze the document data. ITA is a method for extracting the independent topics from the document data by using the Independent Component Analysis (ICA). ICA is a technique in the signal processing; however, it is difficult to apply the ITA to increasing number of document data. Because ITA must use the all document data so temporal and spatial cost is very high. Therefore, we present Incremental ITA which extracts the independent topics from increasing number of document data. Incremental ITA is a method of updating the independent topics when the document data is added after extracted the independent topics from a just previous the data. In addition, Incremental ITA updates the independent topics when the document data is added. And we show the result applied Incremental ITA to benchmark datasets.Keywords: Text mining, topic extraction, independent, incremental, independent component analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10582444 Similarity Detection in Collaborative Development of Object-Oriented Formal Specifications
Authors: Fathi Taibi, Fouad Mohammed Abbou, Md. Jahangir Alam
Abstract:
The complexity of today-s software systems makes collaborative development necessary to accomplish tasks. Frameworks are necessary to allow developers perform their tasks independently yet collaboratively. Similarity detection is one of the major issues to consider when developing such frameworks. It allows developers to mine existing repositories when developing their own views of a software artifact, and it is necessary for identifying the correspondences between the views to allow merging them and checking their consistency. Due to the importance of the requirements specification stage in software development, this paper proposes a framework for collaborative development of Object- Oriented formal specifications along with a similarity detection approach to support the creation, merging and consistency checking of specifications. The paper also explores the impact of using additional concepts on improving the matching results. Finally, the proposed approach is empirically evaluated.Keywords: Collaborative Development, Formal methods, Object-Oriented, Similarity detection
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14692443 Software Process Improvement: A Organizational Change that Need to be Managed and Motivated
Authors: Marília Guterres Ferreira, Raul Sidnei Wazlawick
Abstract:
As seen in literature, about 70% of the improvement initiatives fail, and a significant number do not even get started. This paper analyses the problem of failing initiatives on Software Process Improvement (SPI), and proposes good practices supported by motivational tools that can help minimizing failures. It elaborates on the hypothesis that human factors are poorly addressed by deployers, especially because implementation guides usually emphasize only technical factors. This research was conducted with SPI deployers and analyses 32 SPI initiatives. The results indicate that although human factors are not commonly highlighted in guidelines, the successful initiatives usually address human factors implicitly. This research shows that practices based on human factors indeed perform a crucial role on successful implantations of SPI, proposes change management as a theoretical framework to introduce those practices in the SPI context and suggests some motivational tools based on SPI deployers experience to support it.
Keywords: Change management, human factors, motivation, software process improvement.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 23492442 Time-Cost-Quality Trade-off Software by using Simplified Genetic Algorithm for Typical Repetitive Construction Projects
Authors: Refaat H. Abd El Razek, Ahmed M. Diab, Sherif M. Hafez, Remon F. Aziz
Abstract:
Time-Cost Optimization "TCO" is one of the greatest challenges in construction project planning and control, since the optimization of either time or cost, would usually be at the expense of the other. Since there is a hidden trade-off relationship between project and cost, it might be difficult to predict whether the total cost would increase or decrease as a result of the schedule compression. Recently third dimension in trade-off analysis is taken into consideration that is quality of the projects. Few of the existing algorithms are applied in a case of construction project with threedimensional trade-off analysis, Time-Cost-Quality relationships. The objective of this paper is to presents the development of a practical software system; that named Automatic Multi-objective Typical Construction Resource Optimization System "AMTCROS". This system incorporates the basic concepts of Line Of Balance "LOB" and Critical Path Method "CPM" in a multi-objective Genetic Algorithms "GAs" model. The main objective of this system is to provide a practical support for typical construction planners who need to optimize resource utilization in order to minimize project cost and duration while maximizing its quality simultaneously. The application of these research developments in planning the typical construction projects holds a strong promise to: 1) Increase the efficiency of resource use in typical construction projects; 2) Reduce construction duration period; 3) Minimize construction cost (direct cost plus indirect cost); and 4) Improve the quality of newly construction projects. A general description of the proposed software for the Time-Cost-Quality Trade-Off "TCQTO" is presented. The main inputs and outputs of the proposed software are outlined. The main subroutines and the inference engine of this software are detailed. The complexity analysis of the software is discussed. In addition, the verification, and complexity of the proposed software are proved and tested using a real case study.
Keywords: Project management, typical (repetitive) large scale projects, line of balance, multi-objective optimization, genetic algorithms, time-cost-quality trade-offs.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 30652441 Low Dimensional Representation of Dorsal Hand Vein Features Using Principle Component Analysis (PCA)
Authors: M.Heenaye-Mamode Khan, R.K. Subramanian, N. A. Mamode Khan
Abstract:
The quest of providing more secure identification system has led to a rise in developing biometric systems. Dorsal hand vein pattern is an emerging biometric which has attracted the attention of many researchers, of late. Different approaches have been used to extract the vein pattern and match them. In this work, Principle Component Analysis (PCA) which is a method that has been successfully applied on human faces and hand geometry is applied on the dorsal hand vein pattern. PCA has been used to obtain eigenveins which is a low dimensional representation of vein pattern features. Low cost CCD cameras were used to obtain the vein images. The extraction of the vein pattern was obtained by applying morphology. We have applied noise reduction filters to enhance the vein patterns. The system has been successfully tested on a database of 200 images using a threshold value of 0.9. The results obtained are encouraging.Keywords: Biometric, Dorsal vein pattern, PCA.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18952440 A New Analytical Approach to Reconstruct Residual Stresses Due to Turning Process
Authors: G.H. Farrahi, S.A. Faghidian, D.J. Smith
Abstract:
A thin layer on the component surface can be found with high tensile residual stresses, due to turning operations, which can dangerously affect the fatigue performance of the component. In this paper an analytical approach is presented to reconstruct the residual stress field from a limited incomplete set of measurements. Airy stress function is used as the primary unknown to directly solve the equilibrium equations and satisfying the boundary conditions. In this new method there exists the flexibility to impose the physical conditions that govern the behavior of residual stress to achieve a meaningful complete stress field. The analysis is also coupled to a least squares approximation and a regularization method to provide stability of the inverse problem. The power of this new method is then demonstrated by analyzing some experimental measurements and achieving a good agreement between the model prediction and the results obtained from residual stress measurement.Keywords: Residual stress, Limited measurements, Inverse problems, Turning process.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14302439 Redundancy Component Matrix and Structural Robustness
Authors: Xinjian Kou, Linlin Li, Yongju Zhou, Jimian Song
Abstract:
We introduce the redundancy matrix that expresses clearly the geometrical/topological configuration of the structure. With the matrix, the redundancy of the structure is resolved into redundant components and assigned to each member or rigid joint. The values of the diagonal elements in the matrix indicates the importance of the corresponding members or rigid joints, and the geometrically correlations can be shown with the non-diagonal elements. If a member or rigid joint failures, reassignment of the redundant components can be calculated with the recursive method given in the paper. By combining the indexes of reliability and redundancy components, we define an index concerning the structural robustness. To further explain the properties of the redundancy matrix, we cited several examples of statically indeterminate structures, including two trusses and a rigid frame. With the examples, some simple results and the properties of the matrix are discussed. The examples also illustrate that the redundancy matrix and the relevant concepts are valuable in structural safety analysis.
Keywords: Structural robustness, structural reliability, redundancy component, redundancy matrix.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 11402438 On the Dynamic Behaviour of a Four-Bar Linkage Driven by a Velocity Controlled DC Motor
Authors: Giovanni Incerti
Abstract:
The dynamic behaviour of a four-bar linkage driven by a velocity controlled DC motor is discussed in the paper. In particular the author presents the results obtained by means of a specifically developed software, which implements the mathematical models of all components of the system (linkage, transmission, electric motor, control devices). The use of this software enables a more efficient design approach, since it allows the designer to check, in a simple and immediate way, the dynamic behaviour of the mechanism, arising from different values of the system parameters.
Keywords: Four-bar linkage, Speed control, Dynamic analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 40842437 A Novel Approach for Tracking of a Mobile Node Based on Particle Filter and Trilateration
Authors: Muhammad Haroon Siddiqui, Muhammad Rehan Khalid
Abstract:
This paper evaluates the performance of a novel algorithm for tracking of a mobile node, interms of execution time and root mean square error (RMSE). Particle Filter algorithm is used to track the mobile node, however a new technique in particle filter algorithm is also proposed to reduce the execution time. The stationary points were calculated through trilateration and finally by averaging the number of points collected for a specific time, whereas tracking is done through trilateration as well as particle filter algorithm. Wi-Fi signal is used to get initial guess of the position of mobile node in x-y coordinates system. Commercially available software “Wireless Mon" was used to read the WiFi signal strength from the WiFi card. Visual Cµ version 6 was used to interact with this software to read only the required data from the log-file generated by “Wireless Mon" software. Results are evaluated through mathematical modeling and MATLAB simulation.Keywords: Particle Filter, Tracking, Wireless Local Area Network, WiFi, Trilateration
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20672436 A Neural Network Approach in Predicting the Blood Glucose Level for Diabetic Patients
Authors: Zarita Zainuddin, Ong Pauline, C. Ardil
Abstract:
Diabetes Mellitus is a chronic metabolic disorder, where the improper management of the blood glucose level in the diabetic patients will lead to the risk of heart attack, kidney disease and renal failure. This paper attempts to enhance the diagnostic accuracy of the advancing blood glucose levels of the diabetic patients, by combining principal component analysis and wavelet neural network. The proposed system makes separate blood glucose prediction in the morning, afternoon, evening and night intervals, using dataset from one patient covering a period of 77 days. Comparisons of the diagnostic accuracy with other neural network models, which use the same dataset are made. The comparison results showed overall improved accuracy, which indicates the effectiveness of this proposed system.Keywords: Diabetes Mellitus, principal component analysis, time-series, wavelet neural network.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 29892435 Personal Digital Assistants for Fieldwork Training in College Campus
Authors: Takaharu Miyoshi, Tadahiko Higuchi
Abstract:
Education supported by mobile computers has been widely done for some time. Teachers have attempted to use mobile computers and to find concrete subjects for student-s fieldwork training in college education. The purpose of this research is to develop software for Personal Digital Assistant (PDA) to conduct fieldwork in our campus, and to report a fieldwork class using PDAs in the curriculum of the Department of Regional Environment Studies.
Keywords: Development of software for PDA, fieldwork training, computer supported education, experiential learning.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 11842434 Roller Guide Design and Manufacturing for Spatial Cylindrical Cams
Authors: Yuan L. Lai, Jui P. Hung, Jian H. Chen
Abstract:
This paper was aimed at developing a computer aided design and manufacturing system for spatial cylindrical cams. In the proposed system, a milling tool with a diameter smaller than that of the roller, instead of the standard cutter for traditional machining process, was used to generate the tool path for spatial cams. To verify the feasibility of the proposed method, a multi-axis machining simulation software was further used to simulate the practical milling operation of spatial cams. It was observed from computer simulation that the tool path of small-sized cutter were within the motion range of a standard cutter, no occurrence of overcutting. Examination of a finished cam component clearly verifies the accuracy of the tool path generated for small-sized milling tool. It is believed that the use of small-sized cutter for the machining of the spatial cylindrical cams can generate a better surface morphology with higher accuracy. The improvement in efficiency and cost for the manufacturing of the spatial cylindrical cam can be expected through the proposed method.Keywords: Cylindrical cams, Computer-aided manufacturing, Tool path.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 34512433 Cognitive Radio Spectrum Management
Authors: Swapnil Singhal, Santosh Kumar Singh
Abstract:
The emerging Cognitive Radio is combo of both the technologies i.e. Radio dynamics and software technology. It involve wireless system with efficient coding, designing, and making them artificial intelligent to take the decision according to the surrounding environment and adopt themselves accordingly, so as to deliver the best QoS. This is the breakthrough from fixed hardware and fixed utilization of the spectrum. This software-defined approach of research is centralized at user-definition and application driven model, various software method are used for the optimization of the wireless communication. This paper focused on the Spectrum allocation technique using genetic algorithm GA to evolve radio, represented by chromosomes. The chromosomes gene represents the adjustable parameters in given radio and by using GA, evolving over the generations, the optimized set of parameters are evolved, as per the requirement of user and availability of the spectrum, in our prototype the gene consist of 6 different parameters, and the best set of parameters are evolved according to the application need and availability of the spectrum holes and thus maintaining best QoS for user, simultaneously maintaining licensed user rights. The analyzing tool Matlab is used for the performance of the prototype.
Keywords: ASDR, Cognitive Radio, QoS, Spectrum.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20542432 Student Records Management System Using Smart Cards and Biometric Technology for Educational Institutions
Authors: Patrick O. Bobbie, Prince S. Attrams
Abstract:
In recent times, the rapid change in new technologies has spurred up the way and manner records are handled in educational institutions. Also, there is a need for reliable access and ease-of use to these records, resulting in increased productivity in organizations. In academic institutions, such benefits help in quality assessments, institutional performance, and assessments of teaching and evaluation methods. Students in educational institutions benefit the most when advanced technologies are deployed in accessing records. This research paper discusses the use of biometric technologies coupled with smartcard technologies to provide a unique way of identifying students and matching their data to financial records to grant them access to restricted areas such as examination halls. The system developed in this paper, has an identity verification component as part of its main functionalities. A systematic software development cycle of analysis, design, coding, testing and support was used. The system provides a secured way of verifying student’s identity and real time verification of financial records. An advanced prototype version of the system has been developed for testing purposes.Keywords: Biometrics, fingerprints, identity-verification, smartcards.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20522431 Effects of Different Meteorological Variables on Reference Evapotranspiration Modeling: Application of Principal Component Analysis
Authors: Akinola Ikudayisi, Josiah Adeyemo
Abstract:
The correct estimation of reference evapotranspiration (ETₒ) is required for effective irrigation water resources planning and management. However, there are some variables that must be considered while estimating and modeling ETₒ. This study therefore determines the multivariate analysis of correlated variables involved in the estimation and modeling of ETₒ at Vaalharts irrigation scheme (VIS) in South Africa using Principal Component Analysis (PCA) technique. Weather and meteorological data between 1994 and 2014 were obtained both from South African Weather Service (SAWS) and Agricultural Research Council (ARC) in South Africa for this study. Average monthly data of minimum and maximum temperature (°C), rainfall (mm), relative humidity (%), and wind speed (m/s) were the inputs to the PCA-based model, while ETₒ is the output. PCA technique was adopted to extract the most important information from the dataset and also to analyze the relationship between the five variables and ETₒ. This is to determine the most significant variables affecting ETₒ estimation at VIS. From the model performances, two principal components with a variance of 82.7% were retained after the eigenvector extraction. The results of the two principal components were compared and the model output shows that minimum temperature, maximum temperature and windspeed are the most important variables in ETₒ estimation and modeling at VIS. In order words, ETₒ increases with temperature and windspeed. Other variables such as rainfall and relative humidity are less important and cannot be used to provide enough information about ETₒ estimation at VIS. The outcome of this study has helped to reduce input variable dimensionality from five to the three most significant variables in ETₒ modelling at VIS, South Africa.
Keywords: Irrigation, principal component analysis, reference evapotranspiration, Vaalharts.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10612430 Numerical Analysis on Rapid Decompression in Conventional Dry Gases using One- Dimensional Mathematical Modeling
Authors: Evgeniy Burlutskiy
Abstract:
The paper presents a one-dimensional transient mathematical model of compressible thermal multi-component gas mixture flows in pipes. The set of the mass, momentum and enthalpy conservation equations for gas phase is solved. Thermo-physical properties of multi-component gas mixture are calculated by solving the Equation of State (EOS) model. The Soave-Redlich-Kwong (SRK-EOS) model is chosen. Gas mixture viscosity is calculated on the basis of the Lee-Gonzales-Eakin (LGE) correlation. Numerical analysis on rapid decompression in conventional dry gases is performed by using the proposed mathematical model. The model is validated on measured values of the decompression wave speed in dry natural gas mixtures. All predictions show excellent agreement with the experimental data at high and low pressure. The presented model predicts the decompression in dry natural gas mixtures much better than GASDECOM and OLGA codes, which are the most frequently-used codes in oil and gas pipeline transport service.Keywords: Mathematical model, Rapid Gas Decompression
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 30092429 Free Vibration Analysis of Gabled Frame Considering Elastic Supports and Semi-Rigid Connections
Authors: A. Shooshtari, A. R. Masoodi, S. Heyrani Moghaddam
Abstract:
Free vibration analysis of a gabled frame with elastic support and semi-rigid connections is performed by using a program in OpenSees software. Natural frequencies and mode shape details of frame are obtained for two states, which are semi-rigid connections and elastic supports, separately. The members of this structure are analyzed as a prismatic nonlinear beam-column element in software. The mass of structure is considered as two equal lumped masses at the head of two columns in horizontal and vertical directions. Note that the degree of freedom, allocated to all nodes, is equal to three. Furthermore, the mode shapes of frame are achieved. Conclusively, the effects of connections and supports flexibility on the natural frequencies and mode shapes of structure are investigated.
Keywords: Natural frequency, mode shape, gabled frame, semi-rigid connection, elastic support, OpenSees software.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 30172428 An Exploration on Competency-Based Curricula in Integrated Circuit Design
Authors: Chih Chin Yang, Chung Shan Sun
Abstract:
In this paper the relationships between professional competences and school curriculain IC design industry are explored. The semi-structured questionnaire survey and focus group interview is the research method. Study participants are graduates of microelectronics engineering professional departments who are currently employed in the IC industry. The IC industries are defined as the electronic component manufacturing industry and optical-electronic component manufacturing industry in the semiconductor industry and optical-electronic material devices, respectively. Study participants selected from IC design industry include IC engineering and electronic & semiconductor engineering. The human training with IC design professional competence in microelectronics engineering professional departments is explored in this research. IC professional competences of human resources in the IC design industry include general intelligence and professional intelligence.
Keywords: IC design, curricula, competence, task, duty.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14952427 Real Time Lidar and Radar High-Level Fusion for Obstacle Detection and Tracking with Evaluation on a Ground Truth
Authors: Hatem Hajri, Mohamed-Cherif Rahal
Abstract:
Both Lidars and Radars are sensors for obstacle detection. While Lidars are very accurate on obstacles positions and less accurate on their velocities, Radars are more precise on obstacles velocities and less precise on their positions. Sensor fusion between Lidar and Radar aims at improving obstacle detection using advantages of the two sensors. The present paper proposes a real-time Lidar/Radar data fusion algorithm for obstacle detection and tracking based on the global nearest neighbour standard filter (GNN). This algorithm is implemented and embedded in an automative vehicle as a component generated by a real-time multisensor software. The benefits of data fusion comparing with the use of a single sensor are illustrated through several tracking scenarios (on a highway and on a bend) and using real-time kinematic sensors mounted on the ego and tracked vehicles as a ground truth.Keywords: Ground truth, Hungarian algorithm, lidar Radar data fusion, global nearest neighbor filter.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9612426 Issues in Spectral Source Separation Techniques for Plant-wide Oscillation Detection and Diagnosis
Authors: A.K. Tangirala, S. Babji
Abstract:
In the last few years, three multivariate spectral analysis techniques namely, Principal Component Analysis (PCA), Independent Component Analysis (ICA) and Non-negative Matrix Factorization (NMF) have emerged as effective tools for oscillation detection and isolation. While the first method is used in determining the number of oscillatory sources, the latter two methods are used to identify source signatures by formulating the detection problem as a source identification problem in the spectral domain. In this paper, we present a critical drawback of the underlying linear (mixing) model which strongly limits the ability of the associated source separation methods to determine the number of sources and/or identify the physical source signatures. It is shown that the assumed mixing model is only valid if each unit of the process gives equal weighting (all-pass filter) to all oscillatory components in its inputs. This is in contrast to the fact that each unit, in general, acts as a filter with non-uniform frequency response. Thus, the model can only facilitate correct identification of a source with a single frequency component, which is again unrealistic. To overcome this deficiency, an iterative post-processing algorithm that correctly identifies the physical source(s) is developed. An additional issue with the existing methods is that they lack a procedure to pre-screen non-oscillatory/noisy measurements which obscure the identification of oscillatory sources. In this regard, a pre-screening procedure is prescribed based on the notion of sparseness index to eliminate the noisy and non-oscillatory measurements from the data set used for analysis.Keywords: non-negative matrix factorization, PCA, source separation, plant-wide diagnosis
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15342425 Meta Model for Optimum Design Objective Function of Steel Frames Subjected to Seismic Loads
Authors: Salah R. Al Zaidee, Ali S. Mahdi
Abstract:
Except for simple problems of statically determinate structures, optimum design problems in structural engineering have implicit objective functions where structural analysis and design are essential within each searching loop. With these implicit functions, the structural engineer is usually enforced to write his/her own computer code for analysis, design, and searching for optimum design among many feasible candidates and cannot take advantage of available software for structural analysis, design, and searching for the optimum solution. The meta-model is a regression model used to transform an implicit objective function into objective one and leads in turn to decouple the structural analysis and design processes from the optimum searching process. With the meta-model, well-known software for structural analysis and design can be used in sequence with optimum searching software. In this paper, the meta-model has been used to develop an explicit objective function for plane steel frames subjected to dead, live, and seismic forces. Frame topology is assumed as predefined based on architectural and functional requirements. Columns and beams sections and different connections details are the main design variables in this study. Columns and beams are grouped to reduce the number of design variables and to make the problem similar to that adopted in engineering practice. Data for the implicit objective function have been generated based on analysis and assessment for many design proposals with CSI SAP software. These data have been used later in SPSS software to develop a pure quadratic nonlinear regression model for the explicit objective function. Good correlations with a coefficient, R2, in the range from 0.88 to 0.99 have been noted between the original implicit functions and the corresponding explicit functions generated with meta-model.
Keywords: Meta-modal, objective function, steel frames, seismic analysis, design.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13332424 GODYS-PC: a Software Package for Modeling,Simulating and Analyzing Dynamic Systems
Authors: Jacek Kuraś, Jacek Lembas, Marek Skomorowski
Abstract:
In this paper, we introduce GODYS-PC software package for modeling, simulating and analyzing dynamic systems. To illustrate the use of GODYS-PC we present a few examples which concern modeling and simulating of engineering systems. In order to compare GODYS-PC with widely used in academia and industry Simulink®, the same examples are provided both in GODYS-PC and Simulink®.Keywords: Modeling, simulating and analyzing dynamicsystems.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14782423 Automatic Removal of Ocular Artifacts using JADE Algorithm and Neural Network
Authors: V Krishnaveni, S Jayaraman, A Gunasekaran, K Ramadoss
Abstract:
The ElectroEncephaloGram (EEG) is useful for clinical diagnosis and biomedical research. EEG signals often contain strong ElectroOculoGram (EOG) artifacts produced by eye movements and eye blinks especially in EEG recorded from frontal channels. These artifacts obscure the underlying brain activity, making its visual or automated inspection difficult. The goal of ocular artifact removal is to remove ocular artifacts from the recorded EEG, leaving the underlying background signals due to brain activity. In recent times, Independent Component Analysis (ICA) algorithms have demonstrated superior potential in obtaining the least dependent source components. In this paper, the independent components are obtained by using the JADE algorithm (best separating algorithm) and are classified into either artifact component or neural component. Neural Network is used for the classification of the obtained independent components. Neural Network requires input features that exactly represent the true character of the input signals so that the neural network could classify the signals based on those key characters that differentiate between various signals. In this work, Auto Regressive (AR) coefficients are used as the input features for classification. Two neural network approaches are used to learn classification rules from EEG data. First, a Polynomial Neural Network (PNN) trained by GMDH (Group Method of Data Handling) algorithm is used and secondly, feed-forward neural network classifier trained by a standard back-propagation algorithm is used for classification and the results show that JADE-FNN performs better than JADEPNN.Keywords: Auto Regressive (AR) Coefficients, Feed Forward Neural Network (FNN), Joint Approximation Diagonalisation of Eigen matrices (JADE) Algorithm, Polynomial Neural Network (PNN).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18892422 Design an Electronic Market Framework Using JADE Environment
Authors: Mohammad Ali Tabarzad, Caro Lucas
Abstract:
The daily growing use of agents in software environments, because of many reasons such as independence and intelligence is not a secret anymore. One of such environments in which there is a prominent job for the agents would be emarketplaces in which a user is able to give those agents the responsibility of buying and selling, instead of searching the emarketplace himself. Making up a framework which has sufficient attention to the required roles and their relations, is the first step of achieving such e-markets. In this paper, we suggest a framework in order to establish such e-markets and we will continue investigating the roles such as seller or buyer and the relations in JADE environment in details.
Keywords: Framework, software agents, e-commerce, e-market.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14412421 On the Standardizing the Metal Die of Punchand Matrix by Mechanical Desktop Software
Authors: A. M. R. Mosalman Yazdi, B. A. R. Mosalman Yazdi
Abstract:
In industry, on of the most important subjects is die and it's characteristics in which for cutting and forming different mechanical pieces, various punch and matrix metal die are used. whereas the common parts which form the main frame die are not often proportion with pieces and dies therefore using a part as socalled common part for frames in specified dimension ranges can decrease the time of designing, occupied space of warehouse and manufacturing costs. Parts in dies with getting uniform in their shape and dimension make common parts of dies. Common parts of punch and matrix metal die are as bolster, guide bush, guide pillar and shank. In this paper the common parts and effective parameters in selecting each of them as the primary information are studied, afterward for selection and design of mechanical parts an introduction and investigation based on the Mech. Desk. software is done hence with developing this software can standardize the metal common parts of punch and matrix. These studies will be so useful for designer in their designing and also using it has with very much advantage for manufactures of products in decreasing occupied spaces by dies.Keywords: Die, Matrix, Punch, Standardize.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13772420 High Dynamic Range Resampling for Software Radio
Authors: Arthur David Snider, Laiq Azam
Abstract:
The classic problem of recovering arbitrary values of a band-limited signal from its samples has an added complication in software radio applications; namely, the resampling calculations inevitably fold aliases of the analog signal back into the original bandwidth. The phenomenon is quantified by the spur-free dynamic range. We demonstrate how a novel application of the Remez (Parks- McClellan) algorithm permits optimal signal recovery and SFDR, far surpassing state-of-the-art resamplers.Keywords: Sampling methods, Signal sampling, Digital radio, Digital-analog conversion.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14062419 Material Characterization and Numerical Simulation of a Rubber Bumper
Authors: Tamás Mankovits, Dávid Huri, Imre Kállai, Imre Kocsis, Tamás Szabó
Abstract:
Non-linear FEM calculations are indispensable when important technical information like operating performance of a rubber component is desired. Rubber bumpers built into air-spring structures may undergo large deformations under load, which in itself shows non-linear behavior. The changing contact range between the parts and the incompressibility of the rubber increases this non-linear behavior further. The material characterization of an elastomeric component is also a demanding engineering task. In this paper a comprehensive investigation is introduced including laboratory measurements, mesh density analysis and complex finite element simulations to obtain the load-displacement curve of the chosen rubber bumper. Contact and friction effects are also taken into consideration. The aim of this research is to elaborate a FEM model which is accurate and competitive for a future shape optimization task.
Keywords: Rubber bumper, finite element analysis, compression test, Mooney-Rivlin material model.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 35842418 Management Software for the Elaboration of an Electronic File in the Pharmaceutical Industry Following Mexican Regulations
Authors: M. Peña Aguilar Juan, Ríos Hernández Ezequiel, R. Valencia Luis
Abstract:
For certification, certain goods of public interest, such as medicines and food, it is required the preparation and delivery of a dossier. For its elaboration, legal and administrative knowledge must be taken, as well as organization of the documents of the process, and an order that allows the file verification. Therefore, a virtual platform was developed to support the process of management and elaboration of the dossier, providing accessibility to the information and interfaces that allow the user to know the status of projects. The development of dossier system on the cloud allows the inclusion of the technical requirements for the software management, including the validation and the manufacturing in the field industry. The platform guides and facilitates the dossier elaboration (report, file or history), considering Mexican legislation and regulations, it also has auxiliary tools for its management. This technological alternative provides organization support for documents and accessibility to the information required to specify the successful development of a dossier. The platform divides into the following modules: System control, catalog, dossier and enterprise management. The modules are designed per the structure required in a dossier in those areas. However, the structure allows for flexibility, as its goal is to become a tool that facilitates and does not obstruct processes. The architecture and development of the software allows flexibility for future work expansion to other fields, this would imply feeding the system with new regulations.
Keywords: Electronic dossier, technologies for management, web software, dossier elaboration, pharmaceutical industry.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1203