Search results for: precomputed simplification
39 GPU-Accelerated Triangle Mesh Simplification Using Parallel Vertex Removal
Authors: Thomas Odaker, Dieter Kranzlmueller, Jens Volkert
Abstract:
We present an approach to triangle mesh simplification designed to be executed on the GPU. We use a quadric error metric to calculate an error value for each vertex of the mesh and order all vertices based on this value. This step is followed by the parallel removal of a number of vertices with the lowest calculated error values. To allow for the parallel removal of multiple vertices we use a set of per-vertex boundaries that prevent mesh foldovers even when simplification operations are performed on neighbouring vertices. We execute multiple iterations of the calculation of the vertex errors, ordering of the error values and removal of vertices until either a desired number of vertices remains in the mesh or a minimum error value is reached. This parallel approach is used to speed up the simplification process while maintaining mesh topology and avoiding foldovers at every step of the simplification.Keywords: Computer graphics, half edge collapse, mesh simplification, precomputed simplification, topology preserving.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 279438 Geometric Simplification Method of Building Energy Model Based on Building Performance Simulation
Authors: Yan Lyu, Yiqun Pan, Zhizhong Huang
Abstract:
In the design stage of a new building, the energy model of this building is often required for the analysis of the performance on energy efficiency. In practice, a certain degree of geometric simplification should be done in the establishment of building energy models, since the detailed geometric features of a real building are hard to be described perfectly in most energy simulation engine, such as ESP-r, eQuest or EnergyPlus. Actually, the detailed description is not necessary when the result with extremely high accuracy is not demanded. Therefore, this paper analyzed the relationship between the error of the simulation result from building energy models and the geometric simplification of the models. Finally, the following two parameters are selected as the indices to characterize the geometric feature of in building energy simulation: the southward projected area and total side surface area of the building. Based on the parameterization method, the simplification from an arbitrary column building to a typical shape (a cuboid) building can be made for energy modeling. The result in this study indicates that no more than 7% prediction error of annual cooling/heating load will be caused by the geometric simplification for those buildings with the ratio of southward projection length to total perimeter of the bottom of 0.25~0.35, which means this method is applicable for building performance simulation.
Keywords: building energy model, simulation, geometric simplification, design, regression
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 62337 Incorporation of Long-Term Redundancy in ECG Time Domain Compression Methods through Curve Simplification and Block-Sorting
Authors: Bachir Boucheham, Youcef Ferdi, Mohamed Chaouki Batouche
Abstract:
We suggest a novel method to incorporate longterm redundancy (LTR) in signal time domain compression methods. The proposition is based on block-sorting and curve simplification. The proposition is illustrated on the ECG signal as a post-processor for the FAN method. Test applications on the new so-obtained FAN+ method using the MIT-BIH database show substantial improvement of the compression ratio-distortion behavior for a higher quality reconstructed signal.Keywords: ECG compression, Long-term redundancy, Block-sorting, Curve Simplification.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 152036 3D Brain Tumor Segmentation Using Level-Sets Method and Meshes Simplification from Volumetric MR Images
Authors: K. Aloui, M. S. Naceur
Abstract:
The main objective of this paper is to provide an efficient tool for delineating brain tumors in three-dimensional magnetic resonance images. To achieve this goal, we use basically a level-sets approach to delineating three-dimensional brain tumors. Then we introduce a compression plan of 3D brain structures based for the meshes simplification, adapted for time to the specific needs of the telemedicine and to the capacities restricted by network communication. We present here the main stages of our system, and preliminary results which are very encouraging for clinical practice.
Keywords: Medical imaging, level-sets, compression, meshess implification, telemedicine.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 213235 Integrating Fast Karnough Map and Modular Neural Networks for Simplification and Realization of Complex Boolean Functions
Authors: Hazem M. El-Bakry
Abstract:
In this paper a new fast simplification method is presented. Such method realizes Karnough map with large number of variables. In order to accelerate the operation of the proposed method, a new approach for fast detection of group of ones is presented. Such approach implemented in the frequency domain. The search operation relies on performing cross correlation in the frequency domain rather than time one. It is proved mathematically and practically that the number of computation steps required for the presented method is less than that needed by conventional cross correlation. Simulation results using MATLAB confirm the theoretical computations. Furthermore, a powerful solution for realization of complex functions is given. The simplified functions are implemented by using a new desigen for neural networks. Neural networks are used because they are fault tolerance and as a result they can recognize signals even with noise or distortion. This is very useful for logic functions used in data and computer communications. Moreover, the implemented functions are realized with minimum amount of components. This is done by using modular neural nets (MNNs) that divide the input space into several homogenous regions. Such approach is applied to implement XOR function, 16 logic functions on one bit level, and 2-bit digital multiplier. Compared to previous non- modular designs, a clear reduction in the order of computations and hardware requirements is achieved.Keywords: Boolean Functions, Simplification, KarnoughMap, Implementation of Logic Functions, Modular NeuralNetworks.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 181334 Integrating Fast Karnough Map and Modular Neural Networks for Simplification and Realization of Complex Boolean Functions
Authors: Hazem M. El-Bakry
Abstract:
In this paper a new fast simplification method is presented. Such method realizes Karnough map with large number of variables. In order to accelerate the operation of the proposed method, a new approach for fast detection of group of ones is presented. Such approach implemented in the frequency domain. The search operation relies on performing cross correlation in the frequency domain rather than time one. It is proved mathematically and practically that the number of computation steps required for the presented method is less than that needed by conventional cross correlation. Simulation results using MATLAB confirm the theoretical computations. Furthermore, a powerful solution for realization of complex functions is given. The simplified functions are implemented by using a new desigen for neural networks. Neural networks are used because they are fault tolerance and as a result they can recognize signals even with noise or distortion. This is very useful for logic functions used in data and computer communications. Moreover, the implemented functions are realized with minimum amount of components. This is done by using modular neural nets (MNNs) that divide the input space into several homogenous regions. Such approach is applied to implement XOR function, 16 logic functions on one bit level, and 2-bit digital multiplier. Compared to previous non- modular designs, a clear reduction in the order of computations and hardware requirements is achieved.
Keywords: Boolean functions, simplification, Karnough map, implementation of logic functions, modular neural networks.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 206933 Development of an Intelligent Tool for Planning the Operation
Authors: T. R. Alencar, P. T. Leite
Abstract:
Several optimization algorithms specifically applied to the problem of Operation Planning of Hydrothermal Power Systems have been developed and are used. Although providing solutions to various problems encountered, these algorithms have some weaknesses, difficulties in convergence, simplification of the original formulation of the problem, or owing to the complexity of the objective function. Thus, this paper presents the development of a computational tool for solving optimization problem identified and to provide the User an easy handling. Adopted as intelligent optimization technique, Genetic Algorithms and programming language Java. First made the modeling of the chromosomes, then implemented the function assessment of the problem and the operators involved, and finally the drafting of the graphical interfaces for access to the User. The program has managed to relate a coherent performance in problem resolution without the need for simplification of the calculations together with the ease of manipulating the parameters of simulation and visualization of output results.Keywords: Energy, Optimization, Hydrothermal Power Systemsand Genetic Algorithms
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 169632 SimplexIS: Evaluating the Impact of e-Gov Simplification Measures in the Information System Architecure
Authors: Bruno Félix, André Vasconcelos, José Tribolet
Abstract:
Nowadays increasingly the population makes use of Information Technology (IT). As such, in recent year the Portuguese government increased its focus on using the IT for improving people-s life and began to develop a set of measures to enable the modernization of the Public Administration, and so reducing the gap between Public Administration and citizens.Thus the Portuguese Government launched the Simplex Program. However these SIMPLEX eGov measures, which have been implemented over the years, present a serious challenge: how to forecast its impact on existing Information Systems Architecture (ISA). Thus, this research is focus in addressing the problem of automating the evaluation of the actual impact of implementation an eGovSimplification and Modernization measures in the Information Systems Architecture. To realize the evaluation we proposes a Framework, which is supported by some key concepts as: Quality Factors, ISA modeling, Multicriteria Approach, Polarity Profile and Quality MetricsKeywords: Information System Architecture, Evaluation, eGov Simplification measure, Multicriteria Evaluation
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 144531 Automatic Map Simplification for Visualization on Mobile Devices
Authors: Hang Yu
Abstract:
The visualization of geographic information on mobile devices has become popular as the widespread use of mobile Internet. The mobility of these devices brings about much convenience to people-s life. By the add-on location-based services of the devices, people can have an access to timely information relevant to their tasks. However, visual analysis of geographic data on mobile devices presents several challenges due to the small display and restricted computing resources. These limitations on the screen size and resources may impair the usability aspects of the visualization applications. In this paper, a variable-scale visualization method is proposed to handle the challenge of small mobile display. By merging multiple scales of information into a single image, the viewer is able to focus on the interesting region, while having a good grasp of the surrounding context. This is essentially visualizing the map through a fisheye lens. However, the fisheye lens induces undesirable geometric distortion in the peripheral, which renders the information meaningless. The proposed solution is to apply map generalization that removes excessive information around the peripheral and an automatic smoothing process to correct the distortion while keeping the local topology consistent. The proposed method is applied on both artificial and real geographical data for evaluation.
Keywords: Map simplification, visualization, mobile devices.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 143530 Finite Element Analysis of Sheet Metal Airbending Using Hyperform LS-DYNA
Authors: Himanshu V. Gajjar, Anish H. Gandhi, Harit K. Raval
Abstract:
Air bending is one of the important metal forming processes, because of its simplicity and large field application. Accuracy of analytical and empirical models reported for the analysis of bending processes is governed by simplifying assumption and do not consider the effect of dynamic parameters. Number of researches is reported on the finite element analysis (FEA) of V-bending, Ubending, and air V-bending processes. FEA of bending is found to be very sensitive to many physical and numerical parameters. FE models must be computationally efficient for practical use. Reported work shows the 3D FEA of air bending process using Hyperform LSDYNA and its comparison with, published 3D FEA results of air bending in Ansys LS-DYNA and experimental results. Observing the planer symmetry and based on the assumption of plane strain condition, air bending problem was modeled in 2D with symmetric boundary condition in width. Stress-strain results of 2D FEA were compared with 3D FEA results and experiments. Simplification of air bending problem from 3D to 2D resulted into tremendous reduction in the solution time with only marginal effect on stressstrain results. FE model simplification by studying the problem symmetry is more efficient and practical approach for solution of more complex large dimensions slow forming processes.Keywords: Air V-bending, Finite element analysis, HyperformLS-DYNA, Planner symmetry.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 320829 Hybrid Collaborative-Context Based Recommendations for Civil Affairs Operations
Authors: Patrick Cummings, Laura Cassani, Deirdre Kelliher
Abstract:
In this paper we present findings from a research effort to apply a hybrid collaborative-context approach for a system focused on Marine Corps civil affairs data collection, aggregation, and analysis called the Marine Civil Information Management System (MARCIMS). The goal of this effort is to provide operators with information to make sense of the interconnectedness of entities and relationships in their area of operation and discover existing data to support civil military operations. Our approach to build a recommendation engine was designed to overcome several technical challenges, including 1) ensuring models were robust to the relatively small amount of data collected by the Marine Corps civil affairs community; 2) finding methods to recommend novel data for which there are no interactions captured; and 3) overcoming confirmation bias by ensuring content was recommended that was relevant for the mission despite being obscure or less well known. We solve this by implementing a combination of collective matrix factorization (CMF) and graph-based random walks to provide recommendations to civil military operations users. We also present a method to resolve the challenge of computation complexity inherent from highly connected nodes through a precomputed process.
Keywords: Recommendation engine, collaborative filtering, context based recommendation, graph analysis, coverage, civil affairs operations, Marine Corps.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 38928 3D Segmentation, Compression and Wireless Transmission of Volumetric Brain MR Images
Authors: K. Aloui, M. S. Naceur
Abstract:
The main objective of this paper is to provide an efficient tool for delineating brain tumors in three-dimensional magnetic resonance images and set up compression-transmit schemes to distribute result to the remote doctor. To achieve this goal, we use basically a level-sets approach to delineating brain tumors in threedimensional. Then introduce a new compression and transmission plan of 3D brain structures based for the meshes simplification, adapted for time to the specific needs of the telemedicine and to the capacities restricted by wireless network communication. We present here the main stages of our system, and preliminary results which are very encouraging for clinical practice.
Keywords: Medical imaging, level-sets, compression, meshessimplification, telemedicine, wireless transmission.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 192827 A Classical Method of Optimizing Manufacturing Systems Using a Number of Industrial Engineering Techniques
Authors: John M. Ikome, Martha E. Ikome, Therese Van Wyk
Abstract:
Productivity optimization of a company can significantly increase the company’s output and productivity which can be in the form of corrective actions of ineffective activities, process simplification, and reduction of variations, responsiveness, and reduction of set-up-time which are all under the classification of waste within the manufacturing environment. Deriving a means to eliminate a number of these issues has a key importance for manufacturing organization. This paper focused on a number of industrial engineering techniques which include a cause and effect diagram, to identify and optimize the method or systems being used. Based on our results, it shows that there are a number of variations within the production processes that can significantly disrupt the expected output.
Keywords: Optimization, fishbone diagram, productivity.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 100026 3D Mesh Coarsening via Uniform Clustering
Authors: Shuhua Lai, Kairui Chen
Abstract:
In this paper, we present a fast and efficient mesh coarsening algorithm for 3D triangular meshes. Theis approach can be applied to very complex 3D meshes of arbitrary topology and with millions of vertices. The algorithm is based on the clustering of the input mesh elements, which divides the faces of an input mesh into a given number of clusters for clustering purpose by approximating the Centroidal Voronoi Tessellation of the input mesh. Once a clustering is achieved, it provides us an efficient way to construct uniform tessellations, and therefore leads to good coarsening of polygonal meshes. With proliferation of 3D scanners, this coarsening algorithm is particularly useful for reverse engineering applications of 3D models, which in many cases are dense, non-uniform, irregular and arbitrary topology. Examples demonstrating effectiveness of the new algorithm are also included in the paper.Keywords: Coarsening, mesh clustering, shape approximation, mesh simplification.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 140425 Establishing a New Simple Formula for Buckling Length Factor (K) of Rigid Frames Columns
Authors: Ehab Hasan Ahmed Hasan Ali
Abstract:
The calculation of buckling length factor (K) for steel frames columns is a major and governing processes to determine the dimensions steel frame columns cross sections during design. The buckling length of steel frames columns has a direct effect on the cost (weight) of using cross section. A new formula is required to determine buckling length factor (K) by simplified way. In this research a new formula for buckling length factor (K) was established to determine by accurate method for a limited interval of columns ends rigidity (GA, GB). The new formula can be used ease to evaluate the buckling length factor without needing to complicated equations or difficult charts.Keywords: Buckling length, New formula, Curve fitting, Simplification, Steel column design.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 225824 A New Block-based NLMS Algorithm and Its Realization in Block Floating Point Format
Authors: Abhijit Mitra
Abstract:
we propose a new normalized LMS (NLMS) algorithm, which gives satisfactory performance in certain applications in comaprison with con-ventional NLMS recursion. This new algorithm can be treated as a block based simplification of NLMS algorithm with significantly reduced number of multi¬ply and accumulate as well as division operations. It is also shown that such a recursion can be easily implemented in block floating point (BFP) arithmetic, treating the implementational issues much efficiently. In particular, the core challenges of a BFP realization to such adaptive filters are mainly considered in this regard. A global upper bound on the step size control parameter of the new algorithm due to BFP implementation is also proposed to prevent overflow in filtering as well as weight updating operations jointly.
Keywords: Adaptive algorithm, Block floating point arithmetic, Implementation issues, Normalized least mean square methods
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 236323 A Numerical Simulation of the Indoor Air Flow
Authors: Karel Frana, Jianshun S. Zhang, Milos Muller
Abstract:
The indoor airflow with a mixed natural/forced convection was numerically calculated using the laminar and turbulent approach. The Boussinesq approximation was considered for a simplification of the mathematical model and calculations. The results obtained, such as mean velocity fields, were successfully compared with experimental PIV flow visualizations. The effect of the distance between the cooled wall and the heat exchanger on the temperature and velocity distributions was calculated. In a room with a simple shape, the computational code OpenFOAM demonstrated an ability to numerically predict flow patterns. Furthermore, numerical techniques, boundary type conditions and the computational grid quality were examined. Calculations using the turbulence model k-omega had a significant effect on the results influencing temperature and velocity distributions.Keywords: natural and forced convections, numerical simulations, indoor airflows.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 320622 Analysis of a WDM System for Tanzania
Authors: Shaban Pazi, Chris Chatwin, Rupert Young, Philip Birch
Abstract:
Internet infrastructures in most places of the world have been supported by the advancement of optical fiber technology, most notably wavelength division multiplexing (WDM) system. Optical technology by means of WDM system has revolutionized long distance data transport and has resulted in high data capacity, cost reductions, extremely low bit error rate, and operational simplification of the overall Internet infrastructure. This paper analyses and compares the system impairments, which occur at data transmission rates of 2.5Gb/s and 10 Gb/s per wavelength channel in our proposed optical WDM system for Internet infrastructure in Tanzania. The results show that the data transmission rate of 2.5 Gb/s has minimum system impairments compared with a rate of 10 Gb/s per wavelength channel, and achieves a sufficient system performance to provide a good Internet access service.Keywords: Internet infrastructure, WDM system, standard single mode fibers, system impairments.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 166121 Tuberculosis Modelling Using Bio-PEPA Approach
Authors: Dalila Hamami, Baghdad Atmani
Abstract:
Modelling is a widely used tool to facilitate the evaluation of disease management. The interest of epidemiological models lies in their ability to explore hypothetical scenarios and provide decision makers with evidence to anticipate the consequences of disease incursion and impact of intervention strategies.
All models are, by nature, simplification of more complex systems. Models that involve diseases can be classified into different categories depending on how they treat the variability, time, space, and structure of the population. Approaches may be different from simple deterministic mathematical models, to complex stochastic simulations spatially explicit.
Thus, epidemiological modelling is now a necessity for epidemiological investigations, surveillance, testing hypotheses and generating follow-up activities necessary to perform complete and appropriate analysis.
The state of the art presented in the following, allows us to position itself to the most appropriate approaches in the epidemiological study.
Keywords: Bio-PEPA, Cellular automata, Epidemiological modelling, multi agent system, ordinary differential equations, PEPA, Process Algebra, Tuberculosis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 215820 A Novel Adaptive E-Learning Model Based on Developed Learner's Styles
Authors: Hazem M. El-Bakry, Ahmed A. Saleh, Taghreed T. Asfour
Abstract:
Adaptive e-learning today gives the student a central role in his own learning process. It allows learners to try things out, participate in courses like never before, and get more out of learning than before. In this paper, an adaptive e-learning model for logic design, simplification of Boolean functions and related fields is presented. Such model presents suitable courses for each student in a dynamic and adaptive manner using existing database and workflow technologies. The main objective of this research work is to provide an adaptive e-learning model based learners' personality using explicit and implicit feedback. To recognize the learner-s, we develop dimensions to decide each individual learning style in order to accommodate different abilities of the users and to develop vital skills. Thus, the proposed model becomes more powerful, user friendly and easy to use and interpret. Finally, it suggests a learning strategy and appropriate electronic media that match the learner-s preference.Keywords: Adaptive learning, Learning styles, Teaching strategies.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 202419 A Parameter-Tuning Framework for Metaheuristics Based on Design of Experiments and Artificial Neural Networks
Authors: Felix Dobslaw
Abstract:
In this paper, a framework for the simplification and standardization of metaheuristic related parameter-tuning by applying a four phase methodology, utilizing Design of Experiments and Artificial Neural Networks, is presented. Metaheuristics are multipurpose problem solvers that are utilized on computational optimization problems for which no efficient problem specific algorithm exist. Their successful application to concrete problems requires the finding of a good initial parameter setting, which is a tedious and time consuming task. Recent research reveals the lack of approach when it comes to this so called parameter-tuning process. In the majority of publications, researchers do have a weak motivation for their respective choices, if any. Because initial parameter settings have a significant impact on the solutions quality, this course of action could lead to suboptimal experimental results, and thereby a fraudulent basis for the drawing of conclusions.Keywords: Parameter-Tuning, Metaheuristics, Design of Experiments, Artificial Neural Networks.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 177618 Parametrization of Piezoelectric Vibration Energy Harvesters for Low Power Embedded Systems
Authors: Yannick Verbelen, Tim Dekegel, Ann Peeters, Klara Stinders, Niek Blondeel, Sam De Winne, An Braeken, Abdellah Touhafi
Abstract:
Matching an embedded electronic application with a cantilever vibration energy harvester remains a difficult endeavour due to the large number of factors influencing the output power. In the presented work, complementary balanced energy harvester parametrization is used as a methodology for simplification of harvester integration in electronic applications. This is achieved by a dual approach consisting of an adaptation of the general parametrization methodology in conjunction with a straight forward harvester benchmarking strategy. For this purpose, the design and implementation of a suitable user friendly cantilever energy harvester benchmarking platform is discussed. Its effectiveness is demonstrated by applying the methodology to a commercially available Mide V21BL vibration energy harvester, with excitation amplitude and frequency as variables.Keywords: Energy harvesting, vibrations, piezoelectric transducers, embedded systems, harvester parametrization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 130917 An Approximate Engineering Method for Aerodynamic Heating Solution around Blunt Body Nose
Authors: Sahar Noori, Seyed Amir Hossein, Mohammad Ebrahimi
Abstract:
This paper is devoted to predict laminar and turbulent heating rates around blunt re-entry spacecraft at hypersonic conditions. Heating calculation of a hypersonic body is normally performed during the critical part of its flight trajectory. The procedure is of an inverse method, where a shock wave is assumed, and the body shape that supports this shock, as well as the flowfield between the shock and body, are calculated. For simplicity the normal momentum equation is replaced with a second order pressure relation; this simplification significantly reduces computation time. The geometries specified in this research, are parabola and ellipsoids which may have conical after bodies. An excellent agreement is observed between the results obtained in this paper and those calculated by others- research. Since this method is much faster than Navier-Stokes solutions, it can be used in preliminary design, parametric study of hypersonic vehicles.Keywords: Aerodynamic Heating, Blunt Body, Hypersonic Flow, Laminar, Turbulent.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 371916 A Method for 3D Mesh Adaptation in FEA
Authors: S. Sfarni, E. Bellenger, J. Fortin, M. Guessasma
Abstract:
The use of the mechanical simulation (in particular the finite element analysis) requires the management of assumptions in order to analyse a real complex system. In finite element analysis (FEA), two modeling steps require assumptions to be able to carry out the computations and to obtain some results: the building of the physical model and the building of the simulation model. The simplification assumptions made on the analysed system in these two steps can generate two kinds of errors: the physical modeling errors (mathematical model, domain simplifications, materials properties, boundary conditions and loads) and the mesh discretization errors. This paper proposes a mesh adaptive method based on the use of an h-adaptive scheme in combination with an error estimator in order to choose the mesh of the simulation model. This method allows us to choose the mesh of the simulation model in order to control the cost and the quality of the finite element analysis.
Keywords: Finite element, discretization errors, adaptivity.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 147715 Research on the Survivability of Embedded Real-time System
Abstract:
Introducing survivability into embedded real-time system (ERTS) can improve the survivability power of the system. This paper mainly discusses about the survivability of ERTS. The first is the survivability origin of ERTS. The second is survivability analysis. According to the definition of survivability based on survivability specification and division of the entire survivability analysis process for ERTS, a survivability analysis profile is presented. The quantitative analysis model of this profile is emphasized and illuminated in detail, the quantifying analysis of system was showed helpful to evaluate system survivability more accurate. The third is platform design of survivability analysis. In terms of the profile, the analysis process is encapsulated and assembled into one platform, on which quantification, standardization and simplification of survivability analysis are all achieved. The fourth is survivability design. According to character of ERTS, strengthened design method is selected to realize system survivability design. Through the analysis of embedded mobile video-on-demand system, intrusion tolerant technology is introduced in whole survivability design.
Keywords: ERTS (embedded real-time system), survivability, quantitative analysis, survivability specification, intrusion tolerant
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 130214 Approximate Bounded Knowledge Extraction Using Type-I Fuzzy Logic
Authors: Syed Muhammad Aqil Burney, Tahseen Ahmed Jilani, C. Ardil
Abstract:
Using neural network we try to model the unknown function f for given input-output data pairs. The connection strength of each neuron is updated through learning. Repeated simulations of crisp neural network produce different values of weight factors that are directly affected by the change of different parameters. We propose the idea that for each neuron in the network, we can obtain quasi-fuzzy weight sets (QFWS) using repeated simulation of the crisp neural network. Such type of fuzzy weight functions may be applied where we have multivariate crisp input that needs to be adjusted after iterative learning, like claim amount distribution analysis. As real data is subjected to noise and uncertainty, therefore, QFWS may be helpful in the simplification of such complex problems. Secondly, these QFWS provide good initial solution for training of fuzzy neural networks with reduced computational complexity.
Keywords: Crisp neural networks, fuzzy systems, extraction of logical rules, quasi-fuzzy numbers.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 173913 Ontology-Driven Generation of Radiation Protection Procedures
Authors: Chamseddine Barki, Salam Labidi, Hanen Boussi Rahmouni
Abstract:
In this article, we present the principle and suitable methodology for the design of a medical ontology that highlights the radiological and dosimetric knowledge, applied in diagnostic radiology and radiation-therapy. Our ontology, which we named «Onto.Rap», is the subject of radiation protection in medical and radiology centers by providing a standardized regulatory oversight. Thanks to its added values of knowledge-sharing, reuse and the ease of maintenance, this ontology tends to solve many problems. Of which we name the confusion between radiological procedures a practitioner might face while performing a patient radiological exam. Adding to it, the difficulties they might have in interpreting applicable patient radioprotection standards. Here, the ontology, thanks to its concepts simplification and expressiveness capabilities, can ensure an efficient classification of radiological procedures. It also provides an explicit representation of the relations between the different components of the studied concept. In fact, an ontology based-radioprotection expert system, when used in radiological center, could implement systematic radioprotection best practices during patient exam and a regulatory compliance service auditing afterwards.Keywords: Ontology, radiology, medicine, knowledge, radiation protection, audit.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 128912 Radiation Usage Impact of on Anti-Nutritional Compounds (Antitrypsin and Phytic Acid) of Livestock and Poultry Foods
Authors: Mohammad Khosravi, Ali Kiani, Behroz Dastar, Parvin Showrang
Abstract:
Review was carried out on important anti-nutritional compounds of livestock and poultry foods and the effect of radiation usage. Nowadays, with advancement in technology, different methods have been considered for the optimum usage of nutrients in livestock and poultry foods. Steaming, extruding, pelleting, and the use of chemicals are the most common and popular methods in food processing. Use of radiation in food processing researches in the livestock and poultry industry is currently highly regarded. Ionizing (electrons, gamma) and non-ionizing beams (microwave and infrared) are the most useable rays in animal food processing. In recent researches, these beams have been used to remove and reduce the anti-nutritional factors and microbial contamination and improve the digestibility of nutrients in poultry and livestock food. The evidence presented will help researchers to recognize techniques of relevance to them. Simplification of some of these techniques, especially in developing countries, must be addressed so that they can be used more widely.
Keywords: Antitrypsin, gamma anti-nutritional components, phytic acid, radiation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 121011 Possibilities for Testing User Experience and User Interface Design on Mobile Devices
Authors: J. Berčík, A. Mravcová, J. Gálová, K. Neomániová
Abstract:
In an era when everything is increasingly digital, consumers are always looking for new options in solutions to their everyday needs. In this context, mobile apps are developing at an exponential pace. One of the fastest growing segments of mobile technologies is, obviously, e-commerce. It can be predicted that mobile commerce will record nearly three times the global growth of e-commerce across all platforms, which indicates its importance in the given segment. The current coronavirus pandemic is also changing many of the existing paradigms both socially, economically, and technologically, which has a major impact on changing consumer behavior and the emphasis on simplification and clarity of mobile solutions. This is the area that User Experience (UX) and User Interface (UI) designers deal with. Their task is to design a sufficiently attractive and interesting solution that will be available on all mobile devices and at the same time will be easy enough for the customer/visitor to get to the destination or to get the necessary information in a few clicks. The basis for changes in UX design can now be obtained not only through online analytical tools, but also through neuromarketing, especially in the case of mobile devices. The paper highlights possibilities for testing UX design applications on mobile devices using a special platform that combines a stationary eye camera (eye tracking) and facial analysis (facial coding).
Keywords: Emotions, mobile design, user experience, visual attention.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 65510 Reliability Verification of the Performance Evaluation of Multiphase Pump
Authors: Joon-Hyung Kim, Him-Chan Lee, Jin-Hyuk Kim, Yong-Kab Lee, Young-Seok Choi
Abstract:
The crude oil in an oil well exists in various phases such as gas, seawater, and sand, as well as oil. Therefore, a phase separator is needed at the front of a single-phase pump for pressurization and transfer. On the other hand, the application of a multiphase pump can provide such advantages as simplification of the equipment structure and cost savings, because there is no need for a phase separation process. Therefore, the crude oil transfer method using a multiphase pump is being applied to recently developed oil wells. Due to this increase in demand, technical demands for the development of multiphase pumps are sharply increasing, but the progress of research into related technologies is insufficient, due to the nature of multiphase pumps that require high levels of skills. This study was conducted to verify the reliability of pump performance evaluation using numerical analysis, which is the basis of the development of a multiphase pump. For this study, a model was designed by selecting the specifications of this study. The performance of the designed model was evaluated through numerical analysis and experiment. The results of the performance evaluation were compared to verify the reliability of the result using numerical analysis.
Keywords: Multiphase pump, Numerical analysis, Experiment, Performance evaluation, Reliability verification.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3133