Search results for: linear image degradation model
4744 Automatic Tuning for a Systemic Model of Banking Originated Losses (SYMBOL) Tool on Multicore
Authors: Ronal Muresano, Andrea Pagano
Abstract:
Nowadays, the mathematical/statistical applications are developed with more complexity and accuracy. However, these precisions and complexities have brought as result that applications need more computational power in order to be executed faster. In this sense, the multicore environments are playing an important role to improve and to optimize the execution time of these applications. These environments allow us the inclusion of more parallelism inside the node. However, to take advantage of this parallelism is not an easy task, because we have to deal with some problems such as: cores communications, data locality, memory sizes (cache and RAM), synchronizations, data dependencies on the model, etc. These issues are becoming more important when we wish to improve the application’s performance and scalability. Hence, this paper describes an optimization method developed for Systemic Model of Banking Originated Losses (SYMBOL) tool developed by the European Commission, which is based on analyzing the application's weakness in order to exploit the advantages of the multicore. All these improvements are done in an automatic and transparent manner with the aim of improving the performance metrics of our tool. Finally, experimental evaluations show the effectiveness of our new optimized version, in which we have achieved a considerable improvement on the execution time. The time has been reduced around 96% for the best case tested, between the original serial version and the automatic parallel version.
Keywords: Algorithm optimization, Bank Failures, OpenMP, Parallel Techniques, Statistical tool.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19024743 Face Recognition Based On Vector Quantization Using Fuzzy Neuro Clustering
Authors: Elizabeth B. Varghese, M. Wilscy
Abstract:
A face recognition system is a computer application for automatically identifying or verifying a person from a digital image or a video frame. A lot of algorithms have been proposed for face recognition. Vector Quantization (VQ) based face recognition is a novel approach for face recognition. Here a new codebook generation for VQ based face recognition using Integrated Adaptive Fuzzy Clustering (IAFC) is proposed. IAFC is a fuzzy neural network which incorporates a fuzzy learning rule into a competitive neural network. The performance of proposed algorithm is demonstrated by using publicly available AT&T database, Yale database, Indian Face database and a small face database, DCSKU database created in our lab. In all the databases the proposed approach got a higher recognition rate than most of the existing methods. In terms of Equal Error Rate (ERR) also the proposed codebook is better than the existing methods.
Keywords: Face Recognition, Vector Quantization, Integrated Adaptive Fuzzy Clustering, Self Organization Map.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22424742 Simulation-Based Diversity Management in Human-Robot Collaborative Scenarios
Authors: Titanilla Komenda, Viktorio Malisa
Abstract:
In this paper, the influence of diversity-related factors on the design of collaborative scenarios is analysed. Based on the evaluation, a framework for simulating human-robot-collaboration is presented that considers both human factors as well as the overall system performance. The implementation of the model is shown on a real-life scenario from industry and validated in terms of traceability, safety and physical limitations. By comparing scenarios that consider diversity with those only meeting system performance, an overall understanding of individually adapted human-robot-collaborative workspaces is reached. A diversity-related guideline for human-robot-collaborations provides a summary of the research and aids in optimizing future applications. Finally, limitations and future amendments of the model are discussed.Keywords: Diversity, human-machine-system, human-robot-collaboration, simulation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13374741 Describing Learning Features of Reusable Resources: A Proposal
Authors: Serena Alvino, Paola Forcheri, Maria Grazia Ierardi, Luigi Sarti
Abstract:
One of the main advantages of the LO paradigm is to allow the availability of good quality, shareable learning material through the Web. The effectiveness of the retrieval process requires a formal description of the resources (metadata) that closely fits the user-s search criteria; in spite of the huge international efforts in this field, educational metadata schemata often fail to fulfil this requirement. This work aims to improve the situation, by the definition of a metadata model capturing specific didactic features of shareable learning resources. It classifies LOs into “teacher-oriented" and “student-oriented" categories, in order to describe the role a LO is to play when it is integrated into the educational process. This article describes the model and a first experimental validation process that has been carried out in a controlled environment.Keywords: Learning object, pedagogical metadata, experimental validation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15454740 Development of Sustainable Building Environmental Model (SBEM) in Hong Kong
Authors: Kwok W. Mui, Ling T. Wong, F. Xiao, Chin T. Cheung, Ho C. Yu
Abstract:
This study addresses a concept of the Sustainable Building Environmental Model (SBEM) developed to optimize energy consumption in air conditioning and ventilation (ACV) systems without any deterioration of indoor environmental quality (IEQ). The SBEM incorporates two main components: an adaptive comfort temperature control module (ACT) and a new carbon dioxide demand control module (nDCV). These two modules take an innovative approach to maintain satisfaction of the Indoor Environmental Quality (IEQ) with optimum energy consumption; they provide a rational basis of effective control. A total of 2133 sets of measurement data of indoor air temperature (Ta), relative humidity (Rh) and carbon dioxide concentration (CO2) were conducted in some Hong Kong offices to investigate the potential of integrating the SBEM. A simulation was used to evaluate the dynamic performance of the energy and air conditioning system with the integration of the SBEM in an air-conditioned building. It allows us make a clear picture of the control strategies and performed any pre-tuned of controllers before utilized in real systems. With the integration of SBEM, it was able to save up to 12.3% in simulation of overall electricity consumption, and maintain the average carbon dioxide concentration within 1000ppm and occupant dissatisfaction in 20%.
Keywords: —Sustainable building environmental model (SBEM), adaptive comfort temperature (ACT), new demand control ventilation (nDCV), energy saving.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18254739 A Numerical Simulation of the Indoor Air Flow
Authors: Karel Frana, Jianshun S. Zhang, Milos Muller
Abstract:
The indoor airflow with a mixed natural/forced convection was numerically calculated using the laminar and turbulent approach. The Boussinesq approximation was considered for a simplification of the mathematical model and calculations. The results obtained, such as mean velocity fields, were successfully compared with experimental PIV flow visualizations. The effect of the distance between the cooled wall and the heat exchanger on the temperature and velocity distributions was calculated. In a room with a simple shape, the computational code OpenFOAM demonstrated an ability to numerically predict flow patterns. Furthermore, numerical techniques, boundary type conditions and the computational grid quality were examined. Calculations using the turbulence model k-omega had a significant effect on the results influencing temperature and velocity distributions.Keywords: natural and forced convections, numerical simulations, indoor airflows.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 32064738 Automated Detection of Alzheimer Disease Using Region Growing technique and Artificial Neural Network
Authors: B. Al-Naami, N. Gharaibeh, A. AlRazzaq Kheshman
Abstract:
Alzheimer is known as the loss of mental functions such as thinking, memory, and reasoning that is severe enough to interfere with a person's daily functioning. The appearance of Alzheimer Disease symptoms (AD) are resulted based on which part of the brain has a variety of infection or damage. In this case, the MRI is the best biomedical instrumentation can be ever used to discover the AD existence. Therefore, this paper proposed a fusion method to distinguish between the normal and (AD) MRIs. In this combined method around 27 MRIs collected from Jordanian Hospitals are analyzed based on the use of Low pass -morphological filters to get the extracted statistical outputs through intensity histogram to be employed by the descriptive box plot. Also, the artificial neural network (ANN) is applied to test the performance of this approach. Finally, the obtained result of t-test with confidence accuracy (95%) has compared with classification accuracy of ANN (100 %). The robust of the developed method can be considered effectively to diagnose and determine the type of AD image.Keywords: Alzheimer disease, Brain MRI analysis, Morphological filter, Box plot, Intensity histogram, ANN.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 31424737 Mathematical Modeling of an Avalanche Release and Estimation of Flow Parameters by Numerical Method
Authors: Mahmoud Zarrini
Abstract:
Avalanche release of snow has been modeled in the present studies. Snow is assumed to be represented by semi-solid and the governing equations have been studied from the concept of continuum approach. The dynamical equations have been solved for two different zones [starting zone and track zone] by using appropriate initial and boundary conditions. Effect of density (ρ), Eddy viscosity (η), Slope angle (θ), Slab depth (R) on the flow parameters have been observed in the present studies. Numerical methods have been employed for computing the non linear differential equations. One of the most interesting and fundamental innovation in the present studies is getting initial condition for the computation of velocity by numerical approach. This information of the velocity has obtained through the concept of fracture mechanics applicable to snow. The results on the flow parameters have found to be in qualitative agreement with the published results.
Keywords: Snow avalanche, fracture mechanics, avalanche velocity, avalanche zones.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17724736 Reduced Order Modeling of Natural Gas Transient Flow in Pipelines
Authors: M. Behbahani-Nejad, Y. Shekari
Abstract:
A reduced order modeling approach for natural gas transient flow in pipelines is presented. The Euler equations are considered as the governing equations and solved numerically using the implicit Steger-Warming flux vector splitting method. Next, the linearized form of the equations is derived and the corresponding eigensystem is obtained. Then, a few dominant flow eigenmodes are used to construct an efficient reduced-order model. A well-known test case is presented to demonstrate the accuracy and the computational efficiency of the proposed method. The results obtained are in good agreement with those of the direct numerical method and field data. Moreover, it is shown that the present reduced-order model is more efficient than the conventional numerical techniques for transient flow analysis of natural gas in pipelines.Keywords: Eigenmode, Natural Gas, Reduced Order Modeling, Transient Flow.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19384735 A Note on the Minimum Cardinality of Critical Sets of Inertias for Irreducible Zero-nonzero Patterns of Order 4
Authors: Ber-Lin Yu, Ting-Zhu Huang
Abstract:
If there exists a nonempty, proper subset S of the set of all (n+1)(n+2)/2 inertias such that S Ôèå i(A) is sufficient for any n×n zero-nonzero pattern A to be inertially arbitrary, then S is called a critical set of inertias for zero-nonzero patterns of order n. If no proper subset of S is a critical set, then S is called a minimal critical set of inertias. In [Kim, Olesky and Driessche, Critical sets of inertias for matrix patterns, Linear and Multilinear Algebra, 57 (3) (2009) 293-306], identifying all minimal critical sets of inertias for n×n zero-nonzero patterns with n ≥ 3 and the minimum cardinality of such a set are posed as two open questions by Kim, Olesky and Driessche. In this note, the minimum cardinality of all critical sets of inertias for 4 × 4 irreducible zero-nonzero patterns is identified.
Keywords: Zero-nonzero pattern, inertia, critical set of inertias, inertially arbitrary.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 11984734 Automatic Segmentation of Thigh Magnetic Resonance Images
Authors: Lorena Urricelqui, Armando Malanda, Arantxa Villanueva
Abstract:
Purpose: To develop a method for automatic segmentation of adipose and muscular tissue in thighs from magnetic resonance images. Materials and methods: Thirty obese women were scanned on a Siemens Impact Expert 1T resonance machine. 1500 images were finally used in the tests. The developed segmentation method is a recursive and multilevel process that makes use of several concepts such as shaped histograms, adaptative thresholding and connectivity. The segmentation process was implemented in Matlab and operates without the need of any user interaction. The whole set of images were segmented with the developed method. An expert radiologist segmented the same set of images following a manual procedure with the aid of the SliceOmatic software (Tomovision). These constituted our 'goal standard'. Results: The number of coincidental pixels of the automatic and manual segmentation procedures was measured. The average results were above 90 % of success in most of the images. Conclusions: The proposed approach allows effective automatic segmentation of MRIs from thighs, comparable to expert manual performance.
Keywords: Segmentation, thigh, magnetic resonance image, fat, muscle.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19054733 Quantification of Periodicities in Fugitive Emission of Gases from Lyari Waterway
Authors: Rana Khalid Naeem, Asif Mansoor
Abstract:
Periodicities in the environmetric time series can be idyllically assessed by utilizing periodic models. In this communication fugitive emission of gases from open sewer channel Lyari which follows periodic behaviour are approximated by employing periodic autoregressive model of order p. The orders of periodic model for each season are selected through the examination of periodic partial autocorrelation or information criteria. The parameters for the selected order of season are estimated individually for each emitted air toxin. Subsequently, adequacies of fitted models are established by examining the properties of the residual for each season. These models are beneficial for schemer and administrative bodies for the improvement of implemented policies to surmount future environmental problems.Keywords: Exchange of Gases, Goodness of Fit, Open Sewer Channel, PAR(p) Models, Periodicities, Season Wise Models.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 12584732 Integrating the Theory of Constraints and Six Sigma in Manufacturing Process Improvement
Authors: Kai Jin, Hyder Abdul-Razzak, Yousri Elkassabgi, Hong Zhou, Aaron Herrera
Abstract:
Six Sigma is a well known discipline that reduces variation using complex statistical tools and the DMAIC model. By integrating Goldratts-s Theory of Constraints, the Five Focusing Points and System Thinking tools, Six Sigma projects can be selected where it can cause more impact in the company. This research defines an integrated model of six sigma and constraint management that shows a step-by-step guide using the original methodologies from each discipline and is evaluated in a case study from the production line of a Automobile engine monoblock V8, resulting in an increase in the line capacity from 18.7 pieces per hour to 22.4 pieces per hour, a reduction of 60% of Work-In-Process and a variation decrease of 0.73%.Keywords: Constraint Management, Manufacturing Process Improvement, Six Sigma, System Thinking.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17634731 Exchanges of Knowledge about Product Configurations using XML Topic Map
Authors: Namchul Do, Jihun Cho
Abstract:
Modeling product configurations needs large amounts of knowledge about technical and marketing restrictions on the product. Previous attempts to automate product configurations concentrate on representations and management of the knowledge for specific domains in fixed and isolated computing environments. Since the knowledge about product configurations is subject to continuous change and hard to express, these attempts often failed to efficiently manage and exchange the knowledge in collaborative product development. In this paper, XML Topic Map (XTM) is introduced to represent and exchange the knowledge about product configurations in collaborative product development. A product configuration model based on XTM along with its merger and inference facilities enables configuration engineers in collaborative product development to manage and exchange their knowledge efficiently. A prototype implementation is also presented to demonstrate the proposed model can be applied to engineering information systems to exchange the product configuration knowledge.
Keywords: Knowledge exchange, product configurations, XML topic map.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13534730 Discrete Polynomial Moments and Savitzky-Golay Smoothing
Authors: Paul O'Leary, Matthew Harker
Abstract:
This paper presents unified theory for local (Savitzky- Golay) and global polynomial smoothing. The algebraic framework can represent any polynomial approximation and is seamless from low degree local, to high degree global approximations. The representation of the smoothing operator as a projection onto orthonormal basis functions enables the computation of: the covariance matrix for noise propagation through the filter; the noise gain and; the frequency response of the polynomial filters. A virtually perfect Gram polynomial basis is synthesized, whereby polynomials of degree d = 1000 can be synthesized without significant errors. The perfect basis ensures that the filters are strictly polynomial preserving. Given n points and a support length ls = 2m + 1 then the smoothing operator is strictly linear phase for the points xi, i = m+1. . . n-m. The method is demonstrated on geometric surfaces data lying on an invariant 2D lattice.Keywords: Gram polynomials, Savitzky-Golay Smoothing, Discrete Polynomial Moments
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 27914729 Air Handling Units Power Consumption Using Generalized Additive Model for Anomaly Detection: A Case Study in a Singapore Campus
Authors: Ju Peng Poh, Jun Yu Charles Lee, Jonathan Chew Hoe Khoo
Abstract:
The emergence of digital twin technology, a digital replica of physical world, has improved the real-time access to data from sensors about the performance of buildings. This digital transformation has opened up many opportunities to improve the management of the building by using the data collected to help monitor consumption patterns and energy leakages. One example is the integration of predictive models for anomaly detection. In this paper, we use the GAM (Generalised Additive Model) for the anomaly detection of Air Handling Units (AHU) power consumption pattern. There is ample research work on the use of GAM for the prediction of power consumption at the office building and nation-wide level. However, there is limited illustration of its anomaly detection capabilities, prescriptive analytics case study, and its integration with the latest development of digital twin technology. In this paper, we applied the general GAM modelling framework on the historical data of the AHU power consumption and cooling load of the building between Jan 2018 to Aug 2019 from an education campus in Singapore to train prediction models that, in turn, yield predicted values and ranges. The historical data are seamlessly extracted from the digital twin for modelling purposes. We enhanced the utility of the GAM model by using it to power a real-time anomaly detection system based on the forward predicted ranges. The magnitude of deviation from the upper and lower bounds of the uncertainty intervals is used to inform and identify anomalous data points, all based on historical data, without explicit intervention from domain experts. Notwithstanding, the domain expert fits in through an optional feedback loop through which iterative data cleansing is performed. After an anomalously high or low level of power consumption detected, a set of rule-based conditions are evaluated in real-time to help determine the next course of action for the facilities manager. The performance of GAM is then compared with other approaches to evaluate its effectiveness. Lastly, we discuss the successfully deployment of this approach for the detection of anomalous power consumption pattern and illustrated with real-world use cases.
Keywords: Anomaly detection, digital twin, Generalised Additive Model, Power Consumption Model.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5024728 Modeling Spatial Distributions of Point and Nonpoint Source Pollution Loadings in the Great Lakes Watersheds
Authors: Chansheng He, Carlo DeMarchi
Abstract:
A physically based, spatially-distributed water quality model is being developed to simulate spatial and temporal distributions of material transport in the Great Lakes Watersheds of the U.S. Multiple databases of meteorology, land use, topography, hydrography, soils, agricultural statistics, and water quality were used to estimate nonpoint source loading potential in the study watersheds. Animal manure production was computed from tabulations of animals by zip code area for the census years of 1987, 1992, 1997, and 2002. Relative chemical loadings for agricultural land use were calculated from fertilizer and pesticide estimates by crop for the same periods. Comparison of these estimates to the monitored total phosphorous load indicates that both point and nonpoint sources are major contributors to the total nutrient loads in the study watersheds, with nonpoint sources being the largest contributor, particularly in the rural watersheds. These estimates are used as the input to the distributed water quality model for simulating pollutant transport through surface and subsurface processes to Great Lakes waters. Visualization and GIS interfaces are developed to visualize the spatial and temporal distribution of the pollutant transport in support of water management programs.
Keywords: Distributed Large Basin Runoff Model, Great LakesWatersheds, nonpoint source pollution, and point sources.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15344727 PSO-based Possibilistic Portfolio Model with Transaction Costs
Authors: Wei Chen, Cui-you Yao, Yue Qiu
Abstract:
This paper deals with a portfolio selection problem based on the possibility theory under the assumption that the returns of assets are LR-type fuzzy numbers. A possibilistic portfolio model with transaction costs is proposed, in which the possibilistic mean value of the return is termed measure of investment return, and the possibilistic variance of the return is termed measure of investment risk. Due to considering transaction costs, the existing traditional optimization algorithms usually fail to find the optimal solution efficiently and heuristic algorithms can be the best method. Therefore, a particle swarm optimization is designed to solve the corresponding optimization problem. At last, a numerical example is given to illustrate our proposed effective means and approaches.Keywords: Possibility theory, portfolio selection, transaction costs, particle swarm optimization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15344726 Low Cost Technique for Measuring Luminance in Biological Systems
Abstract:
In this work, the relationship between the melanin content in a tissue and subsequent absorption of light through that tissue was determined using a digital camera. This technique proved to be simple, cost effective, efficient and reliable. Tissue phantom samples were created using milk and soy sauce to simulate the optical properties of melanin content in human tissue. Increasing the concentration of soy sauce in the milk correlated to an increase in melanin content of an individual. Two methods were employed to measure the light transmitted through the sample. The first was direct measurement of the transmitted intensity using a conventional lux meter. The second method involved correctly calibrating an ordinary digital camera and using image analysis software to calculate the transmitted intensity through the phantom. The results from these methods were then graphically compared to the theoretical relationship between the intensity of transmitted light and the concentration of absorbers in the sample. Conclusions were then drawn about the effectiveness and efficiency of these low cost methods.Keywords: Tissue phantoms, scattering coefficient, albedo, low-cost method.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13014725 Investigation of Dissolution in Diammonium Hydrogen Phosphate Solutions of Gypsum
Authors: Turan Çalban, Nursel Keskin, Sabri Çolak, Soner Kuşlu
Abstract:
Gypsum (CaSO4.2H2O) is a mineral that is found in large quantities in the Turkey and in the World. In this study, the dissolution of this mineral in the diammonium hydrogen phosphate solutions has been studied. The dissolution and dissolution kinetics of gypsum in diammonium hydrogen phosphate solutions will be useful for evaluating of solid wastes containing gypsum. Parameters such as diammonium hydrogen phosphate concentration, temperature and stirring speed affecting on the dissolution rate of the gypsum in diammonium hydrogen phosphate solutions were investigated. In experimental studies have researched effectiveness of the selected parameters. The dissolution of gypsum were examined in two parts at low and high temperatures. The experimental results were successfully correlated by linear regression using Statistica program. Dissolution curves were evaluated shrinking core models for solidfluid systems. The activation energy was found to be 34.58 kJ/mol and 44.45 kJ/mol for the low and the high temperatures. The dissolution of gypsum was controlled by chemical reaction both low temperatures and high temperatures.
Keywords: Diammonium hydrogen phosphate, Dissolution, Gypsum, Kinetics.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20514724 State Estimation Based on Unscented Kalman Filter for Burgers’ Equation
Authors: Takashi Shimizu, Tomoaki Hashimoto
Abstract:
Controlling the flow of fluids is a challenging problem that arises in many fields. Burgers’ equation is a fundamental equation for several flow phenomena such as traffic, shock waves, and turbulence. The optimal feedback control method, so-called model predictive control, has been proposed for Burgers’ equation. However, the model predictive control method is inapplicable to systems whose all state variables are not exactly known. In practical point of view, it is unusual that all the state variables of systems are exactly known, because the state variables of systems are measured through output sensors and limited parts of them can be only available. In fact, it is usual that flow velocities of fluid systems cannot be measured for all spatial domains. Hence, any practical feedback controller for fluid systems must incorporate some type of state estimator. To apply the model predictive control to the fluid systems described by Burgers’ equation, it is needed to establish a state estimation method for Burgers’ equation with limited measurable state variables. To this purpose, we apply unscented Kalman filter for estimating the state variables of fluid systems described by Burgers’ equation. The objective of this study is to establish a state estimation method based on unscented Kalman filter for Burgers’ equation. The effectiveness of the proposed method is verified by numerical simulations.Keywords: State estimation, fluid systems, observer systems, unscented Kalman filter.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 7424723 Weak Instability in Direct Integration Methods for Structural Dynamics
Authors: Shuenn-Yih Chang, Chiu-Li Huang
Abstract:
Three structure-dependent integration methods have been developed for solving equations of motion, which are second-order ordinary differential equations, for structural dynamics and earthquake engineering applications. Although they generally have the same numerical properties, such as explicit formulation, unconditional stability and second-order accuracy, a different performance is found in solving the free vibration response to either linear elastic or nonlinear systems with high frequency modes. The root cause of this different performance in the free vibration responses is analytically explored herein. As a result, it is verified that a weak instability is responsible for the different performance of the integration methods. In general, a weak instability will result in an inaccurate solution or even numerical instability in the free vibration responses of high frequency modes. As a result, a weak instability must be prohibited for time integration methods.Keywords: Dynamic analysis, high frequency, integration method, overshoot, weak instability.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6764722 Performance Analysis of Quantum Cascaded Lasers
Authors: M. B. El_Mashade, I. I. Mahamoud, M. S. El_Tokhy
Abstract:
Improving the performance of the QCL through block diagram as well as mathematical models is the main scope of this paper. In order to enhance the performance of the underlined device, the mathematical model parameters are used in a reliable manner in such a way that the optimum behavior was achieved. These parameters play the central role in specifying the optical characteristics of the considered laser source. Moreover, it is important to have a large amount of radiated power, where increasing the amount of radiated power represents the main hopping process that can be predicted from the behavior of quantum laser devices. It was found that there is a good agreement between the calculated values from our mathematical model and those obtained with VisSim and experimental results. These demonstrate the strength of mplementation of both mathematical and block diagram models.
Keywords: Quantum Cascaded Lasers (QCLs), Modeling, Block Diagram Programming, Intersubband transitions
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14404721 Thrust Vectoring Control of Supersonic Flow Through an Orifice Injector
Authors: Ibrahim Mnafeg, Azgal Abichou, Lotfi Beji
Abstract:
Traditional mechanical control systems in thrust vectoring are efficient in rocket thrust guidance but their costs and their weights are excessive. The fluidic injection in the nozzle divergent constitutes an alternative procedure to achieve the goal. In this paper, we present a 3D analytical model for fluidic injection in a supersonic nozzle integrating an orifice. The fluidic vectoring uses a sonic secondary injection in the divergent. As a result, the flow and interaction between the main and secondary jet has built in order to express the pressure fields from which the forces and thrust vectoring are deduced. Under various separation criteria, the present analytical model results are compared with the existing numerical and experimental data from the literature.Keywords: Flow separation, Fluidic thrust vectoring, Nozzle, Secondary jet, Shock wave.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21294720 Infrared Camera-Based Hand Gesture Space Touch System Implementation of Smart Device Environment
Authors: Yang-Keun Ahn, Kwang-Soon Choi, Young-Choong Park, Kwang-Mo Jung
Abstract:
This paper proposes a method to recognize the tip of a finger and space touch hand gesture using an infrared camera in a smart device environment. The proposed method estimates the tip of a finger with a curvature-based ellipse fitting algorithm, and verifies that the estimated object is indeed a finger with an ellipse fitting rectangular area. The feature extracted from the verified finger tip is used to implement the movement of a mouse and clicking gesture. The proposed algorithm was implemented with an actual smart device to test the proposed method. Empirical parameters were obtained from the keypad software and an image analysis tool for the performance optimization, and a comparative analysis with conventional research showed improved performance with the proposed method.
Keywords: Infrared camera, Hand gesture, Smart device, Space touch.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 23374719 Virtual Conciliation in Colombia: Evaluation of Maturity Level within the Framework of E-Government
Authors: Jenny Paola Forero Pachón, Sonia Cristina Gamboa Sarmiento, Luis Carlos Gómez Flórez
Abstract:
The Colombian government has defined an e-government strategy to take advantage of Information Technologies (IT) in order to contribute to the building of a more efficient, transparent and participative State that provides better services to citizens and businesses. In this regard, the Justice sector is one of the government sectors where IT has generated more expectation considering that the country has a judicial processes backlog. This situation has led to the search for alternative forms of access to justice that speed up the process while providing a low cost for citizens. To this end, the Colombian government has authorized the use of Alternative Dispute Resolution methods (ADR), a remedy where disputes can be resolved more quickly compared to judicial processes while facilitating greater communication between the parties, without recourse to judicial authority. One of these methods is conciliation, which includes a special modality that takes advantage of IT for the development of itself known as virtual conciliation. With this option the conciliation is supported by information systems, applications or platforms and communications are provided through it. This paper evaluates the level of maturity in how the service of virtual conciliation is under the framework of this strategy. This evaluation is carried out considering Shahkooh's 5-phase model for e-government. As a result, it is evident that in the context of conciliation, maturity does not reach the necessary level in the model so that it can be considered as virtual conciliation; therefore, it is necessary to define strategies to maximize the potential of IT in this context.
Keywords: Alternative dispute resolution, e-government, evaluation of maturity, Shahkooh model, virtual conciliation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9494718 An Experimental Investigation of Bond Properties of Reinforcements Embedded in Geopolymer Concrete
Authors: Jee-Sang Kim, Jong Ho Park
Abstract:
Geopolymer concretes are new class of construction materials that have emerged as an alternative to Ordinary Portland cement concrete. Considerable researches have been carried out on material development of geopolymer concrete; however, a few studies have been reported on the structural use of them. This paper presents the bond behaviors of reinforcement embedded in fly ash based geopolymer concrete. The development lengths of reinforcement for various compressive strengths of concrete, 20, 30 and 40 MPa, and reinforcement diameters, 10, 16 and 25 mm, are investigated. Total 27 specimens were manufactured and pull-out test according to EN 10080 was applied to measure bond strength and slips between concrete and reinforcements. The average bond strengths decreased from 23.06MPa to 17.26 MPa, as the diameters of reinforcements increased from 10mm to 25mm. The compressive strength levels of geopolymer concrete showed no significant influence on bond strengths in this study. Also, the bond-slip relations between geopolymer concrete and reinforcement are derived using non-linear regression analysis for various experimental conditions.
Keywords: Bond-slip relation, bond strength, geopolymer concrete, pull-out test.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 34474717 Comparative Analysis of the Public Funding for Greek Universities: An Ordinal DEA/MCDM Approach
Authors: Yiannis Smirlis, Dimitris K. Despotis
Abstract:
This study performs a comparative analysis of the 21 Greek Universities in terms of their public funding, awarded for covering their operating expenditure. First it introduces a DEA/MCDM model that allocates the fund into four expenditure factors in the most favorable way for each university. Then, it presents a common, consensual assessment model to reallocate the amounts, remaining in the same level of total public budget. From the analysis it derives that a number of universities cannot justify the public funding in terms of their size and operational workload. For them, the sufficient reduction of their public funding amount is estimated as a future target. Due to the lack of precise data for a number of expenditure criteria, the analysis is based on a mixed crisp-ordinal data set.Keywords: Data envelopment analysis, Greek universities, operating expenditures, ordinal data.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17674716 Efficiency Based Model for Solar Urban Planning
Authors: Amado, M. P., Amado, A., Poggi, F., Correia de Freitas, J.
Abstract:
Today is widely understood that global energy consumption patterns are directly related to the urban expansion and development process. This expansion is based on the natural growth of human activities and has left most urban areas totally dependent on fossil fuel derived external energy inputs. This status-quo of production, transportation, storage and consumption of energy has become inefficient and is set to become even more so when the continuous increases in energy demand are factored in. The territorial management of land use and related activities is a central component in the search for more efficient models of energy use, models that can meet current and future regional, national and European goals.
In this paper a methodology is developed and discussed with the aim of improving energy efficiency at the municipal level. The development of this methodology is based on the monitoring of energy consumption and its use patterns resulting from the natural dynamism of human activities in the territory and can be utilized to assess sustainability at the local scale. A set of parameters and indicators are defined with the objective of constructing a systemic model based on the optimization, adaptation and innovation of the current energy framework and the associated energy consumption patterns. The use of the model will enable local governments to strike the necessary balance between human activities and economic development and the local and global environment while safeguarding fairness in the energy sector.
Keywords: Solar urban planning, solar smart city, urban development, energy efficiency.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19634715 Simulation of Sloshing-Shear Mixed Shallow Water Waves (II) Numerical Solutions
Authors: Weihao Chung, Iau-Teh Wang, Yu-Hsi Hu
Abstract:
This is the second part of the paper. It, aside from the core subroutine test reported previously, focuses on the simulation of turbulence governed by the full STF Navier-Stokes equations on a large scale. Law of the wall is found plausible in this study as a model of the boundary layer dynamics. Model validations proceed to include velocity profiles of a stationary turbulent Couette flow, pure sloshing flow simulations, and the identification of water-surface inclination due to fluid accelerations. Errors resulting from the irrotational and hydrostatic assumptions are explored when studying a wind-driven water circulation with no shakings. Illustrative examples show that this numerical strategy works for the simulation of sloshing-shear mixed flow in a 3-D rigid rectangular base tank.Keywords: potential flow theory, sloshing flow, space-timefiltering, order of accuracy.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1492