Search results for: classical microbiological analysis
27688 Conceptualising an Open Living Museum beyond Musealization in the Context of a Historic City: Study of Bhaktapur World Heritage Site, Nepal
Authors: Shyam Sunder Kawan
Abstract:
Museums are enclosed buildings encompassing and displaying creative artworks, artefacts, and discoveries for people’s knowledge and observation. In the context of Nepal, museums and exhibition areas are either adaptive to small gallery spaces in residences or ‘neo-classical palatial complexes’ that evolved during the 19th century. This study accepts the sparse occurrence of a diverse range of artworks and expressions in the country's complex cultural manifestations within vivid ethnic groups. This study explores the immense potential of one such prevalence beyond the delimitation of physical boundaries. Taking Bhaktapur World Heritage Site as a case, the study perpetuates its investigation into real-time life activities that this city and its cultural landscapes ensemble. Seeking the ‘musealization’ as an urban process to induce museums into the city precinct, this study anticipates art space into urban spaces to offer a limitless experience for this contemporary world. Unveiling art as an experiential component, this study aims to conceptualize a living heritage as an infinite resource for museum interpretation beyond just educational institute purposes.Keywords: living museum, site museum, museulization, contemporary arts, cultural heritage, historic cities
Procedia PDF Downloads 10327687 Robust Quantum Image Encryption Algorithm Leveraging 3D-BNM Chaotic Maps and Controlled Qubit-Level Operations
Authors: Vivek Verma, Sanjeev Kumar
Abstract:
This study presents a novel quantum image encryption algorithm, using a 3D chaotic map and controlled qubit-level scrambling operations. The newly proposed 3D-BNM chaotic map effectively reduces the degradation of chaotic dynamics resulting from the finite word length effect. It facilitates the generation of highly unpredictable random sequences and enhances chaotic performance. The system’s efficacy is additionally enhanced by the inclusion of a SHA-256 hash function. Initially, classical plain images are converted into their quantum equivalents using the Novel Enhanced Quantum Representation (NEQR) model. The Generalized Quantum Arnold Transformation (GQAT) is then applied to disrupt the coordinate information of the quantum image. Subsequently, to diffuse the pixel values of the scrambled image, XOR operations are performed using pseudorandom sequences generated by the 3D-BNM chaotic map. Furthermore, to enhance the randomness and reduce the correlation among the pixels in the resulting cipher image, a controlled qubit-level scrambling operation is employed. The encryption process utilizes fundamental quantum gates such as C-NOT and CCNOT. Both theoretical and numerical simulations validate the effectiveness of the proposed algorithm against various statistical and differential attacks. Moreover, the proposed encryption algorithm operates with low computational complexity.Keywords: 3D Chaotic map, SHA-256, quantum image encryption, Qubit level scrambling, NEQR
Procedia PDF Downloads 1027686 Performance Evaluation of an Ontology-Based Arabic Sentiment Analysis
Authors: Salima Behdenna, Fatiha Barigou, Ghalem Belalem
Abstract:
Due to the quick increase in the volume of Arabic opinions posted on various social media, Arabic sentiment analysis has become one of the most important areas of research. Compared to English, there is very little works on Arabic sentiment analysis, in particular aspect-based sentiment analysis (ABSA). In ABSA, aspect extraction is the most important task. In this paper, we propose a semantic aspect-based sentiment analysis approach for standard Arabic reviews to extract explicit aspect terms and identify the polarity of the extracted aspects. The proposed approach was evaluated using HAAD datasets. Experiments showed that the proposed approach achieved a good level of performance compared with baseline results. The F-measure was improved by 19% for the aspect term extraction tasks and 55% aspect term polarity task.Keywords: sentiment analysis, opinion mining, Arabic, aspect level, opinion, polarity
Procedia PDF Downloads 16327685 Comparative Analysis of Photovoltaic Systems
Authors: Irtaza M. Syed, Kaameran Raahemifar
Abstract:
This paper presents comparative analysis of photovoltaic systems (PVS) and proposes practical techniques to improve operational efficiency of the PVS. The best engineering and construction practices for PVS are identified and field oriented recommendation are made. Comparative analysis of central and string inverter based, as well as 600 and 1000 VDC PVS are performed. In addition, direct current (DC) and alternating current (AC) photovoltaic (PV) module based systems are compared. Comparison shows that 1000 V DC String Inverters based PVS is the best choice.Keywords: photovoltaic module, photovoltaic systems, operational efficiency improvement, comparative analysis
Procedia PDF Downloads 48527684 Comparative Analysis of the Impact of Urbanization on Land Surface Temperature in the United Arab Emirates
Authors: A. O. Abulibdeh
Abstract:
The aim of this study is to investigate and compare the changes in the Land Surface Temperature (LST) as a function of urbanization, particularly land use/land cover changes, in three cities in the UAE, mainly Abu Dhabi, Dubai, and Al Ain. The scale of this assessment will be at the macro- and micro-levels. At the macro-level, a comparative assessment will take place to compare between the four cities in the UAE. At the micro-level, the study will compare between the effects of different land use/land cover on the LST. This will provide a clear and quantitative city-specific information related to the relationship between urbanization and local spatial intra-urban LST variation in three cities in the UAE. The main objectives of this study are 1) to investigate the development of LST on the macro- and micro-level between and in three cities in the UAE over two decades time period, 2) to examine the impact of different types of land use/land cover on the spatial distribution of LST. Because these three cities are facing harsh arid climate, it is hypothesized that (1) urbanization is affecting and connected to the spatial changes in LST; (2) different land use/land cover have different impact on the LST; and (3) changes in spatial configuration of land use and vegetation concentration over time would control urban microclimate on a city scale and control macroclimate on the country scale. This study will be carried out over a 20-year period (1996-2016) and throughout the whole year. The study will compare between two distinct periods with different thermal characteristics which are the cool/cold period from November to March and warm/hot period between April and October. The best practice research method for this topic is to use remote sensing data to target different aspects of natural and anthropogenic systems impacts. The project will follow classical remote sensing and mapping techniques to investigate the impact of urbanization, mainly changes in land use/land cover, on LST. The investigation in this study will be performed in two stages. Stage one remote sensing data will be used to investigate the impact of urbanization on LST on a macroclimate level where the LST and Urban Heat Island (UHI) will be compared in the three cities using data from the past two decades. Stage two will investigate the impact on microclimate scale by investigating the LST and UHI using a particular land use/land cover type. In both stages, an LST and urban land cover maps will be generated over the study area. The outcome of this study should represent an important contribution to recent urban climate studies, particularly in the UAE. Based on the aim and objectives of this study, the expected outcomes are as follow: i) to determine the increase or decrease of LST as a result of urbanization in these four cities, ii) to determine the effect of different land uses/land covers on increasing or decreasing the LST.Keywords: land use/land cover, global warming, land surface temperature, remote sensing
Procedia PDF Downloads 24827683 Introducing α-Oxoester (COBz) as a Protecting Group for Carbohydrates
Authors: Atul Kumar, Veeranjaneyulu Gannedi, Qazi Naveed Ahmed
Abstract:
Oligosaccharides, which are essential to all cellular organisms, play vital roles in cell recognition, signaling, and are involved in a broad range of biological processes. The chemical synthesis of carbohydrates represents a powerful tool to provide homogeneous glycans. In carbohydrate synthesis, the major concern is the orthogonal protection of hydroxyl groups that can be unmasked independently. Classical protecting groups include benzyl ethers (Bn), which are normally cleaved through hydrogenolysis or by means of metal reduction, and acetate (Ac), benzoate (Bz) or pivaloate esters, which are removed using base promoted hydrolysis. In present work a series of α-Oxoester (COBz) protected saccharides, with divergent base sensitivity profiles against benzoyl (Bz) and acetyl (Ac), were designed and KHSO₅/CH₃COCl in methanol was identified as an easy, mild, selective and efficient deprotecting reagent for their removal in the perspective of carbohydrate synthesis. Timely monitoring of later reagent was advantageous in establishing both sequential as well as simultaneous deprotecting of COBz, Bz, and Ac. The salient feature of our work is its ease to generate different acceptors using designed monosaccharides. In summary, we demonstrated α-Oxoester (COBz) as a new protecting group for carbohydrates and the application of this group for the synthesis of Glycosylphosphatidylinositol (GPI) anchor are in progress.Keywords: α-Oxoester, oligosaccharides, new protecting group, acceptor synthesis, glycosylation
Procedia PDF Downloads 15027682 The Analysis and Simulation of TRACE in the Ultimate Response Guideline for Chinshan BWR/4 Nuclear Power Plant
Authors: J. R. Wang, H. T. Lin, H. C. Chen, C. Shih, S. W. Chen, S. C. Chiang, C. C. Liu
Abstract:
In this research, TRACE model of Chinshan BWR/4 Nuclear Power Plant (NPP) has been developed for the simulation and analysis of Ultimate Response Guideline (URG). The main actions of URG are the depressurization and low pressure water injection of reactor and containment venting. This research focuses to verify the URG efficiency under Fukushima-like conditions. Trace analysis results show that the URG can keep the PCT below the criteria 1088.7 K under Fukushima-like conditions. It indicated that Chinshan NPP was safe.Keywords: BWR, trace, safety analysis, URG
Procedia PDF Downloads 62127681 Simulation Analysis of a Full-Scale Five-Story Building with Vibration Control Dampers
Authors: Naohiro Nakamura
Abstract:
Analysis methods to accurately estimate the behavior of buildings when earthquakes occur is very important for improving the seismic safety of such buildings. Recently, the use of damping devices has increased significantly and there is a particular need to appropriately evaluate the behavior of buildings with such devices during earthquakes in the design stage. At present, however, the accuracy of the analysis evaluations is not sufficient. One reason is that the accuracy of current analysis methods has not been appropriately verified because there is very limited data on the behavior of actual buildings during earthquakes. Many types of shaking table test of large structures are performed at the '3-Dimensional Full-Scale Earthquake Testing Facility' (nicknamed 'E-Defense') operated by the National Research Institute of Earth Science and Disaster Prevention (NIED). In this study, simulations using 3- dimensional analysis models were conducted on shaking table test of a 5-story steel-frame structure with dampers. The results of the analysis correspond favorably to the test results announced afterward by the committee. However, the suitability of the parameters and models used in the analysis and the influence they had on the responses remain unclear. Hence, we conducted additional analysis and studies on these models and parameters. In this paper, outlines of the test are shown and the utilized analysis model is explained. Next, the analysis results are compared with the test results. Then, the additional analyses, concerning with the hysteresis curve of the dampers and the beam-end stiffness of the frame, are investigated.Keywords: three-dimensional analysis, E-defense, full-scale experimen, vibration control damper
Procedia PDF Downloads 19027680 Applications for Additive Manufacturing Technology for Reducing the Weight of Body Parts of Gas Turbine Engines
Authors: Liubov Magerramova, Mikhail Petrov, Vladimir Isakov, Liana Shcherbinina, Suren Gukasyan, Daniil Povalyukhin, Olga Klimova-Korsmik, Darya Volosevich
Abstract:
Aircraft engines are developing along the path of increasing resource, strength, reliability, and safety. The building of gas turbine engine body parts is a complex design and technological task. Particularly complex in the design and manufacturing are the casings of the input stages of helicopter gearboxes and central drives of aircraft engines. Traditional technologies, such as precision casting or isothermal forging, are characterized by significant limitations in parts production. For parts like housing, additive technologies guarantee spatial freedom and limitless or flexible design. This article presents the results of computational and experimental studies. These investigations justify the applicability of additive technologies (AT) to reduce the weight of aircraft housing gearbox parts by up to 32%. This is possible due to geometrical optimization compared to the classical, less flexible manufacturing methods and as-casted aircraft parts with over-insured values of safety factors. Using an example of the body of the input stage of an aircraft gearbox, visualization of the layer-by-layer manufacturing of a part based on thermal deformation was demonstrated.Keywords: additive technologies, gas turbine engines, topological optimization, synthesis process
Procedia PDF Downloads 11627679 Application of Infrared Thermal Imaging, Eye Tracking and Behavioral Analysis for Deception Detection
Authors: Petra Hypšová, Martin Seitl
Abstract:
One of the challenges of forensic psychology is to detect deception during a face-to-face interview. In addition to the classical approaches of monitoring the utterance and its components, detection is also sought by observing behavioral and physiological changes that occur as a result of the increased emotional and cognitive load caused by the production of distorted information. Typical are changes in facial temperature, eye movements and their fixation, pupil dilation, emotional micro-expression, heart rate and its variability. Expanding technological capabilities have opened the space to detect these psychophysiological changes and behavioral manifestations through non-contact technologies that do not interfere with face-to-face interaction. Non-contact deception detection methodology is still in development, and there is a lack of studies that combine multiple non-contact technologies to investigate their accuracy, as well as studies that show how different types of lies produced by different interviewers affect physiological and behavioral changes. The main objective of this study is to apply a specific non-contact technology for deception detection. The next objective is to investigate scenarios in which non-contact deception detection is possible. A series of psychophysiological experiments using infrared thermal imaging, eye tracking and behavioral analysis with FaceReader 9.0 software was used to achieve our goals. In the laboratory experiment, 16 adults (12 women, 4 men) between 18 and 35 years of age (SD = 4.42) were instructed to produce alternating prepared and spontaneous truths and lies. The baseline of each proband was also measured, and its results were compared to the experimental conditions. Because the personality of the examiner (particularly gender and facial appearance) to whom the subject is lying can influence physiological and behavioral changes, the experiment included four different interviewers. The interviewer was represented by a photograph of a face that met the required parameters in terms of gender and facial appearance (i.e., interviewer likability/antipathy) to follow standardized procedures. The subject provided all information to the simulated interviewer. During follow-up analyzes, facial temperature (main ROIs: forehead, cheeks, the tip of the nose, chin, and corners of the eyes), heart rate, emotional expression, intensity and fixation of eye movements and pupil dilation were observed. The results showed that the variables studied varied with respect to the production of prepared truths and lies versus the production of spontaneous truths and lies, as well as the variability of the simulated interviewer. The results also supported the assumption of variability in physiological and behavioural values during the subject's resting state, the so-called baseline, and the production of prepared and spontaneous truths and lies. A series of psychophysiological experiments provided evidence of variability in the areas of interest in the production of truths and lies to different interviewers. The combination of technologies used also led to a comprehensive assessment of the physiological and behavioral changes associated with false and true statements. The study presented here opens the space for further research in the field of lie detection with non-contact technologies.Keywords: emotional expression decoding, eye-tracking, functional infrared thermal imaging, non-contact deception detection, psychophysiological experiment
Procedia PDF Downloads 9927678 Double Layer Security Authentication Model for Automatic Dependent Surveillance-Broadcast
Authors: Buse T. Aydin, Enver Ozdemir
Abstract:
An automatic dependent surveillance-broadcast (ADS-B) system has serious security problems. In this study, a double layer authentication scheme between the aircraft and ground station, aircraft to aircraft, ground station to ATC tower is designed to prevent any unauthorized aircrafts from introducing themselves as friends. This method can be used as a solution to the problem of authentication. The method is a combination of classical cryptographic methods and new generation physical layers. The first layer has employed the embedded key of the aircraft. The embedded key is assumed to installed during the construction of the utility. The other layer is a physical attribute (flight path, distance, etc.) between the aircraft and the ATC tower. We create a mathematical model so that two layers’ information is employed and an aircraft is authenticated as a friend or unknown according to the accuracy of the results of the model. The results of the aircraft are compared with the results of the ATC tower and if the values found by the aircraft and ATC tower match within a certain error margin, we mark the aircraft as friend. As a result, the ADS-B messages coming from this authenticated friendly aircraft will be processed. In this method, even if the embedded key is captured by the unknown aircraft, without the information of the second layer, the unknown aircraft can easily be determined. Overall, in this work, we present a reliable system by adding physical layer in the authentication process.Keywords: ADS-B, authentication, communication with physical layer security, cryptography, identification friend or foe
Procedia PDF Downloads 17927677 The Use of Geographically Weighted Regression for Deforestation Analysis: Case Study in Brazilian Cerrado
Authors: Ana Paula Camelo, Keila Sanches
Abstract:
The Geographically Weighted Regression (GWR) was proposed in geography literature to allow relationship in a regression model to vary over space. In Brazil, the agricultural exploitation of the Cerrado Biome is the main cause of deforestation. In this study, we propose a methodology using geostatistical methods to characterize the spatial dependence of deforestation in the Cerrado based on agricultural production indicators. Therefore, it was used the set of exploratory spatial data analysis tools (ESDA) and confirmatory analysis using GWR. It was made the calibration a non-spatial model, evaluation the nature of the regression curve, election of the variables by stepwise process and multicollinearity analysis. After the evaluation of the non-spatial model was processed the spatial-regression model, statistic evaluation of the intercept and verification of its effect on calibration. In an analysis of Spearman’s correlation the results between deforestation and livestock was +0.783 and with soybeans +0.405. The model presented R²=0.936 and showed a strong spatial dependence of agricultural activity of soybeans associated to maize and cotton crops. The GWR is a very effective tool presenting results closer to the reality of deforestation in the Cerrado when compared with other analysis.Keywords: deforestation, geographically weighted regression, land use, spatial analysis
Procedia PDF Downloads 36327676 Damage Detection in Beams Using Wavelet Analysis
Authors: Goutham Kumar Dogiparti, D. R. Seshu
Abstract:
In the present study, wavelet analysis was used for locating damage in simply supported and cantilever beams. Study was carried out varying different levels and locations of damage. In numerical method, ANSYS software was used for modal analysis of damaged and undamaged beams. The mode shapes obtained from numerical analysis is processed using MATLAB wavelet toolbox to locate damage. Effect of several parameters such as (damage level, location) on the natural frequencies and mode shapes were also studied. The results indicated the potential of wavelets in identifying the damage location.Keywords: damage, detection, beams, wavelets
Procedia PDF Downloads 36527675 Field Oriented Control of Electrical Motor for Efficiency Improvement of Aerial Vehicle
Authors: Francois Defay
Abstract:
Uses of Unmanned aerial vehicle (UAV) are increasing for many applicative cases. Long endurance UAVs are required for inspection or transportation in some deserted places. The global optimization of the efficiency is the aim of the works in ISAE-SUPAERO. From the propulsive part until the motor control, the global optimization can increase significantly the global efficiency. This paper deals with the global improvement of the efficiency of the electrical propulsion for the aerial vehicle. The application case of study is a small airplane of 2kg. A global modelization is presented in order to validate the electrical engine in a complete simulation from aerodynamics to battery. The classical control of the synchronous permanent drive is compared to the field-oriented control which is not yet applied for UAVs. The experimental results presented show an increase of more than 10 percent of the efficiency. A complete modelization and simulation based on Matlab/ Simulink are presented in this paper and compared to the experimental study. Finally this paper presents solutions to increase the endurance of the electrical aerial vehicle and provide models to optimize the global consumption for a specific mission. The next step is to use this model and the control to work with distributed propulsion which is the future for small distance plane.Keywords: electrical propulsion, endurance, field-oriented control, UAV
Procedia PDF Downloads 23727674 The Optimum Mel-Frequency Cepstral Coefficients (MFCCs) Contribution to Iranian Traditional Music Genre Classification by Instrumental Features
Authors: M. Abbasi Layegh, S. Haghipour, K. Athari, R. Khosravi, M. Tafkikialamdari
Abstract:
An approach to find the optimum mel-frequency cepstral coefficients (MFCCs) for the Radif of Mirzâ Ábdollâh, which is the principal emblem and the heart of Persian music, performed by most famous Iranian masters on two Iranian stringed instruments ‘Tar’ and ‘Setar’ is proposed. While investigating the variance of MFCC for each record in themusic database of 1500 gushe of the repertoire belonging to 12 modal systems (dastgâh and âvâz), we have applied the Fuzzy C-Mean clustering algorithm on each of the 12 coefficient and different combinations of those coefficients. We have applied the same experiment while increasing the number of coefficients but the clustering accuracy remained the same. Therefore, we can conclude that the first 7 MFCCs (V-7MFCC) are enough for classification of The Radif of Mirzâ Ábdollâh. Classical machine learning algorithms such as MLP neural networks, K-Nearest Neighbors (KNN), Gaussian Mixture Model (GMM), Hidden Markov Model (HMM) and Support Vector Machine (SVM) have been employed. Finally, it can be realized that SVM shows a better performance in this study.Keywords: radif of Mirzâ Ábdollâh, Gushe, mel frequency cepstral coefficients, fuzzy c-mean clustering algorithm, k-nearest neighbors (KNN), gaussian mixture model (GMM), hidden markov model (HMM), support vector machine (SVM)
Procedia PDF Downloads 44627673 Art Market in Oran: Emergence and Contraintes
Authors: Hirreche Baghdad Mohamed
Abstract:
Our research is linked to cultural policies because the initiation to taste and beauty is a matter for all cultural and educational institutions. It's done by a downstream process (programs, actions, lessons, etc.) that begins at a young age in order to inscribe aesthetic values in memories, imaginations, and practices. Preparing future art lovers probably takes a lot of time. Upstream, continuity is ensured by the "cultural industries" which make cultural products available to actors in the "art market" through professional training, production, dissemination, and sales processes. It turns out that the cultural industries borrow from the "classical" industries the same processes and logic: product, production, marketing, diffusion, profit and profits, supply and demand, the market, the creation of wealth, the entrepreneurship. Today, culture has become a product almost like the others. In the cultural industries system, we protect the rights of authors (owners) and the rights of intermediaries (entrepreneurs of culture), and we provide consumers with an accessible product that meets their needs and expectations. We aim to present an inventory and to reveal, through the speeches of the actors themselves, the processes and modes of operation and deployment of the plastic arts market by showing how it is perceived, imagined, and lived in the city of 'Oran from the 2000s to the present day. However, it is possible to clarify this field of research by looking at previous periods; and even to make comparisons with other regions in Algeria in order to give meaning to practices in various contexts.Keywords: Oran, Algeria, fine art, art market
Procedia PDF Downloads 12227672 Different Sampling Schemes for Semi-Parametric Frailty Model
Authors: Nursel Koyuncu, Nihal Ata Tutkun
Abstract:
Frailty model is a survival model that takes into account the unobserved heterogeneity for exploring the relationship between the survival of an individual and several covariates. In the recent years, proposed survival models become more complex and this feature causes convergence problems especially in large data sets. Therefore selection of sample from these big data sets is very important for estimation of parameters. In sampling literature, some authors have defined new sampling schemes to predict the parameters correctly. For this aim, we try to see the effect of sampling design in semi-parametric frailty model. We conducted a simulation study in R programme to estimate the parameters of semi-parametric frailty model for different sample sizes, censoring rates under classical simple random sampling and ranked set sampling schemes. In the simulation study, we used data set recording 17260 male Civil Servants aged 40–64 years with complete 10-year follow-up as population. Time to death from coronary heart disease is treated as a survival-time and age, systolic blood pressure are used as covariates. We select the 1000 samples from population using different sampling schemes and estimate the parameters. From the simulation study, we concluded that ranked set sampling design performs better than simple random sampling for each scenario.Keywords: frailty model, ranked set sampling, efficiency, simple random sampling
Procedia PDF Downloads 21127671 Application Potential of Selected Tools in Context of Critical Infrastructure Protection and Risk Analysis
Authors: Hromada Martin
Abstract:
Risk analysis is considered as a fundamental aspect relevant for ensuring the level of critical infrastructure protection, where the critical infrastructure is seen as system, asset or its part which is important for maintaining the vital societal functions. Article actually discusses and analyzes the potential application of selected tools of information support for the implementation and within the framework of risk analysis and critical infrastructure protection. Use of the information in relation to their risk analysis can be viewed as a form of simplifying the analytical process. It is clear that these instruments (information support) for these purposes are countless, so they were selected representatives who have already been applied in the selected area of critical infrastructure, or they can be used. All presented fact were the basis for critical infrastructure resilience evaluation methodology development.Keywords: critical infrastructure, protection, resilience, risk analysis
Procedia PDF Downloads 63827670 Use of Gaussian-Euclidean Hybrid Function Based Artificial Immune System for Breast Cancer Diagnosis
Authors: Cuneyt Yucelbas, Seral Ozsen, Sule Yucelbas, Gulay Tezel
Abstract:
Due to the fact that there exist only a small number of complex systems in artificial immune system (AIS) that work out nonlinear problems, nonlinear AIS approaches, among the well-known solution techniques, need to be developed. Gaussian function is usually used as similarity estimation in classification problems and pattern recognition. In this study, diagnosis of breast cancer, the second type of the most widespread cancer in women, was performed with different distance calculation functions that euclidean, gaussian and gaussian-euclidean hybrid function in the clonal selection model of classical AIS on Wisconsin Breast Cancer Dataset (WBCD), which was taken from the University of California, Irvine Machine-Learning Repository. We used 3-fold cross validation method to train and test the dataset. According to the results, the maximum test classification accuracy was reported as 97.35% by using of gaussian-euclidean hybrid function for fold-3. Also, mean of test classification accuracies for all of functions were obtained as 94.78%, 94.45% and 95.31% with use of euclidean, gaussian and gaussian-euclidean, respectively. With these results, gaussian-euclidean hybrid function seems to be a potential distance calculation method, and it may be considered as an alternative distance calculation method for hard nonlinear classification problems.Keywords: artificial immune system, breast cancer diagnosis, Euclidean function, Gaussian function
Procedia PDF Downloads 43527669 Sentence vs. Keyword Content Analysis in Intellectual Capital Disclosures Study
Authors: Martin Surya Mulyadi, Yunita Anwar, Rosinta Ria Panggabean
Abstract:
Major transformations in economic activity from an agricultural economy to knowledge economy have led to an increasing focus on intellectual capital (IC) that has been characterized by continuous innovation, the spread of digital and communication technologies, intangible and human factors. IC is defined as the possession of knowledge and experience, professional knowledge and skill, proper relationships and technological capacities, which when applied will give organizations a competitive advantage. All of IC report/disclosure could be captured from the corporate annual report as it is a communication device that allows a corporation to connect with various external and internal stakeholders. This study was conducted using sentence-content analysis of IC disclosure in the annual report. This research aims to analyze whether the keyword-content analysis is reliable research methodology for IC disclosure related research.Keywords: intellectual capital, intellectual capital disclosure, content analysis, annual report, sentence analysis, keyword analysis
Procedia PDF Downloads 36727668 Market Solvency Capital Requirement Minimization: How Non-linear Solvers Provide Portfolios Complying with Solvency II Regulation
Authors: Abraham Castellanos, Christophe Durville, Sophie Echenim
Abstract:
In this article, a portfolio optimization problem is performed in a Solvency II context: it illustrates how advanced optimization techniques can help to tackle complex operational pain points around the monitoring, control, and stability of Solvency Capital Requirement (SCR). The market SCR of a portfolio is calculated as a combination of SCR sub-modules. These sub-modules are the results of stress-tests on interest rate, equity, property, credit and FX factors, as well as concentration on counter-parties. The market SCR is non convex and non differentiable, which does not make it a natural optimization criteria candidate. In the SCR formulation, correlations between sub-modules are fixed, whereas risk-driven portfolio allocation is usually driven by the dynamics of the actual correlations. Implementing a portfolio construction approach that is efficient on both a regulatory and economic standpoint is not straightforward. Moreover, the challenge for insurance portfolio managers is not only to achieve a minimal SCR to reduce non-invested capital but also to ensure stability of the SCR. Some optimizations have already been performed in the literature, simplifying the standard formula into a quadratic function. But to our knowledge, it is the first time that the standard formula of the market SCR is used in an optimization problem. Two solvers are combined: a bundle algorithm for convex non- differentiable problems, and a BFGS (Broyden-Fletcher-Goldfarb- Shanno)-SQP (Sequential Quadratic Programming) algorithm, to cope with non-convex cases. A market SCR minimization is then performed with historical data. This approach results in significant reduction of the capital requirement, compared to a classical Markowitz approach based on the historical volatility. A comparative analysis of different optimization models (equi-risk-contribution portfolio, minimizing volatility portfolio and minimizing value-at-risk portfolio) is performed and the impact of these strategies on risk measures including market SCR and its sub-modules is evaluated. A lack of diversification of market SCR is observed, specially for equities. This was expected since the market SCR strongly penalizes this type of financial instrument. It was shown that this direct effect of the regulation can be attenuated by implementing constraints in the optimization process or minimizing the market SCR together with the historical volatility, proving the interest of having a portfolio construction approach that can incorporate such features. The present results are further explained by the Market SCR modelling.Keywords: financial risk, numerical optimization, portfolio management, solvency capital requirement
Procedia PDF Downloads 11727667 Vulnerability Assessment of Reinforced Concrete Frames Based on Inelastic Spectral Displacement
Authors: Chao Xu
Abstract:
Selecting ground motion intensity measures reasonably is one of the very important issues to affect the input ground motions selecting and the reliability of vulnerability analysis results. In this paper, inelastic spectral displacement is used as an alternative intensity measure to characterize the ground motion damage potential. The inelastic spectral displacement is calculated based modal pushover analysis and inelastic spectral displacement based incremental dynamic analysis is developed. Probability seismic demand analysis of a six story and an eleven story RC frame are carried out through cloud analysis and advanced incremental dynamic analysis. The sufficiency and efficiency of inelastic spectral displacement are investigated by means of regression and residual analysis, and compared with elastic spectral displacement. Vulnerability curves are developed based on inelastic spectral displacement. The study shows that inelastic spectral displacement reflects the impact of different frequency components with periods larger than fundamental period on inelastic structural response. The damage potential of ground motion on structures with fundamental period prolonging caused by structural soften can be caught by inelastic spectral displacement. To be compared with elastic spectral displacement, inelastic spectral displacement is a more sufficient and efficient intensity measure, which reduces the uncertainty of vulnerability analysis and the impact of input ground motion selection on vulnerability analysis result.Keywords: vulnerability, probability seismic demand analysis, ground motion intensity measure, sufficiency, efficiency, inelastic time history analysis
Procedia PDF Downloads 35427666 Requirement Analysis for Emergency Management Software
Authors: Tomáš Ludík, Jiří Barta, Sabina Chytilová, Josef Navrátil
Abstract:
Emergency management is a discipline of dealing with and avoiding risks. Appropriate emergency management software allows better management of these risks and has a direct influence on reducing potential negative impacts. Although there are several emergency management software products in the Czech Republic, they cover user requirements from the emergency management field only partially. Therefore, the paper focuses on the issues of requirement analysis within development of emergency management software. Analysis of the current state describes the basic features and properties of user requirements for software development as well as basic methods and approaches for gathering these requirements. Then, the paper presents more specific mechanisms for requirement analysis based on chosen software development approach: structured, object-oriented or agile. Based on these experiences it is designed new methodology for requirement analysis. Methodology describes how to map user requirements comprehensively in the field of emergency management and thus reduce misunderstanding between software analyst and emergency manager. Proposed methodology was consulted with department of fire brigade and also has been applied in the requirements analysis for their current emergency management software. The proposed methodology has general character and can be used also in other specific areas during requirement analysis.Keywords: emergency software, methodology, requirement analysis, stakeholders, use case diagram, user stories
Procedia PDF Downloads 54027665 Numerical Pricing of Financial Options under Irrational Exercise Times and Regime-Switching Models
Authors: Mohammad Saber Rohi, Saghar Heidari
Abstract:
In this paper, we studied the pricing problem of American options under a regime-switching model with the possibility of a non-optimal exercise policy (early or late exercise time) which is called an irrational strategy. For this, we consider a Markovmodulated model for the dynamic of the underlying asset as an alternative model to the classical Balck-Scholes-Merton model (BSM) and an intensity-based model for the irrational strategy, to provide more realistic results for American option prices under the irrational behavior in real financial markets. Applying a partial differential equation (PDE) approach, the pricing problem of American options under regime-switching models can be formulated as coupled PDEs. To solve the resulting systems of PDEs in this model, we apply a finite element method as the numerical solving procedure to the resulting variational inequality. Under some appropriate assumptions, we establish the stability of the method and compare its accuracy to some recent works to illustrate the suitability of the proposed model and the accuracy of the applied numerical method for the pricing problem of American options under the regime-switching model with irrational behaviors.Keywords: irrational exercise strategy, rationality parameter, regime-switching model, American option, finite element method, variational inequality
Procedia PDF Downloads 7327664 A Machine Learning Approach for Anomaly Detection in Environmental IoT-Driven Wastewater Purification Systems
Authors: Giovanni Cicceri, Roberta Maisano, Nathalie Morey, Salvatore Distefano
Abstract:
The main goal of this paper is to present a solution for a water purification system based on an Environmental Internet of Things (EIoT) platform to monitor and control water quality and machine learning (ML) models to support decision making and speed up the processes of purification of water. A real case study has been implemented by deploying an EIoT platform and a network of devices, called Gramb meters and belonging to the Gramb project, on wastewater purification systems located in Calabria, south of Italy. The data thus collected are used to control the wastewater quality, detect anomalies and predict the behaviour of the purification system. To this extent, three different statistical and machine learning models have been adopted and thus compared: Autoregressive Integrated Moving Average (ARIMA), Long Short Term Memory (LSTM) autoencoder, and Facebook Prophet (FP). The results demonstrated that the ML solution (LSTM) out-perform classical statistical approaches (ARIMA, FP), in terms of both accuracy, efficiency and effectiveness in monitoring and controlling the wastewater purification processes.Keywords: environmental internet of things, EIoT, machine learning, anomaly detection, environment monitoring
Procedia PDF Downloads 15127663 An Architectural Model for APT Detection
Authors: Nam-Uk Kim, Sung-Hwan Kim, Tai-Myoung Chung
Abstract:
Typical security management systems are not suitable for detecting APT attack, because they cannot draw the big picture from trivial events of security solutions. Although SIEM solutions have security analysis engine for that, their security analysis mechanisms need to be verified in academic field. Although this paper proposes merely an architectural model for APT detection, we will keep studying on correlation analysis mechanism in the future.Keywords: advanced persistent threat, anomaly detection, data mining
Procedia PDF Downloads 52827662 Critical Analysis of Heat Exchanger Cycle for its Maintainability Using Failure Modes and Effect Analysis and Pareto Analysis
Authors: Sayali Vyas, Atharva Desai, Shreyas Badave, Apurv Kulkarni, B. Rajiv
Abstract:
The Failure Modes and Effect Analysis (FMEA) is an efficient evaluation technique to identify potential failures in products, processes, and services. FMEA is designed to identify and prioritize failure modes. It proves to be a useful method for identifying and correcting possible failures at its earliest possible level so that one can avoid consequences of poor performance. In this paper, FMEA tool is used in detection of failures of various components of heat exchanger cycle and to identify critical failures of the components which may hamper the system’s performance. Further, a detailed Pareto analysis is done to find out the most critical components of the cycle, the causes of its failures, and possible recommended actions. This paper can be used as a checklist which will help in maintainability of the system.Keywords: FMEA, heat exchanger cycle, Ishikawa diagram, pareto analysis, RPN (Risk Priority Number)
Procedia PDF Downloads 40227661 Comparison between Pushover Analysis Techniques and Validation of the Simplified Modal Pushover Analysis
Authors: N. F. Hanna, A. M. Haridy
Abstract:
One of the main drawbacks of the Modal Pushover Analysis (MPA) is the need to perform nonlinear time-history analysis, which complicates the analysis method and time. A simplified version of the MPA has been proposed based on the concept of the inelastic deformation ratio. Furthermore, the effect of the higher modes of vibration is considered by assuming linearly-elastic responses, which enables the use of standard elastic response spectrum analysis. In this thesis, the simplified MPA (SMPA) method is applied to determine the target global drift and the inter-story drifts of steel frame building. The effect of the higher vibration modes is considered within the framework of the SMPA. A comprehensive survey about the inelastic deformation ratio is presented. After that, a suitable expression from literature is selected for the inelastic deformation ratio and then implemented in the SMPA. The estimated seismic demands using the SMPA, such as target drift, base shear, and the inter-story drifts, are compared with the seismic responses determined by applying the standard MPA. The accuracy of the estimated seismic demands is validated by comparing with the results obtained by the nonlinear time-history analysis using real earthquake records.Keywords: modal analysis, pushover analysis, seismic performance, target displacement
Procedia PDF Downloads 36127660 Failure Analysis and Fatigue Life Estimation of a Shaft of a Rotary Draw Bending Machine
Authors: B. Engel, Sara Salman Hassan Al-Maeeni
Abstract:
Human consumption of the Earth's resources increases the need for a sustainable development as an important ecological, social, and economic theme. Re-engineering of machine tools, in terms of design and failure analysis, is defined as steps performed on an obsolete machine to return it to a new machine with the warranty that matches the customer requirement. To understand the future fatigue behavior of the used machine components, it is important to investigate the possible causes of machine parts failure through design, surface, and material inspections. In this study, the failure modes of the shaft of the rotary draw bending machine are inspected. Furthermore, stress and deflection analysis of the shaft subjected to combined torsion and bending loads are carried out by an analytical method and compared with a finite element analysis method. The theoretical fatigue strength, correction factors, and fatigue life sustained by the shaft before damaged are estimated by creating a stress-cycle (S-N) diagram. In conclusion, it is seen that the shaft can work in the second life, but it needs some surface treatments to increase the reliability and fatigue life.Keywords: failure analysis, fatigue life, FEM analysis, shaft, stress analysis
Procedia PDF Downloads 30127659 A Proposed Approach for Emotion Lexicon Enrichment
Authors: Amr Mansour Mohsen, Hesham Ahmed Hassan, Amira M. Idrees
Abstract:
Document Analysis is an important research field that aims to gather the information by analyzing the data in documents. As one of the important targets for many fields is to understand what people actually want, sentimental analysis field has been one of the vital fields that are tightly related to the document analysis. This research focuses on analyzing text documents to classify each document according to its opinion. The aim of this research is to detect the emotions from text documents based on enriching the lexicon with adapting their content based on semantic patterns extraction. The proposed approach has been presented, and different experiments are applied by different perspectives to reveal the positive impact of the proposed approach on the classification results.Keywords: document analysis, sentimental analysis, emotion detection, WEKA tool, NRC lexicon
Procedia PDF Downloads 442