Search results for: Jaccard’s similarity measure.
1157 Undecimated Wavelet Transform Based Contrast Enhancement
Authors: Numan Unaldi, Samil Temel, Süleyman Demirci
Abstract:
A novel undecimated wavelet transform based contrast enhancement algorithmis proposed to for both gray scale andcolor images. Contrast enhancement is realized by tuning the magnitude of approximation coefficients at each level with respect to the approximation coefficients of one higher level during the inverse transform phase in a center/surround enhancement sense.The performance of the proposed algorithm is evaluated using a statistical visual contrast measure (VCM). Experimental results on the proposed algorithm show improvement in terms of the VCM.
Keywords: Image enhancement, local contrast enhancement, visual contrast measure.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 27461156 Isolation and Identification Fibrinolytic Protease Endophytic Fungi from Hibiscus Leaves in Shah Alam
Authors: Mohd Sidek Ahmad, Zainon Mohd Noor, Zaidah Zainal Ariffin
Abstract:
Fibrin degradation is an important part in prevention or treatment of intravascular thrombosis and cardiovascular diseases. Plasmin like fibrinolytic enzymes has given new hope to patient with cardiovascular diseases by treating fibrin aggregation related diseases with traditional plasminogen activator which have many side effects. Various researches involving wide range of sources for production of fibrinolytic proteases, from bacteria, fungi, insects and fermented foods. But few have looked into endophytic fungi as a potential source. Sixteen (16) endophytic fungi were isolated from Hibiscus sp. leaves from six different locations in Shah Alam, Selangor. Only two endophytic fungi, FH3 and S13 showed positive fibrinolytic protease activities. FH3 produced 5.78cm and S13 produced 4.48cm on Skim Milk Agar after 4 days of incubation at 27°C. Fibrinolytic activity was observed; 3.87cm and 1.82cm diameter clear zone on fibrin plate of FH3 and S13 respectively. 18srRNA was done for identification of the isolated fungi with positive fibrinolytic protease. S13 had the highest similarity (100%) to that of Penicillium citrinum strain TG2 and FH3 had the highest similarity (99%) to that of Fusarium sp. FW2PhC1, Fusarium sp. 13002, Fusarium sp. 08006, Fusarium equiseti strain Salicorn 8 and Fungal sp. FCASAn-2. Media composition variation showed the effects of carbon nitrogen on protein concentration, where the decrement of 50% of media composition caused drastic decrease in protease of FH3 from 1.081 to 0.056 and also S13 from 2.946 to 0.198.
Keywords: Isolation, identification, fibrinolytic protease, endophytic fungi, Hibiscus leaves.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 32081155 Cumulative Learning based on Dynamic Clustering of Hierarchical Production Rules(HPRs)
Authors: Kamal K.Bharadwaj, Rekha Kandwal
Abstract:
An important structuring mechanism for knowledge bases is building clusters based on the content of their knowledge objects. The objects are clustered based on the principle of maximizing the intraclass similarity and minimizing the interclass similarity. Clustering can also facilitate taxonomy formation, that is, the organization of observations into a hierarchy of classes that group similar events together. Hierarchical representation allows us to easily manage the complexity of knowledge, to view the knowledge at different levels of details, and to focus our attention on the interesting aspects only. One of such efficient and easy to understand systems is Hierarchical Production rule (HPRs) system. A HPR, a standard production rule augmented with generality and specificity information, is of the following form Decision If < condition> Generality
Keywords: Cumulative learning, clustering, data mining, hierarchical production rules.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14381154 CT Medical Images Denoising Based on New Wavelet Thresholding Compared with Curvelet and Contourlet
Authors: Amir Moslemi, Amir Movafeghi, Shahab Moradi
Abstract:
One of the most important challenging factors in medical images is nominated as noise. Image denoising refers to the improvement of a digital medical image that has been infected by Additive White Gaussian Noise (AWGN). The digital medical image or video can be affected by different types of noises. They are impulse noise, Poisson noise and AWGN. Computed tomography (CT) images are subjects to low quality due to the noise. Quality of CT images is dependent on absorbed dose to patients directly in such a way that increase in absorbed radiation, consequently absorbed dose to patients (ADP), enhances the CT images quality. In this manner, noise reduction techniques on purpose of images quality enhancement exposing no excess radiation to patients is one the challenging problems for CT images processing. In this work, noise reduction in CT images was performed using two different directional 2 dimensional (2D) transformations; i.e., Curvelet and Contourlet and Discrete Wavelet Transform (DWT) thresholding methods of BayesShrink and AdaptShrink, compared to each other and we proposed a new threshold in wavelet domain for not only noise reduction but also edge retaining, consequently the proposed method retains the modified coefficients significantly that result good visual quality. Data evaluations were accomplished by using two criterions; namely, peak signal to noise ratio (PSNR) and Structure similarity (Ssim).Keywords: Computed Tomography (CT), noise reduction, curve-let, contour-let, Signal to Noise Peak-Peak Ratio (PSNR), Structure Similarity (Ssim), Absorbed Dose to Patient (ADP).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 29201153 Measurement of Intellectual Capital in an Algerian Company
Authors: S. Brahmi, S. Aitouche, M. D. Mouss
Abstract:
Every modern company should measure the value of its intellectual capital and to report to complement the traditional annual balance sheets. The purpose of this work is to measure the intellectual capital in an Algerian company (or production system) using the Weightless Wealth Tool Kit (WWTK). The results of the measurement of intellectual capital are supplemented by traditional financial ratios. The measurement was applied to the National Company of Wells Services (ENSP) in Hassi Messaoud city, in the south of Algeria. We calculated the intellectual capital (intangible resources) of the ENSP to help the organization to better capitalize on its potential of workers and their know-how. The intangible value of the ENSP is evaluated at 16,936,173,345 DA in 2015.
Keywords: Financial valuation, intangible capital, intellectual capital, intellectual capital measurement.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 11521152 Efficient Frontier - Comparing Different Volatility Estimators
Authors: Tea Poklepović, Zdravka Aljinović, Mario Matković
Abstract:
Modern Portfolio Theory (MPT) according to Markowitz states that investors form mean-variance efficient portfolios which maximizes their utility. Markowitz proposed the standard deviation as a simple measure for portfolio risk and the lower semi-variance as the only risk measure of interest to rational investors. This paper uses a third volatility estimator based on intraday data and compares three efficient frontiers on the Croatian Stock Market. The results show that range-based volatility estimator outperforms both mean-variance and lower semi-variance model.
Keywords: Variance, lower semi-variance, range-based volatility.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 25781151 Similarity Measures and Weighted Fuzzy C-Mean Clustering Algorithm
Authors: Bainian Li, Kongsheng Zhang, Jian Xu
Abstract:
In this paper we study the fuzzy c-mean clustering algorithm combined with principal components method. Demonstratively analysis indicate that the new clustering method is well rather than some clustering algorithms. We also consider the validity of clustering method.
Keywords: FCM algorithm, Principal Components Analysis, Clustervalidity
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17241150 Compromise Ratio Method for Decision Making under Fuzzy Environment using Fuzzy Distance Measure
Authors: Debashree Guha, Debjani Chakraborty
Abstract:
The aim of this paper is to adopt a compromise ratio (CR) methodology for fuzzy multi-attribute single-expert decision making proble. In this paper, the rating of each alternative has been described by linguistic terms, which can be expressed as triangular fuzzy numbers. The compromise ratio method for fuzzy multi-attribute single expert decision making has been considered here by taking the ranking index based on the concept that the chosen alternative should be as close as possible to the ideal solution and as far away as possible from the negative-ideal solution simultaneously. From logical point of view, the distance between two triangular fuzzy numbers also is a fuzzy number, not a crisp value. Therefore a fuzzy distance measure, which is itself a fuzzy number, has been used here to calculate the difference between two triangular fuzzy numbers. Now in this paper, with the help of this fuzzy distance measure, it has been shown that the compromise ratio is a fuzzy number and this eases the problem of the decision maker to take the decision. The computation principle and the procedure of the compromise ratio method have been described in detail in this paper. A comparative analysis of the compromise ratio method previously proposed [1] and the newly adopted method have been illustrated with two numerical examples.
Keywords: Compromise ratio method, Fuzzy multi-attributesingle-expert decision making, Fuzzy number, Linguistic variable
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14111149 Some Solid Transportation Models with Crisp and Rough Costs
Authors: Pradip Kundu, Samarjit Kar, Manoranjan Maiti
Abstract:
In this paper, some practical solid transportation models are formulated considering per trip capacity of each type of conveyances with crisp and rough unit transportation costs. This is applicable for the system in which full vehicles, e.g. trucks, rail coaches are to be booked for transportation of products so that transportation cost is determined on the full of the conveyances. The models with unit transportation costs as rough variables are transformed into deterministic forms using rough chance constrained programming with the help of trust measure. Numerical examples are provided to illustrate the proposed models in crisp environment as well as with unit transportation costs as rough variables.
Keywords: Solid transportation problem, Rough set, Rough variable, Trust measure.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 26221148 A Note on Penalized Power-Divergence Test Statistics
Authors: Aylin Alin
Abstract:
In this paper, penalized power-divergence test statistics have been defined and their exact size properties to test a nested sequence of log-linear models have been compared with ordinary power-divergence test statistics for various penalization, λ and main effect values. Since the ordinary and penalized power-divergence test statistics have the same asymptotic distribution, comparisons have been only made for small and moderate samples. Three-way contingency tables distributed according to a multinomial distribution have been considered. Simulation results reveal that penalized power-divergence test statistics perform much better than their ordinary counterparts.
Keywords: Contingency table, Log-linear models, Penalization, Power-divergence measure, Penalized power-divergence measure.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13161147 Assessment of Multiscale Information for Short Physiological Time Series
Authors: Young-Seok Choi
Abstract:
This paper presents a multiscale information measure of Electroencephalogram (EEG) for analysis with a short data length. A multiscale extension of permutation entropy (MPE) is capable of fully reflecting the dynamical characteristics of EEG across different temporal scales. However, MPE yields an imprecise estimation due to coarse-grained procedure at large scales. We present an improved MPE measure to estimate entropy more accurately with a short time series. By computing entropies of all coarse-grained time series and averaging those at each scale, it leads to the modified MPE (MMPE) which provides an enhanced accuracy as compared to MPE. Simulation and experimental studies confirmed that MMPE has proved its capability over MPE in terms of accuracy.Keywords: Multiscale entropy, permutation entropy, EEG, seizure.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14781146 Creating the Color Panoramic View using Medley of Grayscale and Color Partial Images
Authors: Dr. H. B. Kekre, Sudeep D. Thepade
Abstract:
Panoramic view generation has always offered novel and distinct challenges in the field of image processing. Panoramic view generation is nothing but construction of bigger view mosaic image from set of partial images of the desired view. The paper presents a solution to one of the problems of image seascape formation where some of the partial images are color and others are grayscale. The simplest solution could be to convert all image parts into grayscale images and fusing them to get grayscale image panorama. But in the multihued world, obtaining the colored seascape will always be preferred. This could be achieved by picking colors from the color parts and squirting them in grayscale parts of the seascape. So firstly the grayscale image parts should be colored with help of color image parts and then these parts should be fused to construct the seascape image. The problem of coloring grayscale images has no exact solution. In the proposed technique of panoramic view generation, the job of transferring color traits from reference color image to grayscale image is done by palette based method. In this technique, the color palette is prepared using pixel windows of some degrees taken from color image parts. Then the grayscale image part is divided into pixel windows with same degrees. For every window of grayscale image part the palette is searched and equivalent color values are found, which could be used to color grayscale window. For palette preparation we have used RGB color space and Kekre-s LUV color space. Kekre-s LUV color space gives better quality of coloring. The searching time through color palette is improved over the exhaustive search using Kekre-s fast search technique. After coloring the grayscale image pieces the next job is fusion of all these pieces to obtain panoramic view. For similarity estimation between partial images correlation coefficient is used.Keywords: Panoramic View, Similarity Estimate, Color Transfer, Color Palette, Kekre's Fast Search, Kekre's LUV
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17531145 A Spatial Information Network Traffic Prediction Method Based on Hybrid Model
Authors: Jingling Li, Yi Zhang, Wei Liang, Tao Cui, Jun Li
Abstract:
Compared with terrestrial network, the traffic of spatial information network has both self-similarity and short correlation characteristics. By studying its traffic prediction method, the resource utilization of spatial information network can be improved, and the method can provide an important basis for traffic planning of a spatial information network. In this paper, considering the accuracy and complexity of the algorithm, the spatial information network traffic is decomposed into approximate component with long correlation and detail component with short correlation, and a time series hybrid prediction model based on wavelet decomposition is proposed to predict the spatial network traffic. Firstly, the original traffic data are decomposed to approximate components and detail components by using wavelet decomposition algorithm. According to the autocorrelation and partial correlation smearing and truncation characteristics of each component, the corresponding model (AR/MA/ARMA) of each detail component can be directly established, while the type of approximate component modeling can be established by ARIMA model after smoothing. Finally, the prediction results of the multiple models are fitted to obtain the prediction results of the original data. The method not only considers the self-similarity of a spatial information network, but also takes into account the short correlation caused by network burst information, which is verified by using the measured data of a certain back bone network released by the MAWI working group in 2018. Compared with the typical time series model, the predicted data of hybrid model is closer to the real traffic data and has a smaller relative root means square error, which is more suitable for a spatial information network.
Keywords: Spatial Information Network, Traffic prediction, Wavelet decomposition, Time series model.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6371144 A Comparative Analysis of E-Government Quality Models
Authors: Abdoullah Fath-Allah, Laila Cheikhi, Rafa E. Al-Qutaish, Ali Idri
Abstract:
Many quality models have been used to measure egovernment portals quality. However, the absence of an international consensus for e-government portals quality models results in many differences in terms of quality attributes and measures. The aim of this paper is to compare and analyze the existing e-government quality models proposed in literature (those that are based on ISO standards and those that are not) in order to propose guidelines to build a good and useful e-government portals quality model. Our findings show that, there is no e-government portal quality model based on the new international standard ISO 25010. Besides that, the quality models are not based on a best practice model to allow agencies to both; measure e-government portals quality and identify missing best practices for those portals.
Keywords: E-government, portal, best practices, quality model, ISO, standard, ISO 25010, ISO 9126.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 36491143 Assessing Complexity of Neuronal Multiunit Activity by Information Theoretic Measure
Authors: Young-Seok Choi
Abstract:
This paper provides a quantitative measure of the time-varying multiunit neuronal spiking activity using an entropy based approach. To verify the status embedded in the neuronal activity of a population of neurons, the discrete wavelet transform (DWT) is used to isolate the inherent spiking activity of MUA. Due to the de-correlating property of DWT, the spiking activity would be preserved while reducing the non-spiking component. By evaluating the entropy of the wavelet coefficients of the de-noised MUA, a multiresolution Shannon entropy (MRSE) of the MUA signal is developed. The proposed entropy was tested in the analysis of both simulated noisy MUA and actual MUA recorded from cortex in rodent model. Simulation and experimental results demonstrate that the dynamics of a population can be quantified by using the proposed entropy.
Keywords: Discrete wavelet transform, Entropy, Multiresolution, Multiunit activity.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18441142 Extending the Quantum Entropy to Multidimensional Signal Processing
Authors: Youssef Khmou, Said Safi, Miloud Frikel
Abstract:
This paper treats different aspects of entropy measure in classical information theory and statistical quantum mechanics, it presents the possibility of extending the definition of Von Neumann entropy to image and array processing. In the first part, we generalize the quantum entropy using singular values of arbitrary rectangular matrices to measure the randomness and the quality of denoising operation, this new definition of entropy can be implemented to compare the performance analysis of filtering methods. In the second part, we apply the concept of pure state in quantum formalism to generalize the maximum entropy method for narrowband and farfield source localization problem. Several computer simulation results are illustrated to demonstrate the effectiveness of the proposed techniques.Keywords: Von Neumann entropy, Filtering, array, DoA, Maximum Entropy Method.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 25051141 Efficient Program Slicing Algorithms for Measuring Functional Cohesion and Parallelism
Authors: Jehad Al Dallal
Abstract:
Program slicing is the task of finding all statements in a program that directly or indirectly influence the value of a variable occurrence. The set of statements that can affect the value of a variable at some point in a program is called a program slice. In several software engineering applications, such as program debugging and measuring program cohesion and parallelism, several slices are computed at different program points. In this paper, algorithms are introduced to compute all backward and forward static slices of a computer program by traversing the program representation graph once. The program representation graph used in this paper is called Program Dependence Graph (PDG). We have conducted an experimental comparison study using 25 software modules to show the effectiveness of the introduced algorithm for computing all backward static slices over single-point slicing approaches in computing the parallelism and functional cohesion of program modules. The effectiveness of the algorithm is measured in terms of time execution and number of traversed PDG edges. The comparison study results indicate that using the introduced algorithm considerably saves the slicing time and effort required to measure module parallelism and functional cohesion.
Keywords: Backward slicing, cohesion measure, forward slicing, parallelism measure, program dependence graph, program slicing, static slicing.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14481140 Efficiency Evaluation of E-Commerce Websites
Authors: A. K. Abd El-Aleem, W. F. Abd El-wahed, N. A. Ismail, F. A. Torkey
Abstract:
This study suggests a model of a new set of evaluation criteria that will be used to measure the efficiency of real-world E-commerce websites. Evaluation criteria include design, usability and performance for websites, the Data Envelopment Analysis (DEA) technique has been used to measure the websites efficiency. An efficient Web site is defined as a site that generates the most outputs, using the smallest amount of inputs. Inputs refer to measurements representing the amount of effort required to build, maintain and perform the site. Output is amount of traffic the site generates. These outputs are measured as the average number of daily hits and the average number of daily unique visitors.Keywords: Data Envelopment Analysis, E-commerce, Efficiency.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 41051139 Data Oriented Model of Image: as a Framework for Image Processing
Authors: A. Habibizad Navin, A. Sadighi, M. Naghian Fesharaki, M. Mirnia, M. Teshnelab, R. Keshmiri
Abstract:
This paper presents a new data oriented model of image. Then a representation of it, ADBT, is introduced. The ability of ADBT is clustering, segmentation, measuring similarity of images etc, with desired precision and corresponding speed.
Keywords: Data oriented modelling, image, clustering, segmentation, classification, ADBT and image processing.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17991138 Effects of the Second Entrant in GSM Telecommunication Market in MENA Region
Authors: A.R. Yari, M.R. Sadri
Abstract:
For the first incumbent operator it is very important to understand how to react when the second operator comes to the market. In this paper which is prepared for preliminary study of GSM market in Iran, we have studied five MENA markets according to the similarity point of view. This paper aims at analyzing the impact of second entrants in selected markets on certain marketing key performance indicators (KPI) such as: Market shares (by operator), prepaid share, minutes of use (MoU), Price and average revenue per user (ARPU) (for total market each).Keywords: GSM Market, Second entrant, MENA.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14751137 Measuring of Urban Sustainability in Town Planners Practice
Authors: J. Zagorskas, I. Veteikyte
Abstract:
Physical urban form is recognized to be the media for human transactions. It directly influences the travel demand of people in a specific urban area and the amount of energy used for transportation. Distorted, sprawling form often creates sustainability problems in urban areas. It is declared in EU strategic planning documents that compact urban form and mixed land use pattern must be given the main focus to achieve better sustainability in urban areas, but the methods to measure and compare these characteristics are still not clear. This paper presents the simple methods to measure the spatial characteristics of urban form by analyzing the location and distribution of objects in an urban environment. The extended CA (cellular automata) model is used to simulate urban development scenarios.
Keywords: Cellular automata (CA), Mixed used planning, Spatial analysis, Urban compactness, Geographic information systems (GIS).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 26881136 Using Copulas to Measure Association between Air Pollution and Respiratory Diseases
Authors: Snezhana P. Kostova, Krassi V. Rumchev, Todor Vlaev, Silviya B. Popova
Abstract:
Air pollution is still considered as one of the major environmental and health issues. There is enough research evidence to show a strong relationship between exposure to air contaminants and respiratory illnesses among children and adults. In this paper we used the Copula approach to study a potential relationship between selected air pollutants (PM10 and NO2) and hospital admissions for respiratory diseases. Kendall-s tau and Spearman-s rho rank correlation coefficients are calculated and used in Copula method. This paper demonstrates that copulas can be used to provide additional information as a measure of an association when compared to the standard correlation coefficients. The results find a significant correlation between the selected air pollutants and hospital admissions for most of the selected respiratory illnesses.Keywords: Air pollution, Copula, Respiratory Health.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18271135 Variables for Measuring the Impact of the Social Enterprises in the Field of Community Development
Authors: A. Irudaya Veni Mary, M. Victor Louis Anthuvan, P. Christie, A. Indira
Abstract:
In India, social enterprises are working to create social value in various fields including education; health; women and child development; environment protection and community development. Although social enterprises have brought about tremendous changes in the lives of beneficiaries, the importance of their works is not understood thoroughly. One of the ways to prove themselves is to measure the impact, which in recent times has received much attention. This paper focuses on the study of social value created by the social enterprises in the field of community development. It also aims to put forth a research tool for measuring the social value created by the social enterprises in the field of community development. A close-ended interview schedule was prepared to measure the social value creation and it was administered among 60 beneficiaries of two social enterprises who work in the field of community development. The study results show that the social enterprises have brought four types of impact in the life of their beneficiaries; economic impact, social impact, political impact and cultural impact. This study is limited to the social enterprises those who work towards community development. This empirical finding will enable the reader to understand various types of social value created by the social enterprises working in the field of community development. This study will also serve as guide for social enterprises in community development activities to measure their impact and thereby improve their operation towards the betterment of the society. This paper is derived from an empirical research carried out to describe the different types of social value created by the social enterprises in India.
Keywords: Social enterprise, social entrepreneurs, social impact, social value, tool for social impact measurement.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19171134 A Multi-Attribute Utility Model for Performance Evaluation of Sustainable Banking
Authors: Sonia Rebai, Mohamed Naceur Azaiez, Dhafer Saidane
Abstract:
In this study, we develop a performance evaluation model based on a multi-attribute utility approach aiming at reaching the sustainable banking (SB) status. This model is built accounting for various banks’ stakeholders in a win-win paradigm. In addition, it offers the opportunity for adopting a global measure of performance as an indication of a bank’s sustainability degree. This measure is referred to as banking sustainability performance index (BSPI). This index may constitute a basis for ranking banks. Moreover, it may constitute a bridge between the assessment types of financial and extra-financial rating agencies. A real application is performed on three French banks.
Keywords: Multi-attribute utility theory, Performance, Sustainable banking.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22071133 Different Views and Evaluations of IT Artifacts
Authors: Sameh Al-Natour, Izak Benbasat
Abstract:
The introduction of a multitude of new and interactive e-commerce information technology (IT) artifacts has impacted adoption research. Rather than solely functioning as productivity tools, new IT artifacts assume the roles of interaction mediators and social actors. This paper describes the varying roles assumed by IT artifacts, and proposes and distinguishes between four distinct foci of how the artifacts are evaluated. It further proposes a theoretical model that maps the different views of IT artifacts to four distinct types of evaluations.Keywords: IT adoption, IT artifacts, similarity, social actor.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13521132 Building an Arithmetic Model to Assess Visual Consistency in Townscape
Authors: Dheyaa Hussein, Peter Armstrong
Abstract:
The phenomenon of visual disorder is prominent in contemporary townscapes. This paper provides a theoretical framework for the assessment of visual consistency in townscape in order to achieve more favourable outcomes for users. In this paper, visual consistency refers to the amount of similarity between adjacent components of townscape. The paper investigates parameters which relate to visual consistency in townscape, explores the relationships between them and highlights their significance. The paper uses arithmetic methods from outside the domain of urban design to enable the establishment of an objective approach of assessment which considers subjective indicators including users’ preferences. These methods involve the standard of deviation, colour distance and the distance between points. The paper identifies urban space as a key representative of the visual parameters of townscape. It focuses on its two components, geometry and colour in the evaluation of the visual consistency of townscape. Accordingly, this article proposes four measurements. The first quantifies the number of vertices, which are points in the three-dimensional space that are connected, by lines, to represent the appearance of elements. The second evaluates the visual surroundings of urban space through assessing the location of their vertices. The last two measurements calculate the visual similarity in both vertices and colour in townscape by the calculation of their variation using methods including standard of deviation and colour difference. The proposed quantitative assessment is based on users’ preferences towards these measurements. The paper offers a theoretical basis for a practical tool which can alter the current understanding of architectural form and its application in urban space. This tool is currently under development. The proposed method underpins expert subjective assessment and permits the establishment of a unified framework which adds to creativity by the achievement of a higher level of consistency and satisfaction among the citizens of evolving townscapes.
Keywords: Townscape, Urban Design, Visual Assessment, Visual Consistency.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16351131 Application of Multi-Dimensional Principal Component Analysis to Medical Data
Authors: Naoki Yamamoto, Jun Murakami, Chiharu Okuma, Yutaro Shigeto, Satoko Saito, Takashi Izumi, Nozomi Hayashida
Abstract:
Multi-dimensional principal component analysis (PCA) is the extension of the PCA, which is used widely as the dimensionality reduction technique in multivariate data analysis, to handle multi-dimensional data. To calculate the PCA the singular value decomposition (SVD) is commonly employed by the reason of its numerical stability. The multi-dimensional PCA can be calculated by using the higher-order SVD (HOSVD), which is proposed by Lathauwer et al., similarly with the case of ordinary PCA. In this paper, we apply the multi-dimensional PCA to the multi-dimensional medical data including the functional independence measure (FIM) score, and describe the results of experimental analysis.Keywords: multi-dimensional principal component analysis, higher-order SVD (HOSVD), functional independence measure (FIM), medical data, tensor decomposition
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 25021130 PSO-based Possibilistic Portfolio Model with Transaction Costs
Authors: Wei Chen, Cui-you Yao, Yue Qiu
Abstract:
This paper deals with a portfolio selection problem based on the possibility theory under the assumption that the returns of assets are LR-type fuzzy numbers. A possibilistic portfolio model with transaction costs is proposed, in which the possibilistic mean value of the return is termed measure of investment return, and the possibilistic variance of the return is termed measure of investment risk. Due to considering transaction costs, the existing traditional optimization algorithms usually fail to find the optimal solution efficiently and heuristic algorithms can be the best method. Therefore, a particle swarm optimization is designed to solve the corresponding optimization problem. At last, a numerical example is given to illustrate our proposed effective means and approaches.Keywords: Possibility theory, portfolio selection, transaction costs, particle swarm optimization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15341129 Accurate Dimensional Measurement of 3D Round Holes Based on Stereo Vision
Authors: Zhiguo Ren, Lilong Cai
Abstract:
This paper present an effective method to accurately reconstruct and measure the 3D curve edges of small industrial parts based on stereo vision. To effectively fit the curve of the measured parts using a series of line segments in the images, a strategy from coarse to fine is employed based on multi-scale curve fitting. After reconstructing the 3D curve of a hole through a curved surface, its axis is adjusted so that it is parallel to the Z axis with least squares error and the dimensions of the hole can be calculated on the XY plane easily. Experimental results show that the presented method can accurately measure the dimensions of round holes through a curved surface.
Keywords: Stereo Vision, 3D Round Hole Measurement, Curve Fitting, 3D Curve Reconstruction, Least Squares Error.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16271128 A Study on Reducing Malicious Replies on the Internet: An Approach by Game Theory
Authors: Sanghun Lee
Abstract:
Since the advent of the information era, the Internet has brought various positive effects in everyday life. Nevertheless, recently, problems and side-effects have been noted. Internet witch-trials and spread of pornography are only a few of these problems.In this study, problems and causes of malicious replies on internet boards were analyzed, using the key ideas of game theory. The study provides a mathematical model for the internet reply game to devise three possible plans that could efficiently counteract malicious replies. Furthermore, seven specific measures that comply with one of the three plans were proposed and evaluated according to the importance and utility of each measure using the orthogonal array survey and SPSS conjoint analysis.The conclusion was that the most effective measure would be forbidding unsigned user access to malicious replies. Also notable was that some analytically proposed measures, when implemented, could backfire and encourage malicious replies.Keywords: Conjoint Analysis, Game Theory, Internet, MaliciousReplies, Prisoner's Dilemma
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1842