Search results for: feature points management
1798 Problems and Needs of Frozen Shrimp Industry Small and Medium Enterprises in the Central Region of the Lower Three Provinces
Authors: P. Thepnarintra
Abstract:
Frozen shrimp industry plays an important role in the development of production industry of the country. There has been a continuing development to response the increasing demand; however, there have been some problems in running the enterprises. The purposes of this study are to: 1) investigate problems related to basic factors in operating frozen shrimp industry based on the entrepreneurs’ points of view. The enterprises involved in this study were small and medium industry receiving Thai Frozen Foods Association. 2) Compare the problems of the frozen shrimp industry according to their sizes of operation in 3 provinces of the central region Thailand. Population in this study consisted of 148 managers from 148 frozen shrimp enterprises Thai Frozen Foods Association which 77 were small size and 71 were medium size. The data were analyzed to find percentage, arithmetic mean, standard deviation, and independent sample T-test with the significant hypothesis at .05. The results revealed that the problems of the frozen shrimp industries of both size were in high level. The needs for government supporting were in high level. The comparison of the problems and the basic factors between the small and medium size enterprises showed no statistically significant level. The problems that they mentioned included raw materials, labors, production, marketing, and the need for academic supporting from the government sector.Keywords: Frozen shrimp industry, problems, related to the enterprise, operation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 11031797 An Inductive Coupling Based CMOS Wireless Powering Link for Implantable Biomedical Applications
Authors: Lei Yao, Jia Hao Cheong, Rui-Feng Xue, Minkyu Je
Abstract:
A closed-loop controlled wireless power transmission circuit block for implantable biomedical applications is described in this paper. The circuit consists of one front-end rectifier, power management sub-block including bandgap reference and low drop-out regulators (LDOs) as well as transmission power detection / feedback circuits. Simulation result shows that the front-end rectifier achieves 80% power efficiency with 750-mV single-end peak-to-peak input voltage and 1.28-V output voltage under load current of 4 mA. The power management block can supply 1.8mA average load current under 1V consuming only 12μW power, which is equivalent to 99.3% power efficiency. The wireless power transmission block described in this paper achieves a maximum power efficiency of 80%. The wireless power transmission circuit block is designed and implemented using UMC 65-nm CMOS/RF process. It occupies 1 mm × 1.2 mm silicon area.
Keywords: Implantable biomedical devices, wireless power transfer, LDO, rectifier, closed-loop power control
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22851796 Developing and Implementing Successful Key Performance Indicators
Authors: Marie Mikušová, Viktorie Janečková
Abstract:
Measurement and the following evaluation of performance represent important part of management. The paper focuses on indicators as the basic elements of performance measurement system. It emphasizes a necessity of searching requirements for quality indicators so that they can become part of the useful system. It introduces standpoints for a systematic dividing of indicators so that they have as high as possible informative value of background sources for searching, analysis, designing and using of indicators. It draws attention to requirements for indicators' quality and at the same it deals with some dangers decreasing indicator's informative value. It submits a draft of questions that should be answered at the construction of indicator. It is obvious that particular indicators need to be defined exactly to stimulate the desired behavior in order to attain expected results. In the enclosure a concrete example of the defined indicator in the concrete conditions of a small firm is given. The authors of the paper pay attention to the fact that a quality indicator makes it possible to get to the basic causes of the problem and include the established facts into the company information system. At the same time they emphasize that developing of a quality indicator is a prerequisite for the utilization of the system of measurement in management.Keywords: performance, measurement, firm, indicator
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15561795 Motion Prediction and Motion Vector Cost Reduction during Fast Block Motion Estimation in MCTF
Authors: Karunakar A K, Manohara Pai M M
Abstract:
In 3D-wavelet video coding framework temporal filtering is done along the trajectory of motion using Motion Compensated Temporal Filtering (MCTF). Hence computationally efficient motion estimation technique is the need of MCTF. In this paper a predictive technique is proposed in order to reduce the computational complexity of the MCTF framework, by exploiting the high correlation among the frames in a Group Of Picture (GOP). The proposed technique applies coarse and fine searches of any fast block based motion estimation, only to the first pair of frames in a GOP. The generated motion vectors are supplied to the next consecutive frames, even to subsequent temporal levels and only fine search is carried out around those predicted motion vectors. Hence coarse search is skipped for all the motion estimation in a GOP except for the first pair of frames. The technique has been tested for different fast block based motion estimation algorithms over different standard test sequences using MC-EZBC, a state-of-the-art scalable video coder. The simulation result reveals substantial reduction (i.e. 20.75% to 38.24%) in the number of search points during motion estimation, without compromising the quality of the reconstructed video compared to non-predictive techniques. Since the motion vectors of all the pair of frames in a GOP except the first pair will have value ±1 around the motion vectors of the previous pair of frames, the number of bits required for motion vectors is also reduced by 50%.Keywords: Motion Compensated Temporal Filtering, predictivemotion estimation, lifted wavelet transform, motion vector
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16191794 Fuzzy Group Decision Making for the Assessment of Health-Care Waste Disposal Alternatives in Istanbul
Authors: Mehtap Dursun, E. Ertugrul Karsak, Melis Almula Karadayi
Abstract:
Disposal of health-care waste (HCW) is considered as an important environmental problem especially in large cities. Multiple criteria decision making (MCDM) techniques are apt to deal with quantitative and qualitative considerations of the health-care waste management (HCWM) problems. This research proposes a fuzzy multi-criteria group decision making approach with a multilevel hierarchical structure including qualitative as well as quantitative performance attributes for evaluating HCW disposal alternatives for Istanbul. Using the entropy weighting method, objective weights as well as subjective weights are taken into account to determine the importance weighting of quantitative performance attributes. The results obtained using the proposed methodology are thoroughly analyzed.Keywords: Entropy weighting method, group decision making, health-care waste management, hierarchical fuzzy multi-criteriadecision making
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16871793 The Effect of Reducing Superimposed Dead Load on the Lateral Seismic Deformations of Structures
Authors: H. Alnajajra, A. Touqan, M. Dwaikat
Abstract:
The vast majority of the Middle East countries are prone to earthquakes. Despite that and from a seismic hazard point of view, the higher values of the superimposed dead load intensity of partitions and wearing materials of the constructed reinforced concrete slabs in these countries can increase the earthquake vulnerability of the structures. The primary objective of this paper is to investigate the effect of reducing superimposed dead load on the lateral seismic deformations of structures, the inter-story drifts and the seismic pounding damages. The study utilizes a group of three reinforced concrete structures at three different site conditions. These structures are assumed to be constructed in Nablus city of Palestine, and having superimposed dead load value as 1 kN/m2, 3 kN/m2, and 5 kN/m2, respectively. SAP2000 program, Version 18.1.1, is used to perform the response spectrum analysis to obtain the potential lateral seismic deformations of the studied models. Amazingly, the study points that, at the same site, superimposed dead load has a minor effect on the lateral deflections of the models. This, however, promotes the hypothesis that buildings failed during earthquakes mainly because they were not designed appropriately against gravity loads.
Keywords: Gravity loads, inter-story drifts, lateral seismic deformations, reinforced concrete slabs, response spectrum method, SAP2000, seismic design, seismic pounding, superimposed dead load.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13721792 Rigorous Modeling of Fixed-Bed Reactors Containing Finite Hollow Cylindrical Catalyst with Michaelis-Menten Type of Kinetics
Authors: Mohammad Asif
Abstract:
A large number of chemical, bio-chemical and pollution-control processes use heterogeneous fixed-bed reactors. The use of finite hollow cylindrical catalyst pellets can enhance conversion levels in such reactors. The absence of the pellet core can significantly lower the diffusional resistance associated with the solid phase. This leads to a better utilization of the catalytic material, which is reflected in the higher values for the effectiveness factor, leading ultimately to an enhanced conversion level in the reactor. It is however important to develop a rigorous heterogeneous model for the reactor incorporating the two-dimensional feature of the solid phase owing to the presence of the finite hollow cylindrical catalyst pellet. Presently, heterogeneous models reported in the literature invariably employ one-dimension solid phase models meant for spherical catalyst pellets. The objective of the paper is to present a rigorous model of the fixed-bed reactors containing finite hollow cylindrical catalyst pellets. The reaction kinetics considered here is the widely used Michaelis–Menten kinetics for the liquid-phase bio-chemical reactions. The reaction parameters used here are for the enzymatic degradation of urea. Results indicate that increasing the height to diameter ratio helps to improve the conversion level. On the other hand, decreasing the thickness is apparently not as effective. This could however be explained in terms of the higher void fraction of the bed that causes a smaller amount of the solid phase to be packed in the fixed-bed bio-chemical reactor.
Keywords: Fixed-bed reactor, Finite hollow cylinder, Catalyst pellet, Conversion, Michaelis-Menten kinetics.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15971791 Single-Walled Carbon Nanotube Synthesis by Chemical Vapor Deposition Using Platinum-Group Metal Catalysts
Authors: T. Maruyama, T. Saida, S. Naritsuka, S. Iijima
Abstract:
Single-walled carbon nanotubes (SWCNTs) are generally synthesized by chemical vapor deposition (CVD) using Fe, Co, and Ni as catalysts. However, due to the Ostwald ripening of metal catalysts, the diameter distribution of the grown SWCNTs is considerably wide (>2 nm), which is not suitable for electronics applications. In addition, reduction in the growth temperature is desirable for fabricating SWCNT devices compatible with the LSI process. Herein, we performed SWCNT growth by alcohol catalytic CVD using platinum-group metal catalysts (Pt, Rh, and Pd) because these metals have high melting points, and the reduction in the Ostwald ripening of catalyst particles is expected. Our results revealed that web-like SWCNTs were obtained from Pt and Rh catalysts at growth temperature between 500 °C and 600 °C by optimizing the ethanol pressure. The SWCNT yield from Pd catalysts was considerably low. By decreasing the growth temperature, the diameter and chirality distribution of SWCNTs from Pt and Rh catalysts became small and narrow. In particular, the diameters of most SWCNTs grown using Pt catalysts were below 1 nm and their diameter distribution was considerably narrow. On the contrary, SWCNTs can grow from Rh catalysts even at 300 °C by optimizing the growth condition, which is the lowest temperature recorded for SWCNT growth. Our results demonstrated that platinum-group metals are useful for the growth of small-diameter SWCNTs and facilitate low-temperature growth.
Keywords: Carbon nanotube, chemical vapor deposition, catalyst, Pt, Rh, Pd.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 8591790 Business-Intelligence Mining of Large Decentralized Multimedia Datasets with a Distributed Multi-Agent System
Authors: Karima Qayumi, Alex Norta
Abstract:
The rapid generation of high volume and a broad variety of data from the application of new technologies pose challenges for the generation of business-intelligence. Most organizations and business owners need to extract data from multiple sources and apply analytical methods for the purposes of developing their business. Therefore, the recently decentralized data management environment is relying on a distributed computing paradigm. While data are stored in highly distributed systems, the implementation of distributed data-mining techniques is a challenge. The aim of this technique is to gather knowledge from every domain and all the datasets stemming from distributed resources. As agent technologies offer significant contributions for managing the complexity of distributed systems, we consider this for next-generation data-mining processes. To demonstrate agent-based business intelligence operations, we use agent-oriented modeling techniques to develop a new artifact for mining massive datasets.
Keywords: Agent-oriented modeling, business Intelligence management, distributed data mining, multi-agent system.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13741789 A Spatial Hypergraph Based Semi-Supervised Band Selection Method for Hyperspectral Imagery Semantic Interpretation
Authors: Akrem Sellami, Imed Riadh Farah
Abstract:
Hyperspectral imagery (HSI) typically provides a wealth of information captured in a wide range of the electromagnetic spectrum for each pixel in the image. Hence, a pixel in HSI is a high-dimensional vector of intensities with a large spectral range and a high spectral resolution. Therefore, the semantic interpretation is a challenging task of HSI analysis. We focused in this paper on object classification as HSI semantic interpretation. However, HSI classification still faces some issues, among which are the following: The spatial variability of spectral signatures, the high number of spectral bands, and the high cost of true sample labeling. Therefore, the high number of spectral bands and the low number of training samples pose the problem of the curse of dimensionality. In order to resolve this problem, we propose to introduce the process of dimensionality reduction trying to improve the classification of HSI. The presented approach is a semi-supervised band selection method based on spatial hypergraph embedding model to represent higher order relationships with different weights of the spatial neighbors corresponding to the centroid of pixel. This semi-supervised band selection has been developed to select useful bands for object classification. The presented approach is evaluated on AVIRIS and ROSIS HSIs and compared to other dimensionality reduction methods. The experimental results demonstrate the efficacy of our approach compared to many existing dimensionality reduction methods for HSI classification.Keywords: Hyperspectral image, spatial hypergraph, dimensionality reduction, semantic interpretation, band selection, feature extraction.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 12201788 Influence of Slope Shape and Surface Roughness on the Moving Paths of a Single Rockfall
Authors: Iau-Teh Wang, Chin-Yu Lee
Abstract:
Rockfall is a kind of irregular geological disaster. Its destruction time, space and movements are highly random. The impact force is determined by the way and velocity rocks move. The movement velocity of a rockfall depends on slope gradient of its moving paths, height, slope surface roughness and rock shapes. For effectively mitigate and prevent disasters brought by rockfalls, it is required to precisely calculate the moving paths of a rockfall so as to provide the best protective design. This paper applies Colorado Rockfall Simulation Program (CRSP) as our study tool to discuss the impact of slope shape and surface roughness on the moving paths of a single rockfall. The analytical results showed that the slope, m=1:1, acted as the threshold for rockfall bounce height on a monoclinal slight slope. When JRC ´╝£ 1.2, movement velocity reduced and bounce height increased as JCR increased. If slope fixed and JRC increased, the bounce height of rocks increased gradually with reducing movement velocity. Therefore, the analysis on the moving paths of rockfalls with CRSP could simulate bouncing of falling rocks. By analyzing moving paths, velocity, and bounce height of falling rocks, we could effectively locate impact points of falling rocks on a slope. Such analysis can be served as a reference for future disaster prevention and control.Keywords: Rockfall, Slope Shape, Moving Path, SurfaceRoughness.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 28091787 The Formation of Mutual Understanding in Conversation: An Embodied Approach
Authors: Haruo Okabayashi
Abstract:
The mutual understanding in conversation is very important for human relations. This study investigates the mental function of the formation of mutual understanding between two people in conversation using the embodied approach. Forty people participated in this study. They are divided into pairs randomly. Four conversation situations between two (make/listen to fun or pleasant talk, make/listen to regrettable talk) are set for four minutes each, and the finger plethysmogram (200 Hz) of each participant is measured. As a result, the attractors of the participants who reported “I did not understand my partner” show the collapsed shape, which means the fluctuation of their rhythm is too small to match their partner’s rhythm, and their cross correlation is low. The autonomic balance of both persons tends to resonate during conversation, and both LLEs tend to resonate, too. In human history, in order for human beings as weak mammals to live, they may have been with others; that is, they have brought about resonating characteristics, which is called self-organization. However, the resonant feature sometimes collapses, depending on the lifestyle that the person was formed by himself after birth. It is difficult for people who do not have a lifestyle of mutual gaze to resonate their biological signal waves with others’. These people have features such as anxiety, fatigue, and confusion tendency. Mutual understanding is thought to be formed as a result of cooperation between the features of self-organization of the persons who are talking and the lifestyle indicated by mutual gaze. Such an entanglement phenomenon is called a nonlinear relation. By this research, it is found that the formation of mutual understanding is expressed by the rhythm of a biological signal showing a nonlinear relationship.
Keywords: Embodied approach, finger plethysmogram, mutual understanding, nonlinear phenomenon.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 12991786 The Applicability of the Zipper Strut to Seismic Rehabilitation of Steel Structures
Authors: G. R. Nouri, H. Imani Kalesar, Zahra Ameli
Abstract:
Chevron frames (Inverted-V-braced frames or Vbraced frames) have seismic disadvantages, such as not good exhibit force redistribution capability and compression brace buckles immediately. Researchers developed new design provisions on increasing both the ductility and lateral resistance of these structures in seismic areas. One of these new methods is adding zipper columns, as proposed by Khatib et al. (1988) [2]. Zipper columns are vertical members connecting the intersection points of the braces above the first floor. In this paper applicability of the suspended zipper system to Seismic Rehabilitation of Steel Structures is investigated. The models are 3-, 6-, 9-, and 12-story Inverted-V-braced frames. In this case, it is assumed that the structures must be rehabilitated. For rehabilitation of structures, zipper column is used. The result of researches showed that the suspended zipper system is effective in case of 3-, 6-, and 9-story Inverted-V-braced frames and it would increase lateral resistance of structure up to life safety level. But in case of high-rise buildings (such as 12 story frame), it doesn-t show good performance. For solving this problem, the braced bay can consist of small “units" over the height of the entire structure, which each of them is a zipper-braced bay with a few stories. By using this method the lateral resistance of 12 story Inverted-V-braced frames is increased up to safety life level.Keywords: chevron-braced frames, suspended zipper frames, zipper frames, zipper columns
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22221785 The Key Challenges of the New Bank Regulations
Authors: Petr Teply
Abstract:
The New Basel Capital Accord (Basel II) influences how financial institutions around the world, and especially European Union institutions, determine the amount of capital to reserve. However, as the recent global crisis has shown, the revision of Basel II is needed to reflect current trends, such as increased volatility and correlation, in the world financial markets. The overall objective of Basel II is to increase the safety and soundness of the international financial system. Basel II builds on three main pillars: Pillar I deals with the minimum capital requirements for credit, market and operational risk, Pillar II focuses on the supervisory review process and finally Pillar III promotes market discipline through enhanced disclosure requirements for banks. The aim of this paper is to provide the historical background, key features and impact of Basel II on financial markets. Moreover, we discuss new proposals for international bank regulation (sometimes referred to as Basel III) which include requirements for higher quality, constituency and transparency of banks' capital and risk management, regulation of OTC markets and introduction of new liquidity standards for internationally active banks.
Keywords: Basel II, Basel III, risk management, bank regulation
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20741784 Measuring Banks’ Antifragility via Fuzzy Logic
Authors: Danielle Sandler dos Passos, Helder Coelho, Flávia Mori Sarti
Abstract:
Analysing the world banking sector, we realize that traditional risk measurement methodologies no longer reflect the actual scenario with uncertainty and leave out events that can change the dynamics of markets. Considering this, regulators and financial institutions began to search more realistic models. The aim is to include external influences and interdependencies between agents, to describe and measure the operationalization of these complex systems and their risks in a more coherent and credible way. Within this context, X-Events are more frequent than assumed and, with uncertainties and constant changes, the concept of antifragility starts to gain great prominence in comparison to others methodologies of risk management. It is very useful to analyse whether a system succumbs (fragile), resists (robust) or gets benefits (antifragile) from disorder and stress. Thus, this work proposes the creation of the Banking Antifragility Index (BAI), which is based on the calculation of a triangular fuzzy number – to "quantify" qualitative criteria linked to antifragility.
Keywords: Complex adaptive systems, X-events, risk management, antifragility, banking antifragility index, triangular fuzzy number.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 8971783 An Experimentally Validated Thermo- Mechanical Finite Element Model for Friction Stir Welding in Carbon Steels
Authors: A. H. Kheireddine, A. A. Khalil, A. H. Ammouri, G. T. Kridli, R. F. Hamade
Abstract:
Solidification cracking and hydrogen cracking are some defects generated in the fusion welding of ultrahigh carbon steels. However, friction stir welding (FSW) of such steels, being a solid-state technique, has been demonstrated to alleviate such problems encountered in traditional welding. FSW include different process parameters that must be carefully defined prior processing. These parameters included but not restricted to: tool feed, tool RPM, tool geometry, tool tilt angle. These parameters form a key factor behind avoiding warm holes and voids behind the tool and in achieving a defect-free weld. More importantly, these parameters directly affect the microstructure of the weld and hence the final mechanical properties of weld. For that, 3D finite element (FE) thermo-mechanical model was developed using DEFORM 3D to simulate FSW of carbon steel. At points of interest in the joint, tracking is done for history of critical state variables such as temperature, stresses, and strain rates. Typical results found include the ability to simulate different weld zones. Simulations predictions were successfully compared to experimental FSW tests. It is believed that such a numerical model can be used to optimize FSW processing parameters to favor desirable defect free weld with better mechanical properties.
Keywords: Carbon Steels, DEFORM 3D, FEM, Friction stir welding.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 25711782 Relevance of the Variation in the Angulation of Palatal Throat Form to the Orientation of the Occlusal Plane: A Cephalometric Study
Authors: Sanath Kumar Shetty, Sanya Sinha, K. Kamalakanth Shenoy
Abstract:
The posterior reference for the ala tragal line is a cause of confusion, with different authors suggesting different locations as to the superior, middle or inferior part of the tragus. This study was conducted on 200 subjects to evaluate if any correlation exists between the variation of angulation of palatal throat form and the relative parallelism of occlusal plane to ala-tragal line at different tragal levels. A custom made Occlusal Plane Analyzer was used to check the parallelism between the ala-tragal line and occlusal plane. A lateral cephalogram was shot for each subject to measure the angulation of the palatal throat form. Fisher’s exact test was used to evaluate the correlation between the angulation of the palatal throat form and the relative parallelism of occlusal plane to the ala tragal line. Also, a classification was formulated for the palatal throat form, based on confidence interval. From the results of the study, the inferior part, middle part and superior part of the tragus were seen as the reference points in 49.5%, 32% and 18.5% of the subjects respectively. Class I palatal throat form (41degree-50 degree), Class II palatal throat form (below 41 degree) and Class III palatal throat form (above 50 degree) were seen in 42%, 43% and 15% of the subjects respectively. It was also concluded that there is no significant correlation between the variation in the angulations of the palatal throat form and the relative parallelism of occlusal plane to the ala-tragal line.Keywords: Ala-tragal line, occlusal plane, palatal throat form, cephalometry.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 26451781 Cirrhosis Mortality Prediction as Classification Using Frequent Subgraph Mining
Authors: Abdolghani Ebrahimi, Diego Klabjan, Chenxi Ge, Daniela Ladner, Parker Stride
Abstract:
In this work, we use machine learning and data analysis techniques to predict the one-year mortality of cirrhotic patients. Data from 2,322 patients with liver cirrhosis are collected at a single medical center. Different machine learning models are applied to predict one-year mortality. A comprehensive feature space including demographic information, comorbidity, clinical procedure and laboratory tests is being analyzed. A temporal pattern mining technic called Frequent Subgraph Mining (FSM) is being used. Model for End-stage liver disease (MELD) prediction of mortality is used as a comparator. All of our models statistically significantly outperform the MELD-score model and show an average 10% improvement of the area under the curve (AUC). The FSM technic itself does not improve the model significantly, but FSM, together with a machine learning technique called an ensemble, further improves the model performance. With the abundance of data available in healthcare through electronic health records (EHR), existing predictive models can be refined to identify and treat patients at risk for higher mortality. However, due to the sparsity of the temporal information needed by FSM, the FSM model does not yield significant improvements. Our work applies modern machine learning algorithms and data analysis methods on predicting one-year mortality of cirrhotic patients and builds a model that predicts one-year mortality significantly more accurate than the MELD score. We have also tested the potential of FSM and provided a new perspective of the importance of clinical features.
Keywords: machine learning, liver cirrhosis, subgraph mining, supervised learning
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4491780 An Optimal Load Shedding Approach for Distribution Networks with DGs considering Capacity Deficiency Modelling of Bulked Power Supply
Authors: A. R. Malekpour, A.R. Seifi
Abstract:
This paper discusses a genetic algorithm (GA) based optimal load shedding that can apply for electrical distribution networks with and without dispersed generators (DG). Also, the proposed method has the ability for considering constant and variable capacity deficiency caused by unscheduled outages in the bulked generation and transmission system of bulked power supply. The genetic algorithm (GA) is employed to search for the optimal load shedding strategy in distribution networks considering DGs in two cases of constant and variable modelling of bulked power supply of distribution networks. Electrical power distribution systems have a radial network and unidirectional power flows. With the advent of dispersed generations, the electrical distribution system has a locally looped network and bidirectional power flows. Therefore, installed DG in the electrical distribution systems can cause operational problems and impact on existing operational schemes. Introduction of DGs in electrical distribution systems has introduced many new issues in operational and planning level. Load shedding as one of operational issue has no exempt. The objective is to minimize the sum of curtailed load and also system losses within the frame-work of system operational and security constraints. The proposed method is tested on a radial distribution system with 33 load points for more practical applications.
Keywords: DG, Load shedding, Optimization, Capacity Deficiency Modelling.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17391779 Opponent Color and Curvelet Transform Based Image Retrieval System Using Genetic Algorithm
Authors: Yesubai Rubavathi Charles, Ravi Ramraj
Abstract:
In order to retrieve images efficiently from a large database, a unique method integrating color and texture features using genetic programming has been proposed. Opponent color histogram which gives shadow, shade, and light intensity invariant property is employed in the proposed framework for extracting color features. For texture feature extraction, fast discrete curvelet transform which captures more orientation information at different scales is incorporated to represent curved like edges. The recent scenario in the issues of image retrieval is to reduce the semantic gap between user’s preference and low level features. To address this concern, genetic algorithm combined with relevance feedback is embedded to reduce semantic gap and retrieve user’s preference images. Extensive and comparative experiments have been conducted to evaluate proposed framework for content based image retrieval on two databases, i.e., COIL-100 and Corel-1000. Experimental results clearly show that the proposed system surpassed other existing systems in terms of precision and recall. The proposed work achieves highest performance with average precision of 88.2% on COIL-100 and 76.3% on Corel, the average recall of 69.9% on COIL and 76.3% on Corel. Thus, the experimental results confirm that the proposed content based image retrieval system architecture attains better solution for image retrieval.Keywords: Content based image retrieval, Curvelet transform, Genetic algorithm, Opponent color histogram, Relevance feedback.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18221778 Fake Account Detection in Twitter Based on Minimum Weighted Feature set
Authors: Ahmed El Azab, Amira M. Idrees, Mahmoud A. Mahmoud, Hesham Hefny
Abstract:
Social networking sites such as Twitter and Facebook attracts over 500 million users across the world, for those users, their social life, even their practical life, has become interrelated. Their interaction with social networking has affected their life forever. Accordingly, social networking sites have become among the main channels that are responsible for vast dissemination of different kinds of information during real time events. This popularity in Social networking has led to different problems including the possibility of exposing incorrect information to their users through fake accounts which results to the spread of malicious content during life events. This situation can result to a huge damage in the real world to the society in general including citizens, business entities, and others. In this paper, we present a classification method for detecting the fake accounts on Twitter. The study determines the minimized set of the main factors that influence the detection of the fake accounts on Twitter, and then the determined factors are applied using different classification techniques. A comparison of the results of these techniques has been performed and the most accurate algorithm is selected according to the accuracy of the results. The study has been compared with different recent researches in the same area; this comparison has proved the accuracy of the proposed study. We claim that this study can be continuously applied on Twitter social network to automatically detect the fake accounts; moreover, the study can be applied on different social network sites such as Facebook with minor changes according to the nature of the social network which are discussed in this paper.Keywords: Fake accounts detection, classification algorithms, twitter accounts analysis, features based techniques.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 58381777 Unlocking Tourism Value through a Tourist Experience Management Paradigm
Authors: Siphiwe P. Mandina, Tinashe Shamuyashe
Abstract:
Tourism has become a topical issue amongst academics and practitioners due to its potential to contribute significantly towards an economy’s GDP. The problem underpinning this research is the fact that the major attraction, Victoria Falls, is being marketed in neighboring countries like South Africa, Botswana and Zambia with tour operators providing just day trips to the Victoria Falls. This has deprived Zimbabwe of income from tourism with tourists making day trips and actually not spending nights in Zimbabwe. This therefore calls for cutting edge marketing strategies that are superior to or inimitable by competing nations such as South Africa and Zambia. This study proposes a shift towards an experience management paradigm in the tourism sector. A qualitative research was adopted for this study, and findings of this study were generalized across different tourism contexts, therefore making the survey based research design more appropriate. The target population for this study is tourists visiting Zimbabwe over the period 2016 and ZTA visitor database acquired from the Department of Immigration will form the sampling frame for the purposes of this study.
Keywords: Competitiveness, tourist arrivals, tourist experience, Zimbabwe.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10961776 Application of Voltage Stability Indices for Proper Placement of STATCOM under Load Increase Scenario
Authors: A. S. Telang, P. P. Bedekar
Abstract:
In today’s world, electrical energy has become an indispensable component of all aspects of modern human life. Reliability, security and stability are the key aspects of any power system. Failure to meet any of these three aspects results into a great impediment to modern life. Modern power systems are being subjected to heavily stressed conditions leading to voltage stability problems. If the voltage stability problems are not mitigated properly through proper voltage stability assessment methods, cascading events may occur which may lead to voltage collapse or blackout events. Modern FACTS devices like STATCOM are one of the measures to overcome the blackout problems. As these devices are very costly, they must be installed properly at suitable locations, mostly at weak bus. Line voltage stability indices such as FVSI, Lmn and LQP play important role for identification of a weak bus. This paper presents evaluation of these line stability indices for the assessment of reliable information about the closeness of the power system to voltage collapse. PSAT is a user-friendly MATLAB toolbox, of which CPF is an important feature which has been extensively used for the placement of STATCOM to assess the stability. Novelty of the present research work lies in that the active and reactive load has been changed simultaneously at all the load buses under consideration. MATLAB code has been developed for the same and tested successfully on various standard IEEE test systems. The results for standard IEEE14 bus test system, specifically, are presented in this paper.
Keywords: Voltage stability analysis, voltage collapse, PSAT, CPF, VSI, FVSI, Lmn, LQP.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17831775 Sustainable Intensification of Agriculture in Victoria’s Food Bowl: Optimizing Productivity with the use of Decision-Support Tools
Authors: M. Johnson, R. Faggian, V. Sposito
Abstract:
A participatory and engaged approach is key in connecting agricultural managers to sustainable agricultural systems to support and optimize production in Victoria’s food bowl. A sustainable intensification (SI) approach is well documented globally, but participation rates amongst Victorian farmers is fragmentary, and key outcomes and implementation strategies are poorly understood. Improvement in decision-support management tools and a greater understanding of the productivity gains available upon implementation of SI is necessary. This paper reviews the current understanding and uptake of SI practices amongst farmers in one of Victoria’s premier food producing regions, the Goulburn Broken; and it spatially analyses the potential for this region to adapt to climate change and optimize food production. A Geographical Information Systems (GIS) approach is taken to develop an interactive decision-support tool that can be accessible to on-ground agricultural managers. The tool encompasses multiple criteria analysis (MCA) that identifies factors during the construction phase of the tool, using expert witnesses and regional knowledge, framed within an Analytical Hierarchy Process. Given the complexities of the interrelations between each of the key outcomes, this participatory approach, in which local realities and factors inform the key outcomes and help to strategies for a particular region, results in a robust strategy for sustainably intensifying production in key food producing regions. The creation of an interactive, locally embedded, decision-support management and education tool can help to close the gap between farmer knowledge and production, increase on-farm adoption of sustainable farming strategies and techniques, and optimize farm productivity.
Keywords: Agriculture, decision-support management tools, GIS, sustainable intensification.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 8491774 Application of Artificial Neural Network to Classification Surface Water Quality
Authors: S. Wechmongkhonkon, N.Poomtong, S. Areerachakul
Abstract:
Water quality is a subject of ongoing concern. Deterioration of water quality has initiated serious management efforts in many countries. This study endeavors to automatically classify water quality. The water quality classes are evaluated using 6 factor indices. These factors are pH value (pH), Dissolved Oxygen (DO), Biochemical Oxygen Demand (BOD), Nitrate Nitrogen (NO3N), Ammonia Nitrogen (NH3N) and Total Coliform (TColiform). The methodology involves applying data mining techniques using multilayer perceptron (MLP) neural network models. The data consisted of 11 sites of canals in Dusit district in Bangkok, Thailand. The data is obtained from the Department of Drainage and Sewerage Bangkok Metropolitan Administration during 2007-2011. The results of multilayer perceptron neural network exhibit a high accuracy multilayer perception rate at 96.52% in classifying the water quality of Dusit district canal in Bangkok Subsequently, this encouraging result could be applied with plan and management source of water quality.Keywords: artificial neural network, classification, surface water quality
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 32091773 Elliptical Features Extraction Using Eigen Values of Covariance Matrices, Hough Transform and Raster Scan Algorithms
Authors: J. Prakash, K. Rajesh
Abstract:
In this paper, we introduce a new method for elliptical object identification. The proposed method adopts a hybrid scheme which consists of Eigen values of covariance matrices, Circular Hough transform and Bresenham-s raster scan algorithms. In this approach we use the fact that the large Eigen values and small Eigen values of covariance matrices are associated with the major and minor axial lengths of the ellipse. The centre location of the ellipse can be identified using circular Hough transform (CHT). Sparse matrix technique is used to perform CHT. Since sparse matrices squeeze zero elements and contain a small number of nonzero elements they provide an advantage of matrix storage space and computational time. Neighborhood suppression scheme is used to find the valid Hough peaks. The accurate position of circumference pixels is identified using raster scan algorithm which uses the geometrical symmetry property. This method does not require the evaluation of tangents or curvature of edge contours, which are generally very sensitive to noise working conditions. The proposed method has the advantages of small storage, high speed and accuracy in identifying the feature. The new method has been tested on both synthetic and real images. Several experiments have been conducted on various images with considerable background noise to reveal the efficacy and robustness. Experimental results about the accuracy of the proposed method, comparisons with Hough transform and its variants and other tangential based methods are reported.Keywords: Circular Hough transform, covariance matrix, Eigen values, ellipse detection, raster scan algorithm.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 26411772 Effects of Sea Water Level Fluctuations on Seismic Response of Jacket Type Offshore Platforms
Authors: M. Rad, M. Dolatshahi Pirooz, M. Esmayili
Abstract:
To understand the seismic behavior of the offshore structures, the dynamic interaction of the water-structure-soil should be assessed. In this regard the role of the water dynamic properties in magnifying or reducing of the effects of earthquake induced motions on offshore structures haven't been investigated in precise manner in available literature. In this paper the sea water level fluctuations effects on the seismic behavior of a sample of offshore structures has been investigated by emphasizing on the water-structure interaction phenomenon. For this purpose a two dimensional finite element model of offshore structures as well as surrounded water has been developed using ANSYS software. The effect of soil interaction with embedded pile foundation has been imposed by using a series of nonlinear springs in horizontal and vertical directions in soil-piles contact points. In the model, the earthquake induced motions have been applied on springs and consequently the motions propagated upward to the structure and surrounded water. As a result of numerical study, the horizontal deformations of the offshore deck as well as internal force and buckling coefficient in structural elements have been recorded and controlled with and without water presence. In part of study a parametric study has been accomplished on sea water level fluctuations and effect of this parameter has been studied on the aforementioned numerical results.Keywords: Fluid-Structure Interaction, Jacket, Sea Water Level, Seismic Loading.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20851771 Establishing a Change Management Model for Precision Machinery Industry in Taiwan
Authors: Feng-Tsung Cheng, Shu-Li Wang, Mei-Fang Wu, Hui-Yu Chuang
Abstract:
The rapid development technology and widespread Internet make business environment changing a lot. In order to stand in the global market and to keep subsistence, “changing” is unspoken rule for the company’s survival. The purpose of this paper is building up change model by using SWOT, strategy map, KPI and change management theory. The research findings indicate that the company needs to deal with employee’s resistance emotion firstly before building up change model. The ways of providing performance appraisal reward, consulting and counseling mechanisms that will great help to achieve reducing staff negative emotions and motivate staff’s efficiencies also. To revise strategy map, modify corporate culture, and improve internal operational processes which is based on change model. Through the change model, the increasing growth rate of net income helps company to achieve the goals and be a leading brand of precision machinery industry.
Keywords: Organizational change, SWOT analysis, strategy maps, performance indicators.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17991770 The Advantages of Integration for Social Systems – Evidence from the Automobile Industry
Authors: Waldemiro Francisco Sorte Junior
Abstract:
The Japanese integrative approach to social systems can be observed in supply chain management as well as in the relationship between public and private sectors. Both the Lean Production System and the Developmental State Model are characterized by efforts towards the achievement of mutual goals, resulting in initiatives for capacity building which emphasize the system level. In Brazil, although organizations undertake efforts to build capabilities at the individual and organizational levels, the system level is being neglected. Fieldwork data confirmed the findings of other studies in terms of the lack of integration in supply chain management in the Brazilian automobile industry. Moreover, due to the absence of an active role of the Brazilian state in its relationship with the private sector, automakers are not fully exploiting the opportunities in the domestic and regional markets. For promoting a higher level of economic growth as well as to increase the degree of spill-over of technologies and techniques, a more integrative approach is needed.Keywords: Integration, Lean Production System, DevelopmentalState Model, Brazilian automobile industry.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15281769 A Case Study on Appearance Based Feature Extraction Techniques and Their Susceptibility to Image Degradations for the Task of Face Recognition
Authors: Vitomir Struc, Nikola Pavesic
Abstract:
Over the past decades, automatic face recognition has become a highly active research area, mainly due to the countless application possibilities in both the private as well as the public sector. Numerous algorithms have been proposed in the literature to cope with the problem of face recognition, nevertheless, a group of methods commonly referred to as appearance based have emerged as the dominant solution to the face recognition problem. Many comparative studies concerned with the performance of appearance based methods have already been presented in the literature, not rarely with inconclusive and often with contradictory results. No consent has been reached within the scientific community regarding the relative ranking of the efficiency of appearance based methods for the face recognition task, let alone regarding their susceptibility to appearance changes induced by various environmental factors. To tackle these open issues, this paper assess the performance of the three dominant appearance based methods: principal component analysis, linear discriminant analysis and independent component analysis, and compares them on equal footing (i.e., with the same preprocessing procedure, with optimized parameters for the best possible performance, etc.) in face verification experiments on the publicly available XM2VTS database. In addition to the comparative analysis on the XM2VTS database, ten degraded versions of the database are also employed in the experiments to evaluate the susceptibility of the appearance based methods on various image degradations which can occur in "real-life" operating conditions. Our experimental results suggest that linear discriminant analysis ensures the most consistent verification rates across the tested databases.
Keywords: Biometrics, face recognition, appearance based methods, image degradations, the XM2VTS database.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2284