Search results for: Novelty Measure.
307 Consumer Online Shopping Behavior: The Effect of Internet Marketing Environment, Product Characteristics, Familiarity and Confidence, and Promotional Offer
Authors: Norazah Mohd Suki, Norbayah Mohd Suki
Abstract:
Online shopping enables consumers to search for information and purchase products or services through direct interaction with online store. This study aims to examine the effect of Internet marketing environment, product characteristics, familiarity and confidence, and promotional offers on consumer online shopping behavior. 200 questionnaires were distributed to the respondents, who are students and staff at a public university in the Federal Territory of Labuan, Malaysia, following simple random sampling as a means of data collection. Multiple regression analysis was used as a statistical measure to determine the strength of the relationship between one dependent variable and a series of other independent variables. Results revealed that familiarity and confidence was found to greatly influence consumer online shopping behavior followed by promotional offers. A clear understanding of consumer online shopping behavior can help marketing managers predict the online shopping rate and evaluate the future growth of online commerce.
Keywords: Internet Marketing Environment, Product Characteristics, Multiple Regression Analysis, Malaysia.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 12068306 Fundamental Equation of Complete Factor Synergetics of Complex Systems with Normalization of Dimension
Authors: Li Zong-Cheng
Abstract:
It is by reason of the unified measure of varieties of resources and the unified processing of the disposal of varieties of resources, that these closely related three of new basic models called the resources assembled node and the disposition integrated node as well as the intelligent organizing node are put forth in this paper; the three closely related quantities of integrative analytical mechanics including the disposal intensity and disposal- weighted intensity as well as the charge of resource charge are set; and then the resources assembled space and the disposition integrated space as well as the intelligent organizing space are put forth. The system of fundamental equations and model of complete factor synergetics is preliminarily approached for the general situation in this paper, to form the analytical base of complete factor synergetics. By the essential variables constituting this system of equations we should set twenty variables respectively with relation to the essential dynamical effect, external synergetic action and internal synergetic action of the system.
Keywords: complex system, disposal of resources, completefactor synergetics, fundamental equation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1419305 Fundamentals of Performance Management in the World of Public Service Organisations
Authors: Daniella Kucsma
Abstract:
The examination of the Public Service Organization’s performance evaluation includes several steps that help public organizations to develop a more efficient system. Public sector organizations have different characteristics than the competitive sector, so it can be stated that other/new elements become more important in their performance processes. The literature in this area is diverse, so highlighting an indicator system can be useful for introducing a system, but it is also worthwhile to measure the specific elements of the organization. In the case of a public service organization, due to the service obligation, it is usually possible to talk about a high number of users, so compliance is more difficult. For the organization, it is an important target to place great emphasis on the increase of service standards and the development of related processes. In this research, the health sector is given a prominent role, as it is a sensitive area where both organizational and individual performance is important for all participants. As a primary step, the content of the strategy is decisive, as this is important for the efficient structure of the process. When designing any system, it is important to review the expectations of the stakeholders, as this is primary when considering the design. The goal of this paper is to build the foundations of a performance management and indexing framework that can help a hospital to provide effective feedback and a direction that is important in assessing and developing a service and can become a management philosophy.
Keywords: health sector, public sector, performance management, strategy
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 544304 Estimation of Vertical Handover Probability in an Integrated UMTS and WLAN Networks
Authors: Diganta Kumar Pathak, Manashjyoti Bhuyan, Vaskar Deka
Abstract:
Vertical Handover(VHO) among different communication technologies ensuring uninterruption and service continuity is one of the most important performance parameter in Heterogenous networks environment. In an integrated Universal Mobile Telecommunicatin System(UMTS) and Wireless Local Area Network(WLAN), WLAN is given an inherent priority over UMTS because of its high data rates with low cost. Therefore mobile users want to be associated with WLAN maximum of the time while roaming, to enjoy best possible services with low cost. That encourages reduction of number of VHO. In this work the reduction of number of VHO with respect to varying number of WLAN Access Points(APs) in an integrated UMTS and WLAN network is investigated through simulation to provide best possible cost effective service to the users. The simulation has been carried out for an area (7800 × 9006)m2 where COST-231 Hata model and 3GPP (TR 101 112 V 3.1.0) specified models are used for WLAN and UMTS path loss models respectively. The handover decision is triggered based on the received signal level as compared to the fade margin. Fade margin gives a probabilistic measure of the reliability of the communication link. A relationship between number of WLAN APs and the number of VHO is also established in this work.
Keywords: VHO, UMTS, WLAN, MT, AP, BS.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2036303 The Application of Distributed Optical Strain Sensing to Measure Rock Bolt Deformation Subject to Bedding Shear
Authors: Thomas P. Roper, Brad Forbes, Jurij Karlovšek
Abstract:
Shear displacement along bedding defects is a well-recognised behaviour when tunnelling and mining in stratified rock. This deformation can affect the durability and integrity of installed rock bolts. In-situ monitoring of rock bolt deformation under bedding shear cannot be accurately derived from traditional strain gauge bolts as sensors are too large and spaced too far apart to accurately assess concentrated displacement along discrete defects. A possible solution to this is the use of fiber optic technologies developed for precision monitoring. Distributed Optic Sensor (DOS) embedded rock bolts were installed in a tunnel project with the aim of measuring the bolt deformation profile under significant shear displacements. This technology successfully measured the 3D strain distribution along the bolts when subjected to bedding shear and resolved the axial and lateral strain constituents in order to determine the deformational geometry of the bolts. The results are compared well with the current visual method for monitoring shear displacement using borescope holes, considering this method as suitable.
Keywords: Distributed optical strain sensing, geotechnical monitoring, rock bolt stain measurement, bedding shear displacement.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 933302 Effective Design Factors for Bicycle-Friendly Streets
Authors: Z. Asadi-Shekari, M. Moeinaddini, M. Zaly Shah, A. Hamzah
Abstract:
Bicycle Level of Service (BLOS) is a measure for evaluating street conditions for cyclists. Currently, various methods are proposed for BLOS. These analytical methods however have some drawbacks: they usually assume cyclists as users that can share street facilities with motorized vehicles, it is not easy to link them to design process and they are not easy to follow. In addition, they only support a narrow range of cycling facilities and may not be applicable for all situations. Along this, the current paper introduces various effective design factors for bicycle-friendly streets. This study considers cyclists as users of streets who have special needs and facilities. Therefore, the key factors that influence BLOS based on different cycling facilities that are proposed by developed guidelines and literature are identified. The combination of these factors presents a complete set of effective design factors for bicycle-friendly streets. In addition, the weight of each factor in existing BLOS models is estimated and these effective factors are ranked based on these weights. These factors and their weights can be used in further studies to propose special bicycle-friendly street design model.Keywords: Bicycle level of service, bicycle-friendly streets, cycling facilities, rating system, urban streets.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2281301 Rheological Characteristics of Ice Slurries Based on Propylene- and Ethylene-Glycol at High Ice Fractions
Authors: Senda Trabelsi, Sébastien Poncet, Michel Poirier
Abstract:
Ice slurries are considered as a promising phase-changing secondary fluids for air-conditioning, packaging or cooling industrial processes. An experimental study has been here carried out to measure the rheological characteristics of ice slurries. Ice slurries consist in a solid phase (flake ice crystals) and a liquid phase. The later is composed of a mixture of liquid water and an additive being here either (1) Propylene-Glycol (PG) or (2) Ethylene-Glycol (EG) used to lower the freezing point of water. Concentrations of 5%, 14% and 24% of both additives are investigated with ice mass fractions ranging from 5% to 85%. The rheological measurements are carried out using a Discovery HR-2 vane-concentric cylinder with four full-length blades. The experimental results show that the behavior of ice slurries is generally non-Newtonian with shear-thinning or shear-thickening behaviors depending on the experimental conditions. In order to determine the consistency and the flow index, the Herschel-Bulkley model is used to describe the behavior of ice slurries. The present results are finally validated against an experimental database found in the literature and the predictions of an Artificial Neural Network model.
Keywords: Ice slurry, propylene-glycol, ethylene-glycol, rheology, artificial neural network.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1127300 Data Integrity: Challenges in Health Information Systems in South Africa
Authors: T. Thulare, M. Herselman, A. Botha
Abstract:
Poor system use, including inappropriate design of health information systems, causes difficulties in communication with patients and increased time spent by healthcare professionals in recording the necessary health information for medical records. System features like pop-up reminders, complex menus, and poor user interfaces can make medical records far more time consuming than paper cards as well as affect decision-making processes. Although errors associated with health information and their real and likely effect on the quality of care and patient safety have been documented for many years, more research is needed to measure the occurrence of these errors and determine the causes to implement solutions. Therefore, the purpose of this paper is to identify data integrity challenges in hospital information systems through a scoping review and based on the results provide recommendations on how to manage these. Only 34 papers were found to be most suitable out of 297 publications initially identified in the field. The results indicated that human and computerized systems are the most common challenges associated with data integrity and factors such as policy, environment, health workforce, and lack of awareness attribute to these challenges but if measures are taken the data integrity challenges can be managed.
Keywords: Data integrity, data integrity challenges, hospital information systems, South Africa.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1376299 Encryption Efficiency Analysis and Security Evaluation of RC6 Block Cipher for Digital Images
Authors: Hossam El-din H. Ahmed, Hamdy M. Kalash, Osama S. Farag Allah
Abstract:
This paper investigates the encryption efficiency of RC6 block cipher application to digital images, providing a new mathematical measure for encryption efficiency, which we will call the encryption quality instead of visual inspection, The encryption quality of RC6 block cipher is investigated among its several design parameters such as word size, number of rounds, and secret key length and the optimal choices for the best values of such design parameters are given. Also, the security analysis of RC6 block cipher for digital images is investigated from strict cryptographic viewpoint. The security estimations of RC6 block cipher for digital images against brute-force, statistical, and differential attacks are explored. Experiments are made to test the security of RC6 block cipher for digital images against all aforementioned types of attacks. Experiments and results verify and prove that RC6 block cipher is highly secure for real-time image encryption from cryptographic viewpoint. Thorough experimental tests are carried out with detailed analysis, demonstrating the high security of RC6 block cipher algorithm. So, RC6 block cipher can be considered to be a real-time secure symmetric encryption for digital images.
Keywords: Block cipher, Image encryption, Encryption quality, and Security analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2425298 A Cuckoo Search with Differential Evolution for Clustering Microarray Gene Expression Data
Authors: M. Pandi, K. Premalatha
Abstract:
A DNA microarray technology is a collection of microscopic DNA spots attached to a solid surface. Scientists use DNA microarrays to measure the expression levels of large numbers of genes simultaneously or to genotype multiple regions of a genome. Elucidating the patterns hidden in gene expression data offers a tremendous opportunity for an enhanced understanding of functional genomics. However, the large number of genes and the complexity of biological networks greatly increase the challenges of comprehending and interpreting the resulting mass of data, which often consists of millions of measurements. It is handled by clustering which reveals the natural structures and identifying the interesting patterns in the underlying data. In this paper, gene based clustering in gene expression data is proposed using Cuckoo Search with Differential Evolution (CS-DE). The experiment results are analyzed with gene expression benchmark datasets. The results show that CS-DE outperforms CS in benchmark datasets. To find the validation of the clustering results, this work is tested with one internal and one external cluster validation indexes.
Keywords: DNA, Microarray, genomics, Cuckoo Search, Differential Evolution, Gene expression data, Clustering.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1483297 Evaluation of Soil Stiffness and Strength for Quality Control of Compacted Earthwork
Authors: A. Sawangsuriya, T. B. Edil
Abstract:
Microstructure and fabric of soils play an important role on structural properties e.g. stiffness and strength of compacted earthwork. Traditional quality control monitoring based on moisturedensity tests neither reflects the variability of soil microstructure nor provides a direct assessment of structural property, which is the ultimate objective of the earthwork quality control. Since stiffness and strength are sensitive to soil microstructure and fabric, any independent test methods that provide simple, rapid, and direct measurement of stiffness and strength are anticipated to provide an effective assessment of compacted earthen materials’ uniformity. In this study, the soil stiffness gauge (SSG) and the dynamic cone penetrometer (DCP) were respectively utilized to measure and monitor the stiffness and strength in companion with traditional moisture-density measurements of various earthen materials used in Thailand road construction projects. The practical earthwork quality control criteria are presented herein in order to assure proper earthwork quality control and uniform structural property of compacted earthworks.Keywords: Dynamic cone penetrometer, moisture content, relative compaction, soil stiffness gauge, structural property.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2325296 The Index of Sustainable Functionality: An Application for Measuring Sustainability
Authors: G.T. Cirella, L. Tao
Abstract:
The index of sustainable functionality (ISF) is an adaptive, multi-criteria technique that is used to measure sustainability; it is a concept that can be transposed to many regions throughout the world. An ISF application of the Southern Regional Organisation of Councils (SouthROC) in South East Queensland (SEQ) – the fastest growing region in Australia – indicated over a 25 year period an increase of over 10% level of functionality from 58.0% to 68.3%. The ISF of SouthROC utilised methodologies that derived from an expert panel based approach. The overall results attained an intermediate level of functionality which amounted to related concerns of economic progress and lack of social awareness. Within the region, a solid basis for future testing by way of measured changes and developed trends can be established. In this regard as management tool, the ISF record offers support for regional sustainability practice and decision making alike. This research adaptively analyses sustainability – a concept that is lacking throughout much of the academic literature and any reciprocal experimentation. This lack of knowledge base has been the emphasis of where future sustainability research can grow from and prove useful in rapidly growing regions. It is the intentions of this research to help further develop the notions of index-based quantitative sustainability.
Keywords: Environmental engineering, index of sustainable functionality, sustainability indicators, sustainable development.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2379295 A Novel Computer Vision Method for Evaluating Deformations of Fibers Cross Section in False Twist Textured Yarns
Authors: Dariush Semnani, Mehdi Ahangareianabhari, Hossein Ghayoor
Abstract:
In recent five decades, textured yarns of polyester fiber produced by false twist method are the most important and mass-produced manmade fibers. There are many parameters of cross section which affect the physical and mechanical properties of textured yarns. These parameters are surface area, perimeter, equivalent diameter, large diameter, small diameter, convexity, stiffness, eccentricity, and hydraulic diameter. These parameters were evaluated by digital image processing techniques. To find trends between production criteria and evaluated parameters of cross section, three criteria of production line have been adjusted and different types of yarns were produced. These criteria are temperature, drafting ratio, and D/Y ratio. Finally the relations between production criteria and cross section parameters were considered. The results showed that the presented technique can recognize and measure the parameters of fiber cross section in acceptable accuracy. Also, the optimum condition of adjustments has been estimated from results of image analysis evaluation.Keywords: Computer Vision, Cross Section Analysis, Fibers Deformation, Textured Yarn
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1647294 Influence of Pressure from Compression Textile Bands: Their Using in the Treatment of Venous Human Leg Ulcers
Authors: Bachir Chemani, Rachid Halfaoui
Abstract:
The aim of study was to evaluate pressure distribution characteristics of the elastic textile bandages using two instrumental techniques: a prototype Instrument and a load Transference. The prototype instrument which simulates shape of real leg has pressure sensors which measure bandage pressure. Using this instrument, the results show that elastic textile bandages presents different pressure distribution characteristics and none produces a uniform distribution around lower limb.
The load transference test procedure is used to determine whether a relationship exists between elastic textile bandage structure and pressure distribution characteristics. The test procedure assesses degree of load, directly transferred through a textile when loads series are applied to bandaging surface. A range of weave fabrics was produced using needle weaving machine and a sewing technique. A textile bandage was developed with optimal characteristics far superior pressure distribution than other bandages. From results, we find that theoretical pressure is not consistent exactly with practical pressure. It is important in this study to make a practical application for specialized nurses in order to verify the results and draw useful conclusions for predicting the use of this type of elastic band.
Keywords: Textile, cotton, pressure, venous ulcers, elastic.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1749293 Treatment of the Modern Management Mechanism of the Debris Flow Processes Expected in the Mletiskhevi
Authors: G. Chakhaia, S. Gogilava, L. Tsulukidze, Z. Laoshvili, I. Khubulava, S. Bosikashvili, T. Gugushvili
Abstract:
The work reviewed and evaluated various genesis debris flow phenomena recently formatted in the Mletiskhevi, accordingly it revealed necessity of treatment modern debris flow against measures. Based on this, it is proposed the debris flow against truncated semi cone shape construction, which elements are contained in the car’s secondary tires. its constituent elements (sections), due to the possibilities of amortization and geometric shapes is effective and sustainable towards debris flow hitting force. The construction is economical, because after crossing the debris flows in the river bed, the riverbed is not cleanable, also the elements of the building are resource saving. For assessment of influence of cohesive debris flow at the construction and evaluation of the construction effectiveness have been implemented calculation in the specific assumptions with approved methodology. According to the calculation, it was established that after passing debris flow in the debris flow construction (in 3 row case) its hitting force reduces 3 times, that causes reduce of debris flow speed and kinetic energy, as well as sedimentation on a certain section of water drain in the lower part of the construction. Based on the analysis and report on the debris flow against construction, it can be said that construction is effective, inexpensive, technically relatively easy-to-reach measure, that’s why its implementation is prospective.
Keywords: Construction, debris flow, sections, theoretical calculation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 399292 Recycling Construction Waste Materials to Reduce the Environmental Pollutants
Authors: Mehrdad Abkenari, Alireza Rezaei, Naghmeh Pournayeb
Abstract:
There have recently been many studies and investments in developed and developing countries regarding the possibility of recycling construction waste, which are still ongoing. Since the term 'construction waste' covers a vast spectrum of materials in constructing buildings, roads and etc., many investigations are required to measure their technical performance in use as well as their time and place of use. Concrete is among the major and fundamental materials used in current construction industry. Along with the rise of population in developing countries, it is desperately required to meet the people's primary need in construction industry and on the other hand, dispose existing wastes for reducing the amount of environmental pollutants. Restrictions of natural resources and environmental pollution are the most important problems encountered by civil engineers. Reusing construction waste is an important and economic approach that not only assists the preservation of environment but also, provides us with primary raw materials. In line with consistent municipal development in disposal and reuse of construction waste, several approaches including, management of construction waste and materials, materials recycling and innovation and new inventions in materials have been predicted. This article has accordingly attempted to study the activities related to recycling of construction wastes and then, stated the economic, quantitative, qualitative and environmental results obtained.
Keywords: Civil engineering, environment, recycling, construction waste.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2930291 Study on Construction of 3D Topography by UAV-Based Images
Authors: Yun-Yao Chi, Chieh-Kai Tsai, Dai-Ling Li
Abstract:
In this paper, a method of fast 3D topography modeling using the high-resolution camera images is studied based on the characteristics of Unmanned Aerial Vehicle (UAV) system for low altitude aerial photogrammetry and the need of three dimensional (3D) urban landscape modeling. Firstly, the existing high-resolution digital camera with special design of overlap images is designed by reconstructing and analyzing the auto-flying paths of UAVs, which improves the self-calibration function to achieve the high precision imaging by software, and further increased the resolution of the imaging system. Secondly, several-angle images including vertical images and oblique images gotten by the UAV system are used for the detail measure of urban land surfaces and the texture extraction. Finally, the aerial photography and 3D topography construction are both developed in campus of Chang-Jung University and in Guerin district area in Tainan, Taiwan, provide authentication model for construction of 3D topography based on combined UAV-based camera images from system. The results demonstrated that the UAV system for low altitude aerial photogrammetry can be used in the construction of 3D topography production, and the technology solution in this paper offers a new, fast, and technical plan for the 3D expression of the city landscape, fine modeling and visualization.
Keywords: 3D, topography, UAV, images.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 802290 Experimental Investigation to Find Transition Temperature of VG-30 Binder
Authors: D. Latha, V. Sunitha, Samson Mathew
Abstract:
In India, most of the pavement is laid by bituminous road and the consumption of binder is high for pavement construction and also modified binders are used to satisfy any specific pavement requirement. Since the binders are visco-elastic material which is having the mechanical properties of binder transition from viscoelastic solid to visco-elastic fluid. In this paper, two different protocols were used to measure the viscosity property of binder using a Brookfield Viscometer and there is a need to find the appropriate mixing and compaction temperatures of various types of binders which can result in complete aggregate coating and adequate field density of HMA mixtures. The aim of this work is to find the transition temperature from Non-Newtonian behavior to Newtonian behavior of the binder by adopting a steady shear protocol and the shear rate ramp protocol. The transition from non-Newtonian to Newtonian can occur through an increase of temperature and shear of the material. The test has been conducted for unmodified binder VG 30. The transition temperature was found in the unmodified binder VG is 120oC. Therefore, the application of both modified binder and unmodified binder in the pavement construction needs to be studied properly by considering temperature and traffic loading factors of the respective project site.Keywords: Unmodified and modified binders, Brookfield Viscometer, transition temperature, steady shear, shear rate protocol.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1875289 Enhancing the Performance of Wireless Sensor Networks Using Low Power Design
Authors: N. Mahendran, R. Madhuranthi
Abstract:
Wireless sensor networks (WSNs), are constantly in demand to process information more rapidly with less energy and area cost. Presently, processor based solutions have difficult to achieve high processing speed with low-power consumption. This paper presents a simple and accurate data processing scheme for low power wireless sensor node, based on reduced number of processing element (PE). The presented model provides a simple recursive structure (SRS) to process the sampled data in the wireless sensor environment and to reduce the power consumption in wireless sensor node. Based on this model, to process the incoming samples and produce a smaller amount of data sufficient to reconstruct the original signal. The ModelSim simulator used to simulate SRS structure. Functional simulation is carried out for the validation of the presented architecture. Xilinx Power Estimator (XPE) tool is used to measure the power consumption. The experimental results show the average power consumption of 91 mW; this is 42% improvement compared to the folded tree architecture.Keywords: Power consumption, energy efficiency, low power WSN node, recursive structure, sleep/wake scheduling.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1014288 ISC–Intelligent Subspace Clustering, A Density Based Clustering Approach for High Dimensional Dataset
Authors: Sunita Jahirabadkar, Parag Kulkarni
Abstract:
Many real-world data sets consist of a very high dimensional feature space. Most clustering techniques use the distance or similarity between objects as a measure to build clusters. But in high dimensional spaces, distances between points become relatively uniform. In such cases, density based approaches may give better results. Subspace Clustering algorithms automatically identify lower dimensional subspaces of the higher dimensional feature space in which clusters exist. In this paper, we propose a new clustering algorithm, ISC – Intelligent Subspace Clustering, which tries to overcome three major limitations of the existing state-of-art techniques. ISC determines the input parameter such as є – distance at various levels of Subspace Clustering which helps in finding meaningful clusters. The uniform parameters approach is not suitable for different kind of databases. ISC implements dynamic and adaptive determination of Meaningful clustering parameters based on hierarchical filtering approach. Third and most important feature of ISC is the ability of incremental learning and dynamic inclusion and exclusions of subspaces which lead to better cluster formation.
Keywords: Density based clustering, high dimensional data, subspace clustering, dynamic parameter setting.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2018287 Improved Feature Processing for Iris Biometric Authentication System
Authors: Somnath Dey, Debasis Samanta
Abstract:
Iris-based biometric authentication is gaining importance in recent times. Iris biometric processing however, is a complex process and computationally very expensive. In the overall processing of iris biometric in an iris-based biometric authentication system, feature processing is an important task. In feature processing, we extract iris features, which are ultimately used in matching. Since there is a large number of iris features and computational time increases as the number of features increases, it is therefore a challenge to develop an iris processing system with as few as possible number of features and at the same time without compromising the correctness. In this paper, we address this issue and present an approach to feature extraction and feature matching process. We apply Daubechies D4 wavelet with 4 levels to extract features from iris images. These features are encoded with 2 bits by quantizing into 4 quantization levels. With our proposed approach it is possible to represent an iris template with only 304 bits, whereas existing approaches require as many as 1024 bits. In addition, we assign different weights to different iris region to compare two iris templates which significantly increases the accuracy. Further, we match the iris template based on a weighted similarity measure. Experimental results on several iris databases substantiate the efficacy of our approach.Keywords: Iris recognition, biometric, feature processing, patternrecognition, pattern matching.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2139286 A Novel Nucleus-Based Classifier for Discrimination of Osteoclasts and Mesenchymal Precursor Cells in Mouse Bone Marrow Cultures
Authors: Andreas Heindl, Alexander K. Seewald, Martin Schepelmann, Radu Rogojanu, Giovanna Bises, Theresia Thalhammer, Isabella Ellinger
Abstract:
Bone remodeling occurs by the balanced action of bone resorbing osteoclasts (OC) and bone-building osteoblasts. Increased bone resorption by excessive OC activity contributes to malignant and non-malignant diseases including osteoporosis. To study OC differentiation and function, OC formed in in vitro cultures are currently counted manually, a tedious procedure which is prone to inter-observer differences. Aiming for an automated OC-quantification system, classification of OC and precursor cells was done on fluorescence microscope images based on the distinct appearance of fluorescent nuclei. Following ellipse fitting to nuclei, a combination of eight features enabled clustering of OC and precursor cell nuclei. After evaluating different machine-learning techniques, LOGREG achieved 74% correctly classified OC and precursor cell nuclei, outperforming human experts (best expert: 55%). In combination with the automated detection of total cell areas, this system allows to measure various cell parameters and most importantly to quantify proteins involved in osteoclastogenesis.Keywords: osteoclasts, machine learning, ellipse fitting.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1913285 Exergy Analysis of a Solar Humidification- Dehumidification Desalination Unit
Authors: Mohammed A. Elhaj, Jamal S. Yassin
Abstract:
This paper presents the exergy analysis of a desalination unit using humidification-dehumidification process. Here, this unit is considered as a thermal system with three main components, which are the heating unit by using a solar collector, the evaporator or the humidifier, and the condenser or the dehumidifier. In these components the exergy is a measure of the quality or grade of energy and it can be destroyed in them. According to the second law of thermodynamics this destroyed part is due to irreversibilities which must be determined to obtain the exergetic efficiency of the system. In the current paper a computer program has been developed using visual basic to determine the exergy destruction and the exergetic efficiencies of the components of the desalination unit at variable operation conditions such as feed water temperature, outlet air temperature, air to feed water mass ratio and salinity, in addition to cooling water mass flow rate and inlet temperature, as well as quantity of solar irradiance. The results obtained indicate that the exergy efficiency of the humidifier increases by increasing the mass ratio and decreasing the outlet air temperature. In the other hand the exergy efficiency of the condenser increases with the increase of this ratio and also with the increase of the outlet air temperature.
Keywords: Exergy analysis, desalination, solar, humidifier, condenser.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2471284 Measuring Process Component Design on Achieving Managerial Goals
Authors: Eakong Atiptamvaree, Twittie Senivongse
Abstract:
Process-oriented software development is a new software development paradigm in which software design is modeled by a business process which is in turn translated into a process execution language for execution. The building blocks of this paradigm are software units that are composed together to work according to the flow of the business process. This new paradigm still exhibits the characteristic of the applications built with the traditional software component technology. This paper discusses an approach to apply a traditional technique for software component fabrication to the design of process-oriented software units, called process components. These process components result from decomposing a business process of a particular application domain into subprocesses, and these process components can be reused to design the business processes of other application domains. The decomposition considers five managerial goals, namely cost effectiveness, ease of assembly, customization, reusability, and maintainability. The paper presents how to design or decompose process components from a business process model and measure some technical features of the design that would affect the managerial goals. A comparison between the measurement values from different designs can tell which process component design is more appropriate for the managerial goals that have been set. The proposed approach can be applied in Web Services environment which accommodates process-oriented software development.Keywords: Business Process Model, Managerial Goals, ProcessComponent.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1513283 Simultaneous Optimization of Machining Parameters and Tool Geometry Specifications in Turning Operation of AISI1045 Steel
Authors: Farhad Kolahan, Mohsen Manoochehri, Abbas Hosseini
Abstract:
Machining is an important manufacturing process used to produce a wide variety of metallic parts. Among various machining processes, turning is one of the most important one which is employed to shape cylindrical parts. In turning, the quality of finished product is measured in terms of surface roughness. In turn, surface quality is determined by machining parameters and tool geometry specifications. The main objective of this study is to simultaneously model and optimize machining parameters and tool geometry in order to improve the surface roughness for AISI1045 steel. Several levels of machining parameters and tool geometry specifications are considered as input parameters. The surface roughness is selected as process output measure of performance. A Taguchi approach is employed to gather experimental data. Then, based on signal-to-noise (S/N) ratio, the best sets of cutting parameters and tool geometry specifications have been determined. Using these parameters values, the surface roughness of AISI1045 steel parts may be minimized. Experimental results are provided to illustrate the effectiveness of the proposed approach.
Keywords: Taguchi method, turning parameters, tool geometry specifications, S/N ratio, statistical analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2325282 Analysis of Lead Time Delays in Supply Chain: A Case Study
Authors: Abdel-Aziz M. Mohamed, Nermeen Coutry
Abstract:
Lead time is a critical measure of a supply chain's performance. It impacts both the customer satisfactions as well as the total cost of inventory. This paper presents the result of a study on the analysis of the customer order lead-time for a multinational company. In the study, the lead time was divided into three stages respectively: order entry, order fulfillment, and order delivery. A sample of size 2,425 order lines was extracted from the company's records to use for this study. The sample data entails information regarding customer orders from the time of order entry until order delivery. Data regarding the lead time of each stage for different orders were also provided. Summary statistics on lead time data reveals that about 30% of the orders were delivered later than the scheduled due date. The result of the multiple linear regression analysis technique revealed that component type, logistics parameter, order size and the customer type have significant impacts on lead time. Data analysis on the stages of lead time indicates that stage 2 consumed over 50% of the lead time. Pareto analysis was made to study the reasons for the customer order delay in each stage. Recommendation was given to resolve the problem.Keywords: Lead time reduction, customer satisfaction, service quality, statistical analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6689281 A Decision Support System Based on Leprosy Scales
Authors: Dennys Robson Girardi, Hugo Bulegon, Claudia Maria Moro Barra
Abstract:
Leprosy is an infectious disease caused by Mycobacterium Leprae, this disease, generally, compromises the neural fibers, leading to the development of disability. Disabilities are changes that limit daily activities or social life of a normal individual. When comes to leprosy, the study of disability considered the functional limitation (physical disabilities), the limitation of activity and social participation, which are measured respectively by the scales: EHF, SALSA and PARTICIPATION SCALE. The objective of this work is to propose an on-line monitoring of leprosy patients, which is based on information scales EHF, SALSA and PARTICIPATION SCALE. It is expected that the proposed system is applied in monitoring the patient during treatment and after healing therapy of the disease. The correlations that the system is between the scales create a variety of information, presented the state of the patient and full of changes or reductions in disability. The system provides reports with information from each of the scales and the relationships that exist between them. This way, health professionals, with access to patient information, can intervene with techniques for the Prevention of Disability. Through the automated scale, the system shows the level of the patient and allows the patient, or the responsible, to take a preventive measure. With an online system, it is possible take the assessments and monitor patients from anywhere.Keywords: Leprosy, Medical Informatics, Decision SupportSystem, Disability.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2048280 A Method for Iris Recognition Based on 1D Coiflet Wavelet
Authors: Agus Harjoko, Sri Hartati, Henry Dwiyasa
Abstract:
There have been numerous implementations of security system using biometric, especially for identification and verification cases. An example of pattern used in biometric is the iris pattern in human eye. The iris pattern is considered unique for each person. The use of iris pattern poses problems in encoding the human iris. In this research, an efficient iris recognition method is proposed. In the proposed method the iris segmentation is based on the observation that the pupil has lower intensity than the iris, and the iris has lower intensity than the sclera. By detecting the boundary between the pupil and the iris and the boundary between the iris and the sclera, the iris area can be separated from pupil and sclera. A step is taken to reduce the effect of eyelashes and specular reflection of pupil. Then the four levels Coiflet wavelet transform is applied to the extracted iris image. The modified Hamming distance is employed to measure the similarity between two irises. This research yields the identification success rate of 84.25% for the CASIA version 1.0 database. The method gives an accuracy of 77.78% for the left eyes of MMU 1 database and 86.67% for the right eyes. The time required for the encoding process, from the segmentation until the iris code is generated, is 0.7096 seconds. These results show that the accuracy and speed of the method is better than many other methods.Keywords: Biometric, iris recognition, wavelet transform.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1906279 Enhancing the Performance of H.264/AVC in Adaptive Group of Pictures Mode Using Octagon and Square Search Pattern
Authors: S. Sowmyayani, P. Arockia Jansi Rani
Abstract:
This paper integrates Octagon and Square Search pattern (OCTSS) motion estimation algorithm into H.264/AVC (Advanced Video Coding) video codec in Adaptive Group of Pictures (AGOP) mode. AGOP structure is computed based on scene change in the video sequence. Octagon and square search pattern block-based motion estimation method is implemented in inter-prediction process of H.264/AVC. Both these methods reduce bit rate and computational complexity while maintaining the quality of the video sequence respectively. Experiments are conducted for different types of video sequence. The results substantially proved that the bit rate, computation time and PSNR gain achieved by the proposed method is better than the existing H.264/AVC with fixed GOP and AGOP. With a marginal gain in quality of 0.28dB and average gain in bitrate of 132.87kbps, the proposed method reduces the average computation time by 27.31 minutes when compared to the existing state-of-art H.264/AVC video codec.Keywords: Block Distortion Measure, Block Matching Algorithms, H.264/AVC, Motion estimation, Search patterns, Shot cut detection.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1731278 A Complexity-Based Approach in Image Compression using Neural Networks
Authors: Hadi Veisi, Mansour Jamzad
Abstract:
In this paper we present an adaptive method for image compression that is based on complexity level of the image. The basic compressor/de-compressor structure of this method is a multilayer perceptron artificial neural network. In adaptive approach different Back-Propagation artificial neural networks are used as compressor and de-compressor and this is done by dividing the image into blocks, computing the complexity of each block and then selecting one network for each block according to its complexity value. Three complexity measure methods, called Entropy, Activity and Pattern-based are used to determine the level of complexity in image blocks and their ability in complexity estimation are evaluated and compared. In training and evaluation, each image block is assigned to a network based on its complexity value. Best-SNR is another alternative in selecting compressor network for image blocks in evolution phase which chooses one of the trained networks such that results best SNR in compressing the input image block. In our evaluations, best results are obtained when overlapping the blocks is allowed and choosing the networks in compressor is based on the Best-SNR. In this case, the results demonstrate superiority of this method comparing with previous similar works and JPEG standard coding.Keywords: Adaptive image compression, Image complexity, Multi-layer perceptron neural network, JPEG Standard, PSNR.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2222