Search results for: digital terrain models
1095 Utilization of Schnerr-Sauer Cavitation Model for Simulation of Cavitation Inception and Super Cavitation
Authors: Mohammadreza Nezamirad, Azadeh Yazdi, Sepideh Amirahmadian, Nasim Sabetpour, Amirmasoud Hamedi
Abstract:
In this study, the Reynolds-Stress-Navier-Stokes framework is utilized to investigate the flow inside the diesel injector nozzle. The flow is assumed to be multiphase as the formation of vapor by pressure drop is visualized. For pressure and velocity linkage, the coupled algorithm is used. Since the cavitation phenomenon inherently is unsteady, the quasi-steady approach is utilized for saving time and resources in the current study. Schnerr-Sauer cavitation model is used, which was capable of predicting flow behavior both at the initial and final steps of the cavitation process. Two different turbulent models were used in this study to clarify which one is more capable in predicting cavitation inception and super-cavitation. It was found that K-ε was more compatible with the Shnerr-Sauer cavitation model; therefore, the mentioned model is used for the rest of this study.
Keywords: CFD, RANS, cavitation, fuel, injector.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5471094 New Methods for E-Commerce Databases Designing in Semantic Web Systems (Modern Systems)
Authors: Karim Heidari, Serajodin Katebi, Ali Reza Mahdavi Far
Abstract:
The purpose of this paper is to study Database Models to use them efficiently in E-commerce websites. In this paper we are going to find a method which can save and retrieve information in Ecommerce websites. Thus, semantic web applications can work with, and we are also going to study different technologies of E-commerce databases and we know that one of the most important deficits in semantic web is the shortage of semantic data, since most of the information is still stored in relational databases, we present an approach to map legacy data stored in relational databases into the Semantic Web using virtually any modern RDF query language, as long as it is closed within RDF. To achieve this goal we study XML structures for relational data bases of old websites and eventually we will come up one level over XML and look for a map from relational model (RDM) to RDF. Noting that a large number of semantic webs get advantage of relational model, opening the ways which can be converted to XML and RDF in modern systems (semantic web) is important.Keywords: E-Commerce, Semantic Web, Database, XML, RDF.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 23121093 Using Self Organizing Feature Maps for Classification in RGB Images
Authors: Hassan Masoumi, Ahad Salimi, Nazanin Barhemmat, Babak Gholami
Abstract:
Artificial neural networks have gained a lot of interest as empirical models for their powerful representational capacity, multi input and output mapping characteristics. In fact, most feedforward networks with nonlinear nodal functions have been proved to be universal approximates. In this paper, we propose a new supervised method for color image classification based on selforganizing feature maps (SOFM). This algorithm is based on competitive learning. The method partitions the input space using self-organizing feature maps to introduce the concept of local neighborhoods. Our image classification system entered into RGB image. Experiments with simulated data showed that separability of classes increased when increasing training time. In additional, the result shows proposed algorithms are effective for color image classification.Keywords: Classification, SOFM, neural network, RGB images.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 23191092 Automated Textile Defect Recognition System Using Computer Vision and Artificial Neural Networks
Authors: Atiqul Islam, Shamim Akhter, Tumnun E. Mursalin
Abstract:
Least Development Countries (LDC) like Bangladesh, whose 25% revenue earning is achieved from Textile export, requires producing less defective textile for minimizing production cost and time. Inspection processes done on these industries are mostly manual and time consuming. To reduce error on identifying fabric defects requires more automotive and accurate inspection process. Considering this lacking, this research implements a Textile Defect Recognizer which uses computer vision methodology with the combination of multi-layer neural networks to identify four classifications of textile defects. The recognizer, suitable for LDC countries, identifies the fabric defects within economical cost and produces less error prone inspection system in real time. In order to generate input set for the neural network, primarily the recognizer captures digital fabric images by image acquisition device and converts the RGB images into binary images by restoration process and local threshold techniques. Later, the output of the processed image, the area of the faulty portion, the number of objects of the image and the sharp factor of the image, are feed backed as an input layer to the neural network which uses back propagation algorithm to compute the weighted factors and generates the desired classifications of defects as an output.Keywords: Computer vision, image acquisition device, machine vision, multi-layer neural networks.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 33001091 Dynamic Amplification Factors of Some City Bridges
Authors: I. Paeglite, A. Paeglitis
Abstract:
Paper presents a study about dynamic effects obtained from the dynamic load testing of the city highway bridges in Latvia carried out from 2005 to 2012. 9 prestressed concrete bridges and 4 composite bridges were considered. 11 of 13 bridges were designed according to the Eurocodes but two according to the previous structural codes used in Latvia (SNIP 2.05.03-84). The dynamic properties of the bridges were obtained by heavy vehicle passing the bridge roadway with different driving speeds and with or without even pavement. The obtained values of the Dynamic amplification factor (DAF) and the bridge natural frequency were analyzed and compared to the values of built-in traffic load models provided in Eurocode 1. The actual DAF values for even bridge pavement in the most cases are smaller than the value adopted in Eurocode 1. Vehicle speed for uneven pavements significantly influence Dynamic amplification factor values.
Keywords: Bridge, dynamic effects, load testing, dynamic amplification factor.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 30001090 Fast Factored DCT-LMS Speech Enhancement for Performance Enhancement of Digital Hearing Aid
Authors: Sunitha. S.L., V. Udayashankara
Abstract:
Background noise is particularly damaging to speech intelligibility for people with hearing loss especially for sensorineural loss patients. Several investigations on speech intelligibility have demonstrated sensorineural loss patients need 5-15 dB higher SNR than the normal hearing subjects. This paper describes Discrete Cosine Transform Power Normalized Least Mean Square algorithm to improve the SNR and to reduce the convergence rate of the LMS for Sensory neural loss patients. Since it requires only real arithmetic, it establishes the faster convergence rate as compare to time domain LMS and also this transformation improves the eigenvalue distribution of the input autocorrelation matrix of the LMS filter. The DCT has good ortho-normal, separable, and energy compaction property. Although the DCT does not separate frequencies, it is a powerful signal decorrelator. It is a real valued function and thus can be effectively used in real-time operation. The advantages of DCT-LMS as compared to standard LMS algorithm are shown via SNR and eigenvalue ratio computations. . Exploiting the symmetry of the basis functions, the DCT transform matrix [AN] can be factored into a series of ±1 butterflies and rotation angles. This factorization results in one of the fastest DCT implementation. There are different ways to obtain factorizations. This work uses the fast factored DCT algorithm developed by Chen and company. The computer simulations results show superior convergence characteristics of the proposed algorithm by improving the SNR at least 10 dB for input SNR less than and equal to 0 dB, faster convergence speed and better time and frequency characteristics.Keywords: Hearing Impairment, DCT Adaptive filter, Sensorineural loss patients, Convergence rate.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21711089 3D Object Model Reconstruction Based on Polywogs Wavelet Network Parametrization
Authors: Mohamed Othmani, Yassine Khlifi
Abstract:
This paper presents a technique for compact three dimensional (3D) object model reconstruction using wavelet networks. It consists to transform an input surface vertices into signals,and uses wavelet network parameters for signal approximations. To prove this, we use a wavelet network architecture founded on several mother wavelet families. POLYnomials WindOwed with Gaussians (POLYWOG) wavelet families are used to maximize the probability to select the best wavelets which ensure the good generalization of the network. To achieve a better reconstruction, the network is trained several iterations to optimize the wavelet network parameters until the error criterion is small enough. Experimental results will shown that our proposed technique can effectively reconstruct an irregular 3D object models when using the optimized wavelet network parameters. We will prove that an accurateness reconstruction depends on the best choice of the mother wavelets.Keywords: 3D object, optimization, parametrization, Polywog wavelets, reconstruction, wavelet networks.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15011088 Contact Drying Simulation of Particulate Materials: A Comprehensive Approach
Authors: Marco Intelvi, Apolinar Picado, Joaquín Martínez
Abstract:
In this work, simulation algorithms for contact drying of agitated particulate materials under vacuum and at atmospheric pressure were developed. The implementation of algorithms gives a predictive estimation of drying rate curves and bulk bed temperature during contact drying. The calculations are based on the penetration model to describe the drying process, where all process parameters such as heat and mass transfer coefficients, effective bed properties, gas and liquid phase properties are estimated with proper correlations. Simulation results were compared with experimental data from the literature. In both cases, simulation results were in good agreement with experimental data. Few deviations were identified and the limitations of the predictive capabilities of the models are discussed. The programs give a good insight of the drying behaviour of the analysed powders.Keywords: Agitated bed, Atmospheric pressure, Penetrationmodel, Vacuum
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22431087 Dynamic Soil Structure Interaction in Buildings
Authors: Shreya Thusoo, Karan Modi, Ankit Kumar Jha, Rajesh Kumar
Abstract:
Since the evolution of computational tools and simulation software, there has been considerable increase in research on Soil Structure Interaction (SSI) to decrease the computational time and increase accuracy in the results. To aid the designer with a proper understanding of the response of structure in different soil types, the presented paper compares the deformation, shear stress, acceleration and other parameters of multi-storey building for a specific input ground motion using Response-spectrum Analysis (RSA) method. The response of all the models of different heights have been compared in different soil types. Finite Element Simulation software, ANSYS, has been used for all the computational purposes. Overall, higher response is observed with SSI, while it increases with decreasing stiffness of soil.
Keywords: Soil-structure interaction, response-spectrum analysis, finite element method, multi-storey buildings.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22821086 An Evolutionary Algorithm for Optimal Fuel-Type Configurations in Car Lines
Authors: Charalampos Saridakis, Stelios Tsafarakis
Abstract:
Although environmental concern is on the rise across Europe, current market data indicate that adoption rates of environmentally friendly vehicles remain extremely low. Against this background, the aim of this paper is to a) assess preferences of European consumers for clean-fuel cars and their characteristics and b) design car lines that optimize the combination of fuel types among models in the line-up. In this direction, the authors introduce a new evolutionary mechanism and implement it to stated-preference data derived from a large-scale choice-based conjoint experiment that measures consumer preferences for various factors affecting clean-fuel vehicle (CFV) adoption. The proposed two-step methodology provides interesting insights into how new and existing fuel-types can be combined in a car line that maximizes customer satisfaction.Keywords: Clean-fuel vehicles, product line design, conjoint analysis, choice experiment, differential evolution.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10001085 Reduction of Content of Lead and Zinc from Wastewater by Using of Metallurgical Waste
Authors: L. Rozumová, J. Seidlerová
Abstract:
The aim of this paper was to study the sorption properties of a blast furnace sludge used as the sorbent. The sorbent was utilized for reduction of content of lead and zinc ions. Sorbent utilized in this work was obtained from metallurgical industry from process of wet gas treatment in iron production. The blast furnace sludge was characterized by X-Ray diffraction, scanning electron microscopy, and XRFS spectroscopy. Sorption experiments were conducted in batch mode. The sorption of metal ions in the sludge was determined by correlation of adsorption isotherm models. The adsorption of lead and zinc ions was best fitted with Langmuir adsorption isotherms. The adsorption capacity of lead and zinc ions was 53.8 mg.g-1 and 10.7 mg.g-1, respectively. The results indicated that blast furnace sludge could be effectively used as secondary material and could be also employed as a low-cost alternative for the removal of heavy metals ions from wastewater.
Keywords: Blast furnace sludge, lead, zinc, sorption.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9061084 On Formalizing Predefined OCL Properties
Authors: Meryem Lamrani, Younès El Amrani, Aziz Ettouhami
Abstract:
The ability of UML to handle the modeling process of complex industrial software applications has increased its popularity to the extent of becoming the de-facto language in serving the design purpose. Although, its rich graphical notation naturally oriented towards the object-oriented concept, facilitates the understandability, it hardly successes to report all domainspecific aspects in a satisfactory way. OCL, as the standard language for expressing additional constraints on UML models, has great potential to help improve expressiveness. Unfortunately, it suffers from a weak formalism due to its poor semantic resulting in many obstacles towards the build of tools support and thus its application in the industry field. For this reason, many researches were established to formalize OCL expressions using a more rigorous approach. Our contribution join this work in a complementary way since it focuses specifically on OCL predefined properties which constitute an important part in the construction of OCL expressions. Using formal methods, we mainly succeed in expressing rigorously OCL predefined functions.
Keywords: Formal methods, Z, OCL, predefined properties, metamodel types.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 7721083 A Consideration on the Offset Frontal Impact Modeling Using Spring-Mass Model
Authors: Jaemoon Lim
Abstract:
To construct the lumped spring-mass model considering the occupants for the offset frontal crash, the SISAME software and the NHTSA test data were used. The data on 56 kph 40% offset frontal vehicle to deformable barrier crash test of a MY2007 Mazda 6 4-door sedan were obtained from NHTSA test database. The overall behaviors of B-pillar and engine of simulation models agreed very well with the test data. The trends of accelerations at the driver and passenger head were similar but big differences in peak values. The differences of peak values caused the large errors of the HIC36 and 3 ms chest g’s. To predict well the behaviors of dummies, the spring-mass model for the offset frontal crash needs to be improved.Keywords: Chest g’s, HIC36, lumped spring-mass model, offset frontal impact, SISAME.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 26681082 Cardiovascular Modeling Software Tools in Medicine
Authors: J. Fernandez, R. Fernandez de Canete, J. Perea-Paizal, J. C. Ramos-Diaz
Abstract:
The high prevalence of cardiovascular diseases has provoked a raising interest in the development of mathematical models in order to evaluate the cardiovascular function both under physiological and pathological conditions. In this paper, a physical model of the cardiovascular system with intrinsic regulation is presented and implemented by using the object-oriented Modelica simulation software tools. For this task, a multi-compartmental system previously validated with physiological data has been built, based on the interconnection of cardiovascular elements such as resistances, capacitances and pumping among others, by following an electrohydraulic analogy. The results obtained under both physiological and pathological scenarios provide an easy interpretative key to analyze the hemodynamic behavior of the patient. The described approach represents a valuable tool in the teaching of physiology for graduate medical and nursing students among others.
Keywords: Cardiovascular system, Modelica simulation software, physical modeling, teaching tool.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 12521081 Turbulence Modeling and Wave-Current Interactions
Authors: A.-C. Bennis, F. Dumas, F. Ardhuin, B. Blanke
Abstract:
The mechanics of rip currents are complex, involving interactions between waves, currents, water levels and the bathymetry, that present particular challenges for numerical models. Here, the effects of a grid-spacing dependent horizontal mixing on the wave-current interactions are studied. Near the shore, wave rays diverge from channels towards bar crests because of refraction by topography and currents, in a way that depends on the rip current intensity which is itself modulated by the horizontal mixing. At low resolution with the grid-spacing dependent horizontal mixing, the wave motion is the same for both coupling modes because the wave deviation by the currents is weak. In high resolution case, however, classical results are found with the stabilizing effect of the flow by feedback of waves on currents. Lastly, wave-current interactions and the horizontal mixing strongly affect the intensity of the three-dimensional rip velocity.
Keywords: Numerical modeling, Rip currents, Turbulence modeling, Wave-current interactions.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22071080 Conceptual Multidimensional Model
Authors: Manpreet Singh, Parvinder Singh, Suman
Abstract:
The data is available in abundance in any business organization. It includes the records for finance, maintenance, inventory, progress reports etc. As the time progresses, the data keep on accumulating and the challenge is to extract the information from this data bank. Knowledge discovery from these large and complex databases is the key problem of this era. Data mining and machine learning techniques are needed which can scale to the size of the problems and can be customized to the application of business. For the development of accurate and required information for particular problem, business analyst needs to develop multidimensional models which give the reliable information so that they can take right decision for particular problem. If the multidimensional model does not possess the advance features, the accuracy cannot be expected. The present work involves the development of a Multidimensional data model incorporating advance features. The criterion of computation is based on the data precision and to include slowly change time dimension. The final results are displayed in graphical form.Keywords: Multidimensional, data precision.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14581079 Automatic Text Summarization
Authors: Mohamed Abdel Fattah, Fuji Ren
Abstract:
This work proposes an approach to address automatic text summarization. This approach is a trainable summarizer, which takes into account several features, including sentence position, positive keyword, negative keyword, sentence centrality, sentence resemblance to the title, sentence inclusion of name entity, sentence inclusion of numerical data, sentence relative length, Bushy path of the sentence and aggregated similarity for each sentence to generate summaries. First we investigate the effect of each sentence feature on the summarization task. Then we use all features score function to train genetic algorithm (GA) and mathematical regression (MR) models to obtain a suitable combination of feature weights. The proposed approach performance is measured at several compression rates on a data corpus composed of 100 English religious articles. The results of the proposed approach are promising.Keywords: Automatic Summarization, Genetic Algorithm, Mathematical Regression, Text Features.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 23361078 Adaptive Gaussian Mixture Model for Skin Color Segmentation
Authors: Reza Hassanpour, Asadollah Shahbahrami, Stephan Wong
Abstract:
Skin color based tracking techniques often assume a static skin color model obtained either from an offline set of library images or the first few frames of a video stream. These models can show a weak performance in presence of changing lighting or imaging conditions. We propose an adaptive skin color model based on the Gaussian mixture model to handle the changing conditions. Initial estimation of the number and weights of skin color clusters are obtained using a modified form of the general Expectation maximization algorithm, The model adapts to changes in imaging conditions and refines the model parameters dynamically using spatial and temporal constraints. Experimental results show that the method can be used in effectively tracking of hand and face regions.Keywords: Face detection, Segmentation, Tracking, Gaussian Mixture Model, Adaptation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 24151077 A Communication Signal Recognition Algorithm Based on Holder Coefficient Characteristics
Authors: Hui Zhang, Ye Tian, Fang Ye, Ziming Guo
Abstract:
Communication signal modulation recognition technology is one of the key technologies in the field of modern information warfare. At present, communication signal automatic modulation recognition methods are mainly divided into two major categories. One is the maximum likelihood hypothesis testing method based on decision theory, the other is a statistical pattern recognition method based on feature extraction. Now, the most commonly used is a statistical pattern recognition method, which includes feature extraction and classifier design. With the increasingly complex electromagnetic environment of communications, how to effectively extract the features of various signals at low signal-to-noise ratio (SNR) is a hot topic for scholars in various countries. To solve this problem, this paper proposes a feature extraction algorithm for the communication signal based on the improved Holder cloud feature. And the extreme learning machine (ELM) is used which aims at the problem of the real-time in the modern warfare to classify the extracted features. The algorithm extracts the digital features of the improved cloud model without deterministic information in a low SNR environment, and uses the improved cloud model to obtain more stable Holder cloud features and the performance of the algorithm is improved. This algorithm addresses the problem that a simple feature extraction algorithm based on Holder coefficient feature is difficult to recognize at low SNR, and it also has a better recognition accuracy. The results of simulations show that the approach in this paper still has a good classification result at low SNR, even when the SNR is -15dB, the recognition accuracy still reaches 76%.Keywords: Communication signal, feature extraction, holder coefficient, improved cloud model.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 7081076 Line Balancing in the Hard Disk Drive Process Using Simulation Techniques
Authors: Teerapun Saeheaw, Nivit Charoenchai, Wichai Chattinnawat
Abstract:
Simulation model is an easy way to build up models to represent real life scenarios, to identify bottlenecks and to enhance system performance. Using a valid simulation model may give several advantages in creating better manufacturing design in order to improve the system performances. This paper presents result of implementing a simulation model to design hard disk drive manufacturing process by applying line balancing to improve both productivity and quality of hard disk drive process. The line balance efficiency showed 86% decrease in work in process, output was increased by an average of 80%, average time in the system was decreased 86% and waiting time was decreased 90%.Keywords: line balancing, arena, hard disk drive process, simulation, work in process (WIP)
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21641075 The Long Run Relationship between Exports and Imports in South Africa: Evidence from Cointegration Analysis
Authors: Sagaren Pillay
Abstract:
This study empirically examines the long run equilibrium relationship between South Africa’s exports and imports using quarterly data from 1985 to 2012. The theoretical framework used for the study is based on Johansen’s Maximum Likelihood cointegration technique which tests for both the existence and number of cointegration vectors that exists. The study finds that both the series are integrated of order one and are cointegrated. A statistically significant cointegrating relationship is found to exist between exports and imports. The study models this unique linear and lagged relationship using a Vector Error Correction Model (VECM). The findings of the study confirm the existence of a long run equilibrium relationship between exports and imports.
Keywords: Cointegration lagged, linear, maximum likelihood, vector error correction model.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 27841074 Signal-to-Noise Ratio Improvement of EMCCD Cameras
Authors: Wen W. Zhang, Qian Chen, Bei B. Zhou, Wei J. He
Abstract:
Over the past years, the EMCCD has had a profound influence on photon starved imaging applications relying on its unique multiplication register based on the impact ionization effect in the silicon. High signal-to-noise ratio (SNR) means high image quality. Thus, SNR improvement is important for the EMCCD. This work analyzes the SNR performance of an EMCCD with gain off and on. In each mode, simplified SNR models are established for different integration times. The SNR curves are divided into readout noise (or CIC) region and shot noise region by integration time. Theoretical SNR values comparing long frame integration and frame adding in each region are presented and discussed to figure out which method is more effective. In order to further improve the SNR performance, pixel binning is introduced into the EMCCD. The results show that pixel binning does obviously improve the SNR performance, but at the expensive of the spatial resolution.Keywords: EMCCD, SNR improvement, pixel binning
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 28751073 A Pilot Study of Robot Reminiscence in Dementia Care
Authors: Ryuji Yamazaki, Masahiro Kochi, Weiran Zhu, Hiroko Kase
Abstract:
In care for older adults, behavioral and psychological symptoms of dementia (BPSD) like agitation and aggression are distressing for patients and their caretakers, often resulting in premature institutionalization with increased costs of care. To improve mood and mitigate symptoms, as a non-pharmaceutical approach, emotion-oriented therapy like reminiscence work is adopted in face-to-face communication. Telecommunication support is expected to be provided by robotic media as a bridge for digital divide for those with dementia and facilitate social interaction both verbally and nonverbally. The purpose of this case study is to explore the conditions in which robotic media can effectively attract attention from older adults with dementia and promote their well-being. As a pilot study, we introduced the pillow-phone Hugvie®, a huggable humanly shaped communication medium to five residents with dementia at a care facility, to investigate how the following conditions work for the elderly when they use the medium; 1) no sound, 2) radio, non-interactive, 3) daily conversation, and 4) reminiscence work. As a result, under condition 4, reminiscence work, the five participants kept concentration in interacting with the medium for a longer duration than other conditions. In condition 4, they also showed larger amount of utterances than under other conditions. These results indicate that providing topics related to personal histories through robotic media could affect communication positively and should, therefore, be further investigated. In addition, the issue of ethical implications by using persuasive technology that affects emotions and behaviors of older adults is also discussed.
Keywords: BPSD, reminiscence, tactile telecommunication, utterances.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 11591072 MarginDistillation: Distillation for Face Recognition Neural Networks with Margin-Based Softmax
Authors: Svitov David, Alyamkin Sergey
Abstract:
The usage of convolutional neural networks (CNNs) in conjunction with the margin-based softmax approach demonstrates the state-of-the-art performance for the face recognition problem. Recently, lightweight neural network models trained with the margin-based softmax have been introduced for the face identification task for edge devices. In this paper, we propose a distillation method for lightweight neural network architectures that outperforms other known methods for the face recognition task on LFW, AgeDB-30 and Megaface datasets. The idea of the proposed method is to use class centers from the teacher network for the student network. Then the student network is trained to get the same angles between the class centers and face embeddings predicted by the teacher network.Keywords: ArcFace, distillation, face recognition, margin-based softmax.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6311071 Comparison of Central Light Reflex Width-to-Retinal Vessel Diameter Ratio between Glaucoma and Normal Eyes by Using Edge Detection Technique
Authors: P. Siriarchawatana, K. Leungchavaphongse, N. Covavisaruch, K. Rojananuangnit, P. Boondaeng, N. Panyayingyong
Abstract:
Glaucoma is a disease that causes visual loss in adults. Glaucoma causes damage to the optic nerve and its overall pathophysiology is still not fully understood. Vasculopathy may be one of the possible causes of nerve damage. Photographic imaging of retinal vessels by fundus camera during eye examination may complement clinical management. This paper presents an innovation for measuring central light reflex width-to-retinal vessel diameter ratio (CRR) from digital retinal photographs. Using our edge detection technique, CRRs from glaucoma and normal eyes were compared to examine differences and associations. CRRs were evaluated on fundus photographs of participants from Mettapracharak (Wat Raikhing) Hospital in Nakhon Pathom, Thailand. Fifty-five photographs from normal eyes and twenty-one photographs from glaucoma eyes were included. Participants with hypertension were excluded. In each photograph, CRRs from four retinal vessels, including arteries and veins in the inferotemporal and superotemporal regions, were quantified using edge detection technique. From our finding, mean CRRs of all four retinal arteries and veins were significantly higher in persons with glaucoma than in those without glaucoma (0.34 vs. 0.32, p < 0.05 for inferotemporal vein, 0.33 vs. 0.30, p < 0.01 for inferotemporal artery, 0.34 vs. 0.31, p < 0.01 for superotemporal vein, and 0.33 vs. 0.30, p < 0.05 for superotemporal artery). From these results, an increase in CRRs of retinal vessels, as quantitatively measured from fundus photographs, could be associated with glaucoma.
Keywords: Glaucoma, retinal vessel, central light reflex, image processing, fundus photograph, edge detection.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10871070 Co-Administration Effects of Conjugated Linoleic Acid and L-Carnitine on Weight Gain and Biochemical Profile in Diet Induced Obese Rats
Authors: Maryam Nazari, Majid Karandish, Alihossein Saberi
Abstract:
Obesity as a global health challenge motivates pharmaceutical industries to produce anti-obesity drugs. However, effectiveness of these agents is remained unclear. Because of popularity of dietary supplements, the aim of this study was tp investigate the effects of Conjugated Linoleic Acid (CLA) and L-carnitine (LC) on serum glucose, triglyceride, cholesterol and weight changes in diet induced obese rats. 48 male Wistar rats were randomly divided into two groups: Normal fat diet (n=8), and High fat diet (HFD) (n=32). After eight weeks, the second group which was maintained on HFD until the end of study, was subdivided into four categories: a) 500 mg Corn Oil (as control group), b) 500 mg CLA, c) 200 mg LC, d) 500 mg CLA+ 200 mg LC.All doses are planned per kg body weights, which were administered by oral gavage for four weeks. Body weights were measured and recorded weekly by means of a digital scale. At the end of the study, blood samples were collected for biochemical markers measurement. SPSS Version 16 was used for statistical analysis. At the end of 8th week, a significant difference in weight was observed between HFD and NFD group. After 12 weeks, LC significantly reduced weight gain by 4.2%. Trend of weight gain in CLA and CLA+LC groups was insignificantly decelerated. CLA+LC reduced triglyceride level significantly, but just CLA had significant influence on total cholesterol and insignificant decreasing effect on FBS. Our results showed that an obesogenic diet in a relative short time led to obesity and dyslipidemia which can be modified by LC and CLA to some extent.
Keywords: Conjugated linoleic acid, high fat diet, L-carnitine, obesity.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9421069 Numerical Evaluation of Nusselt Number on the Hot Wall in Square Enclosure Filled with Nanofluid
Authors: A. Ghafouri, A. Falavand Jozaei, M. Salari
Abstract:
In this paper, effects of using Alumina-water nanofluid on the rate of heat transfer have been investigated numerically. Physical model is a square enclosure with insulated top and bottom horizontal walls, while the vertical walls are kept at different constant temperatures. Two appropriate models are used to evaluate the viscosity and thermal conductivity of nanofluid. The governing stream-vorticity equations are solved using a second order central finite difference scheme, coupled to the conservation of mass and energy. The study has been carried out for the Richardson number 0.1 to 10 and the solid volume fraction 0 to 0.04. Results are presented by isotherms lines, average Nusselt number and normalized Nusselt number in different range of φ and Ri for forced, combined and natural convection dominated regime. It is found that higher heat transfer rate is predicted when the effects of nanoparticle is taken into account.
Keywords: Nanofluid, Heat Transfer Enhancement, Square Enclosure, Nusselt number.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 23361068 Hypergraph Models of Metabolism
Authors: Nicole Pearcy, Jonathan J. Crofts, Nadia Chuzhanova
Abstract:
In this paper, we employ a directed hypergraph model to investigate the extent to which environmental variability influences the set of available biochemical reactions within a living cell. Such an approach avoids the limitations of the usual complex network formalism by allowing for the multilateral relationships (i.e. connections involving more than two nodes) that naturally occur within many biological processes. More specifically, we extend the concept of network reciprocity to complex hyper-networks, thus enabling us to characterise a network in terms of the existence of mutual hyper-connections, which may be considered a proxy for metabolic network complexity. To demonstrate these ideas, we study 115 metabolic hyper-networks of bacteria, each of which can be classified into one of 6 increasingly varied habitats. In particular, we found that reciprocity increases significantly with increased environmental variability, supporting the view that organism adaptability leads to increased complexities in the resultant biochemical networks.
Keywords: Complexity, hypergraphs, reciprocity, metabolism.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 24551067 Performance Evaluation of ROI Extraction Models from Stationary Images
Authors: K.V. Sridhar, Varun Gunnala, K.S.R Krishna Prasad
Abstract:
In this paper three basic approaches and different methods under each of them for extracting region of interest (ROI) from stationary images are explored. The results obtained for each of the proposed methods are shown, and it is demonstrated where each method outperforms the other. Two main problems in ROI extraction: the channel selection problem and the saliency reversal problem are discussed and how best these two are addressed by various methods is also seen. The basic approaches are 1) Saliency based approach 2) Wavelet based approach 3) Clustering based approach. The saliency approach performs well on images containing objects of high saturation and brightness. The wavelet based approach performs well on natural scene images that contain regions of distinct textures. The mean shift clustering approach partitions the image into regions according to the density distribution of pixel intensities. The experimental results of various methodologies show that each technique performs at different acceptable levels for various types of images.Keywords: clustering, ROI, saliency, wavelets.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14091066 Dissolution Leaching Kinetics of Ulexite in Disodium Hydrogen Phosphate Solutions
Authors: Betül Özgenç Kaya, Soner Kuslu, Sabri Çolak, Turan Çalban
Abstract:
Ulexite (Na2O.2CaO.5B2O3.16H2O) is boron mineral that is found in large quantities in the Turkey and world. In this study, the dissolution of this mineral in the disodium hydrogen phosphate solutions has been studied. Temperature, concentration, stirring speed, solid liquid ratio and particle size were selected as parameters. The experimental results were successfully correlated by linear regression using Statistica program. Dissolution curves were evaluated shrinking core models for solid-fluid systems. It was observed that increase in the reaction temperature and decrease in the solid/liquid ratio causes an increase the dissolution rate of ulexite. The activation energy was found to be 63.4 kJ/mol. The leaching of ulexite was controlled by chemical reaction.
Keywords: Disodium hydrogen phosphate, Leaching kinetics, Ulexite.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2151