Search results for: Frame Analyses Method
7774 Seismic Vulnerability Mitigation of Non-Engineered Buildings
Authors: Muhammad Tariq A. Chaudhary
Abstract:
The tremendous loss of life that resulted in the aftermath of recent earthquakes in developing countries is mostly due to the collapse of non-engineered and semi-engineered building structures. Such structures are used as houses, schools, primary healthcare centers and government offices. These building are classified structurally into two categories viz. non-engineered and semi-engineered. Non-engineered structures include: adobe, unreinforced masonry (URM) and wood buildings. Semi-engineered buildings are mostly low-rise (up to 3 story) light concrete frame structures or masonry bearing walls with reinforced concrete slab. This paper presents an overview of the typical damage observed in non-engineered structures and their most likely causes in the past earthquakes with specific emphasis on the performance of such structures in the 2005 Kashmir earthquake. It is demonstrated that seismic performance of these structures can be improved from life-safety viewpoint by adopting simple low-cost modifications to the existing construction practices. Incorporation of some of these practices in the reconstruction efforts after the 2005 Kashmir earthquake are examined in the last section for mitigating seismic risk hazard.
Keywords: Kashmir earthquake, non-engineered buildings, seismic hazard, structural details, structural strengthening.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 29417773 Fast Search for MPEG Video Clips Using Adjacent Pixel Intensity Difference Quantization Histogram Feature
Authors: Feifei Lee, Qiu Chen, Koji Kotani, Tadahiro Ohmi
Abstract:
In this paper, we propose a novel fast search algorithm for short MPEG video clips from video database. This algorithm is based on the adjacent pixel intensity difference quantization (APIDQ) algorithm, which had been reliably applied to human face recognition previously. An APIDQ histogram is utilized as the feature vector of the frame image. Instead of fully decompressed video frames, partially decoded data, namely DC images are utilized. Combined with active search [4], a temporal pruning algorithm, fast and robust video search can be realized. The proposed search algorithm has been evaluated by 6 hours of video to search for given 200 MPEG video clips which each length is 15 seconds. Experimental results show the proposed algorithm can detect the similar video clip in merely 80ms, and Equal Error Rate (ERR) of 3 % is achieved, which is more accurately and robust than conventional fast video search algorithm.
Keywords: Fast search, adjacent pixel intensity difference quantization (APIDQ), DC image, histogram feature.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15787772 Dynamic Fast Tracing and Smoothing Technique for Geiger-Muller Dosimeter
Authors: M. Ebrahimi Shohani, S. M. Taheri, S. M. Golgoun
Abstract:
Environmental radiation dosimeter is a kind of detector that measures the dose of the radiation area. Dosimeter registers the radiation and converts it to the dose according to the calibration parameters. The limit of a dose is different at each radiation area and this limit should be notified and reported to the user and health physics department. The stochastic nature of radiation is the reason for the fluctuation of any gamma detector dosimetry. In this research we investigated Geiger-Muller type of dosimeter and tried to improve the dose measurement. Geiger-Muller dosimeter is a counter that converts registered radiation to the dose. Therefore, for better data analysis, it is necessary to apply an algorithm to smooth statistical variations of registered radiation. We proposed a method to smooth these fluctuations much more and also proposed a dynamic way to trace rapid changes of radiations. Results show that our method is fast and reliable method in comparison the traditional method.
Keywords: Geiger-Muller, radiation detection, smoothing algorithms, dosimeter, dose calculation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4657771 Investigation of the Effect of Teaching a Thinking and Research Lesson by Cooperative and Traditional Methods on the Creativity of Sixth Grade Students
Authors: Faroogh Khakzad, Marzieh Dehghani, Elahe Hejazi
Abstract:
The present study investigates the effect of teaching a Thinking and Research lesson by cooperative and traditional methods on the creativity of sixth-grade students in Piranshahr province. The statistical society includes all the sixth-grade students of Piranshahr province. The sample of this studytable was selected by available sampling from among male elementary schools of Piranshahr. They were randomly assigned into two groups of cooperative teaching method and traditional teaching method. The design of the study is quasi-experimental with a control group. In this study, to assess students’ creativity, Abedi’s creativity questionnaire was used. Based on Cronbach’s alpha coefficient, the reliability of the factor flow was 0.74, innovation was 0.61, flexibility was 0.63, and expansion was 0.68. To analyze the data, t-test, univariate and multivariate covariance analysis were used for evaluation of the difference of means and the pretest and posttest scores. The findings of the research showed that cooperative teaching method does not significantly increase creativity (p > 0.05). Moreover, cooperative teaching method was found to have significant effect on flow factor (p < 0.05), but in innovation and expansion factors no significant effect was observed (p < 0.05).
Keywords: Cooperative teaching method, traditional teaching method, creativity, flow, innovation, flexibility, expansion, thinking and research lesson.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6917770 Fuzzy Types Clustering for Microarray Data
Authors: Seo Young Kim, Tai Myong Choi
Abstract:
The main goal of microarray experiments is to quantify the expression of every object on a slide as precisely as possible, with a further goal of clustering the objects. Recently, many studies have discussed clustering issues involving similar patterns of gene expression. This paper presents an application of fuzzy-type methods for clustering DNA microarray data that can be applied to typical comparisons. Clustering and analyses were performed on microarray and simulated data. The results show that fuzzy-possibility c-means clustering substantially improves the findings obtained by others.Keywords: Clustering, microarray data, Fuzzy-type clustering, Validation
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15207769 Numerical Modelling of Dry Stone Masonry Structures Based on Finite-Discrete Element Method
Authors: Ž. Nikolić, H. Smoljanović, N. Živaljić
Abstract:
This paper presents numerical model based on finite-discrete element method for analysis of the structural response of dry stone masonry structures under static and dynamic loads. More precisely, each discrete stone block is discretized by finite elements. Material non-linearity including fracture and fragmentation of discrete elements as well as cyclic behavior during dynamic load are considered through contact elements which are implemented within a finite element mesh. The application of the model was conducted on several examples of these structures. The performed analysis shows high accuracy of the numerical results in comparison with the experimental ones and demonstrates the potential of the finite-discrete element method for modelling of the response of dry stone masonry structures.Keywords: Finite-discrete element method, dry stone masonry structures, static load, dynamic load.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16167768 A Collusion-Resistant Distributed Signature Delegation Based on Anonymous Mobile Agent
Authors: Omaima Bamasak
Abstract:
This paper presents a novel method that allows an agent host to delegate its signing power to an anonymous mobile agent in such away that the mobile agent does not reveal any information about its host-s identity and, at the same time, can be authenticated by the service host, hence, ensuring fairness of service provision. The solution introduces a verification server to verify the signature generated by the mobile agent in such a way that even if colluding with the service host, both parties will not get more information than what they already have. The solution incorporates three methods: Agent Signature Key Generation method, Agent Signature Generation method, Agent Signature Verification method. The most notable feature of the solution is that, in addition to allowing secure and anonymous signature delegation, it enables tracking of malicious mobile agents when a service host is attacked. The security properties of the proposed solution are analyzed, and the solution is compared with the most related work.Keywords: Anonymous signature delegation, collusion resistance, e-commerce fairness, mobile agent security.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14447767 Analysis and Remediation of Fecal Coliform Bacteria Pollution in Selected Surface Water Bodies of Enugu State of Nigeria
Authors: Chime Charles C., Ikechukwu Alexander Okorie, Ekanem E.J., Kagbu J. A.
Abstract:
The assessment of surface waters in Enugu metropolis for fecal coliform bacteria was undertaken. Enugu urban was divided into three areas (A1, A2 and A3), and fecal coliform bacteria analysed in the surface waters found in these areas for four years (2005-2008). The plate count method was used for the analyses. Data generated were subjected to statistical tests involving; Normality test, Homogeneity of variance test, correlation test, and tolerance limit test. The influence of seasonality and pollution trends were investigated using time series plots. Results from the tolerance limit test at 95% coverage with 95% confidence, and with respect to EU maximum permissible concentration show that the three areas suffer from fecal coliform pollution. To this end, remediation procedure involving the use of saw-dust extracts from three woods namely; Chlorophora-Excelsa (C-Excelsa),Khayan-Senegalensis,(CSenegalensis) and Erythrophylum-Ivorensis (E-Ivorensis) in controlling the coliforms was studied. Results show that mixture of the acetone extracts of the woods show the most effective antibacterial inhibitory activities (26.00mm zone of inhibition) against E-coli. Methanol extract mixture of the three woods gave best inhibitory activity (26.00mm zone of inhibition) against S-areus, and 25.00mm zones of inhibition against E-Aerogenes. The aqueous extracts mixture gave acceptable zones of inhibitions against the three bacteria organisms.Keywords: Coliform bacteria, Pollution, Remediation, Saw-dust
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20407766 A Character Detection Method for Ancient Yi Books Based on Connected Components and Regressive Character Segmentation
Authors: Xu Han, Shanxiong Chen, Shiyu Zhu, Xiaoyu Lin, Fujia Zhao, Dingwang Wang
Abstract:
Character detection is an important issue for character recognition of ancient Yi books. The accuracy of detection directly affects the recognition effect of ancient Yi books. Considering the complex layout, the lack of standard typesetting and the mixed arrangement between images and texts, we propose a character detection method for ancient Yi books based on connected components and regressive character segmentation. First, the scanned images of ancient Yi books are preprocessed with nonlocal mean filtering, and then a modified local adaptive threshold binarization algorithm is used to obtain the binary images to segment the foreground and background for the images. Second, the non-text areas are removed by the method based on connected components. Finally, the single character in the ancient Yi books is segmented by our method. The experimental results show that the method can effectively separate the text areas and non-text areas for ancient Yi books and achieve higher accuracy and recall rate in the experiment of character detection, and effectively solve the problem of character detection and segmentation in character recognition of ancient books.Keywords: Computing methodologies, interest point, salient region detections, image segmentation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 8647765 A Video-Based Observation and Analysis Method to Assess Human Movement and Behaviour in Crowded Areas
Authors: Shahrol Mohamaddan, Keith Case, Ana Sakura Zainal Abidin
Abstract:
Human movement in the real world provides important information for developing human behaviour models and simulations. However, it is difficult to assess ‘real’ human behaviour since there is no established method available. As part of the AUNTSUE (Accessibility and User Needs in Transport – Sustainable Urban Environments) project, this research aimed to propose a method to assess human movement and behaviour in crowded areas. The method is based on the three major steps of video recording, conceptual behavior modelling and video analysis. The focus is on individual human movement and behaviour in normal situations (panic situations are not considered) and the interactions between individuals in localized areas. Emphasis is placed on gaining knowledge of characteristics of human movement and behaviour in the real world that can be modelled in the virtual environment.
Keywords: Video observation, Human movement, Behaviour, Crowds, Ergonomics, AUNT-SUE
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22447764 Designing Pictogram for Food Portion Size
Authors: Y.C. Liu, S.J. Lu, Y.C. Weng, H. Su
Abstract:
The objective of this paper is to investigate a new approach based on the idea of pictograms for food portion size. This approach adopts the model of the United States Pharmacopeia- Drug Information (USP-DI). The representation of each food portion size composed of three parts: frame, the connotation of dietary portion sizes and layout. To investigate users- comprehension based on this approach, two experiments were conducted, included 122 Taiwanese people, 60 male and 62 female with ages between 16 and 64 (divided into age groups of 16-30, 31-45 and 46-64). In Experiment 1, the mean correcting rate of the understanding level of food items is 48.54% (S.D.= 95.08) and the mean response time 2.89sec (S.D.=2.14). The difference on the correct rates for different age groups is significant (P*=0.00<0.05). In Experiment 2, the correcting rate of selecting the right life-size measurement aid is 65.02% (S.D.=21.31). The result showed the potential of the approach for certain food potion sizes. Issues raised for discussions including comprehension on numerous food varieties in an open environment, selection of photograph or drawing, reasons of different correcting rates for the measurement aid. This research also could be used for those interested in systematic and pictorial representation of dietary portion size information.Keywords: Comprehension, Food Portion Size, Model of DietaryInformation, Pictogram Design, USP-DI.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19357763 Development of In Situ Permeability Test Using Constant Discharge Method for Sandy Soils
Authors: A. Rifa’i, Y. Takeshita, M. Komatsu
Abstract:
The post-rain puddles problem that occurs in the first yard of Prambanan Temple are often disturbing visitor activity. A poodle layer and a drainage system had ever built to avoid such a problem, but puddles still did not stop appearing after rain. Permeability parameter needs to be determined by using a simpler procedure to find exact method of solution. The instrument modelling was proposed according to the development of field permeability testing instrument. This experiment used a proposed Constant Discharge method. Constant Discharge method used a tube poured with constant water flow from unsaturated until saturated soil condition. Volumetric water content (θ) were monitored by soil moisture measurement device. The results were correlations between k and θ which were drawn by numerical approach from Van Genutchen model. Parameters θr optimum value obtained from the test was at very dry soil. Coefficient of permeability with a density of 19.8 kN/m3 for unsaturated conditions was in range of 3 x 10-6 cm/sec (Sr=68%) until 9.98 x 10-4 cm/sec (Sr=82%). The equipment and testing procedure developed in this research was quite effective, simple and easy to be implemented on determining field soil permeability coefficient value of sandy soil. Using constant discharge method in proposed permeability test, value of permeability coefficient under unsaturated condition can be obtained without establish soil water characteristic curve.
Keywords: Constant discharge method, in situ permeability test, sandy soil, unsaturated conditions.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 34587762 An Iterative Method for the Symmetric Arrowhead Solution of Matrix Equation
Authors: Minghui Wang, Luping Xu, Juntao Zhang
Abstract:
In this paper, according to the classical algorithm LSQR for solving the least-squares problem, an iterative method is proposed for least-squares solution of constrained matrix equation. By using the Kronecker product, the matrix-form LSQR is presented to obtain the like-minimum norm and minimum norm solutions in a constrained matrix set for the symmetric arrowhead matrices. Finally, numerical examples are also given to investigate the performance.Keywords: Symmetric arrowhead matrix, iterative method, like-minimum norm, minimum norm, Algorithm LSQR.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14087761 Characterization of the O.ul-mS952 Intron:A Potential Molecular Marker to Distinguish Between Ophiostoma Ulmi and Ophiostoma Novo-Ulmi Subsp. Americana
Authors: Mohamed Hafez, Georg Hausner
Abstract:
The full length mitochondrial small subunit ribosomal (mt-rns) gene has been characterized for Ophiostoma novo-ulmi subspecies americana. The gene was also characterized for Ophiostoma ulmi and a group II intron was noted in the mt-rns gene of O. ulmi. The insertion in the mt-rns gene is at position S952 and it is a group IIB1 intron that encodes a double motif LAGLIDADG homing endonuclease from an open reading frame located within a loop of domain III. Secondary structure models for the mt-rns RNA of O. novo-ulmi subsp. americana and O. ulmi were generated to place the intron within the context of the ribosomal RNA. The in vivo splicing of the O.ul-mS952 group II intron was confirmed with reverse transcription-PCR. A survey of 182 strains of Dutch Elm Diseases causing agents showed that the mS952 intron was absent in what is considered to be the more aggressive species O. novo-ulmi but present in strains of the less aggressive O. ulmi. This observation suggests that the O.ul-mS952 intron can be used as a PCR-based molecular marker to discriminate between O. ulmi and O. novo-ulmi subsp. americana.Keywords: Dutch Elm Disease, group II introns, mtDNA, species identification
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14577760 Application of Fuzzy Neural Network for Image Tumor Description
Authors: Nahla Ibraheem Jabbar, Monica Mehrotra
Abstract:
This paper used a fuzzy kohonen neural network for medical image segmentation. Image segmentation plays a important role in the many of medical imaging applications by automating or facilitating the diagnostic. The paper analyses the tumor by extraction of the features of (area, entropy, means and standard deviation).These measurements gives a description for a tumor.
Keywords: FCM, features extraction, medical image processing, neural network, segmentation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21087759 A Sequential Pattern Mining Method Based On Sequential Interestingness
Authors: Shigeaki Sakurai, Youichi Kitahara, Ryohei Orihara
Abstract:
Sequential mining methods efficiently discover all frequent sequential patterns included in sequential data. These methods use the support, which is the previous criterion that satisfies the Apriori property, to evaluate the frequency. However, the discovered patterns do not always correspond to the interests of analysts, because the patterns are common and the analysts cannot get new knowledge from the patterns. The paper proposes a new criterion, namely, the sequential interestingness, to discover sequential patterns that are more attractive for the analysts. The paper shows that the criterion satisfies the Apriori property and how the criterion is related to the support. Also, the paper proposes an efficient sequential mining method based on the proposed criterion. Lastly, the paper shows the effectiveness of the proposed method by applying the method to two kinds of sequential data.
Keywords: Sequential mining, Support, Confidence, Apriori property
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 12757758 Shape Optimization of Permanent Magnet Motors Using the Reduced Basis Technique
Authors: A. Jabbari, M. Shakeri, A. Nabavi
Abstract:
In this paper, a tooth shape optimization method for cogging torque reduction in Permanent Magnet (PM) motors is developed by using the Reduced Basis Technique (RBT) coupled by Finite Element Analysis (FEA) and Design of Experiments (DOE) methods. The primary objective of the method is to reduce the enormous number of design variables required to define the tooth shape. RBT is a weighted combination of several basis shapes. The aim of the method is to find the best combination using the weights for each tooth shape as the design variables. A multi-level design process is developed to find suitable basis shapes or trial shapes at each level that can be used in the reduced basis technique. Each level is treated as a separated optimization problem until the required objective – minimum cogging torque – is achieved. The process is started with geometrically simple basis shapes that are defined by their shape co-ordinates. The experimental design of Taguchi method is used to build the approximation model and to perform optimization. This method is demonstrated on the tooth shape optimization of a 8-poles/12-slots PM motor.Keywords: PM motor, cogging torque, tooth shape optimization, RBT, FEA, DOE.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 25027757 Pd(II) Complex with 4-Bromo-Bis-Hydroxymethyl Phenol and Nicotinamide: Synthesis and Spectral Analysis
Authors: Özlen Altun, Zeliha Yoruç
Abstract:
In the present study, the reactions involving 4-bromo-2,6-bis-hydroxymethyl-phenol (BBHMP) and nicotinamide (NA) in the presence of Pd(II) ions were investigated. Optimum conditions for the reactions were established as pH = 7 and λ = 450 nm. According to absorbance measurements, the molar ratio of BBHMP: NA: Pd2+ was found to be 1: 2: 2. As a result of physicochemical, spectrophotometric and thermal analyses, the reactions of BBHMP and NA with Pd(II) are complexation reactions and one molecule of BBHMP and two molecules of NA react with two molecules of the Pd(II) ion.
Keywords: Nicotinamide, 4-bromo-2, 6-bis-hydroxymethyl-phenol, Pd(II), spectral analysis, synthesis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 7287756 A Finite Element/Finite Volume Method for Dam-Break Flows over Deformable Beds
Authors: Alia Alghosoun, Ashraf Osman, Mohammed Seaid
Abstract:
A coupled two-layer finite volume/finite element method was proposed for solving dam-break flow problem over deformable beds. The governing equations consist of the well-balanced two-layer shallow water equations for the water flow and a linear elastic model for the bed deformations. Deformations in the topography can be caused by a brutal localized force or simply by a class of sliding displacements on the bathymetry. This deformation in the bed is a source of perturbations, on the water surface generating water waves which propagate with different amplitudes and frequencies. Coupling conditions at the interface are also investigated in the current study and two mesh procedure is proposed for the transfer of information through the interface. In the present work a new procedure is implemented at the soil-water interface using the finite element and two-layer finite volume meshes with a conservative distribution of the forces at their intersections. The finite element method employs quadratic elements in an unstructured triangular mesh and the finite volume method uses the Rusanove to reconstruct the numerical fluxes. The numerical coupled method is highly efficient, accurate, well balanced, and it can handle complex geometries as well as rapidly varying flows. Numerical results are presented for several test examples of dam-break flows over deformable beds. Mesh convergence study is performed for both methods, the overall model provides new insight into the problems at minimal computational cost.Keywords: Dam-break flows, deformable beds, finite element method, finite volume method, linear elasticity, Shallow water equations.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9127755 Occurrence of Foreign Matter in Food: Applied Identification Method - Association of Official Agricultural Chemists (AOAC) and Food and Drug Administration (FDA)
Authors: E. C. Mattos, V. S. M. G. Daros, R. Dal Col, A. L. Nascimento
Abstract:
The aim of this study is to present the results of a retrospective survey on the foreign matter found in foods analyzed at the Adolfo Lutz Institute, from July 2001 to July 2015. All the analyses were conducted according to the official methods described on Association of Official Agricultural Chemists (AOAC) for the micro analytical procedures and Food and Drug Administration (FDA) for the macro analytical procedures. The results showed flours, cereals and derivatives such as baking and pasta products were the types of food where foreign matters were found more frequently followed by condiments and teas. Fragments of stored grains insects, its larvae, nets, excrement, dead mites and rodent excrement were the most foreign matter found in food. Besides, foreign matters that can cause a physical risk to the consumer’s health such as metal, stones, glass, wood were found but rarely. Miscellaneous (shell, sand, dirt and seeds) were also reported. There are a lot of extraneous materials that are considered unavoidable since are something inherent to the product itself, such as insect fragments in grains. In contrast, there are avoidable extraneous materials that are less tolerated because it is preventable with the Good Manufacturing Practice. The conclusion of this work is that although most extraneous materials found in food are considered unavoidable it is necessary to keep the Good Manufacturing Practice throughout the food processing as well as maintaining a constant surveillance of the production process in order to avoid accidents that may lead to occurrence of these extraneous materials in food.Keywords: Food contamination, extraneous materials, foreign matter, surveillance.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 37017754 Human Tracking across Heterogeneous Systems Based On Mobile Agent Technologies
Authors: Tappei Yotsumoto, Atsushi Nomura, Kozo Tanigawa, Kenichi Takahashi, Takao Kawamura, Kazunori Sugahara
Abstract:
In a human tracking system, expanding a monitoring range of one system is complicating the management of devices and increasing its cost. Therefore, we propose a method to realize a wide-range human tracking by connecting small systems. In this paper, we examined an agent deploy method and information contents across the heterogeneous human tracking systems. By implementing the proposed method, we can construct a human tracking system across heterogeneous systems, and the system can track a target continuously between systems.
Keywords: Human tracking system, mobile agent, monitoring.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15167753 Design Systems and the Need for a Usability Method: Assessing the Fitness of Components and Interaction Patterns in Design Systems Using Atmosphere Methodology
Authors: P. Johansson, S. Mardh
Abstract:
The present study proposes a usability test method, Atmosphere, to assess the fitness of components and interaction patterns of design systems. The method covers the user’s perception of the components of the system, the efficiency of the logic of the interaction patterns, perceived ease of use as well as the user’s understanding of the intended outcome of interactions. These aspects are assessed by combining measures of first impression, visual affordance and expectancy. The method was applied to a design system developed for the design of an electronic health record system. The study was conducted involving 15 healthcare personnel. It could be concluded that the Atmosphere method provides tangible data that enable human-computer interaction practitioners to analyze and categorize components and patterns based on perceived usability, success rate of identifying interactive components and success rate of understanding components and interaction patterns intended outcome.
Keywords: atomic design, atmosphere methodology, design system, expectancy testing, first impression testing, usability testing, visual affordance testing
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 8517752 Discovery of Time Series Event Patterns based on Time Constraints from Textual Data
Authors: Shigeaki Sakurai, Ken Ueno, Ryohei Orihara
Abstract:
This paper proposes a method that discovers time series event patterns from textual data with time information. The patterns are composed of sequences of events and each event is extracted from the textual data, where an event is characteristic content included in the textual data such as a company name, an action, and an impression of a customer. The method introduces 7 types of time constraints based on the analysis of the textual data. The method also evaluates these constraints when the frequency of a time series event pattern is calculated. We can flexibly define the time constraints for interesting combinations of events and can discover valid time series event patterns which satisfy these conditions. The paper applies the method to daily business reports collected by a sales force automation system and verifies its effectiveness through numerical experiments.
Keywords: Text mining, sequential mining, time constraints, daily business reports.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14877751 In-Situ EBSD Observations of Bending for Single-Crystalline Pure Copper
Authors: Takashi Sakai, Saori Yoshikawa, Hideo Morimoto
Abstract:
To understand the material characteristics of singleand poly-crystals of pure copper, the respective relationships between crystallographic orientations and microstructures, and the bending and mechanical properties were examined. And texture distribution is also analyzed. A bending test is performed in a SEM apparatus and while its behaviors are observed in situ. Furthermore, some analytical results related to crystal direction maps, inverse pole figures, and textures were obtained from EBSD analyses.
Keywords: Pure Copper, Bending, Single Crystal, SEM-EBSD Analysis, Texture, Microstructure
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18457750 Seismic Behavior Evaluation of Semi-Rigid Steel Frames with Knee Bracing by Modal Pushover Analysis (MPA)
Authors: Farzan Namvari, Panam Zarfam
Abstract:
Nowadays use of a new structural bracing system called 'Knee Bracing System' have taken the specialists attention too much. On the other hand nonlinear static analysis procedures in estimate structures performance in earthquake time have taken attention too much. One of these procedure is modal pushover analysis (MPA) procedure. The accuracy of MPA procedure for simple steel moment resisting frame has been verified and considered in Chintanapakdee and Chopra-s article in 2003. Since the accuracy of MPA procedure has not verified for semi-rigid steel frames with knee bracing, we are going to get through with this matter in this study. For this purpose, the selected structures are four frames with different heights, 5 to 20 stories, will be designed according to AISC criteria. Then MPA procedure is used for the same frames with different rigidity percentiles of connections. The results of seismic responses are compared with dynamic nonlinear response history analysis as exact procedure and accuracy of MPA procedure is evaluated. It seems that MPA procedure accuracy will come down by reduction of the rigidity percentiles of semi-rigid connections.Keywords: Knee Bracing, Modal Pushover Analysis, SeismicBehavior, Semi-Rigid Connections.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21327749 Performance Evaluation of Acoustic-Spectrographic Voice Identification Method in Native and Non-Native Speech
Authors: E. Krasnova, E. Bulgakova, V. Shchemelinin
Abstract:
The paper deals with acoustic-spectrographic voice identification method in terms of its performance in non-native language speech. Performance evaluation is conducted by comparing the result of the analysis of recordings containing native language speech with recordings that contain foreign language speech. Our research is based on Tajik and Russian speech of Tajik native speakers due to the character of the criminal situation with drug trafficking. We propose a pilot experiment that represents a primary attempt enter the field.Keywords: Speaker identification, acoustic-spectrographic method, non-native speech.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 8657748 Determination of Optimum Length of Framesand Number of Vectors to Compress ECG Signals
Authors: Rafet Akdeniz, Pınar Tüfekçi, B.Sıddık Yarman
Abstract:
In this study, to compress ECG signals, KLT (Karhunen- Loeve Transform) method has been used. The purpose of this method is to perform effective ECG coding by a correlation between the length of frames and the number of vectors of ECG signals.Keywords: ECG Compression, EKG Compression.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14847747 Least Squares Method Identification of Corona Current-Voltage Characteristics and Electromagnetic Field in Electrostatic Precipitator
Authors: H. Nouri, I. E. Achouri, A. Grimes, H. Ait Said, M. Aissou, Y. Zebboudj
Abstract:
This paper aims to analysis the behavior of DC corona discharge in wire-to-plate electrostatic precipitators (ESP). Currentvoltage curves are particularly analyzed. Experimental results show that discharge current is strongly affected by the applied voltage. The proposed method of current identification is to use the method of least squares. Least squares problems that of into two categories: linear or ordinary least squares and non-linear least squares, depending on whether or not the residuals are linear in all unknowns. The linear least-squares problem occurs in statistical regression analysis; it has a closed-form solution. A closed-form solution (or closed form expression) is any formula that can be evaluated in a finite number of standard operations. The non-linear problem has no closed-form solution and is usually solved by iterative.Keywords: Electrostatic precipitator, current-voltage characteristics, Least Squares method, electric field, magnetic field.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20967746 The Economic Cost of Health and Safety in Work Places: An Approach on the Costs Calculating Model
Authors: Efat Lali Dastjerdi, Hassan Sadeghi Naeini, Hadi Sanjari
Abstract:
One of the important steps in a safety and risk management system is the economical evaluation of occupational accident and diseases costs in order to decrease accidents from reoccurring in the workplace. This study proposed a plausible method for calculating occupational accident costs and illnesses in work place. This method design for cost estimation takes into account both the personnel, organizational level as well as the community level especially intended for an Iranian work place. The research indicates that a using systematic method for calculating costs which also provides risk evaluation can help managers to plan correctly the investment in health and safety measures. Using this method is that not only is it comprehensive, easy and practical and could be applied in practice by a manager within a short period of time but it also shows the importance of accident costs as well as calculates the real cost of an accident and illnesses.
Keywords: ost calculating model, Economics of health and Safety.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22307745 A Holistic Workflow Modeling Method for Business Process Redesign
Authors: Heejung Lee
Abstract:
In a highly competitive environment, it becomes more important to shorten the whole business process while delivering or even enhancing the business value to the customers and suppliers. Although the workflow management systems receive much attention for its capacity to practically support the business process enactment, the effective workflow modeling method remain still challenging and the high degree of process complexity makes it more difficult to gain the short lead time. This paper presents a workflow structuring method in a holistic way that can reduce the process complexity using activity-needs and formal concept analysis, which eventually enhances the key performance such as quality, delivery, and cost in business process.
Keywords: Workflow management, reengineering, formal concept analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1949