Search results for: metallurgical image processing
3246 The Brain’s Attenuation Coefficient as a Potential Estimator of Temperature Elevation during Intracranial High Intensity Focused Ultrasound Procedures
Authors: Daniel Dahis, Haim Azhari
Abstract:
Noninvasive image-guided intracranial treatments using high intensity focused ultrasound (HIFU) are on the course of translation into clinical applications. They include, among others, tumor ablation, hyperthermia, and blood-brain-barrier (BBB) penetration. Since many of these procedures are associated with local temperature elevation, thermal monitoring is essential. MRI constitutes an imaging method with high spatial resolution and thermal mapping capacity. It is the currently leading modality for temperature guidance, commonly under the name MRgHIFU (magnetic-resonance guided HIFU). Nevertheless, MRI is a very expensive non-portable modality which jeopardizes its accessibility. Ultrasonic thermal monitoring, on the other hand, could provide a modular, cost-effective alternative with higher temporal resolution and accessibility. In order to assess the feasibility of ultrasonic brain thermal monitoring, this study investigated the usage of brain tissue attenuation coefficient (AC) temporal changes as potential estimators of thermal changes. Newton's law of cooling describes a temporal exponential decay behavior for the temperature of a heated object immersed in a relatively cold surrounding. Similarly, in the case of cerebral HIFU treatments, the temperature in the region of interest, i.e., focal zone, is suggested to follow the same law. Thus, it was hypothesized that the AC of the irradiated tissue may follow a temporal exponential behavior during cool down regime. Three ex-vivo bovine brain tissue specimens were inserted into plastic containers along with four thermocouple probes in each sample. The containers were placed inside a specially built ultrasonic tomograph and scanned at room temperature. The corresponding pixel-averaged AC was acquired for each specimen and used as a reference. Subsequently, the containers were placed in a beaker containing hot water and gradually heated to about 45ᵒC. They were then repeatedly rescanned during cool down using ultrasonic through-transmission raster trajectory until reaching about 30ᵒC. From the obtained images, the normalized AC and its temporal derivative as a function of temperature and time were registered. The results have demonstrated high correlation (R² > 0.92) between both the brain AC and its temporal derivative to temperature. This indicates the validity of the hypothesis and the possibility of obtaining brain tissue temperature estimation from the temporal AC thermal changes. It is important to note that each brain yielded different AC values and slopes. This implies that a calibration step is required for each specimen. Thus, for a practical acoustic monitoring of the brain, two steps are suggested. The first step consists of simply measuring the AC at normal body temperature. The second step entails measuring the AC after small temperature elevation. In face of the urging need for a more accessible thermal monitoring technique for brain treatments, the proposed methodology enables a cost-effective high temporal resolution acoustical temperature estimation during HIFU treatments.Keywords: attenuation coefficient, brain, HIFU, image-guidance, temperature
Procedia PDF Downloads 1613245 Improving the Accuracy of Stress Intensity Factors Obtained by Scaled Boundary Finite Element Method on Hybrid Quadtree Meshes
Authors: Adrian W. Egger, Savvas P. Triantafyllou, Eleni N. Chatzi
Abstract:
The scaled boundary finite element method (SBFEM) is a semi-analytical numerical method, which introduces a scaling center in each element’s domain, thus transitioning from a Cartesian reference frame to one resembling polar coordinates. Consequently, an analytical solution is achieved in radial direction, implying that only the boundary need be discretized. The only limitation imposed on the resulting polygonal elements is that they remain star-convex. Further arbitrary p- or h-refinement may be applied locally in a mesh. The polygonal nature of SBFEM elements has been exploited in quadtree meshes to alleviate all issues conventionally associated with hanging nodes. Furthermore, since in 2D this results in only 16 possible cell configurations, these are precomputed in order to accelerate the forward analysis significantly. Any cells, which are clipped to accommodate the domain geometry, must be computed conventionally. However, since SBFEM permits polygonal elements, significantly coarser meshes at comparable accuracy levels are obtained when compared with conventional quadtree analysis, further increasing the computational efficiency of this scheme. The generalized stress intensity factors (gSIFs) are computed by exploiting the semi-analytical solution in radial direction. This is initiated by placing the scaling center of the element containing the crack at the crack tip. Taking an analytical limit of this element’s stress field as it approaches the crack tip, delivers an expression for the singular stress field. By applying the problem specific boundary conditions, the geometry correction factor is obtained, and the gSIFs are then evaluated based on their formal definition. Since the SBFEM solution is constructed as a power series, not unlike mode superposition in FEM, the two modes contributing to the singular response of the element can be easily identified in post-processing. Compared to the extended finite element method (XFEM) this approach is highly convenient, since neither enrichment terms nor a priori knowledge of the singularity is required. Computation of the gSIFs by SBFEM permits exceptional accuracy, however, when combined with hybrid quadtrees employing linear elements, this does not always hold. Nevertheless, it has been shown that crack propagation schemes are highly effective even given very coarse discretization since they only rely on the ratio of mode one to mode two gSIFs. The absolute values of the gSIFs may still be subject to large errors. Hence, we propose a post-processing scheme, which minimizes the error resulting from the approximation space of the cracked element, thus limiting the error in the gSIFs to the discretization error of the quadtree mesh. This is achieved by h- and/or p-refinement of the cracked element, which elevates the amount of modes present in the solution. The resulting numerical description of the element is highly accurate, with the main error source now stemming from its boundary displacement solution. Numerical examples show that this post-processing procedure can significantly improve the accuracy of the computed gSIFs with negligible computational cost even on coarse meshes resulting from hybrid quadtrees.Keywords: linear elastic fracture mechanics, generalized stress intensity factors, scaled finite element method, hybrid quadtrees
Procedia PDF Downloads 1463244 Studying the Effect of Nanoclays on the Mechanical Properties of Polypropylene/Polyamide Nanocomposites
Authors: Benalia Kouini, Aicha Serier
Abstract:
Nanocomposites based on polypropylene/polyamide 66 (PP/PA66) nanoblends containing organophilic montmorillonite (OMMT) and maleic anhydride grafted polypropylene (PP-g-MAH) were prepared by melt compounding method followed by injection molding. Two different types of nanoclays were used in this work. DELLITE LVF is the untreated nanoclay and DELLITE 67G is the treated one. The morphology of the nanocomposites was studied using the XR diffraction (XRD). The results indicate that the incorporation of treated nanoclay has a significant effect on the impact strength of PP/PA66 nanocomposites. Furthermore, it was found that XRD results revealed the intercalation, exfoliation of nanaclays of nanocomposites.Keywords: nNanoclay, Nanocomposites, Polypropylene, Polyamide, melt processing, mechanical properties.
Procedia PDF Downloads 3543243 Effects of Animal Metaphor on Consumer Response to Product Advertising
Authors: Wen-Hsien Huang, Hsu-Ting Hsu
Abstract:
While advertisers often use animal metaphors to promote product performance, representing through the use of a product image together with an animal-like messenger to imply the undesirable health states of not using the product, the effect of such metaphors on persuasion remains unclear. The current research addresses this issue by investigating how consumers perceive and react to animal metaphor advertising in the context of product promotion. Three studies are carried out using field and experimental data. The findings demonstrate that animal metaphor ads are less persuasive than non-metaphor ads and that ads with animal-like messengers (as opposed to human messengers) activate stronger dehumanization perceptions, which in turn lead to lower product choice, product evaluation and purchase intention, regardless of whether the animal metaphors are presented visually in the picture or verbally in the headline. Furthermore, when the metaphorical pairing includes a more disliked animal, consumer reaction was less favorable. The implications of the findings for advertisers considering the use of animalized messengers are discussed.Keywords: animal metaphor, dehumanization, product evaluation, health communication
Procedia PDF Downloads 843242 Simultaneous Measurement of Wave Pressure and Wind Speed with the Specific Instrument and the Unit of Measurement Description
Authors: Branimir Jurun, Elza Jurun
Abstract:
The focus of this paper is the description of an instrument called 'Quattuor 45' and defining of wave pressure measurement. Special attention is given to measurement of wave pressure created by the wind speed increasing obtained with the instrument 'Quattuor 45' in the investigated area. The study begins with respect to theoretical attitudes and numerous up to date investigations related to the waves approaching the coast. The detailed schematic view of the instrument is enriched with pictures from ground plan and side view. Horizontal stability of the instrument is achieved by mooring which relies on two concrete blocks. Vertical wave peak monitoring is ensured by one float above the instrument. The synthesis of horizontal stability and vertical wave peak monitoring allows to create a representative database for wave pressure measuring. Instrument ‘Quattuor 45' is named according to the way the database is received. Namely, the electronic part of the instrument consists of the main chip ‘Arduino', its memory, four load cells with the appropriate modules and the wind speed sensor 'Anemometers'. The 'Arduino' chip is programmed to store two data from each load cell and two data from the anemometer on SD card each second. The next part of the research is dedicated to data processing. All measured results are stored automatically in the database and after that detailed processing is carried out in the MS Excel. The result of the wave pressure measurement is synthesized by the unit of measurement kN/m². This paper also suggests a graphical presentation of the results by multi-line graph. The wave pressure is presented on the left vertical axis, while the wind speed is shown on the right vertical axis. The time of measurement is displayed on the horizontal axis. The paper proposes an algorithm for wind speed measurements showing the results for two characteristic winds in the Adriatic Sea, called 'Bura' and 'Jugo'. The first of them is the northern wind that reaches high speeds, causing low and extremely steep waves, where the pressure of the wave is relatively weak. On the other hand, the southern wind 'Jugo' has a lower speed than the northern wind, but due to its constant duration and constant speed maintenance, it causes extremely long and high waves that cause extremely high wave pressure.Keywords: instrument, measuring unit, waves pressure metering, wind seed measurement
Procedia PDF Downloads 1983241 Automatic LV Segmentation with K-means Clustering and Graph Searching on Cardiac MRI
Authors: Hae-Yeoun Lee
Abstract:
Quantification of cardiac function is performed by calculating blood volume and ejection fraction in routine clinical practice. However, these works have been performed by manual contouring,which requires computational costs and varies on the observer. In this paper, an automatic left ventricle segmentation algorithm on cardiac magnetic resonance images (MRI) is presented. Using knowledge on cardiac MRI, a K-mean clustering technique is applied to segment blood region on a coil-sensitivity corrected image. Then, a graph searching technique is used to correct segmentation errors from coil distortion and noises. Finally, blood volume and ejection fraction are calculated. Using cardiac MRI from 15 subjects, the presented algorithm is tested and compared with manual contouring by experts to show outstanding performance.Keywords: cardiac MRI, graph searching, left ventricle segmentation, K-means clustering
Procedia PDF Downloads 3993240 Facial Emotion Recognition with Convolutional Neural Network Based Architecture
Authors: Koray U. Erbas
Abstract:
Neural networks are appealing for many applications since they are able to learn complex non-linear relationships between input and output data. As the number of neurons and layers in a neural network increase, it is possible to represent more complex relationships with automatically extracted features. Nowadays Deep Neural Networks (DNNs) are widely used in Computer Vision problems such as; classification, object detection, segmentation image editing etc. In this work, Facial Emotion Recognition task is performed by proposed Convolutional Neural Network (CNN)-based DNN architecture using FER2013 Dataset. Moreover, the effects of different hyperparameters (activation function, kernel size, initializer, batch size and network size) are investigated and ablation study results for Pooling Layer, Dropout and Batch Normalization are presented.Keywords: convolutional neural network, deep learning, deep learning based FER, facial emotion recognition
Procedia PDF Downloads 2743239 A Semiparametric Approach to Estimate the Mode of Continuous Multivariate Data
Authors: Tiee-Jian Wu, Chih-Yuan Hsu
Abstract:
Mode estimation is an important task, because it has applications to data from a wide variety of sources. We propose a semi-parametric approach to estimate the mode of an unknown continuous multivariate density function. Our approach is based on a weighted average of a parametric density estimate using the Box-Cox transform and a non-parametric kernel density estimate. Our semi-parametric mode estimate improves both the parametric- and non-parametric- mode estimates. Specifically, our mode estimate solves the non-consistency problem of parametric mode estimates (at large sample sizes) and reduces the variability of non-parametric mode estimates (at small sample sizes). The performance of our method at practical sample sizes is demonstrated by simulation examples and two real examples from the fields of climatology and image recognition.Keywords: Box-Cox transform, density estimation, mode seeking, semiparametric method
Procedia PDF Downloads 2853238 Application of Unmanned Aerial Vehicle in Urban Rail Transit Intelligent Inspection
Authors: Xinglu Nie, Feifei Tang, Chuntao Wei, Zhimin Ruan, Qianhong Zhu
Abstract:
Current method of manual-style inspection can not fully meet the requirement of the urban rail transit security in China. In this paper, an intelligent inspection method using unmanned aerial vehicle (UAV) is utilized. A series of orthophoto of rail transit monitored area was collected by UAV, image correction and registration were operated among multi-phase images, then the change detection was used to detect the changes, judging the engineering activities and human activities that may become potential threats to the security of urban rail. Not only qualitative judgment, but also quantitative judgment of changes in the security control area can be provided by this method, which improves the objectives and efficiency of the patrol results. The No.6 line of Chongqing Municipality was taken as an example to verify the validation of this method.Keywords: rail transit, control of protected areas, intelligent inspection, UAV, change detection
Procedia PDF Downloads 3703237 Dyeing with Natural Dye from Pterocarpus indicus Extract Using Eco-Friendly Mordants
Authors: Ploysai Ohama, Nuttawadee Hanchengchai, Thiva Saksri
Abstract:
Natural dye extracted from Pterocarpus indicus was applied to a cotton fabric and silk yarn by dyeing processing different eco-friendly mordants. Analytical studies such as UV–VIS spectrophotometry and gravimetric analysis were performed on the extracts. The color of each dyed material was investigated in terms of the CIELAB (L*, a* and b*) and K/S values. Cotton fabric dyed without mordants had a shade of greenish-brown, while those post-mordanted with selected eco-friendly mordants such as alum, lemon juice and limewater result in a variety of brown and darker color shade of fabric.Keywords: natural dyes, plant materials, dyeing, mordant
Procedia PDF Downloads 4153236 Extending Image Captioning to Video Captioning Using Encoder-Decoder
Authors: Sikiru Ademola Adewale, Joe Thomas, Bolanle Hafiz Matti, Tosin Ige
Abstract:
This project demonstrates the implementation and use of an encoder-decoder model to perform a many-to-many mapping of video data to text captions. The many-to-many mapping occurs via an input temporal sequence of video frames to an output sequence of words to form a caption sentence. Data preprocessing, model construction, and model training are discussed. Caption correctness is evaluated using 2-gram BLEU scores across the different splits of the dataset. Specific examples of output captions were shown to demonstrate model generality over the video temporal dimension. Predicted captions were shown to generalize over video action, even in instances where the video scene changed dramatically. Model architecture changes are discussed to improve sentence grammar and correctness.Keywords: decoder, encoder, many-to-many mapping, video captioning, 2-gram BLEU
Procedia PDF Downloads 1083235 Computational Tool for Surface Electromyography Analysis; an Easy Way for Non-Engineers
Authors: Fabiano Araujo Soares, Sauro Emerick Salomoni, Joao Paulo Lima da Silva, Igor Luiz Moura, Adson Ferreira da Rocha
Abstract:
This paper presents a tool developed in the Matlab platform. It was developed to simplify the analysis of surface electromyography signals (S-EMG) in a way accessible to users that are not familiarized with signal processing procedures. The tool receives data by commands in window fields and generates results as graphics and excel tables. The underlying math of each S-EMG estimator is presented. Setup window and result graphics are presented. The tool was presented to four non-engineer users and all of them managed to appropriately use it after a 5 minutes instruction period.Keywords: S-EMG estimators, electromyography, surface electromyography, ARV, RMS, MDF, MNF, CV
Procedia PDF Downloads 5593234 Residential Architecture and Its Representation in Movies: Bangkok's Spatial Research in the Study of Thai Cinematography
Authors: Janis Matvejs
Abstract:
Visual representation of a city creates unique perspectives that allow to interpret the urban environment and enable to understand a space that is culturally created and territorially organized. Residential complexes are an essential part of cities and cinema is a specific representation form of these areas. There has been very little research done on exploring how these areas are depicted in the Thai movies. The aim of this research is to interpret the discourse of residential areas of Bangkok throughout the 20th and 21st centuries and to examine essential changes in the residential structure. Specific cinematic formal techniques in relation to the urban image were used. The movie review results were compared with changes in Bangkok’s residential development. Movie analysis displayed that residential areas are frequently used in Thai cinematography and they make up an integral part of the urban visual perception.Keywords: Bangkok, cinema, residential area, representation, visual perception
Procedia PDF Downloads 1943233 Adaptive Few-Shot Deep Metric Learning
Authors: Wentian Shi, Daming Shi, Maysam Orouskhani, Feng Tian
Abstract:
Whereas currently the most prevalent deep learning methods require a large amount of data for training, few-shot learning tries to learn a model from limited data without extensive retraining. In this paper, we present a loss function based on triplet loss for solving few-shot problem using metric based learning. Instead of setting the margin distance in triplet loss as a constant number empirically, we propose an adaptive margin distance strategy to obtain the appropriate margin distance automatically. We implement the strategy in the deep siamese network for deep metric embedding, by utilizing an optimization approach by penalizing the worst case and rewarding the best. Our experiments on image recognition and co-segmentation model demonstrate that using our proposed triplet loss with adaptive margin distance can significantly improve the performance.Keywords: few-shot learning, triplet network, adaptive margin, deep learning
Procedia PDF Downloads 1713232 Effect of Friction Parameters on the Residual Bagging Behaviors of Denim Fabrics
Authors: M. Gazzah, B. Jaouachi, F. Sakli
Abstract:
This research focuses on the yarn-to-yarn and metal-to-fabric friction effects on the residual bagging behavior expressed by residual bagging height, volume and recovery of some denim fabrics. The results show, that both residual bagging height and residual bagging volume, which is determined using image analysis method, are significantly affected due to the most influential fabric parameter variations, the weft yarns density and the mean frictional coefficients. After the applied number of fatigue cycles, the findings revealed that the weft yarn rigidity contributes on fabric bagging behavior accurately. Among the tested samples, our results show that the elastic fabrics present a high recovery ability to give low bagging height and volume values.Keywords: bagging recovery, denim fabric, metal-to-fabric friction, residual bagging height, yarn-to-yarn friction
Procedia PDF Downloads 5773231 The Nature of the Complicated Fabric Textures: How to Represent in Primary Visual Cortex
Authors: J. L. Liu, L. Wang, B. Zhu, J. Zhou, W. D. Gao
Abstract:
Fabric textures are very common in our daily life. However, we never explore the representation of fabric textures from neuroscience view. Theoretical studies suggest that primary visual cortex (V1) uses a sparse code to efficiently represent natural images. However, how the simple cells in V1 encode the artificial textures is still a mystery. So, here we will take fabric texture as stimulus to study the response of independent component analysis that is established to model the receptive field of simple cells in V1. Experimental results based on 140 classical fabric images indicate that the receptive fields of simple cells have obvious selectivity in orientation, frequency, and phase when drifting gratings are used to determine their tuning properties. Additionally, the distribution of optimal orientation and frequency shows that the patch size selected from each original fabric image has a significant effect on the frequency selectivity.Keywords: fabric texture, receptive filed, simple cell, spare coding
Procedia PDF Downloads 4753230 The Idea of Reputation in a Post-Truth Era
Authors: Karen Armstrong
Abstract:
This paper considers the importance of acquiring, cultivating, and protecting one’s personal online reputation in a post-truth era. Although the idea of the individual is essential psychological construct, the concept necessarily now includes our online reputation. The idea of this online reputation has expanded to become almost more important than any other factor in terms of our professional, social and psychological development. The discussion will first consider philosophical ideas of the self, followed by an examination of underlying concepts of perception and interpretation in a post-truth world. Then, the idea of the recent shift to a consideration of posted images, through words and photos, in the construction of self, will be discussed. Next, the relation between private personal life and exterior social life, including our reputation in a variety of realms will be addressed. This will include the adoption of specific strategies and behaviors, which facilitate accuracy, currency and necessary modifications with regard to our online reputation. Finally, specific ways in which we can negotiate the fluid dynamic between reputation, and inner and outer selves to optimum effect will conclude the discussion.Keywords: image, post-truth, privacy, reputation, surveillance
Procedia PDF Downloads 2553229 Screening of Antagonistic/Synergistic Effect between Lactic Acid Bacteria (LAB) and Yeast Strains Isolated from Kefir
Authors: Mihriban Korukluoglu, Goksen Arik, Cagla Erdogan, Selen Kocakoglu
Abstract:
Kefir is a traditional fermented refreshing beverage which is known for its valuable and beneficial properties for human health. Mainly yeast species, lactic acid bacteria (LAB) strains and fewer acetic acid bacteria strains live together in a natural matrix named “kefir grain”, which is formed from various proteins and polysaccharides. Different microbial species live together in slimy kefir grain and it has been thought that synergetic effect could take place between microorganisms, which belong to different genera and species. In this research, yeast and LAB were isolated from kefir samples obtained from Uludag University Food Engineering Department. The cell morphology of isolates was screened by microscopic examination. Gram reactions of bacteria isolates were determined by Gram staining method, and as well catalase activity was examined. After observing the microscopic/morphological and physical, enzymatic properties of all isolates, they were divided into the groups as LAB and/or yeast according to their physicochemical responses to the applied examinations. As part of this research, the antagonistic/synergistic efficacy of the identified five LAB and five yeast strains to each other were determined individually by disk diffusion method. The antagonistic or synergistic effect is one of the most important properties in a co-culture system that different microorganisms are living together. The synergistic effect should be promoted, whereas the antagonistic effect is prevented to provide effective culture for fermentation of kefir. The aim of this study was to determine microbial interactions between identified yeast and LAB strains, and whether their effect is antagonistic or synergistic. Thus, if there is a strain which inhibits or retards the growth of other strains found in Kefir microflora, this circumstance shows the presence of antagonistic effect in the medium. Such negative influence should be prevented, whereas the microorganisms which have synergistic effect on each other should be promoted by combining them in kefir grain. Standardisation is the most desired property for industrial production. Each microorganism found in the microbial flora of a kefir grain should be identified individually. The members of the microbial community found in the glue-like kefir grain may be redesigned as a starter culture regarding efficacy of each microorganism to another in kefir processing. The main aim of this research was to shed light on more effective production of kefir grain and to contribute a standardisation of kefir processing in the food industry.Keywords: antagonistic effect, kefir, lactic acid bacteria (LAB), synergistic, yeast
Procedia PDF Downloads 2803228 Handling, Exporting and Archiving Automated Mineralogy Data Using TESCAN TIMA
Authors: Marek Dosbaba
Abstract:
Within the mining sector, SEM-based Automated Mineralogy (AM) has been the standard application for quickly and efficiently handling mineral processing tasks. Over the last decade, the trend has been to analyze larger numbers of samples, often with a higher level of detail. This has necessitated a shift from interactive sample analysis performed by an operator using a SEM, to an increased reliance on offline processing to analyze and report the data. In response to this trend, TESCAN TIMA Mineral Analyzer is designed to quickly create a virtual copy of the studied samples, thereby preserving all the necessary information. Depending on the selected data acquisition mode, TESCAN TIMA can perform hyperspectral mapping and save an X-ray spectrum for each pixel or segment, respectively. This approach allows the user to browse through elemental distribution maps of all elements detectable by means of energy dispersive spectroscopy. Re-evaluation of the existing data for the presence of previously unconsidered elements is possible without the need to repeat the analysis. Additional tiers of data such as a secondary electron or cathodoluminescence images can also be recorded. To take full advantage of these information-rich datasets, TIMA utilizes a new archiving tool introduced by TESCAN. The dataset size can be reduced for long-term storage and all information can be recovered on-demand in case of renewed interest. TESCAN TIMA is optimized for network storage of its datasets because of the larger data storage capacity of servers compared to local drives, which also allows multiple users to access the data remotely. This goes hand in hand with the support of remote control for the entire data acquisition process. TESCAN also brings a newly extended open-source data format that allows other applications to extract, process and report AM data. This offers the ability to link TIMA data to large databases feeding plant performance dashboards or geometallurgical models. The traditional tabular particle-by-particle or grain-by-grain export process is preserved and can be customized with scripts to include user-defined particle/grain properties.Keywords: Tescan, electron microscopy, mineralogy, SEM, automated mineralogy, database, TESCAN TIMA, open format, archiving, big data
Procedia PDF Downloads 1103227 Leveraging Unannotated Data to Improve Question Answering for French Contract Analysis
Authors: Touila Ahmed, Elie Louis, Hamza Gharbi
Abstract:
State of the art question answering models have recently shown impressive performance especially in a zero-shot setting. This approach is particularly useful when confronted with a highly diverse domain such as the legal field, in which it is increasingly difficult to have a dataset covering every notion and concept. In this work, we propose a flexible generative question answering approach to contract analysis as well as a weakly supervised procedure to leverage unannotated data and boost our models’ performance in general, and their zero-shot performance in particular.Keywords: question answering, contract analysis, zero-shot, natural language processing, generative models, self-supervision
Procedia PDF Downloads 1943226 The Satisfaction of International Tourists toward Thai Economy and Bangkok's Attributes
Authors: Ladaporn Pithuk
Abstract:
This research attempts to explore the satisfaction of international tourists toward Thai economy and Bangkok attributes. Due to tourism industry provides high rate of revenue for Thailand, and the outcome from this business drives every sections of Thailand. Unfortunately, some incidents in the country, such as some turmoil, have ruined the city’s image which obviously impacts to tourism industry. Hence, this survey was established to better understand the tourist’s satisfaction in these matters. The size of this research was 400 international tourists who visit Bangkok, Thailand during the 1st – 20th March 2009 and age between 20 – 65 years. The results reveal that tourists satisfy with all of Bangkok’s attributes including general attractions, heritage attraction, maintenance factors and cultural attraction. Also, tourists’ perception toward Thai politics is significantly related to their satisfaction of Bangkok’s attributes but their perception toward Thai economy is not significantly correlated to their satisfaction of Bangkok’s attributes.Keywords: Bangkok’s attributes, satisfaction of international tourists, Thai economy, and tourism industry
Procedia PDF Downloads 2773225 Quantitative Characterization of Single Orifice Hydraulic Flat Spray Nozzle
Authors: Y. C. Khoo, W. T. Lai
Abstract:
The single orifice hydraulic flat spray nozzle was evaluated with two global imaging techniques to characterize various aspects of the resulting spray. The two techniques were high resolution flow visualization and Particle Image Velocimetry (PIV). A CCD camera with 29 million pixels was used to capture shadowgraph images to realize ligament formation and collapse as well as droplet interaction. Quantitative analysis was performed to give the sizing information of the droplets and ligaments. This camera was then applied with a PIV system to evaluate the overall velocity field of the spray, from nozzle exit to droplet discharge. PIV images were further post-processed to determine the inclusion angle of the spray. The results from those investigations provided significant quantitative understanding of the spray structure. Based on the quantitative results, detailed understanding of the spray behavior was achieved.Keywords: spray, flow visualization, PIV, shadowgraph, quantitative sizing, velocity field
Procedia PDF Downloads 3823224 A Method to Evaluate and Compare Web Information Extractors
Authors: Patricia Jiménez, Rafael Corchuelo, Hassan A. Sleiman
Abstract:
Web mining is gaining importance at an increasing pace. Currently, there are many complementary research topics under this umbrella. Their common theme is that they all focus on applying knowledge discovery techniques to data that is gathered from the Web. Sometimes, these data are relatively easy to gather, chiefly when it comes from server logs. Unfortunately, there are cases in which the data to be mined is the data that is displayed on a web document. In such cases, it is necessary to apply a pre-processing step to first extract the information of interest from the web documents. Such pre-processing steps are performed using so-called information extractors, which are software components that are typically configured by means of rules that are tailored to extracting the information of interest from a web page and structuring it according to a pre-defined schema. Paramount to getting good mining results is that the technique used to extract the source information is exact, which requires to evaluate and compare the different proposals in the literature from an empirical point of view. According to Google Scholar, about 4 200 papers on information extraction have been published during the last decade. Unfortunately, they were not evaluated within a homogeneous framework, which leads to difficulties to compare them empirically. In this paper, we report on an original information extraction evaluation method. Our contribution is three-fold: a) this is the first attempt to provide an evaluation method for proposals that work on semi-structured documents; the little existing work on this topic focuses on proposals that work on free text, which has little to do with extracting information from semi-structured documents. b) It provides a method that relies on statistically sound tests to support the conclusions drawn; the previous work does not provide clear guidelines or recommend statistically sound tests, but rather a survey that collects many features to take into account as well as related work; c) We provide a novel method to compute the performance measures regarding unsupervised proposals; otherwise they would require the intervention of a user to compute them by using the annotations on the evaluation sets and the information extracted. Our contributions will definitely help researchers in this area make sure that they have advanced the state of the art not only conceptually, but from an empirical point of view; it will also help practitioners make informed decisions on which proposal is the most adequate for a particular problem. This conference is a good forum to discuss on our ideas so that we can spread them to help improve the evaluation of information extraction proposals and gather valuable feedback from other researchers.Keywords: web information extractors, information extraction evaluation method, Google scholar, web
Procedia PDF Downloads 2483223 Approaches to Promote Healthy Recreation Activities for Elderly Tourists at Bang Nam Phueng Floating Market, Prapradeang District, Samutprakarn Province
Authors: Sasitorn Chetanont
Abstract:
The objectives of this study are to find out the approaches to promote healthy recreation activities for elderly tourists and develop Bang Nam Phueng Floating Market to be a health tourism attraction. The research methodology was to analyze internal and external situations according to MP-MF and the MC-STEPS principles. As for the results of this study the researcher found that the healthy recreational activities for elderly tourists could be divided in 7 groups; travelling Bang Nam Phueng Floating Market activity, homestay relaxation, arts center platform activity, healthy massage activity, paying homage to a Buddha image activity, herbal joss-stick home activity, making local desserts and food activity.Keywords: elderly tourists, recreation activities, Bang Nam Phueng Floating Market, health tourism
Procedia PDF Downloads 4203222 Linguistic Analysis of Argumentation Structures in Georgian Political Speeches
Authors: Mariam Matiashvili
Abstract:
Argumentation is an integral part of our daily communications - formal or informal. Argumentative reasoning, techniques, and language tools are used both in personal conversations and in the business environment. Verbalization of the opinions requires the use of extraordinary syntactic-pragmatic structural quantities - arguments that add credibility to the statement. The study of argumentative structures allows us to identify the linguistic features that make the text argumentative. Knowing what elements make up an argumentative text in a particular language helps the users of that language improve their skills. Also, natural language processing (NLP) has become especially relevant recently. In this context, one of the main emphases is on the computational processing of argumentative texts, which will enable the automatic recognition and analysis of large volumes of textual data. The research deals with the linguistic analysis of the argumentative structures of Georgian political speeches - particularly the linguistic structure, characteristics, and functions of the parts of the argumentative text - claims, support, and attack statements. The research aims to describe the linguistic cues that give the sentence a judgmental/controversial character and helps to identify reasoning parts of the argumentative text. The empirical data comes from the Georgian Political Corpus, particularly TV debates. Consequently, the texts are of a dialogical nature, representing a discussion between two or more people (most often between a journalist and a politician). The research uses the following approaches to identify and analyze the argumentative structures Lexical Classification & Analysis - Identify lexical items that are relevant in argumentative texts creating process - Creating the lexicon of argumentation (presents groups of words gathered from a semantic point of view); Grammatical Analysis and Classification - means grammatical analysis of the words and phrases identified based on the arguing lexicon. Argumentation Schemas - Describe and identify the Argumentation Schemes that are most likely used in Georgian Political Speeches. As a final step, we analyzed the relations between the above mentioned components. For example, If an identified argument scheme is “Argument from Analogy”, identified lexical items semantically express analogy too, and they are most likely adverbs in Georgian. As a result, we created the lexicon with the words that play a significant role in creating Georgian argumentative structures. Linguistic analysis has shown that verbs play a crucial role in creating argumentative structures.Keywords: georgian, argumentation schemas, argumentation structures, argumentation lexicon
Procedia PDF Downloads 743221 An Enhanced Support Vector Machine Based Approach for Sentiment Classification of Arabic Tweets of Different Dialects
Authors: Gehad S. Kaseb, Mona F. Ahmed
Abstract:
Arabic Sentiment Analysis (SA) is one of the most common research fields with many open areas. Few studies apply SA to Arabic dialects. This paper proposes different pre-processing steps and a modified methodology to improve the accuracy using normal Support Vector Machine (SVM) classification. The paper works on two datasets, Arabic Sentiment Tweets Dataset (ASTD) and Extended Arabic Tweets Sentiment Dataset (Extended-AATSD), which are publicly available for academic use. The results show that the classification accuracy approaches 86%.Keywords: Arabic, classification, sentiment analysis, tweets
Procedia PDF Downloads 1493220 Quantitative Wide-Field Swept-Source Optical Coherence Tomography Angiography and Visual Outcomes in Retinal Artery Occlusion
Authors: Yifan Lu, Ying Cui, Ying Zhu, Edward S. Lu, Rebecca Zeng, Rohan Bajaj, Raviv Katz, Rongrong Le, Jay C. Wang, John B. Miller
Abstract:
Purpose: Retinal artery occlusion (RAO) is an ophthalmic emergency that can lead to poor visual outcome and is associated with an increased risk of cerebral stroke and cardiovascular events. Fluorescein angiography (FA) is the traditional diagnostic tool for RAO; however, wide-field swept-source optical coherence tomography angiography (WF SS-OCTA), as a nascent imaging technology, is able to provide quick and non-invasive angiographic information with a wide field of view. In this study, we looked for associations between OCT-A vascular metrics and visual acuity in patients with prior diagnosis of RAO. Methods: Patients with diagnoses of central retinal artery occlusion (CRAO) or branched retinal artery occlusion (BRAO) were included. A 6mm x 6mm Angio and a 15mm x 15mm AngioPlex Montage OCT-A image were obtained for both eyes in each patient using the Zeiss Plex Elite 9000 WF SS-OCTA device. Each 6mm x 6mm image was divided into nine Early Treatment Diabetic Retinopathy Study (ETDRS) subfields. The average measurement of the central foveal subfield, inner ring, and outer ring was calculated for each parameter. Non-perfusion area (NPA) was manually measured using 15mm x 15mm Montage images. A linear regression model was utilized to identify a correlation between the imaging metrics and visual acuity. A P-value less than 0.05 was considered to be statistically significant. Results: Twenty-five subjects were included in the study. For RAO eyes, there was a statistically significant negative correlation between vision and retinal thickness as well as superficial capillary plexus vessel density (SCP VD). A negative correlation was found between vision and deep capillary plexus vessel density (DCP VD) without statistical significance. There was a positive correlation between vision and choroidal thickness as well as choroidal volume without statistical significance. No statistically significant correlation was found between vision and the above metrics in contralateral eyes. For NPA measurements, no significant correlation was found between vision and NPA. Conclusions: This is the first study to our best knowledge to investigate the utility of WF SS-OCTA in RAO and to demonstrate correlations between various retinal vascular imaging metrics and visual outcomes. Further investigations should explore the associations between these imaging findings and cardiovascular risk as RAO patients are at elevated risk for symptomatic stroke. The results of this study provide a basis to understand the structural changes involved in visual outcomes in RAO. Furthermore, they may help guide management of RAO and prevention of cerebral stroke and cardiovascular accidents in patients with RAO.Keywords: OCTA, swept-source OCT, retinal artery occlusion, Zeiss Plex Elite
Procedia PDF Downloads 1393219 Spark Plasma Sintering/Synthesis of Alumina-Graphene Composites
Authors: Nikoloz Jalabadze, Roin Chedia, Lili Nadaraia, Levan Khundadze
Abstract:
Nanocrystalline materials in powder condition can be manufactured by a number of different methods, however manufacture of composite materials product in the same nanocrystalline state is still a problem because the processes of compaction and synthesis of nanocrystalline powders go with intensive growth of particles – the process which promotes formation of pieces in an ordinary crystalline state instead of being crystallized in the desirable nanocrystalline state. To date spark plasma sintering (SPS) has been considered as the most promising and energy efficient method for producing dense bodies of composite materials. An advantage of the SPS method in comparison with other methods is mainly low temperature and short time of the sintering procedure. That finally gives an opportunity to obtain dense material with nanocrystalline structure. Graphene has recently garnered significant interest as a reinforcing phase in composite materials because of its excellent electrical, thermal and mechanical properties. Graphene nanoplatelets (GNPs) in particular have attracted much interest as reinforcements for ceramic matrix composites (mostly in Al2O3, Si3N4, TiO2, ZrB2 a. c.). SPS has been shown to fully densify a variety of ceramic systems effectively including Al2O3 and often with improvements in mechanical and functional behavior. Alumina consolidated by SPS has been shown to have superior hardness, fracture toughness, plasticity and optical translucency compared to conventionally processed alumina. Knowledge of how GNPs influence sintering behavior is important to effectively process and manufacture process. In this study, the effects of GNPs on the SPS processing of Al2O3 are investigated by systematically varying sintering temperature, holding time and pressure. Our experiments showed that SPS process is also appropriate for the synthesis of nanocrystalline powders of alumina-graphene composites. Depending on the size of the molds, it is possible to obtain different amount of nanopowders. Investigation of the structure, physical-chemical, mechanical and performance properties of the elaborated composite materials was performed. The results of this study provide a fundamental understanding of the effects of GNP on sintering behavior, thereby providing a foundation for future optimization of the processing of these promising nanocomposite systems.Keywords: alumina oxide, ceramic matrix composites, graphene nanoplatelets, spark-plasma sintering
Procedia PDF Downloads 3763218 Processing Mild versus Strong Violations in Music: A Pilot Study Using Event-Related Potentials
Authors: Marie-Eve Joret, Marijn Van Vliet, Flavio Camarrone, Marc M. Van Hulle
Abstract:
Event-related potentials (ERPs) provide evidence that the human brain can process and understand music at a pre-attentive level. Music-specific ERPs include the Early Right Anterior Negativity (ERAN) and a late Negativity (N5). This study aims to further investigate this issue using two types of syntactic manipulations in music: mild violations, containing no out-of-key tones and strong violations, containing out-of-key tones. We will examine whether both manipulations will elicit the same ERPs.Keywords: ERAN ERPs, Music, N5, P3, ERPs, Music, N5 component, P3 component
Procedia PDF Downloads 2763217 N400 Investigation of Semantic Priming Effect to Symbolic Pictures in Text
Authors: Thomas Ousterhout
Abstract:
The purpose of this study was to investigate if incorporating meaningful pictures of gestures and facial expressions in short sentences of text could supplement the text with enough semantic information to produce and N400 effect when probe words incongruent to the picture were subsequently presented. Event-related potentials (ERPs) were recorded from a 14-channel commercial grade EEG headset while subjects performed congruent/incongruent reaction time discrimination tasks. Since pictures of meaningful gestures have been shown to be semantically processed in the brain in a similar manner as words are, it is believed that pictures will add supplementary information to text just as the inclusion of their equivalent synonymous word would. The hypothesis is that when subjects read the text/picture mixed sentences, they will process the images and words just like in face-to-face communication and therefore probe words incongruent to the image will produce an N400.Keywords: EEG, ERP, N400, semantics, congruency, facilitation, Emotiv
Procedia PDF Downloads 258