Search results for: incidental information processing
9823 Mapping of Arenga Pinnata Tree Using Remote Sensing
Authors: Zulkiflee Abd Latif, Sitinor Atikah Nordin, Alawi Sulaiman
Abstract:
Different tree species possess different and various benefits. Arenga Pinnata tree species own several potential uses that is valuable for the economy and the country. Mapping vegetation using remote sensing technique involves various process, techniques and consideration. Using satellite imagery, this method enables the access of inaccessible area and with the availability of near infra-red band; it is useful in vegetation analysis, especially in identifying tree species. Pixel-based and object-based classification technique is used as a method in this study. Pixel-based classification technique used in this study divided into unsupervised and supervised classification. Object based classification technique becomes more popular another alternative method in classification process. Using spectral, texture, color and other information, to classify the target make object-based classification is a promising technique for classification. Classification of Arenga Pinnata trees is overlaid with elevation, slope and aspect, soil and river data and several other data to give information regarding the tree character and living environment. This paper will present the utilization of remote sensing technique in order to map Arenga Pinnata tree speciesKeywords: Arenga Pinnata, pixel-based classification, object-based classification, remote sensing
Procedia PDF Downloads 3809822 Other-Generated Disclosure: A Challenge to Privacy on Social Network Sites
Authors: Tharntip Tawnie Chutikulrungsee, Oliver Kisalay Burmeister, Maumita Bhattacharya, Dragana Calic
Abstract:
Sharing on social network sites (SNSs) has rapidly emerged as a new social norm and has become a global phenomenon. Billions of users reveal not only their own information (self disclosure) but also information about others (other-generated disclosure), resulting in a risk and a serious threat to either personal or informational privacy. Self-disclosure (SD) has been extensively researched in the literature, particularly regarding control of individual and existing privacy management. However, far too little attention has been paid to other-generated disclosure (OGD), especially by insiders. OGD has a strong influence on self-presentation, self-image, and electronic word of mouth (eWOM). Moreover, OGD is more credible and less likely manipulated than SD, but lacks privacy control and legal protection to some extent. This article examines OGD in depth, ranging from motivation to both online and offline impacts, based upon lived experiences from both ‘the disclosed’ and ‘the discloser’. Using purposive sampling, this phenomenological study involves an online survey and in-depth interviews. The findings report the influence of peer disclosure as well as users’ strategies to mitigate privacy issues. This article also calls attention to the challenge of OGD privacy and inadequacies in the law related to privacy protection in the digital domain.Keywords: facebook, online privacy, other-generated disclosure, social networks sites (SNSs)
Procedia PDF Downloads 2519821 Applying Multiple Kinect on the Development of a Rapid 3D Mannequin Scan Platform
Authors: Shih-Wen Hsiao, Yi-Cheng Tsao
Abstract:
In the field of reverse engineering and creative industries, applying 3D scanning process to obtain geometric forms of the objects is a mature and common technique. For instance, organic objects such as faces and nonorganic objects such as products could be scanned to acquire the geometric information for further application. However, although the data resolution of 3D scanning device is increasing and there are more and more abundant complementary applications, the penetration rate of 3D scanning for the public is still limited by the relative high price of the devices. On the other hand, Kinect, released by Microsoft, is known for its powerful functions, considerably low price, and complete technology and database support. Therefore, related studies can be done with the applying of Kinect under acceptable cost and data precision. Due to the fact that Kinect utilizes optical mechanism to extracting depth information, limitations are found due to the reason of the straight path of the light. Thus, various angles are required sequentially to obtain the complete 3D information of the object when applying a single Kinect for 3D scanning. The integration process which combines the 3D data from different angles by certain algorithms is also required. This sequential scanning process costs much time and the complex integration process often encounter some technical problems. Therefore, this paper aimed to apply multiple Kinects simultaneously on the field of developing a rapid 3D mannequin scan platform and proposed suggestions on the number and angles of Kinects. In the content, a method of establishing the coordination based on the relation between mannequin and the specifications of Kinect is proposed, and a suggestion of angles and number of Kinects is also described. An experiment of applying multiple Kinect on the scanning of 3D mannequin is constructed by Microsoft API, and the results show that the time required for scanning and technical threshold can be reduced in the industries of fashion and garment design.Keywords: 3D scan, depth sensor, fashion and garment design, mannequin, multiple Kinect sensor
Procedia PDF Downloads 3669820 An Accurate Computer-Aided Diagnosis: CAD System for Diagnosis of Aortic Enlargement by Using Convolutional Neural Networks
Authors: Mahdi Bazarganigilani
Abstract:
Aortic enlargement, also known as an aortic aneurysm, can occur when the walls of the aorta become weak. This disease can become deadly if overlooked and undiagnosed. In this paper, a computer-aided diagnosis (CAD) system was introduced to accurately diagnose aortic enlargement from chest x-ray images. An enhanced convolutional neural network (CNN) was employed and then trained by transfer learning by using three different main areas from the original images. The areas included the left lung, heart, and right lung. The accuracy of the system was then evaluated on 1001 samples by using 4-fold cross-validation. A promising accuracy of 90% was achieved in terms of the F-measure indicator. The results showed using different areas from the original image in the training phase of CNN could increase the accuracy of predictions. This encouraged the author to evaluate this method on a larger dataset and even on different CAD systems for further enhancement of this methodology.Keywords: computer-aided diagnosis systems, aortic enlargement, chest X-ray, image processing, convolutional neural networks
Procedia PDF Downloads 1629819 Screening of Congenital Heart Diseases with Fetal Phonocardiography
Authors: F. Kovács, K. Kádár, G. Hosszú, Á. T. Balogh, T. Zsedrovits, N. Kersner, A. Nagy, Gy. Jeney
Abstract:
The paper presents a novel screening method to indicate congenital heart diseases (CHD), which otherwise could remain undetected because of their low level. Therefore, not belonging to the high-risk population, the pregnancies are not subject to the regular fetal monitoring with ultrasound echocardiography. Based on the fact that CHD is a morphological defect of the heart causing turbulent blood flow, the turbulence appears as a murmur, which can be detected by fetal phonocardiography (fPCG). The proposed method applies measurements on the maternal abdomen and from the recorded sound signal a sophisticated processing determines the fetal heart murmur. The paper describes the problems and the additional advantages of the fPCG method including the possibility of measurements at home and its combination with the prescribed regular cardiotocographic (CTG) monitoring. The proposed screening process implemented on a telemedicine system provides an enhanced safety against hidden cardiac diseases.Keywords: cardiac murmurs, fetal phonocardiography, screening of CHDs, telemedicine system
Procedia PDF Downloads 3329818 Relationship Financing: A Process of Interpretative Phenomenological Analysis
Authors: Y. Fandja, O. Colot, M. Croquet
Abstract:
Small and medium-sized firms (SMEs) face difficulties in accessing bank credit. Bank credit is actually the main source of external financing for SMEs. In general, SMEs are risky businesses because of the potential opacity maintained by the leader in the management of affairs, the agency conflicts between business owners and third-party funders and the potential opportunism of the leader due to the incompleteness of the contracts. These elements accentuate the problems of information asymmetries between SMEs and bankers leading to capital rationing. Moreover, the last economic crisis reinforced this rationing of capital. However, a long-term relationship between SMEs and their bank would enable the latter to accumulate a set of relevant information allowing the reduction of information asymmetry and, consequently, the reduction of credit rationing. The objective of this research is to investigate the lived experience of SMEs loan officers in their relationships with their clients in order to understand how these relationships can affect the financing structure of these SMEs. To carry out this research, an Interpretative Phenomenological Analysis is implemented. This approach is part of the constructivist paradigm and refers to the subjective narratives of the individual rather than to an objective description of the facts. The role of the researcher is to explore the lived experience of the interviewees and to try to understand the meaning they give to this experience. Currently, several sixty-minute semi-structured interviews with loan officers for SMEs have been conducted. The analysis of the content of these interviews brought out three main themes. First, the relationship between the credit officer and the company manager is complex because the credit officer is not aware of establishing a personal relationship with his client. Second; the emotional involvement in the bank financing decision is present and third, the trust in the relationship between the credit officer and his client is very important. The originality of this research is to use the interpretative phenomenological analysis more specific to psychology and sociology in order to approach in a different way the problem of the financing of SMEs through their particular relations with the bankers.Keywords: financing structure, interpretative phenomenological analysis, relationship financing, SME
Procedia PDF Downloads 1599817 Implementation of an Open Source ERP for SMEs in the Automotive Sector in Peru: A Case Study
Authors: Gerson E. Cornejo, Luis A. Gamarra, David S. Mauricio
Abstract:
The Enterprise Resource Planning Systems (ERP) allows the integration of all the business processes of the functional areas of the companies, in order to automate and standardize the processes, obtain accurate information and improve decision making in time real. In Peru, 79% of medium and small companies (SMEs) do not use any management software, this is because it is believed that ERPs are expensive, complex and difficult to implement. However, for more than 20 years there have been Open Source ERPs, which are more accessible and have the same benefit as proprietary ERPs, but there is little information on the implementation process. In this work is made a case of study, in order to show the implementation process of an Open Source ERP, Odoo, based on the ASAP methodology (Accelerated SAP) and applied to a company of corrective and preventive maintenance services of vehicles. The ERP allowed the SME to standardize its business processes, increase its productivity, reducing up to 40% certain processes. The study of this case shows that it is feasible and profitable to implement an Open Source ERP in SMEs in the Automotive Sector of Peru. In addition, it is shown that the ASAP methodology is adequate to carry out Open Source ERPs implementation projects.Keywords: ASAP, automotive sector, ERP implementation, open source
Procedia PDF Downloads 3369816 UWB Channel Estimation Using an Efficient Sub-Nyquist Sampling Scheme
Authors: Yaacoub Tina, Youssef Roua, Radoi Emanuel, Burel Gilles
Abstract:
Recently, low-complexity sub-Nyquist sampling schemes based on the Finite Rate of Innovation (FRI) theory have been introduced to sample parametric signals at minimum rates. The multichannel modulating waveforms (MCMW) is such an efficient scheme, where the received signal is mixed with an appropriate set of arbitrary waveforms, integrated and sampled at rates far below the Nyquist rate. In this paper, the MCMW scheme is adapted to the special case of ultra wideband (UWB) channel estimation, characterized by dense multipaths. First, an appropriate structure, which accounts for the bandpass spectrum feature of UWB signals, is defined. Then, a novel approach to decrease the number of processing channels and reduce the complexity of this sampling scheme is presented. Finally, the proposed concepts are validated by simulation results, obtained with real filters, in the framework of a coherent Rake receiver.Keywords: coherent rake receiver, finite rate of innovation, sub-nyquist sampling, ultra wideband
Procedia PDF Downloads 2569815 Communication Barriers in Disaster Risk Management
Authors: Pooja Pandey
Abstract:
The role of communication plays an integral part in the management of any disaster, whether natural or human-induced, both require effective and strategic delivery of information. The way any information is conveyed carries the most weight while dealing with the disaster. Hence, integrating communication strategies in disaster risk management (DRM) are extensively acknowledged however, these integration and planning are missing from the practical books. Researchers are continuously exploring integrated DRM and have established substantial vents between research and implementation of the strategies (gaps between science and policy). For this reason, this paper reviews the communication barriers that obstruct effective management of the disaster. Communication between first responders (government agencies, police, medical services) and the public (people directly affected by the disaster) is most critical and lacks proper delivery during a disaster. And these challenges can only be resolved if the foundation of the problem is properly dealt with, which is resolving the issues within the organizations. Through this study, it was found that it is necessary to build the communication gap between the organizations themselves as most of the hindrances occur during the mitigation, preparedness, response and recovery phase of the disaster. The study is concluded with the main aim to review the communication barriers within and at the organizational, technological, and social levels that impact effective DRM. In the end, some suggestions are made to strengthen the knowledge for future improvement in communication between the responders and their organizations.Keywords: communication, organization, barriers, first responders, disaster risk management
Procedia PDF Downloads 3009814 One Health Approach: The Importance of Improving the Identification of Waterborne Bacteria in Austrian Water
Authors: Aurora Gitto, Philipp Proksch
Abstract:
The presence of various microorganisms (bacteria, fungi) in surface water and groundwater represents an important issue for human health worldwide. The matrix-assisted laser desorption/ionization-time of flight mass spectrometry (MALDI-TOF-MS) has emerged as a promising and reliable tool for bacteria identification in clinical diagnostic microbiology and environmental strains thanks to an ionization technique that uses a laser energy absorbing matrix to create ions from large molecules with minimal fragmentation. The study aims first to conceptualise and set up library information and create a comprehensive database of MALDI-TOF-MS spectra from environmental water samples. The samples were analysed over a year (2021-2022) using membrane filtration methodology (0.45 μm and 0.22 μm) and then isolated on R2A agar for a period of 5 days and Yeast extract agar growing at 22 °C up to 4 days and 37 °C for 48 hours. The undetected organisms by MALDI-TOF-MS were analysed by PCR and then sequenced. The information obtained by the sequencing was further implemented in the MALDI-TOF-MS library. Among the culturable bacteria, the results show how the incubator temperature affects the growth of some genera instead of others, as demonstrated by Pseudomonas sp., which grows at 22 °C, compared to Bacillus sp., which is abundant at 37 °C. The bacteria community shows a variation in composition also between the media used, as demonstrated with R2A agar which has been defined by a higher presence of organisms not detected compared to YEA. Interesting is the variability of the Genus over one year of sampling and how the seasonality impacts the bacteria community; in fact, in some sampling locations, we observed how the composition changed, moving from winter to spring and summer. In conclusion, the bacteria community in groundwater and river bank filtration represents important information that needs to be added to the library to simplify future water quality analysis but mainly to prevent potential risks to human health.Keywords: water quality, MALDI-TOF-MS, sequencing, library
Procedia PDF Downloads 839813 Syntax and Words as Evolutionary Characters in Comparative Linguistics
Authors: Nancy Retzlaff, Sarah J. Berkemer, Trudie Strauss
Abstract:
In the last couple of decades, the advent of digitalization of any kind of data was probably one of the major advances in all fields of study. This paves the way for also analysing these data even though they might come from disciplines where there was no initial computational necessity to do so. Especially in linguistics, one can find a rather manual tradition. Still when considering studies that involve the history of language families it is hard to overlook the striking similarities to bioinformatics (phylogenetic) approaches. Alignments of words are such a fairly well studied example of an application of bioinformatics methods to historical linguistics. In this paper we will not only consider alignments of strings, i.e., words in this case, but also alignments of syntax trees of selected Indo-European languages. Based on initial, crude alignments, a sophisticated scoring model is trained on both letters and syntactic features. The aim is to gain a better understanding on which features in two languages are related, i.e., most likely to have the same root. Initially, all words in two languages are pre-aligned with a basic scoring model that primarily selects consonants and adjusts them before fitting in the vowels. Mixture models are subsequently used to filter ‘good’ alignments depending on the alignment length and the number of inserted gaps. Using these selected word alignments it is possible to perform tree alignments of the given syntax trees and consequently find sentences that correspond rather well to each other across languages. The syntax alignments are then filtered for meaningful scores—’good’ scores contain evolutionary information and are therefore used to train the sophisticated scoring model. Further iterations of alignments and training steps are performed until the scoring model saturates, i.e., barely changes anymore. A better evaluation of the trained scoring model and its function in containing evolutionary meaningful information will be given. An assessment of sentence alignment compared to possible phrase structure will also be provided. The method described here may have its flaws because of limited prior information. This, however, may offer a good starting point to study languages where only little prior knowledge is available and a detailed, unbiased study is needed.Keywords: alignments, bioinformatics, comparative linguistics, historical linguistics, statistical methods
Procedia PDF Downloads 1549812 Contribution to the Study of Automatic Epileptiform Pattern Recognition in Long Term EEG Signals
Authors: Christine F. Boos, Fernando M. Azevedo
Abstract:
Electroencephalogram (EEG) is a record of the electrical activity of the brain that has many applications, such as monitoring alertness, coma and brain death; locating damaged areas of the brain after head injury, stroke and tumor; monitoring anesthesia depth; researching physiology and sleep disorders; researching epilepsy and localizing the seizure focus. Epilepsy is a chronic condition, or a group of diseases of high prevalence, still poorly explained by science and whose diagnosis is still predominantly clinical. The EEG recording is considered an important test for epilepsy investigation and its visual analysis is very often applied for clinical confirmation of epilepsy diagnosis. Moreover, this EEG analysis can also be used to help define the types of epileptic syndrome, determine epileptiform zone, assist in the planning of drug treatment and provide additional information about the feasibility of surgical intervention. In the context of diagnosis confirmation the analysis is made using long term EEG recordings with at least 24 hours long and acquired by a minimum of 24 electrodes in which the neurophysiologists perform a thorough visual evaluation of EEG screens in search of specific electrographic patterns called epileptiform discharges. Considering that the EEG screens usually display 10 seconds of the recording, the neurophysiologist has to evaluate 360 screens per hour of EEG or a minimum of 8,640 screens per long term EEG recording. Analyzing thousands of EEG screens in search patterns that have a maximum duration of 200 ms is a very time consuming, complex and exhaustive task. Because of this, over the years several studies have proposed automated methodologies that could facilitate the neurophysiologists’ task of identifying epileptiform discharges and a large number of methodologies used neural networks for the pattern classification. One of the differences between all of these methodologies is the type of input stimuli presented to the networks, i.e., how the EEG signal is introduced in the network. Five types of input stimuli have been commonly found in literature: raw EEG signal, morphological descriptors (i.e. parameters related to the signal’s morphology), Fast Fourier Transform (FFT) spectrum, Short-Time Fourier Transform (STFT) spectrograms and Wavelet Transform features. This study evaluates the application of these five types of input stimuli and compares the classification results of neural networks that were implemented using each of these inputs. The performance of using raw signal varied between 43 and 84% efficiency. The results of FFT spectrum and STFT spectrograms were quite similar with average efficiency being 73 and 77%, respectively. The efficiency of Wavelet Transform features varied between 57 and 81% while the descriptors presented efficiency values between 62 and 93%. After simulations we could observe that the best results were achieved when either morphological descriptors or Wavelet features were used as input stimuli.Keywords: Artificial neural network, electroencephalogram signal, pattern recognition, signal processing
Procedia PDF Downloads 5289811 Regulating Information Asymmetries at Online Platforms for Short-Term Vacation Rental in European Union– Legal Conondrum Continues
Authors: Vesna Lukovic
Abstract:
Online platforms as new business models play an important role in today’s economy and the functioning of the EU’s internal market. In the travel industry, algorithms used by online platforms for short-stay accommodation provide suggestions and price information to travelers. Those suggestions and recommendations are displayed in search results via recommendation (ranking) systems. There has been a growing consensus that the current legal framework was not sufficient to resolve problems arising from platform practices. In order to enhance the potential of the EU’s Single Market, smaller businesses should be protected, and their rights strengthened vis-à-vis large online platforms. The Regulation (EU) 2019/1150 of the European Parliament and of the Council on promoting fairness and transparency for business users of online intermediation services aims to level the playing field in that respect. This research looks at Airbnb through the lenses of this regulation. The research explores key determinants and finds that although regulation is an important step in the right direction, it is not enough. It does not entail sufficient clarity obligations that would make online platforms an intermediary service which both accommodation providers and travelers could use with ease.Keywords: algorithm, online platforms, ranking, consumers, EU regulation
Procedia PDF Downloads 1309810 Monitoring Prospective Sites for Water Harvesting Structures Using Remote Sensing and Geographic Information Systems-Based Modeling in Egypt
Authors: Shereif. H. Mahmoud
Abstract:
Egypt has limited water resources, and it will be under water stress by the year 2030. Therefore, Egypt should consider natural and non-conventional water resources to overcome such a problem. Rain harvesting is one solution. This Paper presents a geographic information system (GIS) methodology - based on decision support system (DSS) that uses remote sensing data, filed survey, and GIS to identify potential RWH areas. The input into the DSS includes a map of rainfall surplus, slope, potential runoff coefficient (PRC), land cover/use, soil texture. In addition, the outputs are map showing potential sites for RWH. Identifying suitable RWH sites implemented in the ArcGIS model environment using the model builder of ArcGIS 10.1. Based on Analytical hierarchy process (AHP) analysis taking into account five layers, the spatial extents of RWH suitability areas identified using Multi-Criteria Evaluation (MCE). The suitability model generated a suitability map for RWH with four suitability classes, i.e. Excellent, Moderate, Poor, and unsuitable. The spatial distribution of the suitability map showed that the excellent suitable areas for RWH concentrated in the northern part of Egypt. According to their averages, 3.24% of the total area have excellent and good suitability for RWH, while 45.04 % and 51.48 % of the total area are moderate and unsuitable suitability, respectively. The majority of the areas with excellent suitability have slopes between 2 and 8% and with an intensively cultivated area. The major soil type in the excellent suitable area is loam and the rainfall range from 100 up to 200 mm. Validation of the used technique depends on comparing existing RWH structures locations with the generated suitability map using proximity analysis tool of ArcGIS 10.1. The result shows that most of exiting RWH structures categorized as successful.Keywords: rainwater harvesting (RWH), geographic information system (GIS), analytical hierarchy process (AHP), multi-criteria evaluation (MCE), decision support system (DSS)
Procedia PDF Downloads 3609809 Assessment of Barriers Preventing Recycling Practices among Bars and Eateries in Central South Africa
Authors: Jana Vermaas, Carien Denner
Abstract:
Waste has become a global issue and the management regarding it a priority. Some of the main problems in South Africa (SA) include: (1) the lack of information and education, (2) waste collection services, (3) reusing and recycling is not encouraged, (4) illegal dumping, and the biggest problem of all (5) the lack of waste related regulations and enforcement by the government and municipalities. In SA, there are provinces such as Gauteng and the Western Cape that have some recycling programs in place, but nothing yet in the central part of the country. By identifying the barriers preventing these businesses from recycling, the local municipalities and recycling services could create a solution. Owners or employees of eateries and bars completed a self-administered questionnaire. Information were obtained on knowledge of recycling, participation in recycling and to which extent, barriers that prevent them from recycling and motives that would encourage recycling. The data obtained from the questionnaire indicated that most (98%) participants knew only the basics, that recycling is a process of converting waste materials into new materials and objects. Further knowledge questions indicated that individuals were not educated about recycling as almost half (49%) of the participants believe that they can’t reuse plastic bottles. They do not understand which items of their waste could be re-used or recycled. They had limited knowledge about the recycling opportunities or practices in the area. Only a small number (34%) were involved in recycling or sustainable practices. Many did not even know of any collection services or buy-back centres in their vicinity. Most of the participants (94%) indicated that they would be willing to recycle if it would have a financial benefit. Many also stated that they would be more willing to recycle if the recyclable waste will be collected from their establishment, on a regular basis. The enforcement of recycling by municipalities or government by awarding fines for waste offenders was indicated as a significant motive. It could be concluded that the most significant barrier is knowledge and lack of information. These businesses do not comprehend the impact that they can have with their recycling contributions, not only on the environment, but also on the consumers that they serve. Another barrier is the lack of collection services. There are currently no government or municipal services for the collection of recyclable waste. All waste are taken to landfills. Many of the larger recycling initiatives and companies do not reach as far as central SA. Therefore, the buy-back component of recycling is not present.Keywords: eateries, recycling, sustainable practices, waste
Procedia PDF Downloads 2959808 The Cases Studies of Eyewitness Misidentifications during Criminal Investigation in Taiwan
Authors: Chih Hung Shih
Abstract:
Eyewitness identification is one of the efficient information to identify suspects during criminal investigation. However eyewitness identification is improved frequently, inaccurate and plays vital roles in wrongful convictions. Most eyewitness misidentifications are made during police criminal investigation stage and then accepted by juries. Four failure investigation case studies in Taiwan are conduct to demonstrate how misidentifications are caused during the police investigation context. The result shows that there are several common grounds among these cases: (1) investigators lacked for knowledge about eyewitness memory so that they couldn’t evaluate the validity of the eyewitnesses’ accounts and identifications, (2) eyewitnesses were always asked to filter out several suspects during the investigation, and received investigation information which contaminated the eyewitnesses’ memory, (3) one to one live individual identifications were made in most of cases, (4) eyewitness identifications were always used to support the hypotheses of investigators, and exaggerated theirs powers when conform with the investigation lines, (5) the eyewitnesses’ confidence didn’t t reflect the validity of their identifications , but always influence the investigators’ beliefs for the identifications, (6) the investigators overestimated the power of the eyewitness identifications and ignore the inconsistency with other evidence. Recommendations have been proposed for future academic research and police practice of eyewitness identification in Taiwan.Keywords: criminal investigation, eyewitness identification, investigative bias, investigative failures
Procedia PDF Downloads 2449807 Developing Indicators in System Mapping Process Through Science-Based Visual Tools
Authors: Cristian Matti, Valerie Fowles, Eva Enyedi, Piotr Pogorzelski
Abstract:
The system mapping process can be defined as a knowledge service where a team of facilitators, experts and practitioners facilitate a guided conversation, enable the exchange of information and support an iterative curation process. System mapping processes rely on science-based tools to introduce and simplify a variety of components and concepts of socio-technical systems through metaphors while facilitating an interactive dialogue process to enable the design of co-created maps. System maps work then as “artifacts” to provide information and focus the conversation into specific areas around the defined challenge and related decision-making process. Knowledge management facilitates the curation of that data gathered during the system mapping sessions through practices of documentation and subsequent knowledge co-production for which common practices from data science are applied to identify new patterns, hidden insights, recurrent loops and unexpected elements. This study presents empirical evidence on the application of these techniques to explore mechanisms by which visual tools provide guiding principles to portray system components, key variables and types of data through the lens of climate change. In addition, data science facilitates the structuring of elements that allow the analysis of layers of information through affinity and clustering analysis and, therefore, develop simple indicators for supporting the decision-making process. This paper addresses methodological and empirical elements on the horizontal learning process that integrate system mapping through visual tools, interpretation, cognitive transformation and analysis. The process is designed to introduce practitioners to simple iterative and inclusive processes that create actionable knowledge and enable a shared understanding of the system in which they are embedded.Keywords: indicators, knowledge management, system mapping, visual tools
Procedia PDF Downloads 1959806 Automatic Detection and Classification of Diabetic Retinopathy Using Retinal Fundus Images
Authors: A. Biran, P. Sobhe Bidari, A. Almazroe, V. Lakshminarayanan, K. Raahemifar
Abstract:
Diabetic Retinopathy (DR) is a severe retinal disease which is caused by diabetes mellitus. It leads to blindness when it progress to proliferative level. Early indications of DR are the appearance of microaneurysms, hemorrhages and hard exudates. In this paper, an automatic algorithm for detection of DR has been proposed. The algorithm is based on combination of several image processing techniques including Circular Hough Transform (CHT), Contrast Limited Adaptive Histogram Equalization (CLAHE), Gabor filter and thresholding. Also, Support Vector Machine (SVM) Classifier is used to classify retinal images to normal or abnormal cases including non-proliferative or proliferative DR. The proposed method has been tested on images selected from Structured Analysis of the Retinal (STARE) database using MATLAB code. The method is perfectly able to detect DR. The sensitivity specificity and accuracy of this approach are 90%, 87.5%, and 91.4% respectively.Keywords: diabetic retinopathy, fundus images, STARE, Gabor filter, support vector machine
Procedia PDF Downloads 2949805 Adaption Model for Building Agile Pronunciation Dictionaries Using Phonemic Distance Measurements
Authors: Akella Amarendra Babu, Rama Devi Yellasiri, Natukula Sainath
Abstract:
Where human beings can easily learn and adopt pronunciation variations, machines need training before put into use. Also humans keep minimum vocabulary and their pronunciation variations are stored in front-end of their memory for ready reference, while machines keep the entire pronunciation dictionary for ready reference. Supervised methods are used for preparation of pronunciation dictionaries which take large amounts of manual effort, cost, time and are not suitable for real time use. This paper presents an unsupervised adaptation model for building agile and dynamic pronunciation dictionaries online. These methods mimic human approach in learning the new pronunciations in real time. A new algorithm for measuring sound distances called Dynamic Phone Warping is presented and tested. Performance of the system is measured using an adaptation model and the precision metrics is found to be better than 86 percent.Keywords: pronunciation variations, dynamic programming, machine learning, natural language processing
Procedia PDF Downloads 1769804 Nonparametric Copula Approximations
Authors: Serge Provost, Yishan Zang
Abstract:
Copulas are currently utilized in finance, reliability theory, machine learning, signal processing, geodesy, hydrology and biostatistics, among several other fields of scientific investigation. It follows from Sklar's theorem that the joint distribution function of a multidimensional random vector can be expressed in terms of its associated copula and marginals. Since marginal distributions can easily be determined by making use of a variety of techniques, we address the problem of securing the distribution of the copula. This will be done by using several approaches. For example, we will obtain bivariate least-squares approximations of the empirical copulas, modify the kernel density estimation technique and propose a criterion for selecting appropriate bandwidths, differentiate linearized empirical copulas, secure Bernstein polynomial approximations of suitable degrees, and apply a corollary to Sklar's result. Illustrative examples involving actual observations will be presented. The proposed methodologies will as well be applied to a sample generated from a known copula distribution in order to validate their effectiveness.Keywords: copulas, Bernstein polynomial approximation, least-squares polynomial approximation, kernel density estimation, density approximation
Procedia PDF Downloads 749803 Polarimetric Synthetic Aperture Radar Data Classification Using Support Vector Machine and Mahalanobis Distance
Authors: Najoua El Hajjaji El Idrissi, Necip Gokhan Kasapoglu
Abstract:
Polarimetric Synthetic Aperture Radar-based imaging is a powerful technique used for earth observation and classification of surfaces. Forest evolution has been one of the vital areas of attention for the remote sensing experts. The information about forest areas can be achieved by remote sensing, whether by using active radars or optical instruments. However, due to several weather constraints, such as cloud cover, limited information can be recovered using optical data and for that reason, Polarimetric Synthetic Aperture Radar (PolSAR) is used as a powerful tool for forestry inventory. In this [14paper, we applied support vector machine (SVM) and Mahalanobis distance to the fully polarimetric AIRSAR P, L, C-bands data from the Nezer forest areas, the classification is based in the separation of different tree ages. The classification results were evaluated and the results show that the SVM performs better than the Mahalanobis distance and SVM achieves approximately 75% accuracy. This result proves that SVM classification can be used as a useful method to evaluate fully polarimetric SAR data with sufficient value of accuracy.Keywords: classification, synthetic aperture radar, SAR polarimetry, support vector machine, mahalanobis distance
Procedia PDF Downloads 1339802 Exploring Syntactic and Semantic Features for Text-Based Authorship Attribution
Authors: Haiyan Wu, Ying Liu, Shaoyun Shi
Abstract:
Authorship attribution is to extract features to identify authors of anonymous documents. Many previous works on authorship attribution focus on statistical style features (e.g., sentence/word length), content features (e.g., frequent words, n-grams). Modeling these features by regression or some transparent machine learning methods gives a portrait of the authors' writing style. But these methods do not capture the syntactic (e.g., dependency relationship) or semantic (e.g., topics) information. In recent years, some researchers model syntactic trees or latent semantic information by neural networks. However, few works take them together. Besides, predictions by neural networks are difficult to explain, which is vital in authorship attribution tasks. In this paper, we not only utilize the statistical style and content features but also take advantage of both syntactic and semantic features. Different from an end-to-end neural model, feature selection and prediction are two steps in our method. An attentive n-gram network is utilized to select useful features, and logistic regression is applied to give prediction and understandable representation of writing style. Experiments show that our extracted features can improve the state-of-the-art methods on three benchmark datasets.Keywords: authorship attribution, attention mechanism, syntactic feature, feature extraction
Procedia PDF Downloads 1369801 Radio Frequency Identification Chips in Colour Preference Tracking
Authors: A. Ballard
Abstract:
The ability to track goods and products en route in the delivery system, in the warehouse, and on the top floor is a huge advantage to shippers and retailers. Recently the emergence of radio frequency identification (RFID) technology has enabled this better than ever before. However, a significant problem exists in that RFID technology depends on the quality of the information stored for each tagged product. Because of the profusion of names for colours, it is very difficult to ascertain that stored values are recognised by all users who view the product visually. This paper reports the findings of a study in which 50 consumers and 50 logistics workers were shown colour swatches and asked to choose the name of the colour from a multiple choice list. They were then asked to match consumer products, including toasters, jumpers, and toothbrushes, with the identifying inventory information available for each one. The findings show that the ability to match colours was significantly stronger with the color swatches than with the consumer products and that while logistics professionals made more frequent correct identification than the consumers, their results were still unsatisfactorily low. Based on these findings, a proposed universal model of colour identification numbers has been developed.Keywords: consumer preferences, supply chain logistics, radio frequency identification, RFID, colour preference
Procedia PDF Downloads 1219800 A Robust Digital Image Watermarking Against Geometrical Attack Based on Hybrid Scheme
Authors: M. Samadzadeh Mahabadi, J. Shanbehzadeh
Abstract:
This paper presents a hybrid digital image-watermarking scheme, which is robust against varieties of attacks and geometric distortions. The image content is represented by important feature points obtained by an image-texture-based adaptive Harris corner detector. These feature points are extracted from LL2 of 2-D discrete wavelet transform which are obtained by using the Harris-Laplacian detector. We calculate the Fourier transform of circular regions around these points. The amplitude of this transform is rotation invariant. The experimental results demonstrate the robustness of the proposed method against the geometric distortions and various common image processing operations such as JPEG compression, colour reduction, Gaussian filtering, median filtering, and rotation.Keywords: digital watermarking, geometric distortions, geometrical attack, Harris Laplace, important feature points, rotation, scale invariant feature
Procedia PDF Downloads 5019799 Accumulation of Heavy Metals in Safflower (Carthamus tinctorius L.)
Authors: Violina R. Angelova, Mariana N. Perifanova-Nemska, Galina P. Uzunova, Elitsa N. Kolentsova
Abstract:
Comparative research has been conducted to allow us to determine the accumulation of heavy metals (Pb, Zn and Cd) in the vegetative and reproductive organs of safflower, and to identify the possibility of its growth on soils contaminated by heavy metals and efficacy for phytoremediation. The experiment was performed on an agricultural field contaminated by the Non-Ferrous-Metal Works (MFMW) near Plovdiv, Bulgaria. The experimental plots were situated at different distances (0.1, 0.5, 2.0, and 15 km) from the source of pollution. The contents of heavy metals in plant materials (roots, stems, leaves, seeds) were determined. The quality of safflower oils (heavy metals and fatty acid composition) was also determined. The quantitative measurements were carried out with inductively-coupled plasma (ICP). Safflower is a plant that is tolerant to heavy metals and can be referred to the hyperaccumulators of lead and cadmium and the accumulators of zinc. The plant can be successfully used in the phytoremediation of heavy metal contaminated soils. The processing of safflower seeds into oil and the use of the obtained oil will greatly reduce the cost of phytoremediation.Keywords: heavy metals, accumulation, safflower, polluted soils, phytoremediation
Procedia PDF Downloads 2639798 Screen Method of Distributed Cooperative Navigation Factors for Unmanned Aerial Vehicle Swarm
Authors: Can Zhang, Qun Li, Yonglin Lei, Zhi Zhu, Dong Guo
Abstract:
Aiming at the problem of factor screen in distributed collaborative navigation of dense UAV swarm, an efficient distributed collaborative navigation factor screen method is proposed. The method considered the balance between computing load and positioning accuracy. The proposed algorithm utilized the factor graph model to implement a distributed collaborative navigation algorithm. The GNSS information of the UAV itself and the ranging information between the UAVs are used as the positioning factors. In this distributed scheme, a local factor graph is established for each UAV. The positioning factors of nodes with good geometric position distribution and small variance are selected to participate in the navigation calculation. To demonstrate and verify the proposed methods, the simulation and experiments in different scenarios are performed in this research. Simulation results show that the proposed scheme achieves a good balance between the computing load and positioning accuracy in the distributed cooperative navigation calculation of UAV swarm. This proposed algorithm has important theoretical and practical value for both industry and academic areas.Keywords: screen method, cooperative positioning system, UAV swarm, factor graph, cooperative navigation
Procedia PDF Downloads 809797 A Study of the Views of Information Technologies Teachers regarding In-Service Training
Authors: Halit Arslan, Ismail Sahin, Ahmet Oguz Akturk, Ismail Celik
Abstract:
Today, the means of following the developments in the area of science and technology is to keep up with the pace of the advancements in this area. As is in every profession, apart from their personal efforts, the training of teachers in the period after they start their careers is only possible through in-service training. The aim of the present study is to determine the views of Information Technologies (IT) teachers regarding the in-service training courses organized by the Ministry of National Education. In this study, in which quantitative research methods and techniques were employed, the views of 196 IT teachers were collected by using the “Views on In-service Training” questionnaire developed by the authors of the paper. Independent groups t-test was used to determine whether the views of IT teachers regarding in-service training differed depending on gender, age and professional seniority. One-way analysis of variance (ANOVA) was used to investigate whether the views of IT teachers regarding in-service training differed depending on the number of in-service training courses they joined and the type of in-service training course they wanted to take. According to the findings obtained in the study, the views of IT teachers on in-service training did not show a significant difference depending on gender and age, whereas those views differed depending on professional seniority, the number of in-service training courses they joined and the type of in-service training course they wanted to take.Keywords: in-service training, IT teachers, professional development, personal development
Procedia PDF Downloads 4179796 Scheduling Building Projects: The Chronographical Modeling Concept
Authors: Adel Francis
Abstract:
Most of scheduling methods and software apply the critical path logic. This logic schedule activities, apply constraints between these activities and try to optimize and level the allocated resources. The extensive use of this logic produces a complex an erroneous network hard to present, follow and update. Planning and management building projects should tackle the coordination of works and the management of limited spaces, traffic, and supplies. Activities cannot be performed without the resources available and resources cannot be used beyond the capacity of workplaces. Otherwise, workspace congestion will negatively affect the flow of works. The objective of the space planning is to link the spatial and temporal aspects, promote efficient use of the site, define optimal site occupancy rates, and ensures suitable rotation of the workforce in the different spaces. The Chronographic scheduling modelling belongs to this category and models construction operations as well as their processes, logical constraints, association and organizational models, which help to better illustrate the schedule information using multiple flexible approaches. The model defined three categories of areas (punctual, surface and linear) and four different layers (space creation, systems, closing off space, finishing, and reduction of space). The Chronographical modelling is a more complete communication method, having the ability to alternate from one visual approach to another by manipulation of graphics via a set of parameters and their associated values. Each individual approach can help to schedule a certain project type or specialty. Visual communication can also be improved through layering, sheeting, juxtaposition, alterations, and permutations, allowing for groupings, hierarchies, and classification of project information. In this way, graphic representation becomes a living, transformable image, showing valuable information in a clear and comprehensible manner, simplifying the site management while simultaneously utilizing the visual space as efficiently as possible.Keywords: building projects, chronographic modelling, CPM, critical path, precedence diagram, scheduling
Procedia PDF Downloads 1559795 The Polarization on Twitter and COVID-19 Vaccination in Brazil
Authors: Giselda Cristina Ferreira, Carlos Alberto Kamienski, Ana Lígia Scott
Abstract:
The COVID-19 pandemic has enhanced the anti-vaccination movement in Brazil, supported by unscientific theories and false news and the possibility of wide communication through social networks such as Twitter, Facebook, and YouTube. The World Health Organization (WHO) classified the large volume of information on the subject against COVID-19 as an Infodemic. In this paper, we present a protocol to identify polarizing users (called polarizers) and study the profiles of Brazilian polarizers on Twitter (renamed to X some weeks ago). We analyzed polarizing interactions on Twitter (in Portuguese) to identify the main polarizers and how the conflicts they caused influenced the COVID-19 vaccination rate throughout the pandemic. This protocol uses data from this social network, graph theory, Java, and R-studio scripts to model and analyze the data. The information about the vaccination rate was obtained in a public database for the government called OpenDataSus. The results present the profiles of Twitter’s Polarizer (political position, gender, professional activity, immunization opinions). We observed that social and political events influenced the participation of these different profiles in conflicts and the vaccination rate.Keywords: Twitter, polarization, vaccine, Brazil
Procedia PDF Downloads 759794 Fact-checking and Political Polarization in an Emerging Democracy
Authors: Eric Agyekum, Dominic Asitanga
Abstract:
Ghana is widely considered asa beacon of democracy in sub-Saharan Africa. With a relatively free media, the country was ranked30thin the world and third in Africaon the 2021 Press Freedom Index. Despite the democratic gains, itis one of the most politically polarized nations in the world. Ghana’spolitical division is evident in the current hunglegislature, where each of the two dominant political parties has 137 members, with an independent member occupying the remaining one seat. Misinformation and fake newsthrive in systems with acuteideological and political differences(Imelda et al, 2021; Azzimonti&Fernandes, 2018; Spohr, 2017) and Ghana is no exception. The information disorder problem has been exacerbatedby the COVID-19 pandemic, with its attendant conspiracy theories and speculations, making it difficult for the media and fact-checking organizations to verifyall claims and flag false information. In Ghana, fact-checking agencies like Ghana Fact, Dubawa Ghana, and some mainstream news media organizations have been fact-checking political claims, COVID-19 conspiracy theories, and many others. However, it is not clear if the audience consumeand attach prominence to these fact-checked stories or even visit the websites of the fact-checking agencies to read the content. Nekmat (2020) opine that though the literature on fact-checking suggest that fact-checked stories can alter readers’ beliefs, very few studies have investigated the patronage and the potential of fact-checks to deter users from sharing false news with others, particularly on social media. In response to Nekmat, this study has been initiated to examine the perception and attitude of the audience in Ghana towards fact-checks. Anchored on the principles of the nudge theory, this study will investigate how fact-checked stories alters readers’ behavioural patterns. A survey will be conducted to collect data from sampled members of the Ghanaian society.Keywords: fact-checking, information disorder, nudge theory, political polarization
Procedia PDF Downloads 142