Search results for: vector information
10182 Development of a Method to Prepare In-School Tactile Guide Maps for Visually Impaired School Children
Authors: K. Doi, T. Nishimura, M. Kawano, H. Fujimoto, Y. Tanaka, M. Sawada, S. Oouchi, T. Kaneko, K. Kanamori
Abstract:
As part of reasonable accommodation for people with disabilities in Japan, which has ratified the Convention on the Rights of Persons with Disabilities, tactile guide maps are necessary. Such maps can enable visually impaired children to attend schools of special needs education (visual impairments) to grasp the arrangement of classrooms on their school campuses. However, it takes many years to be able to use a tactile guide map without difficulty. Thus, information support, in which audio information is added in addition to tactile information, is required. In the present research, a method to prepare an in-school tactile guide map with an additional audio reading function was developed. This map can enable visually impaired school children attending schools of special needs education (visual impairments) to grasp the arrangement of classrooms on their school campuses.Keywords: accessible design, visually impaired, braille, tactile map, in-school tactile guide map
Procedia PDF Downloads 36410181 3D Receiver Operator Characteristic Histogram
Authors: Xiaoli Zhang, Xiongfei Li, Yuncong Feng
Abstract:
ROC curves, as a widely used evaluating tool in machine learning field, are the tradeoff of true positive rate and negative rate. However, they are blamed for ignoring some vital information in the evaluation process, such as the amount of information about the target that each instance carries, predicted score given by each classification model to each instance. Hence, in this paper, a new classification performance method is proposed by extending the Receiver Operator Characteristic (ROC) curves to 3D space, which is denoted as 3D ROC Histogram. In the histogram, theKeywords: classification, performance evaluation, receiver operating characteristic histogram, hardness prediction
Procedia PDF Downloads 31510180 The Factors of Supply Chain Collaboration
Authors: Ghada Soltane
Abstract:
The objective of this study was to identify factors impacting supply chain collaboration. a quantitative study was carried out on a sample of 84 Tunisian industrial companies. To verify the research hypotheses and test the direct effect of these factors on supply chain collaboration a multiple regression method was used using SPSS 26 software. The results show that there are four factors direct effects that affect supply chain collaboration in a meaningful and positive way, including: trust, engagement, information sharing and information qualityKeywords: supply chain collaboration, factors of collaboration, principal component analysis, multiple regression
Procedia PDF Downloads 5110179 In-situ Acoustic Emission Analysis of a Polymer Electrolyte Membrane Water Electrolyser
Authors: M. Maier, I. Dedigama, J. Majasan, Y. Wu, Q. Meyer, L. Castanheira, G. Hinds, P. R. Shearing, D. J. L. Brett
Abstract:
Increasing the efficiency of electrolyser technology is commonly seen as one of the main challenges on the way to the Hydrogen Economy. There is a significant lack of understanding of the different states of operation of polymer electrolyte membrane water electrolysers (PEMWE) and how these influence the overall efficiency. This in particular means the two-phase flow through the membrane, gas diffusion layers (GDL) and flow channels. In order to increase the efficiency of PEMWE and facilitate their spread as commercial hydrogen production technology, new analytic approaches have to be found. Acoustic emission (AE) offers the possibility to analyse the processes within a PEMWE in a non-destructive, fast and cheap in-situ way. This work describes the generation and analysis of AE data coming from a PEM water electrolyser, for, to the best of our knowledge, the first time in literature. Different experiments are carried out. Each experiment is designed so that only specific physical processes occur and AE solely related to one process can be measured. Therefore, a range of experimental conditions is used to induce different flow regimes within flow channels and GDL. The resulting AE data is first separated into different events, which are defined by exceeding the noise threshold. Each acoustic event consists of a number of consequent peaks and ends when the wave diminishes under the noise threshold. For all these acoustic events the following key attributes are extracted: maximum peak amplitude, duration, number of peaks, peaks before the maximum, average intensity of a peak and time till the maximum is reached. Each event is then expressed as a vector containing the normalized values for all criteria. Principal Component Analysis is performed on the resulting data, which orders the criteria by the eigenvalues of their covariance matrix. This can be used as an easy way of determining which criteria convey the most information on the acoustic data. In the following, the data is ordered in the two- or three-dimensional space formed by the most relevant criteria axes. By finding spaces in the two- or three-dimensional space only occupied by acoustic events originating from one of the three experiments it is possible to relate physical processes to certain acoustic patterns. Due to the complex nature of the AE data modern machine learning techniques are needed to recognize these patterns in-situ. Using the AE data produced before allows to train a self-learning algorithm and develop an analytical tool to diagnose different operational states in a PEMWE. Combining this technique with the measurement of polarization curves and electrochemical impedance spectroscopy allows for in-situ optimization and recognition of suboptimal states of operation.Keywords: acoustic emission, gas diffusion layers, in-situ diagnosis, PEM water electrolyser
Procedia PDF Downloads 15710178 Spatial Patterns and Temporal Evolution of Octopus Abundance in the Mauritanian Zone
Authors: Dedah Ahmed Babou, Nicolas Bez
Abstract:
The Min-Max autocorrelation factor (MAF) approach makes it possible to express in a space formed by spatially independent factors, spatiotemporal observations. These factors are ordered in decreasing order of spatial autocorrelation. The starting observations are thus expressed in the space formed by these factors according to temporal coordinates. Each vector of temporal coefficients expresses the temporal evolution of the weight of the corresponding factor. Applying this approach has enabled us to achieve the following results: (i) Define a spatially orthogonal space in which the projections of the raw data are determined; (ii) Define a limit threshold for the factors with the strongest structures in order to analyze the weight, and the temporal evolution of these different structures (iii) Study the correlation between the temporal evolution of the persistent spatial structures and that of the observed average abundance (iv) Propose prototypes of campaigns reflecting a high vs. low abundance (v) Propose a classification of campaigns that highlights seasonal and/or temporal similarities. These results were obtained by analyzing the octopus yield during the scientific campaigns of the oceanographic vessel Al Awam during the period 1989-2017 in the Mauritanian exclusive economic zone.Keywords: spatiotemporal , autocorrelation, kriging, variogram, Octopus vulgaris
Procedia PDF Downloads 14810177 The Impact of Information Technology Monitoring on Employee Theft and Productivity
Authors: Ajayi Oluwasola Felix
Abstract:
This paper examines how firm investments in technology-based employee monitoring impact both misconduct and productivity. We use unique and detailed theft and sales data from 392 restaurant locations from five firms that adopt a theft monitoring information technology (IT) product. We use difference-in-differences (DD) models with staggered adoption dates to estimate the treatment effect of IT monitoring on theft and productivity. We find significant treatment effects in reduced theft and improved productivity that appear to be primarily driven by changed worker behavior rather than worker turnover. We examine four mechanisms that may drive this productivity result: economic and cognitive multitasking, fairness-based motivation, and perceived increases of general oversight. The observed productivity results represent substantial financial benefits to both firms and the legitimate tip-based earnings of workers. Our results suggest that employee misconduct is not solely a function of individual differences in ethics or morality, but can also be influenced by managerial policies that can benefit both firms and employees.Keywords: information technology, monitoring, misconduct, employee theft
Procedia PDF Downloads 42210176 Environmental, Social and Corporate Governance Reporting With Regard to Best Practices of Companies Listed on the Warsaw Stock Exchange - Selected Problems
Authors: Katarzyna Olejko
Abstract:
The need to redefine the goals and adapt the operational activities carried out in accordance with the concept of sustainable management to these goals results in the increasing importance of information on the company's activities perceived from the perspective of the effectiveness and efficiency of environmental goals implementation. The narrow scope of reporting data on a company's impact on the environment is not adequate to meet the information needs of modern investors. Reporting obligations are therefore imposed on companies in order to increase the effectiveness of corporate governance and to improve the process of assessing the achievement of environmental goals. The non-financial reporting obligations introduced in Polish legislation increased the scope of reported information. However, the lack of detailed guidelines on the method of reporting resulted in a large diversification of the scope of non-financial information, making it impossible to compare the data presented by companies. The source of information regarding the level of the implementation of standards in Environmental, social and corporate governance (ESG) is the report on compliance with best practices published by the Warsaw Stock Exchange. The document Best Practices of Warsaw Stock Exchange (WSE) Listed Companies (2021), amended by the WSE in 2021, includes the rules applicable to this area (ESG). The aim of this article is to present the level of compliance with good practices in the area of ESG by selected companies listed on the Warsaw Stock Exchange The research carried out as part of this study, which was based on information from reports on the compliance with good practices of companies listed on the Warsaw Stock Exchange that was made available in the good practice scanner, have revealed that good practices in the ESG area are implemented by companies to a limited extent. The level of their application in comparison with other rules is definitely lower. The lack of experience and clear guidelines on ESG reporting may cause some confusion, which is why conscious investors and reporting companies themselves are pinning their hopes on the Corporate Sustainability Reporting Directive (CSRD) adopted by European Parliament.Keywords: reporting, ESG, corporate governance, best practices
Procedia PDF Downloads 7510175 Mechanical Prosthesis Controlled by Brain-Computer Interface
Authors: Tianyu Cao, KIRA (Ruizhi Zhao)
Abstract:
The purpose of our research is to study the possibility of people with physical disabilities manipulating mechanical prostheses through brain-computer interface (BCI) technology. The brain-machine interface (BCI) of the neural prosthesis records signals from neurons and uses mathematical modeling to decode them, converting desired movements into body movements. In order to improve the patient's neural control, the prosthesis is given a natural feeling. It records data from sensitive areas from the body to the prosthetic limb and encodes signals in the form of electrical stimulation to the brain. In our research, the brain-computer interface (BCI) is a bridge connecting patients’ cognition and the real world, allowing information to interact with each other. The efficient work between the two is achieved through external devices. The flow of information is controlled by BCI’s ability to record neuronal signals and decode signals, which are converted into device control. In this way, we could encode information and then send it to the brain through electrical stimulation, which has significant medical application.Keywords: biomedical engineering, brain-computer interface, prosthesis, neural control
Procedia PDF Downloads 18110174 FlexPoints: Efficient Algorithm for Detection of Electrocardiogram Characteristic Points
Authors: Daniel Bulanda, Janusz A. Starzyk, Adrian Horzyk
Abstract:
The electrocardiogram (ECG) is one of the most commonly used medical tests, essential for correct diagnosis and treatment of the patient. While ECG devices generate a huge amount of data, only a small part of them carries valuable medical information. To deal with this problem, many compression algorithms and filters have been developed over the past years. However, the rapid development of new machine learning techniques poses new challenges. To address this class of problems, we created the FlexPoints algorithm that searches for characteristic points on the ECG signal and ignores all other points that do not carry relevant medical information. The conducted experiments proved that the presented algorithm can significantly reduce the number of data points which represents ECG signal without losing valuable medical information. These sparse but essential characteristic points (flex points) can be a perfect input for some modern machine learning models, which works much better using flex points as an input instead of raw data or data compressed by many popular algorithms.Keywords: characteristic points, electrocardiogram, ECG, machine learning, signal compression
Procedia PDF Downloads 16410173 Corporate Governance, Performance, and Financial Reporting Quality of Listed Manufacturing Firms in Nigeria
Authors: Jamila Garba Audu, Shehu Usman Hassan
Abstract:
The widespread failure in the financial information quality has created the need to improve the financial information quality and to strengthen the control of managers by setting up good firms structures. Published accounting information in financial statements is required to provide various users - shareholders, employees, suppliers, creditors, financial analysts, stockbrokers and government agencies – with timely and reliable information useful for making prudent, effective and efficient decisions. The relationship between corporate governance and performance to financial reporting quality is imperative; this is because despite rapid researches in this area the findings obtained from these studies are constantly inconclusive. Data for the study were extracted from the firms’ annual reports and accounts. After running the OLS regression, a robustness test was conducted for the validity of statistical inferences; the data was empirically tested. A multiple regression was employed to test the model as a technique for data analysis. The results from the analysis revealed a negative association between all the regressors and financial reporting quality except the performance of listed manufacturing firms in Nigeria. This indicates that corporate governance plays a significant role in mitigating earnings management and improving financial reporting quality while performance does not. The study recommended among others that the composition of audit committee should be made in accordance with the provision for code of corporate governance which is not more than six (6) members with at least one (1) financial expert.Keywords: corporate governance, financial reporting quality, manufacturing firms, Nigeria, performance
Procedia PDF Downloads 24810172 An Automated System for the Detection of Citrus Greening Disease Based on Visual Descriptors
Authors: Sidra Naeem, Ayesha Naeem, Sahar Rahim, Nadia Nawaz Qadri
Abstract:
Citrus greening is a bacterial disease that causes considerable damage to citrus fruits worldwide. Efficient method for this disease detection must be carried out to minimize the production loss. This paper presents a pattern recognition system that comprises three stages for the detection of citrus greening from Orange leaves: segmentation, feature extraction and classification. Image segmentation is accomplished by adaptive thresholding. The feature extraction stage comprises of three visual descriptors i.e. shape, color and texture. From shape feature we have used asymmetry index, from color feature we have used histogram of Cb component from YCbCr domain and from texture feature we have used local binary pattern. Classification was done using support vector machines and k nearest neighbors. The best performances of the system is Accuracy = 88.02% and AUROC = 90.1% was achieved by automatic segmented images. Our experiments validate that: (1). Segmentation is an imperative preprocessing step for computer assisted diagnosis of citrus greening, and (2). The combination of shape, color and texture features form a complementary set towards the identification of citrus greening disease.Keywords: citrus greening, pattern recognition, feature extraction, classification
Procedia PDF Downloads 18510171 Introduction of Robust Multivariate Process Capability Indices
Authors: Behrooz Khalilloo, Hamid Shahriari, Emad Roghanian
Abstract:
Process capability indices (PCIs) are important concepts of statistical quality control and measure the capability of processes and how much processes are meeting certain specifications. An important issue in statistical quality control is parameter estimation. Under the assumption of multivariate normality, the distribution parameters, mean vector and variance-covariance matrix must be estimated, when they are unknown. Classic estimation methods like method of moment estimation (MME) or maximum likelihood estimation (MLE) makes good estimation of the population parameters when data are not contaminated. But when outliers exist in the data, MME and MLE make weak estimators of the population parameters. So we need some estimators which have good estimation in the presence of outliers. In this work robust M-estimators for estimating these parameters are used and based on robust parameter estimators, robust process capability indices are introduced. The performances of these robust estimators in the presence of outliers and their effects on process capability indices are evaluated by real and simulated multivariate data. The results indicate that the proposed robust capability indices perform much better than the existing process capability indices.Keywords: multivariate process capability indices, robust M-estimator, outlier, multivariate quality control, statistical quality control
Procedia PDF Downloads 28410170 The Pro-Active Public Relations of Faculty of Management Science, Suan Sunandha Rajabhat University
Authors: Kanyakorn Sujarittnetikarn, Surangkana Pipatchokchaiyo
Abstract:
The objective of this research was to study the pro-active public relations of according to the characteristic of Faculty of Management Science, Suan Sunandha Rajabhat University. The sample group for this research report was students from 4 year curriculum and continued / extended curriculum, made a random distribution proportion as follows: a group of 400 students who are working while studying and a group of non – working students. The tools used in this research were questionnaires, asking about the acknowledgement of public relations information of Faculty of Management Science in the academic year 2007. The result found that friends were the most influential in choosing the education institute. The differences of method to receive information of non-working student and working student were the entertainment magazine which was interested mostly by working students and they preferred to search the information on the website after 24:00 O’clock. However, the non-working students preferred 21:00-24:00 O’clock the most.Keywords: development guidelines systems, faculty of management science, public relation planning, proactive public relations
Procedia PDF Downloads 29010169 Effect of Fiscal Policy on Growth in India
Authors: Parma Chakravartti
Abstract:
The impact of government spending and taxation on economic growth has remained a central issue of fiscal policy analysis. There is a wide range of opinions over the strength of fiscal policy’s effect on macroeconomic variables. It can be argued that the impact of fiscal policy depends on the structure and economic condition of the economy. This study makes an attempt to examine the effect of fiscal policy shocks on growth in India using the structural vector autoregressive model (SVAR), considering data from 1950 to 2019. The study finds that government spending is an important instrument of growth in India, where the share of revenue expenditure to capital expenditure plays a key role. The optimum composition of total expenditure is important for growth and it is not necessarily true that capital expenditure multiplier is more than revenue expenditure multiplier. The study also finds that the impact of public economic activities on private economic activities for both consumption expenditure and gross capital formation of government crowds in private consumption expenditure and private gross capital formation, respectively, thus indicating that government expenditure complements private expenditure in India.Keywords: government spending, fiscal policy, multiplier, growth
Procedia PDF Downloads 13410168 Hybrid Feature Selection Method for Sentiment Classification of Movie Reviews
Authors: Vishnu Goyal, Basant Agarwal
Abstract:
Sentiment analysis research provides methods for identifying the people’s opinion written in blogs, reviews, social networking websites etc. Sentiment analysis is to understand what opinion people have about any given entity, object or thing. Sentiment analysis research can be broadly categorised into three types of approaches i.e. semantic orientation, machine learning and lexicon based approaches. Feature selection methods improve the performance of the machine learning algorithms by eliminating the irrelevant features. Information gain feature selection method has been considered best method for sentiment analysis; however, it has the drawback of selection of threshold. Therefore, in this paper, we propose a hybrid feature selection methods comprising of information gain and proposed feature selection method. Initially, features are selected using Information Gain (IG) and further more noisy features are eliminated using the proposed feature selection method. Experimental results show the efficiency of the proposed feature selection methods.Keywords: feature selection, sentiment analysis, hybrid feature selection
Procedia PDF Downloads 34110167 Development of a Secured Telemedical System Using Biometric Feature
Authors: O. Iyare, A. H. Afolayan, O. T. Oluwadare, B. K. Alese
Abstract:
Access to advanced medical services has been one of the medical challenges faced by our present society especially in distant geographical locations which may be inaccessible. Then the need for telemedicine arises through which live videos of a doctor can be streamed to a patient located anywhere in the world at any time. Patients’ medical records contain very sensitive information which should not be made accessible to unauthorized people in order to protect privacy, integrity and confidentiality. This research work focuses on a more robust security measure which is biometric (fingerprint) as a form of access control to data of patients by the medical specialist/practitioner.Keywords: biometrics, telemedicine, privacy, patient information
Procedia PDF Downloads 29010166 New Types of Fitness Equipment for Seniors-Based on Beginning Movement Load Training
Authors: Chia-Chi Chen, Tai-Sheng Huang
Abstract:
Ageing society has been spread around the world. The global population is not only ageing but also declining. The structure of population has changed, which has a significant impact on both the economies and industries. Thus, how to be a healthy senior citizen to relieve the burden to the family and society will be a popular issue. Although fitness equipment manufacturing industry has been mature, the ageing population is still increasing. Therefore, this study aims to design an innovative style of fitness equipment for senior citizens, based on BMLT presented by Dr. Koyama Hirofumi. The analysis of current fitness equipment on the market and the future trend will be applied in the study. With the coming of information age, senior citizens in the future are the users of information product for sure, and the new style of fitness equipment will be combined with information technology as well. Through this study, it is believed to design an innovative style of fitness equipment for seniors and help them live heartier and happier lives.Keywords: aging society, BMLT (Beginning Movement Load Training), seniors, new style of fitness equipment
Procedia PDF Downloads 21910165 Classification of Coughing and Breathing Activities Using Wearable and a Light-Weight DL Model
Authors: Subham Ghosh, Arnab Nandi
Abstract:
Background: The proliferation of Wireless Body Area Networks (WBAN) and Internet of Things (IoT) applications demonstrates the potential for continuous monitoring of physical changes in the body. These technologies are vital for health monitoring tasks, such as identifying coughing and breathing activities, which are necessary for disease diagnosis and management. Monitoring activities such as coughing and deep breathing can provide valuable insights into a variety of medical issues. Wearable radio-based antenna sensors, which are lightweight and easy to incorporate into clothing or portable goods, provide continuous monitoring. This mobility gives it a substantial advantage over stationary environmental sensors like as cameras and radar, which are constrained to certain places. Furthermore, using compressive techniques provides benefits such as reduced data transmission speeds and memory needs. These wearable sensors offer more advanced and diverse health monitoring capabilities. Methodology: This study analyzes the feasibility of using a semi-flexible antenna operating at 2.4 GHz (ISM band) and positioned around the neck and near the mouth to identify three activities: coughing, deep breathing, and idleness. Vector network analyzer (VNA) is used to collect time-varying complex reflection coefficient data from perturbed antenna nearfield. The reflection coefficient (S11) conveys nuanced information caused by simultaneous variations in the nearfield radiation of three activities across time. The signatures are sparsely represented with gaussian windowed Gabor spectrograms. The Gabor spectrogram is used as a sparse representation approach, which reassigns the ridges of the spectrogram images to improve their resolution and focus on essential components. The antenna is biocompatible in terms of specific absorption rate (SAR). The sparsely represented Gabor spectrogram pictures are fed into a lightweight deep learning (DL) model for feature extraction and classification. Two antenna locations are investigated in order to determine the most effective localization for three different activities. Findings: Cross-validation techniques were used on data from both locations. Due to the complex form of the recorded S11, separate analyzes and assessments were performed on the magnitude, phase, and their combination. The combination of magnitude and phase fared better than the separate analyses. Various sliding window sizes, ranging from 1 to 5 seconds, were tested to find the best window for activity classification. It was discovered that a neck-mounted design was effective at detecting the three unique behaviors.Keywords: activity recognition, antenna, deep-learning, time-frequency
Procedia PDF Downloads 1610164 Cell-Cell Interactions in Diseased Conditions Revealed by Three Dimensional and Intravital Two Photon Microscope: From Visualization to Quantification
Authors: Satoshi Nishimura
Abstract:
Although much information has been garnered from the genomes of humans and mice, it remains difficult to extend that information to explain physiological and pathological phenomena. This is because the processes underlying life are by nature stochastic and fluctuate with time. Thus, we developed novel "in vivo molecular imaging" method based on single and two-photon microscopy. We visualized and analyzed many life phenomena, including common adult diseases. We integrated the knowledge obtained, and established new models that will serve as the basis for new minimally invasive therapeutic approaches.Keywords: two photon microscope, intravital visualization, thrombus, artery
Procedia PDF Downloads 37310163 Competition, Stability, and Economic Growth: A Causality Approach
Authors: Mahvish Anwaar
Abstract:
Research Question: In this paper, we explore the causal relationship between banking competition, banking stability, and economic growth. Research Findings: The unbalanced panel data starting from 2000 to 2018 is collected to analyze the causality among banking competition, banking stability, and economic growth. The main focus of the study is to check the direction of causality among selected variables. The results of the study support the demand following, supply leading, feedback, and neutrality hypothesis conditional to different measures of banking competition, banking stability, and economic growth. Theoretical Implication: Jayakumar, Pradhan, Dash, Maradana, and Gaurav (2018) proposed a theoretical model of the causal relationship between banking competition, banking stability, and economic growth by using different indicators. So, we empirically test the proposed indicators in our study. This study makes a contribution to the literature by showing the defined relationship between developing and developed countries. Policy Implications: The study covers various policy implications regarding investors to analyze how to properly manage their finances, and government agencies will take help from the present study to find the best and most suitable policies by examining how the economy can grow concerning its finances.Keywords: competition, stability, economic growth, vector auto-regression, granger causality
Procedia PDF Downloads 6410162 A Study of ZY3 Satellite Digital Elevation Model Verification and Refinement with Shuttle Radar Topography Mission
Authors: Bo Wang
Abstract:
As the first high-resolution civil optical satellite, ZY-3 satellite is able to obtain high-resolution multi-view images with three linear array sensors. The images can be used to generate Digital Elevation Models (DEM) through dense matching of stereo images. However, due to the clouds, forest, water and buildings covered on the images, there are some problems in the dense matching results such as outliers and areas failed to be matched (matching holes). This paper introduced an algorithm to verify the accuracy of DEM that generated by ZY-3 satellite with Shuttle Radar Topography Mission (SRTM). Since the accuracy of SRTM (Internal accuracy: 5 m; External accuracy: 15 m) is relatively uniform in the worldwide, it may be used to improve the accuracy of ZY-3 DEM. Based on the analysis of mass DEM and SRTM data, the processing can be divided into two aspects. The registration of ZY-3 DEM and SRTM can be firstly performed using the conjugate line features and area features matched between these two datasets. Then the ZY-3 DEM can be refined by eliminating the matching outliers and filling the matching holes. The matching outliers can be eliminated based on the statistics on Local Vector Binning (LVB). The matching holes can be filled by the elevation interpolated from SRTM. Some works are also conducted for the accuracy statistics of the ZY-3 DEM.Keywords: ZY-3 satellite imagery, DEM, SRTM, refinement
Procedia PDF Downloads 34510161 A Bibliometric Analysis on Filter Bubble
Authors: Misbah Fatma, Anam Saiyeda
Abstract:
This analysis charts the introduction and expansion of research into the filter bubble phenomena over the last 10 years using a large dataset of academic publications. This bibliometric study demonstrates how interdisciplinary filter bubble research is. The identification of key authors and organizations leading the filter bubble study sheds information on collaborative networks and knowledge transfer. Relevant papers are organized based on themes including algorithmic bias, polarisation, social media, and ethical implications through a systematic examination of the literature. In order to shed light on how these patterns have changed over time, the study plots their historical history. The study also looks at how research is distributed globally, showing geographic patterns and discrepancies in scholarly output. The results of this bibliometric analysis let us fully comprehend the development and reach of filter bubble research. This study offers insights into the ongoing discussion surrounding information personalization and its implications for societal discourse, democratic participation, and the potential risks to an informed citizenry by exposing dominant themes, interdisciplinary collaborations, and geographic patterns. In order to solve the problems caused by filter bubbles and to advance a more diverse and inclusive information environment, this analysis is essential for scholars and researchers.Keywords: bibliometric analysis, social media, social networking, algorithmic personalization, self-selection, content moderation policies and limited access to information, recommender system and polarization
Procedia PDF Downloads 11910160 Usability Testing on Information Design through Single-Lens Wearable Device
Authors: Jae-Hyun Choi, Sung-Soo Bae, Sangyoung Yoon, Hong-Ku Yun, Jiyoung Kwahk
Abstract:
This study was conducted to investigate the effect of ocular dominance on recognition performance using a single-lens smart display designed for cycling. A total of 36 bicycle riders who have been cycling consistently were recruited and participated in the experiment. The participants were asked to perform tasks riding a bicycle on a stationary stand for safety reasons. Independent variables of interest include ocular dominance, bike usage, age group, and information layout. Recognition time (i.e., the time required to identify specific information measured with an eye-tracker), error rate (i.e. false answer or failure to identify the information in 5 seconds), and user preference scores were measured and statistical tests were conducted to identify significant results. Recognition time and error ratio showed significant difference by ocular dominance factor, while the preference score did not. Recognition time was faster when the single-lens see-through display on the dominant eye (average 1.12sec) than on the non-dominant eye (average 1.38sec). Error ratio of the information recognition task was significantly lower when the see-through display was worn on the dominant eye (average 4.86%) than on the non-dominant eye (average 14.04%). The interaction effect of ocular dominance and age group was significant with respect to recognition time and error ratio. The recognition time of the users in their 40s was significantly longer than the other age groups when the display was placed on the non-dominant eye, while no difference was observed on the dominant eye. Error ratio also showed the same pattern. Although no difference was observed for the main effect of ocular dominance and bike usage, the interaction effect between the two variables was significant with respect to preference score. Preference score of daily bike users was higher when the display was placed on the dominant eye, whereas participants who use bikes for leisure purposes showed the opposite preference patterns. It was found more effective and efficient to wear a see-through display on the dominant eye than on the non-dominant eye, although user preference was not affected by ocular dominance. It is recommended to wear a see-through display on the dominant eye since it is safer by helping the user recognize the presented information faster and more accurately, even if the user may not notice the difference.Keywords: eye tracking, information recognition, ocular dominance, smart headware, wearable device
Procedia PDF Downloads 27310159 Analyzing Semantic Feature Using Multiple Information Sources for Reviews Summarization
Authors: Yu Hung Chiang, Hei Chia Wang
Abstract:
Nowadays, tourism has become a part of life. Before reserving hotels, customers need some information, which the most important source is online reviews, about hotels to help them make decisions. Due to the dramatic growing of online reviews, it is impossible for tourists to read all reviews manually. Therefore, designing an automatic review analysis system, which summarizes reviews, is necessary for them. The main purpose of the system is to understand the opinion of reviews, which may be positive or negative. In other words, the system would analyze whether the customers who visited the hotel like it or not. Using sentiment analysis methods will help the system achieve the purpose. In sentiment analysis methods, the targets of opinion (here they are called the feature) should be recognized to clarify the polarity of the opinion because polarity of the opinion may be ambiguous. Hence, the study proposes an unsupervised method using Part-Of-Speech pattern and multi-lexicons sentiment analysis to summarize all reviews. We expect this method can help customers search what they want information as well as make decisions efficiently.Keywords: text mining, sentiment analysis, product feature extraction, multi-lexicons
Procedia PDF Downloads 33110158 Performance Estimation of Two Port Multiple-Input and Multiple-Output Antenna for Wireless Local Area Network Applications
Authors: Radha Tomar, Satish K. Jain, Manish Panchal, P. S. Rathore
Abstract:
In the presented work, inset fed microstrip patch antenna (IFMPA) based two port MIMO Antenna system has been proposed, which is suitable for wireless local area network (WLAN) applications. IFMPA has been designed, optimized for 2.4 GHz and applied for MIMO formation. The optimized parameters of the proposed IFMPA have been used for fabrication of antenna and two port MIMO in a laboratory. Fabrication of the designed MIMO antenna has been done and tested experimentally for performance parameters like Envelope Correlation Coefficient (ECC), Mean Effective Gain (MEG), Directive Gain (DG), Channel Capacity Loss (CCL), Multiplexing Efficiency (ME) etc and results are compared with simulated parameters extracted with simulated S parameters to validate the results. The simulated and experimentally measured plots and numerical values of these MIMO performance parameters resembles very much with each other. This shows the success of MIMO antenna design methodology.Keywords: multiple-input and multiple-output, wireless local area network, vector network analyzer, envelope correlation coefficient
Procedia PDF Downloads 5710157 Flood-prone Urban Area Mapping Using Machine Learning, a Case Sudy of M'sila City (Algeria)
Authors: Medjadj Tarek, Ghribi Hayet
Abstract:
This study aims to develop a flood sensitivity assessment tool using machine learning (ML) techniques and geographic information system (GIS). The importance of this study is integrating the geographic information systems (GIS) and machine learning (ML) techniques for mapping flood risks, which help decision-makers to identify the most vulnerable areas and take the necessary precautions to face this type of natural disaster. To reach this goal, we will study the case of the city of M'sila, which is among the areas most vulnerable to floods. This study drew a map of flood-prone areas based on the methodology where we have made a comparison between 3 machine learning algorithms: the xGboost model, the Random Forest algorithm and the K Nearest Neighbour algorithm. Each of them gave an accuracy respectively of 97.92 - 95 - 93.75. In the process of mapping flood-prone areas, the first model was relied upon, which gave the greatest accuracy (xGboost).Keywords: Geographic information systems (GIS), machine learning (ML), emergency mapping, flood disaster management
Procedia PDF Downloads 9510156 Building Information Modeling and Its Application in the State of Kuwait
Authors: Michael Gerges, Ograbe Ahiakwo, Martin Jaeger, Ahmad Asaad
Abstract:
Recent advances of Building Information Modeling (BIM) especially in the Middle East have increased remarkably. Dubai has been taking a lead on this by making it mandatory for BIM to be adopted for all projects that involve complex architecture designs. This is because BIM is a dynamic process that assists all stakeholders in monitoring the project status throughout different project phases with great transparency. It focuses on utilizing information technology to improve collaboration among project participants during the entire life cycle of the project from the initial design, to the supply chain, resource allocation, construction and all productivity requirements. In view of this trend, the paper examines the extent of applying BIM in the State of Kuwait, by exploring practitioners’ perspectives on BIM, especially their perspectives on main barriers and main advantages. To this end structured interviews were carried out based on questionnaires and with a range of different construction professionals. The results revealed that practitioners perceive improved communication and mitigated project risks by encouraged collaboration between project participants. However, it was also observed that the full implementation of BIM in the State of Kuwait requires concerted efforts to make clients demanding BIM, counteract resistance to change among construction professionals and offer more training for design team members. This paper forms part of an on-going research effort on BIM and its application in the State of Kuwait and it is on this basis that further research on the topic is proposed.Keywords: building information modeling, BIM, construction industry, Kuwait
Procedia PDF Downloads 37810155 Flood Disaster Prevention and Mitigation in Nigeria Using Geographic Information System
Authors: Dinebari Akpee, Friday Aabe Gaage, Florence Fred Nwaigwu
Abstract:
Natural disasters like flood affect many parts of the world including developing countries like Nigeria. As a result, many human lives are lost, properties damaged and so much money is lost in infrastructure damages. These hazards and losses can be mitigated and reduced by providing reliable spatial information to the generality of the people through about flood risks through flood inundation maps. Flood inundation maps are very crucial for emergency action plans, urban planning, ecological studies and insurance rates. Nigeria experience her worst flood in her entire history this year. Many cities were submerged and completely under water due to torrential rainfall. Poor city planning, lack of effective development control among others contributes to the problem too. Geographic information system (GIS) can be used to visualize the extent of flooding, analyze flood maps to produce flood damaged estimation maps and flood risk maps. In this research, the under listed steps were taken in preparation of flood risk maps for the study area: (1) Digitization of topographic data and preparation of digital elevation model using ArcGIS (2) Flood simulation using hydraulic model and integration and (3) Integration of the first two steps to produce flood risk maps. The results shows that GIS can play crucial role in Flood disaster control and mitigation.Keywords: flood disaster, risk maps, geographic information system, hazards
Procedia PDF Downloads 22910154 An Exploratory Study to Appraise the Current Challenges and Limitations Faced in Applying and Integrating the Historic Building Information Modelling Concept for the Management of Historic Buildings
Authors: Oluwatosin Adewale
Abstract:
The sustainability of built heritage has become a relevant issue in recent years due to the social and economic values associated with these buildings. Heritage buildings provide a means for human perception of culture and represent a legacy of long-existing history; they define the local character of the social world and provide a vital connection to the past with their associated aesthetical and communal benefits. The identified values of heritage buildings have increased the importance of conservation and the lifecycle management of these buildings. The recent developments of digital design technology in engineering and the built environment have led to the adoption of Building Information Modelling (BIM) by the Architecture, Engineering, Construction, and Operations (AECO) industry. BIM provides a platform for the lifecycle management of a construction project through effective collaboration among stakeholders and the analysis of a digital information model. This growth in digital design technology has also made its way into the field of architectural heritage management in the form of Historic Building Information Modelling (HBIM). A reverse engineering process for digital documentation of heritage assets that draws upon similar information management processes as the BIM process. However, despite the several scientific and technical contributions made to the development of the HBIM process, it doesn't remain easy to integrate at the most practical level of heritage asset management. The main objective identified under the scope of the study is to review the limitations and challenges faced by heritage management professionals in adopting an HBIM-based asset management procedure for historic building projects. This paper uses an exploratory study in the form of semi-structured interviews to investigate the research problem. A purposive sample of heritage industry experts and professionals were selected to take part in a semi-structured interview to appraise some of the limitations and challenges they have faced with the integration of HBIM into their project workflows. The findings from this study will present the challenges and limitations faced in applying and integrating the HBIM concept for the management of historic buildings.Keywords: building information modelling, built heritage, heritage asset management, historic building information modelling, lifecycle management
Procedia PDF Downloads 10310153 Enterprise Information Portal Features: Results of Content Analysis Literature Review
Authors: Michal Krčál
Abstract:
Since their introduction in 1990’s, Enterprise Information Portals (EIPs) were investigated from different perspectives (e.g. project management, technology acceptance, IS success). However, no systematic literature review was produced to systematize both the research efforts and the technology itself. This paper reports first results of an extent systematic literature review study focused on research of EIPs and its categorization, specifically it reports a conceptual model of EIP features. The previous attempt to categorize EIP features was published in 2002. For the purpose of the literature review, content of 89 articles was analyzed in order to identify and categorize features of EIPs. The methodology of the literature review was as follows. Firstly, search queries in major indexing databases (Web of Science and SCOPUS) were used. The results of queries were analyzed according to their usability for the goal of the study. Then, full-texts were coded in Atlas.ti according to previously established coding scheme. The codes were categorized and the conceptual model of EIP features was created.Keywords: enterprise information portal, content analysis, features, systematic literature review
Procedia PDF Downloads 298