Search results for: data communication
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 27135

Search results for: data communication

25455 Use of WhatsApp Messenger for Optimal Healthcare Operational Communication during the COVID-19 Pandemic

Authors: Josiah O. Carter, Charlotte Hayden, Elizabeth Arthurs

Abstract:

Background: During the COVID-19 pandemic, hospital management policies have changed frequently and rapidly. This has created novel challenges in keeping the workforce abreast of these changes to enable them to deliver safe and effective care. Traditional communication methods, e.g. email, do not keep pace with the rapidly changing environment in the hospital, resulting in inaccurate, irrelevant, or outdated information being communicated, resulting in inefficiencies in patient care. Methods: The creation of a WhatsApp messaging group within the medical division at the Bristol Royal Infirmary has enabled senior clinicians and the hospital management team to update the medical workforce in real-time. It has two primary functions: (1) To enable dissemination of a concise, important operational summary. This comprises information on bed status and infection control procedural changes. It is fed directly from a daily critical incident briefing (2) To facilitate a monthly scheduled question and answer (Q&A) session for junior doctors to clarify issues with clinical directors, rota, and management staff. Additional ad-hoc updates are sent out for time-critical information; otherwise, it mainly functions as a broadcast-only group to prevent important information from being lost amongst other communication. All junior doctors within the medical division were invited to join the group. At present, the group comprises 131 participants, of which 10 are administrative staff (rota coordinators, management staff & clinical directors); the remaining 121 are junior clinicians working within the medical division. An electronic survey via Microsoft forms was sent out to junior doctors via the WhatsApp group and via email to assess its utilisation and effectiveness with the aim of quality improvements. Results: Of the 121 group participants, 19 completed the questionnaire (response rate 15.7%). Of these, 16/19 (84.2%) used it regularly, and 12/19 (63.2%) rated it as the most useful source for reliable updates relating to the hospital response to the COVID-19 pandemic, whereas only 2/19 (10.5%) found the trust intranet and the trust COVID-19 operational email update most useful. Respondents rated the WhatsApp group more useful as an information source (mean score 7.7/10) than as a means of providing feedback to management staff (mean score 6.3/10). Qualitative feedback suggested information around ward closures and changes to COVID cohorting, along with updates on staffing issues, were most useful. Respondents also noted the Q&A sessions were an efficient way of relaying feedback about management decisions but that it would be preferable if these sessions could be delivered more frequently. Discussion: During the current global COVID-19 pandemic, there is an increased need for rapid dissemination of critical information within NHS trusts; this includes communication between junior doctors, managers, and senior clinicians. The versatility of WhatsApp permits a variety of functions allowing for regular updates, the dissemination of time-critical information, and enables conversing and feedback. The project has demonstrated that reserved and well-managed use of a WhatsApp group is a welcome, efficient and practical means of communication between the senior management team and the junior medical workforce.

Keywords: communication, COVID-19, hospital management, WhatsApp

Procedia PDF Downloads 105
25454 Employee Job Performance and Supervisor Workplace Gossip Employee Job Engagement's Mediation Effect

Authors: Pphakamani Irvine Dlamini

Abstract:

The impact of supervisory gossip on subordinate work performance was investigated in this paper. The paper postulated that supervisory gossip, both bad and positive, has an impact on employee job engagement, which in turn has an impact on employee job performance. Data was collected from 238 employees and supervisors from the Mpumalanga Government Municipality in South Africa using a dyadic study approach. Employees responded to questions on supervisor gossip and job engagement, while supervisors responded to questions about employee work performance. Three waves of data gathering were carried out. Favourable superior gossip had a positive and substantial effect on employee job engagement, which increased employee job performance, according to the study, but negative superior gossip had a positive but insignificant effect on employee job engagement. The multicultural aspect of the municipality, as well as causation concerns and frequent method biases connected with research design, hampered the study. After successfully disentangling the supervisor-subordinate reciprocal communication web using Social Exchange Theory (SET), the study suggests that managers should instil effective ways for using both positive and negative gossip in the workplace to achieve favourable employee outcomes. Positive gossip creates workplace rivalry and competition, but negative gossip creates tension, stress, and mistrust among employees. This study attempted to assess the implication of supervisor gossip on employee job engagement and performance in the public service sector, whose employees are characterised by high job security as compared to their peers in the private sector.

Keywords: worlplace gossip, supervisor, employee engagement, LMX

Procedia PDF Downloads 117
25453 Second Language Development with an Intercultural Approach: A Pilot Program Applied to Higher Education Students from a Escuela Normal in Atequiza, Mexico

Authors: Frida C. Jaime Franco, C. Paulina Navarro Núñez, R. Jacob Sánchez Nájera

Abstract:

The importance of developing multi-language abilities in our global society is noteworthy. However, the necessity, interest, and consciousness of the significance that the development of another language represents, apart from the mother tongue, is not always the same in all contexts as it is in multicultural communities, especially in rural higher education institutions immersed in small communities. Leading opportunities for digital interaction among learners from Mexico and abroad partners represents scaffolding towards, not only language skills development but also intercultural communicative competences (ICC). This study leads us to consider what should be the best approach to work while applying a program of ICC integrated into the practice of EFL. While analyzing the roots of the language, it is possible to obtain the main objective of learning another language, to communicate with a functional purpose, as well as attaching social practices to the learning process, giving a result of functionality and significance to the target language. Hence, the collateral impact that collaborative learning leads to, aims to contribute to a better global understanding as well as a means of self and other cultural awareness through intercultural communication. While communicating through the target language by online collaboration among students in platforms of long-distance communication, language is used as a tool of interaction to broaden students’ perspectives reaching a substantial improvement with the help of their differences. This process should consider the application of the target language in the inquiry of sociocultural information, expecting the learners to integrate communicative skills to handle cultural differentiation at the same time they apply the knowledge of their target language in a real scenario of communication, despite being through virtual resources.

Keywords: collaborative learning, communicative approach, culture, interaction, interculturalism, target language, virtual partnership

Procedia PDF Downloads 121
25452 Bayesian Borrowing Methods for Count Data: Analysis of Incontinence Episodes in Patients with Overactive Bladder

Authors: Akalu Banbeta, Emmanuel Lesaffre, Reynaldo Martina, Joost Van Rosmalen

Abstract:

Including data from previous studies (historical data) in the analysis of the current study may reduce the sample size requirement and/or increase the power of analysis. The most common example is incorporating historical control data in the analysis of a current clinical trial. However, this only applies when the historical control dataare similar enough to the current control data. Recently, several Bayesian approaches for incorporating historical data have been proposed, such as the meta-analytic-predictive (MAP) prior and the modified power prior (MPP) both for single control as well as for multiple historical control arms. Here, we examine the performance of the MAP and the MPP approaches for the analysis of (over-dispersed) count data. To this end, we propose a computational method for the MPP approach for the Poisson and the negative binomial models. We conducted an extensive simulation study to assess the performance of Bayesian approaches. Additionally, we illustrate our approaches on an overactive bladder data set. For similar data across the control arms, the MPP approach outperformed the MAP approach with respect to thestatistical power. When the means across the control arms are different, the MPP yielded a slightly inflated type I error (TIE) rate, whereas the MAP did not. In contrast, when the dispersion parameters are different, the MAP gave an inflated TIE rate, whereas the MPP did not.We conclude that the MPP approach is more promising than the MAP approach for incorporating historical count data.

Keywords: count data, meta-analytic prior, negative binomial, poisson

Procedia PDF Downloads 112
25451 Strategic Citizen Participation in Applied Planning Investigations: How Planners Use Etic and Emic Community Input Perspectives to Fill-in the Gaps in Their Analysis

Authors: John Gaber

Abstract:

Planners regularly use citizen input as empirical data to help them better understand community issues they know very little about. This type of community data is based on the lived experiences of local residents and is known as "emic" data. What is becoming more common practice for planners is their use of data from local experts and stakeholders (known as "etic" data or the outsider perspective) to help them fill in the gaps in their analysis of applied planning research projects. Utilizing international Health Impact Assessment (HIA) data, I look at who planners invite to their citizen input investigations. Research presented in this paper shows that planners access a wide range of emic and etic community perspectives in their search for the “community’s view.” The paper concludes with how planners can chart out a new empirical path in their execution of emic/etic citizen participation strategies in their applied planning research projects.

Keywords: citizen participation, emic data, etic data, Health Impact Assessment (HIA)

Procedia PDF Downloads 481
25450 Data Augmentation for Automatic Graphical User Interface Generation Based on Generative Adversarial Network

Authors: Xulu Yao, Moi Hoon Yap, Yanlong Zhang

Abstract:

As a branch of artificial neural network, deep learning is widely used in the field of image recognition, but the lack of its dataset leads to imperfect model learning. By analysing the data scale requirements of deep learning and aiming at the application in GUI generation, it is found that the collection of GUI dataset is a time-consuming and labor-consuming project, which is difficult to meet the needs of current deep learning network. To solve this problem, this paper proposes a semi-supervised deep learning model that relies on the original small-scale datasets to produce a large number of reliable data sets. By combining the cyclic neural network with the generated countermeasure network, the cyclic neural network can learn the sequence relationship and characteristics of data, make the generated countermeasure network generate reasonable data, and then expand the Rico dataset. Relying on the network structure, the characteristics of collected data can be well analysed, and a large number of reasonable data can be generated according to these characteristics. After data processing, a reliable dataset for model training can be formed, which alleviates the problem of dataset shortage in deep learning.

Keywords: GUI, deep learning, GAN, data augmentation

Procedia PDF Downloads 176
25449 Modelling Rainfall-Induced Shallow Landslides in the Northern New South Wales

Authors: S. Ravindran, Y.Liu, I. Gratchev, D.Jeng

Abstract:

Rainfall-induced shallow landslides are more common in the northern New South Wales (NSW), Australia. From 2009 to 2017, around 105 rainfall-induced landslides occurred along the road corridors and caused temporary road closures in the northern NSW. Rainfall causing shallow landslides has different distributions of rainfall varying from uniform, normal, decreasing to increasing rainfall intensity. The duration of rainfall varied from one day to 18 days according to historical data. The objective of this research is to analyse slope instability of some of the sites in the northern NSW by varying cumulative rainfall using SLOPE/W and SEEP/W and compare with field data of rainfall causing shallow landslides. The rainfall data and topographical data from public authorities and soil data obtained from laboratory tests will be used for this modelling. There is a likelihood of shallow landslides if the cumulative rainfall is between 100 mm to 400 mm in accordance with field data.

Keywords: landslides, modelling, rainfall, suction

Procedia PDF Downloads 169
25448 Machine Learning-Enabled Classification of Climbing Using Small Data

Authors: Nicholas Milburn, Yu Liang, Dalei Wu

Abstract:

Athlete performance scoring within the climbing do-main presents interesting challenges as the sport does not have an objective way to assign skill. Assessing skill levels within any sport is valuable as it can be used to mark progress while training, and it can help an athlete choose appropriate climbs to attempt. Machine learning-based methods are popular for complex problems like this. The dataset available was composed of dynamic force data recorded during climbing; however, this dataset came with challenges such as data scarcity, imbalance, and it was temporally heterogeneous. Investigated solutions to these challenges include data augmentation, temporal normalization, conversion of time series to the spectral domain, and cross validation strategies. The investigated solutions to the classification problem included light weight machine classifiers KNN and SVM as well as the deep learning with CNN. The best performing model had an 80% accuracy. In conclusion, there seems to be enough information within climbing force data to accurately categorize climbers by skill.

Keywords: classification, climbing, data imbalance, data scarcity, machine learning, time sequence

Procedia PDF Downloads 138
25447 Analysis of Expression Data Using Unsupervised Techniques

Authors: M. A. I Perera, C. R. Wijesinghe, A. R. Weerasinghe

Abstract:

his study was conducted to review and identify the unsupervised techniques that can be employed to analyze gene expression data in order to identify better subtypes of tumors. Identifying subtypes of cancer help in improving the efficacy and reducing the toxicity of the treatments by identifying clues to find target therapeutics. Process of gene expression data analysis described under three steps as preprocessing, clustering, and cluster validation. Feature selection is important since the genomic data are high dimensional with a large number of features compared to samples. Hierarchical clustering and K Means are often used in the analysis of gene expression data. There are several cluster validation techniques used in validating the clusters. Heatmaps are an effective external validation method that allows comparing the identified classes with clinical variables and visual analysis of the classes.

Keywords: cancer subtypes, gene expression data analysis, clustering, cluster validation

Procedia PDF Downloads 144
25446 Split Health System for Diabetes Care in Urban Area: Experience from an Action Research Project in an Urban Poor Neighborhood in Bengaluru

Authors: T. S. Beerenahally, S. Amruthavalli, C. M. Munegowda, Leelavathi, Nagarathna

Abstract:

Introduction: In majority of urban India, the health system is split between different authorities being responsible for the health care of urban population. We believe that, apart from poor awareness and financial barriers to care, there are other health system barriers which affect quality and access to care for people with diabetes. In this paper, we attempted to identify health system complexity that determines access to public health system for diabetes care in KG Halli, a poor urban neighborhood in Bengaluru. The KG Halli has been a locus of a health systems research from 2009 to 2015. Methodology: The source of data is from the observational field-notes written by research team as part of urban health action research project (UHARP). Field notes included data from the community and the public primary care center. The data was generated by the community health assistants and the other research team members during regular home visits and interaction with individuals who self-reported to be diabetic over four years as part of UHARP. Results: It emerged during data analysis that the patients were not keen on utilizing primary public health center for many reasons. Patient has felt that the service provided at the center was not integrated. There was lack of availability of medicines, with a regular stock out of medicines in a year and laboratory service for investigation was limited. Many of them said that the time given by the providers was not sufficient and there was also a feeling of providers not listening to them attentively. The power dynamics played a huge role in communication. Only the consultation was available for free of cost at the public primary care center. The patient had to spend for the investigations and the major portion for medicine. Conclusion: Diabetes is a chronic disease that poses an important emerging public health concern. Most of the financial burden is borne by the family as the public facilities have failed to provide free care in India. Our study indicated various factors including individual beliefs, stigma and financial constraints affecting compliance to diabetes care.

Keywords: diabetes care, disintegrated health system, quality of care, urban health

Procedia PDF Downloads 155
25445 Learning Analytics in a HiFlex Learning Environment

Authors: Matthew Montebello

Abstract:

Student engagement within a virtual learning environment generates masses of data points that can significantly contribute to the learning analytics that lead to decision support. Ideally, similar data is collected during student interaction with a physical learning space, and as a consequence, data is present at a large scale, even in relatively small classes. In this paper, we report of such an occurrence during classes held in a HiFlex modality as we investigate the advantages of adopting such a methodology. We plan to take full advantage of the learner-generated data in an attempt to further enhance the effectiveness of the adopted learning environment. This could shed crucial light on operating modalities that higher education institutions around the world will switch to in a post-COVID era.

Keywords: HiFlex, big data in higher education, learning analytics, virtual learning environment

Procedia PDF Downloads 193
25444 An Analysis of Humanitarian Data Management of Polish Non-Governmental Organizations in Ukraine Since February 2022 and Its Relevance for Ukrainian Humanitarian Data Ecosystem

Authors: Renata Kurpiewska-Korbut

Abstract:

Making an assumption that the use and sharing of data generated in humanitarian action constitute a core function of humanitarian organizations, the paper analyzes the position of the largest Polish humanitarian non-governmental organizations in the humanitarian data ecosystem in Ukraine and their approach to non-personal and personal data management since February of 2022. Both expert interviews and document analysis of non-profit organizations providing a direct response in the Ukrainian crisis context, i.e., the Polish Humanitarian Action, Caritas, Polish Medical Mission, Polish Red Cross, and the Polish Center for International Aid and the applicability of theoretical perspective of contingency theory – with its central point that the context or specific set of conditions determining the way of behavior and the choice of methods of action – help to examine the significance of data complexity and adaptive approach to data management by relief organizations in the humanitarian supply chain network. The purpose of this study is to determine how the existence of well-established and accurate internal procedures and good practices of using and sharing data (including safeguards for sensitive data) by the surveyed organizations with comparable human and technological capabilities are implemented and adjusted to Ukrainian humanitarian settings and data infrastructure. The study also poses a fundamental question of whether this crisis experience will have a determining effect on their future performance. The obtained finding indicate that Polish humanitarian organizations in Ukraine, which have their own unique code of conduct and effective managerial data practices determined by contingencies, have limited influence on improving the situational awareness of other assistance providers in the data ecosystem despite their attempts to undertake interagency work in the area of data sharing.

Keywords: humanitarian data ecosystem, humanitarian data management, polish NGOs, Ukraine

Procedia PDF Downloads 87
25443 An Approach for Estimation in Hierarchical Clustered Data Applicable to Rare Diseases

Authors: Daniel C. Bonzo

Abstract:

Practical considerations lead to the use of unit of analysis within subjects, e.g., bleeding episodes or treatment-related adverse events, in rare disease settings. This is coupled with data augmentation techniques such as extrapolation to enlarge the subject base. In general, one can think about extrapolation of data as extending information and conclusions from one estimand to another estimand. This approach induces hierarchichal clustered data with varying cluster sizes. Extrapolation of clinical trial data is being accepted increasingly by regulatory agencies as a means of generating data in diverse situations during drug development process. Under certain circumstances, data can be extrapolated to a different population, a different but related indication, and different but similar product. We consider here the problem of estimation (point and interval) using a mixed-models approach under an extrapolation. It is proposed that estimators (point and interval) be constructed using weighting schemes for the clusters, e.g., equally weighted and with weights proportional to cluster size. Simulated data generated under varying scenarios are then used to evaluate the performance of this approach. In conclusion, the evaluation result showed that the approach is a useful means for improving statistical inference in rare disease settings and thus aids not only signal detection but risk-benefit evaluation as well.

Keywords: clustered data, estimand, extrapolation, mixed model

Procedia PDF Downloads 129
25442 A Dual-Polarized Wideband Probe for Near-Field Antenna Measurement

Authors: K. S. Sruthi

Abstract:

Antennas are one of the most important parts of a communication chain. They are used for both communication and calibration purposes. New developments in probe technologies have enabled near-field probes with much larger bandwidth. The objective of this paper is to design, simulate and fabricate a dual polarized wide band inverted quad ridged shape horn antenna which can be used as measurement probe for near field measurements. The inverted quad-ridged horn antenna probe not only provides measurement in the much wider range but also provides dual-polarization measurement thus enabling antenna developers to measure UWB, UHF, VHF antennas more precisely and at lower cost. The antenna is designed to meet the characteristics such as high gain, light weight, linearly polarized with suppressed side lobes for near-field measurement applications. The proposed antenna is simulated with commercially available packages such as Ansoft HFSS. The antenna gives a moderate gain over operating range while delivering a wide bandwidth.

Keywords: near-field antenna measurement, inverted quad-ridge horn antenna, wideband Antennas, dual polarized antennas, ansoft HFSS

Procedia PDF Downloads 414
25441 Digital Twin Strategies and Technologies for Modern Supply Chains

Authors: Mayank Sharma, Anubhaw Kumar, Siddharth Desai, Ankit Tomar

Abstract:

With the advent of cost-effective hardware and communication technologies, the scope of digitalising operations within a supply chain has tremendously increased. This has provided the opportunity to create digital twins of entire supply chains through the use of Internet-of-Things (IoT) and communication technologies. Adverse events like the COVID-19 pandemic and unpredictable geo-political situations have further warranted the importance of digitalization and remote operability of day-to-day operations at critical nodes. Globalisation, rising consumerism & e-commerce has exponentially increased the complexities of existing supply chains. We discuss here a scalable, future-ready and inclusive framework for creating digital twins developed along with the industry leaders from Cisco, Bosch, Accenture, Intel, Deloitte & IBM. We have proposed field-tested key technologies and frameworks required for creating digital twins. We also present case studies of real-life stable deployments done by us in the supply chains of a few marquee industry leaders.

Keywords: internet-of-things, digital twins, smart factory, industry 4.0, smart manufacturing

Procedia PDF Downloads 85
25440 Hand Gesture Recognition for Sign Language: A New Higher Order Fuzzy HMM Approach

Authors: Saad M. Darwish, Magda M. Madbouly, Murad B. Khorsheed

Abstract:

Sign Languages (SL) are the most accomplished forms of gestural communication. Therefore, their automatic analysis is a real challenge, which is interestingly implied to their lexical and syntactic organization levels. Hidden Markov models (HMM’s) have been used prominently and successfully in speech recognition and, more recently, in handwriting recognition. Consequently, they seem ideal for visual recognition of complex, structured hand gestures such as are found in sign language. In this paper, several results concerning static hand gesture recognition using an algorithm based on Type-2 Fuzzy HMM (T2FHMM) are presented. The features used as observables in the training as well as in the recognition phases are based on Singular Value Decomposition (SVD). SVD is an extension of Eigen decomposition to suit non-square matrices to reduce multi attribute hand gesture data to feature vectors. SVD optimally exposes the geometric structure of a matrix. In our approach, we replace the basic HMM arithmetic operators by some adequate Type-2 fuzzy operators that permits us to relax the additive constraint of probability measures. Therefore, T2FHMMs are able to handle both random and fuzzy uncertainties existing universally in the sequential data. Experimental results show that T2FHMMs can effectively handle noise and dialect uncertainties in hand signals besides a better classification performance than the classical HMMs. The recognition rate of the proposed system is 100% for uniform hand images and 86.21% for cluttered hand images.

Keywords: hand gesture recognition, hand detection, type-2 fuzzy logic, hidden Markov Model

Procedia PDF Downloads 455
25439 Space Time Adaptive Algorithm in Bi-Static Passive Radar Systems for Clutter Mitigation

Authors: D. Venu, N. V. Koteswara Rao

Abstract:

Space – time adaptive processing (STAP) is an effective tool for detecting a moving target in spaceborne or airborne radar systems. Since airborne passive radar systems utilize broadcast, navigation and excellent communication signals to perform various surveillance tasks and also has attracted significant interest from the distinct past, therefore the need of the hour is to have cost effective systems as compared to conventional active radar systems. Moreover, requirements of small number of secondary samples for effective clutter suppression in bi-static passive radar offer abundant illuminator resources for passive surveillance radar systems. This paper presents a framework for incorporating knowledge sources directly in the space-time beam former of airborne adaptive radars. STAP algorithm for clutter mitigation for passive bi-static radar has better quantitation of the reduction in sample size thereby amalgamating the earlier data bank with existing radar data sets. Also, we proposed a novel method to estimate the clutter matrix and perform STAP for efficient clutter suppression based on small sample size. Furthermore, the effectiveness of the proposed algorithm is verified using MATLAB simulations in order to validate STAP algorithm for passive bi-static radar. In conclusion, this study highlights the importance for various applications which augments traditional active radars using cost-effective measures.

Keywords: bistatic radar, clutter, covariance matrix passive radar, STAP

Procedia PDF Downloads 290
25438 Data Hiding by Vector Quantization in Color Image

Authors: Yung Gi Wu

Abstract:

With the growing of computer and network, digital data can be spread to anywhere in the world quickly. In addition, digital data can also be copied or tampered easily so that the security issue becomes an important topic in the protection of digital data. Digital watermark is a method to protect the ownership of digital data. Embedding the watermark will influence the quality certainly. In this paper, Vector Quantization (VQ) is used to embed the watermark into the image to fulfill the goal of data hiding. This kind of watermarking is invisible which means that the users will not conscious the existing of embedded watermark even though the embedded image has tiny difference compared to the original image. Meanwhile, VQ needs a lot of computation burden so that we adopt a fast VQ encoding scheme by partial distortion searching (PDS) and mean approximation scheme to speed up the data hiding process. The watermarks we hide to the image could be gray, bi-level and color images. Texts are also can be regarded as watermark to embed. In order to test the robustness of the system, we adopt Photoshop to fulfill sharpen, cropping and altering to check if the extracted watermark is still recognizable. Experimental results demonstrate that the proposed system can resist the above three kinds of tampering in general cases.

Keywords: data hiding, vector quantization, watermark, color image

Procedia PDF Downloads 358
25437 Twitter Ego Networks and the Capital Markets: A Social Network Analysis Perspective of Market Reactions to Earnings Announcement Events

Authors: Gregory D. Saxton

Abstract:

Networks are everywhere: lunch ties among co-workers, golfing partnerships among employees, interlocking board-of-director connections, Facebook friendship ties, etc. Each network varies in terms of its structure -its size, how inter-connected network members are, and the prevalence of sub-groups and cliques. At the same time, within any given network, some network members will have a more important, more central position on account of their greater number of connections or their capacity as “bridges” connecting members of different network cliques. The logic of network structure and position is at the heart of what is known as social network analysis, and this paper applies this logic to the study of the stock market. Using an array of data analytics and machine learning tools, this study will examine 17 million Twitter messages discussing the stocks of the firms in the S&P 1,500 index in 2018. Each of these 1,500 stocks has a distinct Twitter discussion network that varies in terms of core network characteristics such as size, density, influence, norms and values, level of activity, and embedded resources. The study’s core proposition is that the ultimate effect of any market-relevant information is contingent on the characteristics of the network through which it flows. To test this proposition, this study operationalizes each of the core network characteristics and examines their influence on market reactions to 2018 quarterly earnings announcement events.

Keywords: data analytics, investor-to-investor communication, social network analysis, Twitter

Procedia PDF Downloads 110
25436 Anomaly Detection in a Data Center with a Reconstruction Method Using a Multi-Autoencoders Model

Authors: Victor Breux, Jérôme Boutet, Alain Goret, Viviane Cattin

Abstract:

Early detection of anomalies in data centers is important to reduce downtimes and the costs of periodic maintenance. However, there is little research on this topic and even fewer on the fusion of sensor data for the detection of abnormal events. The goal of this paper is to propose a method for anomaly detection in data centers by combining sensor data (temperature, humidity, power) and deep learning models. The model described in the paper uses one autoencoder per sensor to reconstruct the inputs. The auto-encoders contain Long-Short Term Memory (LSTM) layers and are trained using the normal samples of the relevant sensors selected by correlation analysis. The difference signal between the input and its reconstruction is then used to classify the samples using feature extraction and a random forest classifier. The data measured by the sensors of a data center between January 2019 and May 2020 are used to train the model, while the data between June 2020 and May 2021 are used to assess it. Performances of the model are assessed a posteriori through F1-score by comparing detected anomalies with the data center’s history. The proposed model outperforms the state-of-the-art reconstruction method, which uses only one autoencoder taking multivariate sequences and detects an anomaly with a threshold on the reconstruction error, with an F1-score of 83.60% compared to 24.16%.

Keywords: anomaly detection, autoencoder, data centers, deep learning

Procedia PDF Downloads 185
25435 Emoji, the Language of the Future: An Analysis of the Usage and Understanding of Emoji across User-Groups

Authors: Sakshi Bhalla

Abstract:

On the one hand, given their seemingly simplistic, near universal usage and understanding, emoji are discarded as a potential step back in the evolution of communication. On the other, their effectiveness, pervasiveness, and adaptability across and within contexts are undeniable. In this study, the responses of 40 people (categorized by age) were recorded based on a uniform two-part questionnaire where they were required to a) identify the meaning of 15 emoji when placed in isolation, and b) interpret the meaning of the same 15 emoji when placed in a context-defining posting on Twitter. Their responses were studied on the basis of deviation from their responses that identified the emoji in isolation, as well as the originally intended meaning ascribed to the emoji. Based on an analysis of these results, it was discovered that each of the five age categories uses, understands and perceives emoji differently, which could be attributed to the degree of exposure they have undergone. For example, in the case of the youngest category (aged < 20), it was observed that they were the least accurate at correctly identifying emoji in isolation (~55%). Further, their proclivity to change their response with respect to the context was also the least (~31%). However, an analysis of each of their individual responses showed that these first-borns of social media seem to have reached a point where emojis no longer inspire their most literal meanings to them. The meaning and implication of these emoji have evolved to imply their context-derived meanings, even when placed in isolation. These trends carry forward meaningfully for the other four groups as well. In the case of the oldest category (aged > 35), however, the trends indicated inaccuracy and therefore, a higher incidence of a proclivity to change their responses. When studied in a continuum, the responses indicate that slowly and steadily, emoji are evolving from pictograms to ideograms. That is to suggest that they do not just indicate a one-to-one relation between a singular form and singular meaning. In fact, they communicate increasingly complicated ideas. This is much like the evolution of ancient hieroglyphics on papyrus reed or cuneiform on Sumerian clay tablets, which evolved from simple pictograms to progressively more complex ideograms. This evolution within communication is parallel to and contingent on the simultaneous evolution of communication. What’s astounding is the capacity of humans to leverage different platforms to facilitate such changes. Twiterese, as it is now called, is one of the instances where language is adapting to the demands of the digital world. That it does not have a spoken component, an ostensible grammar, and lacks standardization of use and meaning, as some might suggest, may seem like impediments in qualifying it as the 'language' of the digital world. However, that kind of a declarative remains a function of time, and time alone.

Keywords: communication, emoji, language, Twitter

Procedia PDF Downloads 93
25434 A Performance Analysis Study for Cloud Based ERP Systems

Authors: Burak Erkayman

Abstract:

The manufacturing and service organizations are in the need of using ERP systems to integrate many functions from purchasing to storage, production planning to calculation of costs. Using ERP systems by the integration in the level of information provides companies remarkable advantages in terms of profitability, productivity and efficiency in processes. Cloud computing is one of the most significant changes in information and communication technology. The developments in Cloud Computing attract business world to take advantage of this field. Cloud Computing means much more storage area, more cost saving and faster data transfer rate. In addition to these, it presents new business models, new field of study and practicable solutions for anyone’s use. These developments make inevitable the implementation of ERP systems to cloud environment. In this study, the performance of ERP systems in cloud environment is analyzed through various performance criteria and a comparison between traditional and cloud-ERP systems is presented. At the end of study the transformation and the future of ERP systems is discussed.

Keywords: cloud-ERP, ERP system performance, information system transformation

Procedia PDF Downloads 524
25433 Development of Interactional Competence: Listener Responses of Long-Term Stay Abroad Chinese L1 Speakers in Australian Universities

Authors: Wei Gao

Abstract:

The current study investigates the change of listener responses in social conversations of the second language (L2) speakers who are staying abroad with Chinese L1 speakers in Australian universities and how their long-term stay abroad impacted their design for L2 recipient actions. There is a limited amount of empirical work on L2 English listener response acquisition, particularly regarding the influence of long-term stay abroad in English-speaking countries. Little is known whether the development of L2 listener responses and the improvement of interactional competence is affected by the prolonged residency in the target L2 country. Forty-eight participants were recruited, and they participated in the designed speaking task through Computer-Mediated Communication. Results showed that long-term stay abroad Chinese L1 speakers demonstrated an English-like pattern of listener responses in communication. Long-term stay abroad experience had a significant impact on L2 English listener responses production and organization in social conversation. Long-term stay abroad L1 Chinese speakers had an active and productive response in listenership than their non-stay abroad counterparts in terms of frequency and placement in producing listener responses. However, the L2 English listener response production only occurred to be partial in response tokens, such as backchannels and reactive expressions, also in resumptive openers' employment. This study shows that L2 English listener responses could be acquired during a long-term stay abroad in English-speaking countries but showed partial acquisition in collaborative finishes production. In addition, the most prominent finding was that Chinese L1 speakers changed their overall listener responses pattern from L1 Chinese to L2 English. The study reveals specific interactional changes in English L2 listener responses acquisition. It generates pedagogical implications for cross-cultural communication and L2 pragmatics acquisition during a long-term stay abroad.

Keywords: listener responses, stay abroad, interactional competence, L2 pragmatics acquisition

Procedia PDF Downloads 79
25432 Integration Process and Analytic Interface of different Environmental Open Data Sets with Java/Oracle and R

Authors: Pavel H. Llamocca, Victoria Lopez

Abstract:

The main objective of our work is the comparative analysis of environmental data from Open Data bases, belonging to different governments. This means that you have to integrate data from various different sources. Nowadays, many governments have the intention of publishing thousands of data sets for people and organizations to use them. In this way, the quantity of applications based on Open Data is increasing. However each government has its own procedures to publish its data, and it causes a variety of formats of data sets because there are no international standards to specify the formats of the data sets from Open Data bases. Due to this variety of formats, we must build a data integration process that is able to put together all kind of formats. There are some software tools developed in order to give support to the integration process, e.g. Data Tamer, Data Wrangler. The problem with these tools is that they need data scientist interaction to take part in the integration process as a final step. In our case we don’t want to depend on a data scientist, because environmental data are usually similar and these processes can be automated by programming. The main idea of our tool is to build Hadoop procedures adapted to data sources per each government in order to achieve an automated integration. Our work focus in environment data like temperature, energy consumption, air quality, solar radiation, speeds of wind, etc. Since 2 years, the government of Madrid is publishing its Open Data bases relative to environment indicators in real time. In the same way, other governments have published Open Data sets relative to the environment (like Andalucia or Bilbao). But all of those data sets have different formats and our solution is able to integrate all of them, furthermore it allows the user to make and visualize some analysis over the real-time data. Once the integration task is done, all the data from any government has the same format and the analysis process can be initiated in a computational better way. So the tool presented in this work has two goals: 1. Integration process; and 2. Graphic and analytic interface. As a first approach, the integration process was developed using Java and Oracle and the graphic and analytic interface with Java (jsp). However, in order to open our software tool, as second approach, we also developed an implementation with R language as mature open source technology. R is a really powerful open source programming language that allows us to process and analyze a huge amount of data with high performance. There are also some R libraries for the building of a graphic interface like shiny. A performance comparison between both implementations was made and no significant differences were found. In addition, our work provides with an Official Real-Time Integrated Data Set about Environment Data in Spain to any developer in order that they can build their own applications.

Keywords: open data, R language, data integration, environmental data

Procedia PDF Downloads 306
25431 The Interconnection between Curriculum Development and ICT

Authors: Hanane Sarnou, Sabri Koç

Abstract:

In this paper, the interconnection between curriculum development for basic education and the use of information and communication technologies (ICTs) in the classroom referring to the Licence, Master's and Doctorate (LMD) benefits under such link will be presented and analysed. This study seeks to achieve to what extent LMD, competency-based approach (CBA) and ICTs use are interrelated. Likewise, the data collected from the responses of our teachers and learners who are concerned with LMD impact on their learning and teaching through interviews will be discussed, analysed, and classified. This paper is divided into two sections. The first section is about the curriculum development for basic education and its relation with higher education under the LMD and its link with ICTs in the university while the second section is about the classification of learners’ and teachers’ positive/negative responses concerning their positive or negative attitudes towards the ICT integration. The focus will be on the positive aspects of students’ expectations, opinions and assumptions regarding the integration of ICTs into the classroom under LMD and CBA.

Keywords: LMD system, CBA approach, curriculum development, ICT

Procedia PDF Downloads 410
25430 Identifying the Barriers behind the Lack of Six Sigma Use in Libyan Manufacturing Companies

Authors: Osama Elgadi, Martin Birkett, Wai Ming Cheung

Abstract:

This paper investigates the barriers behind the underutilisation of six sigma in Libyan manufacturing companies (LMCs). A mixed-method methodology is proposed, starting by conducting interviews to collect qualitative data followed by the development of a questionnaire to obtain quantitative data. The focus of this paper is on discussing the findings of the interview stage and how these can be used to further develop the questionnaire stage. The interview results showed that only four key barriers were highlighted as being encountered by LMCs. With a difference in terms of their significance, these factors were identified, and placed in descending order according to their importance, namely: “Lack of top management commitment”, “Lack of training”, “Lack of knowledge about six sigma”, and “Culture effect”. The findings also showed that some barriers which, were found in previous studies of six sigma implementation were not considered as barriers to LMCs but can, in fact, be considered as success factors or enablers for six sigma adoption. These factors were identified as: “sufficiency of time and financial resources”; “customers unsatisfied”; “good communication between all departments in the company”; “we are certain about its results and benefits to our company and unhappy with the current quality system”. These results suggest that LMCs face fewer barriers to adopting six sigma than many well-established global companies operating in other countries and could take advantage of these successful factors by developing and implementing a six sigma framework to improve their product quality and competitiveness.

Keywords: six sigma, barriers, Libyan manufacturing companies, interview

Procedia PDF Downloads 223
25429 Transforming Data into Knowledge: Mathematical and Statistical Innovations in Data Analytics

Authors: Zahid Ullah, Atlas Khan

Abstract:

The rapid growth of data in various domains has created a pressing need for effective methods to transform this data into meaningful knowledge. In this era of big data, mathematical and statistical innovations play a crucial role in unlocking insights and facilitating informed decision-making in data analytics. This abstract aims to explore the transformative potential of these innovations and their impact on converting raw data into actionable knowledge. Drawing upon a comprehensive review of existing literature, this research investigates the cutting-edge mathematical and statistical techniques that enable the conversion of data into knowledge. By evaluating their underlying principles, strengths, and limitations, we aim to identify the most promising innovations in data analytics. To demonstrate the practical applications of these innovations, real-world datasets will be utilized through case studies or simulations. This empirical approach will showcase how mathematical and statistical innovations can extract patterns, trends, and insights from complex data, enabling evidence-based decision-making across diverse domains. Furthermore, a comparative analysis will be conducted to assess the performance, scalability, interpretability, and adaptability of different innovations. By benchmarking against established techniques, we aim to validate the effectiveness and superiority of the proposed mathematical and statistical innovations in data analytics. Ethical considerations surrounding data analytics, such as privacy, security, bias, and fairness, will be addressed throughout the research. Guidelines and best practices will be developed to ensure the responsible and ethical use of mathematical and statistical innovations in data analytics. The expected contributions of this research include advancements in mathematical and statistical sciences, improved data analysis techniques, enhanced decision-making processes, and practical implications for industries and policymakers. The outcomes will guide the adoption and implementation of mathematical and statistical innovations, empowering stakeholders to transform data into actionable knowledge and drive meaningful outcomes.

Keywords: data analytics, mathematical innovations, knowledge extraction, decision-making

Procedia PDF Downloads 69
25428 FCNN-MR: A Parallel Instance Selection Method Based on Fast Condensed Nearest Neighbor Rule

Authors: Lu Si, Jie Yu, Shasha Li, Jun Ma, Lei Luo, Qingbo Wu, Yongqi Ma, Zhengji Liu

Abstract:

Instance selection (IS) technique is used to reduce the data size to improve the performance of data mining methods. Recently, to process very large data set, several proposed methods divide the training set into some disjoint subsets and apply IS algorithms independently to each subset. In this paper, we analyze the limitation of these methods and give our viewpoint about how to divide and conquer in IS procedure. Then, based on fast condensed nearest neighbor (FCNN) rule, we propose a large data sets instance selection method with MapReduce framework. Besides ensuring the prediction accuracy and reduction rate, it has two desirable properties: First, it reduces the work load in the aggregation node; Second and most important, it produces the same result with the sequential version, which other parallel methods cannot achieve. We evaluate the performance of FCNN-MR on one small data set and two large data sets. The experimental results show that it is effective and practical.

Keywords: instance selection, data reduction, MapReduce, kNN

Procedia PDF Downloads 248
25427 User Experience in Relation to Eye Tracking Behaviour in VR Gallery

Authors: Veslava Osinska, Adam Szalach, Dominik Piotrowski

Abstract:

Contemporary VR technologies allow users to explore virtual 3D spaces where they can work, socialize, learn, and play. User's interaction with GUI and the pictures displayed implicate perceptual and also cognitive processes which can be monitored due to neuroadaptive technologies. These modalities provide valuable information about the users' intentions, situational interpretations, and emotional states, to adapt an application or interface accordingly. Virtual galleries outfitted by specialized assets have been designed using the Unity engine BITSCOPE project in the frame of CHIST-ERA IV program. Users interaction with gallery objects implies the questions about his/her visual interests in art works and styles. Moreover, an attention, curiosity, and other emotional states are possible to be monitored and analyzed. Natural gaze behavior data and eye position were recorded by built-in eye-tracking module within HTC Vive headset gogle for VR. Eye gaze results are grouped due to various users’ behavior schemes and the appropriate perpetual-cognitive styles are recognized. Parallelly usability tests and surveys were adapted to identify the basic features of a user-centered interface for the virtual environments across most of the timeline of the project. A total of sixty participants were selected from the distinct faculties of University and secondary schools. Users’ primary knowledge about art and was evaluated during pretest and this way the level of art sensitivity was described. Data were collected during two months. Each participant gave written informed consent before participation. In data analysis reducing the high-dimensional data into a relatively low-dimensional subspace ta non linear algorithms were used such as multidimensional scaling and novel technique technique t-Stochastic Neighbor Embedding. This way it can classify digital art objects by multi modal time characteristics of eye tracking measures and reveal signatures describing selected artworks. Current research establishes the optimal place on aesthetic-utility scale because contemporary interfaces of most applications require to be designed in both functional and aesthetical ways. The study concerns also an analysis of visual experience for subsamples of visitors, differentiated, e.g., in terms of frequency of museum visits, cultural interests. Eye tracking data may also show how to better allocate artefacts and paintings or increase their visibility when possible.

Keywords: eye tracking, VR, UX, visual art, virtual gallery, visual communication

Procedia PDF Downloads 34
25426 Study, Design, Simulation and Fabrication of Microwave Slot Antenna

Authors: Khaled A. Madi, Rema A. Mousbahi, Mostafa B. Abuitbel, Abdualhakim O. Nagi

Abstract:

Antenna perhaps is the most important part of any communication system, it determines the overall efficiency and the direction of radiation of the system. Antennas vary in shape and size on a very wide range. For fast moving vehicles, the antenna should offer as litter aerodynamic resistance as possible. Slot antenna is best suited for this purpose. It offers very little aerodynamic resistance, compact, easy to feed and fabricate. This work presented in this paper deals with the investigation of a half wave slot antenna. The antenna has been studied, analyzed, designed, simulated, fabrication, and tested at the X-band. The field of antenna study is an extremely vast one, and to grasp the fundamentals, two pronged approaches have been used, and the focus was on the fabrication and testing of a slot waveguide directional antenna. Focuses on the design and simulation of slot antennas with an emphasis on optimization of a 9.1 GHz a rectangular waveguide have been used to feed slot antenna. A microwave fed slot antenna used in the communication lab was also simulated. The results have been presented and compared with the expected values, where a good agreement was achieved between the simulation and experimental results.

Keywords: microwave, slot antenna, simulation, fabrication

Procedia PDF Downloads 126