Search results for: healthcare data
23649 Analyzing Competitive Advantage of Internet of Things and Data Analytics in Smart City Context
Authors: Petra Hofmann, Dana Koniel, Jussi Luukkanen, Walter Nieminen, Lea Hannola, Ilkka Donoghue
Abstract:
The Covid-19 pandemic forced people to isolate and become physically less connected. The pandemic hasnot only reshaped people’s behaviours and needs but also accelerated digital transformation (DT). DT of cities has become an imperative with the outlook of converting them into smart cities in the future. Embedding digital infrastructure and smart city initiatives as part of the normal design, construction, and operation of cities provides a unique opportunity to improve connection between people. Internet of Things (IoT) is an emerging technology and one of the drivers in DT. It has disrupted many industries by introducing different services and business models, and IoT solutions are being applied in multiple fields, including smart cities. As IoT and data are fundamentally linked together, IoT solutions can only create value if the data generated by the IoT devices is analysed properly. Extracting relevant conclusions and actionable insights by using established techniques, data analytics contributes significantly to the growth and success of IoT applications and investments. Companies must grasp DT and be prepared to redesign their offerings and business models to remain competitive in today’s marketplace. As there are many IoT solutions available today, the amount of data is tremendous. The challenge for companies is to understand what solutions to focus on and how to prioritise and which data to differentiate from the competition. This paper explains how IoT and data analytics can impact competitive advantage and how companies should approach IoT and data analytics to translate them into concrete offerings and solutions in the smart city context. The study was carried out as a qualitative, literature-based research. A case study is provided to validate the preservation of company’s competitive advantage through smart city solutions. The results of the researchcontribution provide insights into the different factors and considerations related to creating competitive advantage through IoT and data analytics deployment in the smart city context. Furthermore, this paper proposes a framework that merges the factors and considerations with examples of offerings and solutions in smart cities. The data collected through IoT devices, and the intelligent use of it, can create a competitive advantage to companies operating in smart city business. Companies should take into consideration the five forces of competition that shape industries and pay attention to the technological, organisational, and external contexts which define factors for consideration of competitive advantages in the field of IoT and data analytics. Companies that can utilise these key assets in their businesses will most likely conquer the markets and have a strong foothold in the smart city business.Keywords: internet of things, data analytics, smart cities, competitive advantage
Procedia PDF Downloads 9423648 Social Data-Based Users Profiles' Enrichment
Authors: Amel Hannech, Mehdi Adda, Hamid Mcheick
Abstract:
In this paper, we propose a generic model of user profile integrating several elements that may positively impact the research process. We exploit the classical behavior of users and integrate a delimitation process of their research activities into several research sessions enriched with contextual and temporal information, which allows reflecting the current interests of these users in every period of time and infer data freshness. We argue that the annotation of resources gives more transparency on users' needs. It also strengthens social links among resources and users, and can so increase the scope of the user profile. Based on this idea, we integrate the social tagging practice in order to exploit the social users' behavior to enrich their profiles. These profiles are then integrated into a recommendation system in order to predict the interesting personalized items of users allowing to assist them in their researches and further enrich their profiles. In this recommendation, we provide users new research experiences.Keywords: user profiles, topical ontology, contextual information, folksonomies, tags' clusters, data freshness, association rules, data recommendation
Procedia PDF Downloads 26523647 ‘BEST BARK’ Dog Care and Owner Consultation System
Authors: Shalitha Jayasekara, Saluk Bawantha, Dinithi Anupama, Isuru Gunarathne, Pradeepa Bandara, Hansi De Silva
Abstract:
Dogs have been known as "man's best friend" for generations, providing friendship and loyalty to their human counterparts. However, due to people's busy lives, they are unaware of the ailments that can affect their pets. However, in recent years, mobile technologies have had a significant impact on our lives, and with technological improvements, a rule-based expert system allows the end-user to enable new types of healthcare systems. The advent of Android OS-based smartphones with more user-friendly interfaces and lower pricing opens new possibilities for continuous monitoring of pets' health conditions, such as healthy dogs, dangerous ingestions, and swallowed objects. The proposed ‘Best Bark’ Dog care and owner consultation system is a mobile application for dog owners. Four main components for dog owners were implemented after a questionnaire was distributed to the target group of audience and the findings were evaluated. The proposed applications are widely used to provide health and clinical support to dog owners, including suggesting exercise and diet plans and answering queries about their dogs. Additionally, after the owner uploads a photo of the dog, the application provides immediate feedback and a description of the dog's skin disease.Keywords: Convolution Neural Networks, Artificial Neural Networks, Knowledgebase, Sentimental Analysis.
Procedia PDF Downloads 15323646 Impact of External Temperature on the Speleothem Growth in the Moravian Karst
Authors: Frantisek Odvarka
Abstract:
Based on the data from the Moravian Karst, the influence of the calcite speleothem growth by selected meteorological factors was evaluated. External temperature was determined as one of the main factors influencing speleothem growth in Moravian Karst. This factor significantly influences the CO₂ concentration in soil/epikarst, and cave atmosphere in the Moravian Karst and significantly contributes to the changes in the CO₂ partial pressure differences between soil/epikarst and cave atmosphere in Moravian Karst, which determines the drip water supersaturation with respect to the calcite and quantity of precipitated calcite in the Moravian Karst cave environment. External air temperatures and cave air temperatures were measured using a COMET S3120 data logger, which can measure temperatures in the range from -30 to +80 °C with an accuracy of ± 0.4 °C. CO₂ concentrations in the cave and soils were measured with a FT A600 CO₂H Ahlborn probe (value range 0 ppmv to 10,000 ppmv, accuracy 1 ppmv), which was connected to the data logger ALMEMO 2290-4, V5 Ahlborn. The soil temperature was measured with a FHA646E1 Ahlborn probe (temperature range -20 to 70 °C, accuracy ± 0.4 °C) connected to an ALMEMO 2290-4 V5 Ahlborn data logger. The airflow velocities into and out of the cave were monitored by a FVA395 TH4 Thermo anemometer (speed range from 0.05 to 2 m s⁻¹, accuracy ± 0.04 m s⁻¹), which was connected to the ALMEMO 2590-4 V5 Ahlborn data logger for recording. The flow was measured in the lower and upper entrance of the Imperial Cave. The data were analyzed in MS Office Excel 2019 and PHREEQC.Keywords: speleothem growth, carbon dioxide partial pressure, Moravian Karst, external temperature
Procedia PDF Downloads 14423645 Using Data Mining in Automotive Safety
Authors: Carine Cridelich, Pablo Juesas Cano, Emmanuel Ramasso, Noureddine Zerhouni, Bernd Weiler
Abstract:
Safety is one of the most important considerations when buying a new car. While active safety aims at avoiding accidents, passive safety systems such as airbags and seat belts protect the occupant in case of an accident. In addition to legal regulations, organizations like Euro NCAP provide consumers with an independent assessment of the safety performance of cars and drive the development of safety systems in automobile industry. Those ratings are mainly based on injury assessment reference values derived from physical parameters measured in dummies during a car crash test. The components and sub-systems of a safety system are designed to achieve the required restraint performance. Sled tests and other types of tests are then carried out by car makers and their suppliers to confirm the protection level of the safety system. A Knowledge Discovery in Databases (KDD) process is proposed in order to minimize the number of tests. The KDD process is based on the data emerging from sled tests according to Euro NCAP specifications. About 30 parameters of the passive safety systems from different data sources (crash data, dummy protocol) are first analysed together with experts opinions. A procedure is proposed to manage missing data and validated on real data sets. Finally, a procedure is developed to estimate a set of rough initial parameters of the passive system before testing aiming at reducing the number of tests.Keywords: KDD process, passive safety systems, sled test, dummy injury assessment reference values, frontal impact
Procedia PDF Downloads 38223644 Moral Distress among Nurses Working in Hospitals in Jazan: A Cross-Sectional Study
Authors: Hussain Darraj
Abstract:
Background: Healthcare workers, especially nurses, are subjected to a great risk of psychological stress, mostly moral distress. Therefore, it is crucial to address moral distress in nurses. Objectives: The aim of this study is to study the extent of moral distress among hospital nurses in Jazan. Methods: This study used a cross-sectional study design, which included 419 nurses from Jazan hospitals. A questionnaire was used to measure moral distress and its related factors. Results: The average total score for moral distress among the study participants is 134.14, with a standard deviation of 53.94. Moreover, the current study findings indicate that those over the age of 35 years who work as nurse managers, working in critical departments, have the intention to leave a position, have received ethical training or workshops, have provided care for COVID-19 cases, or work in a department with staff shortages are associated with the experience of higher-level moral stress. Conclusion: Nurses are recommended to be provided with ongoing education and resources in order to reduce moral distress and create a positive work atmosphere for nurses. Moreover, the current study sheds light on the importance of organizational support to provide enough resources and staffing in order to reduce moral distress among nurses. Further research is needed to focus on other health professionals and moral distress. Moreover, future studies are also required to explore the strategies to reduce moral distress levels among nurses.Keywords: moral distress, Jazan, nurses, hospital
Procedia PDF Downloads 9023643 Comparative Study of Greenhouse Locations through Satellite Images and Geographic Information System: Methodological Evaluation in Venezuela
Authors: Maria A. Castillo H., Andrés R. Leandro C.
Abstract:
During the last decades, agricultural productivity in Latin America has increased with precision agriculture and more efficient agricultural technologies. The use of automated systems, satellite images, geographic information systems, and tools for data analysis, and artificial intelligence have contributed to making more effective strategic decisions. Twenty years ago, the state of Mérida, located in the Venezuelan Andes, reported the largest area covered by greenhouses in the country, where certified seeds of potatoes, vegetables, ornamentals, and flowers were produced for export and consumption in the central region of the country. In recent years, it is estimated that production under greenhouses has changed, and the area covered has decreased due to different factors, but there are few historical statistical data in sufficient quantity and quality to support this estimate or to be used for analysis and decision making. The objective of this study is to compare data collected about geoposition, use, and covered areas of the greenhouses in 2007 to data available in 2021, as support for the analysis of the current situation of horticultural production in the main municipalities of the state of Mérida. The document presents the development of the work in the diagnosis and integration of geographic coordinates in GIS and data analysis phases. As a result, an evaluation of the process is made, a dashboard is presented with the most relevant data along with the geographical coordinates integrated into GIS, and an analysis of the obtained information is made. Finally, some recommendations for actions are added, and works that expand the information obtained and its geographical traceability over time are proposed. This study contributes to granting greater certainty in the supporting data for the evaluation of social, environmental, and economic sustainability indicators and to make better decisions according to the sustainable development goals in the area under review. At the same time, the methodology provides improvements to the agricultural data collection process that can be extended to other study areas and crops.Keywords: greenhouses, geographic information system, protected agriculture, data analysis, Venezuela
Procedia PDF Downloads 9323642 Modelling Consistency and Change of Social Attitudes in 7 Years of Longitudinal Data
Authors: Paul Campbell, Nicholas Biddle
Abstract:
There is a complex, endogenous relationship between individual circumstances, attitudes, and behaviour. This study uses longitudinal panel data to assess changes in social and political attitudes over a 7-year period. Attitudes are captured with the question 'what is the most important issue facing Australia today', collected at multiple time points in a longitudinal survey of 2200 Australians. Consistency of attitudes, and factors predicting change over time, are assessed. The consistency of responses has methodological implications for data collection, specifically how often such questions ought to be asked of a population. When change in attitude is observed, this study assesses the extent to which individual demographic characteristics, personality traits, and broader societal events predict change.Keywords: attitudes, longitudinal survey analysis, personality, social values
Procedia PDF Downloads 13323641 A Generic Middleware to Instantly Sync Intensive Writes of Heterogeneous Massive Data via Internet
Authors: Haitao Yang, Zhenjiang Ruan, Fei Xu, Lanting Xia
Abstract:
Industry data centers often need to sync data changes reliably and instantly from a large-scale of heterogeneous autonomous relational databases accessed via the not-so-reliable Internet, for which a practical universal sync middle of low maintenance and operation costs is most wanted, but developing such a product and adapting it for various scenarios are a very sophisticated and continuous practice. The authors have been devising, applying, and optimizing a generic sync middleware system, named GSMS since 2006, holding the principles or advantages that the middleware must be SyncML-compliant and transparent to data application layer logic, need not refer to implementation details of databases synced, does not rely on host computer operating systems deployed, and its construction is light weighted and hence, of low cost. A series of ultimate experiments with GSMS sync performance were conducted for a persuasive example of a source relational database that underwent a broad range of write loads, say, from one thousand to one million intensive writes within a few minutes. The tests proved that GSMS has achieved an instant sync level of well below a fraction of millisecond per record sync, and GSMS’ smooth performances under ultimate write loads also showed it is feasible and competent.Keywords: heterogeneous massive data, instantly sync intensive writes, Internet generic middleware design, optimization
Procedia PDF Downloads 12223640 Building Transparent Supply Chains through Digital Tracing
Authors: Penina Orenstein
Abstract:
In today’s world, particularly with COVID-19 a constant worldwide threat, organizations need greater visibility over their supply chains more than ever before, in order to find areas for improvement and greater efficiency, reduce the chances of disruption and stay competitive. The concept of supply chain mapping is one where every process and route is mapped in detail between each vendor and supplier. The simplest method of mapping involves sourcing publicly available data including news and financial information concerning relationships between suppliers. An additional layer of information would be disclosed by large, direct suppliers about their production and logistics sites. While this method has the advantage of not requiring any input from suppliers, it also doesn’t allow for much transparency beyond the first supplier tier and may generate irrelevant data—noise—that must be filtered out to find the actionable data. The primary goal of this research is to build data maps of supply chains by focusing on a layered approach. Using these maps, the secondary goal is to address the question as to whether the supply chain is re-engineered to make improvements, for example, to lower the carbon footprint. Using a drill-down approach, the end result is a comprehensive map detailing the linkages between tier-one, tier-two, and tier-three suppliers super-imposed on a geographical map. The driving force behind this idea is to be able to trace individual parts to the exact site where they’re manufactured. In this way, companies can ensure sustainability practices from the production of raw materials through the finished goods. The approach allows companies to identify and anticipate vulnerabilities in their supply chain. It unlocks predictive analytics capabilities and enables them to act proactively. The research is particularly compelling because it unites network science theory with empirical data and presents the results in a visual, intuitive manner.Keywords: data mining, supply chain, empirical research, data mapping
Procedia PDF Downloads 17523639 Investigating Nurses’ Burnout Experiences on TikTok
Authors: Claire Song
Abstract:
Background: TikTok is an emerging social media platform creating an outlet for nurses to express and communicate their nursing experiences and stress related to nursing. Purpose: This study investigates the lived experiences of nursing burnout shared on TikTok. Method: The cross-sectional content analysis examines the video content, format, type, and quantitative indicators, including the number of likes and comments. Results: A total of 35 videos and 18616 comments were examined, published between November 2020 and May 2023. Combined, these 35 videos received 24859 comments and 1159669 of likes. Most of the videos included nurses, and 12 included nurses in professional attire. Three videos included interviewers in the video, but the rest of the videos were self-recorded. Four themes of nurses’ burnout experiences were identified: 1) high-intensity work environment, 2) negative internal perception, 3) culture of nursing work, and 4) poor teamwork experience. Conclusion: This study explored the description of nurses’ burnout experiences via a creative platform. Social media, such as TikTok, is a valuable outlet for healthcare providers to express and share their experiences. Future research might consider using the social media platform to explore coping strategies and resilience in nurses who experienced burnout.Keywords: burnout, emotional wellbeing, nursing, social media
Procedia PDF Downloads 8623638 Health Belief Model to Predict Sharps Injuries among Health Care Workers at First Level Care Facilities in Rural Pakistan
Authors: Mohammad Tahir Yousafzai, Amna Rehana Siddiqui, Naveed Zafar Janjua
Abstract:
We assessed the frequency and predictors of sharp injuries (SIs) among health care workers (HCWs) at first level care facilities (FLCF) in rural Pakistan. HCWs working at public clinic (PC), privately owned licensed practitioners’ clinic (LPC) and non-licensed practitioners’ clinic (NLC) were interviewed on universal precautions (UPs) and constructs of health belief model (HBM) to assess their association with SIs through negative-binomial regression. From 365 clinics, 485 HCWs were interviewed. Overall annual rate of Sis was 192/100 HCWs/year; 78/100 HCWs among licensed prescribers, 191/100 HCWs among non-licensed prescribers, 248/100 HCWs among qualified assistants, and 321/100 HCWs among non-qualified assistants. Increasing knowledge score about bloodborne pathogens (BBPs) transmission (rate-ratio (RR): 0.93; 95%CI: 0.89–0.96), fewer years of work experience, being a non-licensed prescriber (RR: 2.02; 95%CI: 1.36–2.98) licensed (RR: 2.86; 9%CI: 1.81–4.51) or non-licensed assistant (RR: 2.78; 95%CI: 1.72–4.47) compared to a licensed prescriber, perceived barriers (RR: 1.06;95%CI: 1.03–1.08), and compliance with UPs scores (RR: 0.93; 95%CI: 0.87–0.97) were significant predictors of SIs. Improved knowledge about BBPs, compliance with UPs and reduced barriers to follow UPs could reduce SIs to HCWs.Keywords: health belief model, sharp injuries, needle stick injuries, healthcare workers
Procedia PDF Downloads 31223637 Synoptic Analysis of a Heavy Flood in the Province of Sistan-Va-Balouchestan: Iran January 2020
Authors: N. Pegahfar, P. Ghafarian
Abstract:
In this research, the synoptic weather conditions during the heavy flood of 10-12 January 2020 in the Sistan-va-Balouchestan Province of Iran will be analyzed. To this aim, reanalysis data from the National Centers for Environmental Prediction (NCEP) and National Center for Atmospheric Research (NCAR), NCEP Global Forecasting System (GFS) analysis data, measured data from a surface station together with satellite images from the European Organization for the Exploitation of Meteorological Satellites (EUMETSAT) have been used from 9 to 12 January 2020. Atmospheric parameters both at the lower troposphere and also at the upper part of that have been used, including absolute vorticity, wind velocity, temperature, geopotential height, relative humidity, and precipitation. Results indicated that both lower-level and upper-level currents were strong. In addition, the transport of a large amount of humidity from the Oman Sea and the Red Sea to the south and southeast of Iran (Sistan-va-Balouchestan Province) led to the vast and unexpected precipitation and then a heavy flood.Keywords: Sistan-va-Balouchestn Province, heavy flood, synoptic, analysis data
Procedia PDF Downloads 10223636 Role of Machine Learning in Internet of Things Enabled Smart Cities
Authors: Amit Prakash Singh, Shyamli Singh, Chavi Srivastav
Abstract:
This paper presents the idea of Internet of Thing (IoT) for the infrastructure of smart cities. Internet of Thing has been visualized as a communication prototype that incorporates myriad of digital services. The various component of the smart cities shall be implemented using microprocessor, microcontroller, sensors for network communication and protocols. IoT enabled systems have been devised to support the smart city vision, of which aim is to exploit the currently available precocious communication technologies to support the value-added services for function of the city. Due to volume, variety, and velocity of data, it requires analysis using Big Data concept. This paper presented the various techniques used to analyze big data using machine learning.Keywords: IoT, smart city, embedded systems, sustainable environment
Procedia PDF Downloads 57523635 Machine Learning Classification of Fused Sentinel-1 and Sentinel-2 Image Data Towards Mapping Fruit Plantations in Highly Heterogenous Landscapes
Authors: Yingisani Chabalala, Elhadi Adam, Khalid Adem Ali
Abstract:
Mapping smallholder fruit plantations using optical data is challenging due to morphological landscape heterogeneity and crop types having overlapped spectral signatures. Furthermore, cloud covers limit the use of optical sensing, especially in subtropical climates where they are persistent. This research assessed the effectiveness of Sentinel-1 (S1) and Sentinel-2 (S2) data for mapping fruit trees and co-existing land-use types by using support vector machine (SVM) and random forest (RF) classifiers independently. These classifiers were also applied to fused data from the two sensors. Feature ranks were extracted using the RF mean decrease accuracy (MDA) and forward variable selection (FVS) to identify optimal spectral windows to classify fruit trees. Based on RF MDA and FVS, the SVM classifier resulted in relatively high classification accuracy with overall accuracy (OA) = 0.91.6% and kappa coefficient = 0.91% when applied to the fused satellite data. Application of SVM to S1, S2, S2 selected variables and S1S2 fusion independently produced OA = 27.64, Kappa coefficient = 0.13%; OA= 87%, Kappa coefficient = 86.89%; OA = 69.33, Kappa coefficient = 69. %; OA = 87.01%, Kappa coefficient = 87%, respectively. Results also indicated that the optimal spectral bands for fruit tree mapping are green (B3) and SWIR_2 (B10) for S2, whereas for S1, the vertical-horizontal (VH) polarization band. Including the textural metrics from the VV channel improved crop discrimination and co-existing land use cover types. The fusion approach proved robust and well-suited for accurate smallholder fruit plantation mapping.Keywords: smallholder agriculture, fruit trees, data fusion, precision agriculture
Procedia PDF Downloads 5423634 A Tactic for a Cosmopolitan City Comparison through a Data-Driven Approach: Case of Climate City Networking
Authors: Sombol Mokhles
Abstract:
Tackling climate change requires expanding networking opportunities between a diverse range of cities to accelerate climate actions. Existing climate city networks have limitations in actively engaging “ordinary” cities in networking processes between cities, as they encourage a few powerful cities to be followed by the many “ordinary” cities. To reimagine the networking opportunities between cities beyond global cities, this paper incorporates “cosmopolitan comparison” to expand our knowledge of a diverse range of cities using a data-driven approach. Through a cosmopolitan perspective, a framework is presented on how to utilise large data to expand knowledge of cities beyond global cities to reimagine the existing hierarchical networking practices. The contribution of this framework is beyond urban climate governance but inclusive of different fields which strive for a more inclusive and cosmopolitan comparison attentive to the differences across cities.Keywords: cosmopolitan city comparison, data-driven approach, climate city networking, urban climate governance
Procedia PDF Downloads 11123633 Unusual High Origin and Superficial Course of Radial Artery: A Case Report with Embryological Explanation
Authors: Anasuya Ghosh, Subhramoy Chaudhury
Abstract:
During routine cadaveric dissection at gross anatomy lab of our institution, a radial artery was found with unusual origin and superficial course. Normally the radial artery takes its origin as one of the terminal branches of brachial artery at the level of the neck of radius. It usually lies along the lateral border of fore arm deep to the brachioradialis muscle. While dissecting a 72-year-old Caucasian female cadaver, it was found that the right sided radial artery originated from the upper part of brachial artery of arm, 2 cm below the lower border of teres major muscle, from the lateral aspect of brachial artery. Then the radial artery superficially crossed the brachial artery and median nerve from lateral to medial direction and rested superficially at the cubital fossa. Embryologically, it can be explained as a failure of disappearance, or abnormal persistence of some insignificant embryonic vessels may give rise to this kind of vascular anomalies. As radial artery is one of the most important upper limb arteries, its variation and related complications are clinically significant. This unusual origin and course of radial artery should be kept in mind by all healthcare providers including surgeons and radiologists during routine venipuncture, orthopedic and plastic surgeries of arm, coronary angiographic procedures in radial approach etc. to prevent unwanted complications.Keywords: brachial artery anomalies, brachio-radial artery, high origin radial artery, superficial radial artery
Procedia PDF Downloads 32523632 An Analysis on Clustering Based Gene Selection and Classification for Gene Expression Data
Authors: K. Sathishkumar, V. Thiagarasu
Abstract:
Due to recent advances in DNA microarray technology, it is now feasible to obtain gene expression profiles of tissue samples at relatively low costs. Many scientists around the world use the advantage of this gene profiling to characterize complex biological circumstances and diseases. Microarray techniques that are used in genome-wide gene expression and genome mutation analysis help scientists and physicians in understanding of the pathophysiological mechanisms, in diagnoses and prognoses, and choosing treatment plans. DNA microarray technology has now made it possible to simultaneously monitor the expression levels of thousands of genes during important biological processes and across collections of related samples. Elucidating the patterns hidden in gene expression data offers a tremendous opportunity for an enhanced understanding of functional genomics. However, the large number of genes and the complexity of biological networks greatly increase the challenges of comprehending and interpreting the resulting mass of data, which often consists of millions of measurements. A first step toward addressing this challenge is the use of clustering techniques, which is essential in the data mining process to reveal natural structures and identify interesting patterns in the underlying data. This work presents an analysis of several clustering algorithms proposed to deals with the gene expression data effectively. The existing clustering algorithms like Support Vector Machine (SVM), K-means algorithm and evolutionary algorithm etc. are analyzed thoroughly to identify the advantages and limitations. The performance evaluation of the existing algorithms is carried out to determine the best approach. In order to improve the classification performance of the best approach in terms of Accuracy, Convergence Behavior and processing time, a hybrid clustering based optimization approach has been proposed.Keywords: microarray technology, gene expression data, clustering, gene Selection
Procedia PDF Downloads 32323631 A Theoretical Model for Pattern Extraction in Large Datasets
Authors: Muhammad Usman
Abstract:
Pattern extraction has been done in past to extract hidden and interesting patterns from large datasets. Recently, advancements are being made in these techniques by providing the ability of multi-level mining, effective dimension reduction, advanced evaluation and visualization support. This paper focuses on reviewing the current techniques in literature on the basis of these parameters. Literature review suggests that most of the techniques which provide multi-level mining and dimension reduction, do not handle mixed-type data during the process. Patterns are not extracted using advanced algorithms for large datasets. Moreover, the evaluation of patterns is not done using advanced measures which are suited for high-dimensional data. Techniques which provide visualization support are unable to handle a large number of rules in a small space. We present a theoretical model to handle these issues. The implementation of the model is beyond the scope of this paper.Keywords: association rule mining, data mining, data warehouses, visualization of association rules
Procedia PDF Downloads 22323630 Reaching the Goals of Routine HIV Screening Programs: Quantifying and Implementing an Effective HIV Screening System in Northern Nigeria Facilities Based on Optimal Volume Analysis
Authors: Folajinmi Oluwasina, Towolawi Adetayo, Kate Ssamula, Penninah Iutung, Daniel Reijer
Abstract:
Objective: Routine HIV screening has been promoted as an essential component of efforts to reduce incidence, morbidity, and mortality. The objectives of this study were to identify the optimal annual volume needed to realize the public health goals of HIV screening in the AIDS Healthcare Foundation supported hospitals and establish an implementation process to realize that optimal annual volume. Methods: Starting in 2011 a program was established to routinize HIV screening within communities and government hospitals. In 2016 Five-years of HIV screening data were reviewed to identify the optimal annual proportions of age-eligible patients screened to realize the public health goals of reducing new diagnoses and ending late-stage diagnosis (tracked as concurrent HIV/AIDS diagnosis). Analysis demonstrated that rates of new diagnoses level off when 42% of age-eligible patients were screened, providing a baseline for routine screening efforts; and concurrent HIV/AIDS diagnoses reached statistical zero at screening rates of 70%. Annual facility based targets were re-structured to meet these new target volumes. Restructuring efforts focused on right-sizing HIV screening programs to align and transition programs to integrated HIV screening within standard medical care and treatment. Results: Over one million patients were screened for HIV during the five years; 16, 033 new HIV diagnoses and access to care and treatment made successfully for 82 % (13,206), and concurrent diagnosis rates went from 32.26% to 25.27%. While screening rates increased by 104.7% over the 5-years, volume analysis demonstrated that rates need to further increase by 62.52% to reach desired 20% baseline and more than double to reach optimal annual screening volume. In 2011 facility targets for HIV screening were increased to reflect volume analysis, and in that third year, 12 of the 19 facilities reached or exceeded new baseline targets. Conclusions and Recommendation: Quantifying targets against routine HIV screening goals identified optimal annual screening volume and allowed facilities to scale their program size and allocate resources accordingly. The program transitioned from utilizing non-evidence based annual volume increases to establishing annual targets based on optimal volume analysis. This has allowed efforts to be evaluated on the ability to realize quantified goals related to the public health value of HIV screening. Optimal volume analysis helps to determine the size of an HIV screening program. It is a public health tool, not a tool to determine if an individual patient should receive screening.Keywords: HIV screening, optimal volume, HIV diagnosis, routine
Procedia PDF Downloads 26323629 Design of Data Management Software System Supporting Rendezvous and Docking with Various Spaceships
Authors: Zhan Panpan, Lu Lan, Sun Yong, He Xiongwen, Yan Dong, Gu Ming
Abstract:
The function of the two spacecraft docking network, the communication and control of a docking target with various spacecrafts is realized in the space lab data management system. In order to solve the problem of the complex data communication mode between the space lab and various spaceships, and the problem of software reuse caused by non-standard protocol, a data management software system supporting rendezvous and docking with various spaceships has been designed. The software system is based on CCSDS Spcecraft Onboard Interface Service(SOIS). It consists of Software Driver Layer, Middleware Layer and Appliaction Layer. The Software Driver Layer hides the various device interfaces using the uniform device driver framework. The Middleware Layer is divided into three lays, including transfer layer, application support layer and system business layer. The communication of space lab plaform bus and the docking bus is realized in transfer layer. Application support layer provides the inter tasks communitaion and the function of unified time management for the software system. The data management software functions are realized in system business layer, which contains telemetry management service, telecontrol management service, flight status management service, rendezvous and docking management service and so on. The Appliaction Layer accomplishes the space lab data management system defined tasks using the standard interface supplied by the Middleware Layer. On the basis of layered architecture, rendezvous and docking tasks and the rendezvous and docking management service are independent in the software system. The rendezvous and docking tasks will be activated and executed according to the different spaceships. In this way, the communication management functions in the independent flight mode, the combination mode of the manned spaceship and the combination mode of the cargo spaceship are achieved separately. The software architecture designed standard appliction interface for the services in each layer. Different requirements of the space lab can be supported by the use of standard services per layer, and the scalability and flexibility of the data management software can be effectively improved. It can also dynamically expand the number and adapt to the protocol of visiting spaceships. The software system has been applied in the data management subsystem of the space lab, and has been verified in the flight of the space lab. The research results of this paper can provide the basis for the design of the data manage system in the future space station.Keywords: space lab, rendezvous and docking, data management, software system
Procedia PDF Downloads 36823628 Poster for Sickle Cell Disease and Barriers to Care in South Yorkshire from 2017 to 2023
Authors: Amardass Dhami, Clare Samuelson
Abstract:
Background: Sickle cell disease (SCD) is a complex, multisystem condition that significantly impacts patients' quality of life, characterized by acute illness episodes, progressive organ damage, and reduced life expectancy. In the UK, over 13,000 individuals are affected, with South Yorkshire having the fifth highest prevalence, including approximately 800 patients. Retinal complications in SCD can manifest as either proliferative or non-proliferative disease, with proliferative changes being more prevalent. These retinal issues can cause significant morbidity, including visual loss and increased care requirements, underscoring the need for regular monitoring. An integrated approach was applied to ensure timely interventions, ultimately enhancing patient outcomes and reduce ‘did not attend’ rates. Aim: To assess the factors which may influence attendance to Haematology and Ophthalmology Clinics with attention towards levels of deprivation towards non-attendance. Method : A retrospective study on 84 eligible patients, from the regional tertiary Centre for Sickle Cell Care (Sheffield Teaching Hospital) from 2017 to 2023. The study focused on the incidence of sickle cell eye disease, specifically examining the outcomes of patients who attended the combined haematology and ophthalmology clinics. Patients who did not attend either clinic were excluded from the analysis to ensure a clear understanding of the combined clinic's impact. This data was then compared with the United Kingdom’s Index of Multiple Deprivation (IMD) datasets to assess if inequalities of care affected this population. Results: The study concluded that the effectiveness of combining haematology and ophthalmology clinics was reduced following the intervention. The DNA rates increased to 40% for the haematology clinic. Additionally, a significant proportion of the cohort was classified as residing in areas of deprivation, suggesting a possible link between socioeconomic factors and non-attendance rates Conclusion: These findings underscore the challenges of integrating care for SCD patients, particularly in relation to socioeconomic barriers. Despite the intent to streamline care and improve patient outcomes, the increase in DNA rates points to the need for further investigation into the underlying causes of non-attendance. Addressing these issues, especially in deprived areas, could enhance the effectiveness of combined clinics and ensure that patients receive the necessary monitoring and interventions for their eye health and overall well-being. Future strategies may need to focus on improving accessibility, outreach, and support for patients to mitigate the impact of socioeconomic factors on healthcare attendance.Keywords: south yorkshire, sickle cell anemia, deprivation, factors, haematology
Procedia PDF Downloads 1423627 The Wear Recognition on Guide Surface Based on the Feature of Radar Graph
Authors: Youhang Zhou, Weimin Zeng, Qi Xie
Abstract:
Abstract: In order to solve the wear recognition problem of the machine tool guide surface, a new machine tool guide surface recognition method based on the radar-graph barycentre feature is presented in this paper. Firstly, the gray mean value, skewness, projection variance, flat degrees and kurtosis features of the guide surface image data are defined as primary characteristics. Secondly, data Visualization technology based on radar graph is used. The visual barycentre graphical feature is demonstrated based on the radar plot of multi-dimensional data. Thirdly, a classifier based on the support vector machine technology is used, the radar-graph barycentre feature and wear original feature are put into the classifier separately for classification and comparative analysis of classification and experiment results. The calculation and experimental results show that the method based on the radar-graph barycentre feature can detect the guide surface effectively.Keywords: guide surface, wear defects, feature extraction, data visualization
Procedia PDF Downloads 51923626 Aggregation Scheduling Algorithms in Wireless Sensor Networks
Authors: Min Kyung An
Abstract:
In Wireless Sensor Networks which consist of tiny wireless sensor nodes with limited battery power, one of the most fundamental applications is data aggregation which collects nearby environmental conditions and aggregates the data to a designated destination, called a sink node. Important issues concerning the data aggregation are time efficiency and energy consumption due to its limited energy, and therefore, the related problem, named Minimum Latency Aggregation Scheduling (MLAS), has been the focus of many researchers. Its objective is to compute the minimum latency schedule, that is, to compute a schedule with the minimum number of timeslots, such that the sink node can receive the aggregated data from all the other nodes without any collision or interference. For the problem, the two interference models, the graph model and the more realistic physical interference model known as Signal-to-Interference-Noise-Ratio (SINR), have been adopted with different power models, uniform-power and non-uniform power (with power control or without power control), and different antenna models, omni-directional antenna and directional antenna models. In this survey article, as the problem has proven to be NP-hard, we present and compare several state-of-the-art approximation algorithms in various models on the basis of latency as its performance measure.Keywords: data aggregation, convergecast, gathering, approximation, interference, omni-directional, directional
Procedia PDF Downloads 22923625 Reliable and Energy-Aware Data Forwarding under Sink-Hole Attack in Wireless Sensor Networks
Authors: Ebrahim Alrashed
Abstract:
Wireless sensor networks are vulnerable to attacks from adversaries attempting to disrupt their operations. Sink-hole attacks are a type of attack where an adversary node drops data forwarded through it and hence affecting the reliability and accuracy of the network. Since sensor nodes have limited battery power, it is essential that any solution to the sinkhole attack problem be very energy-aware. In this paper, we present a reliable and energy efficient scheme to forward data from source nodes to the base station while under sink-hole attack. The scheme also detects sink-hole attack nodes and avoid paths that includes them.Keywords: energy-aware routing, reliability, sink-hole attack, WSN
Procedia PDF Downloads 39623624 A Near-Optimal Domain Independent Approach for Detecting Approximate Duplicates
Authors: Abdelaziz Fellah, Allaoua Maamir
Abstract:
We propose a domain-independent merging-cluster filter approach complemented with a set of algorithms for identifying approximate duplicate entities efficiently and accurately within a single and across multiple data sources. The near-optimal merging-cluster filter (MCF) approach is based on the Monge-Elkan well-tuned algorithm and extended with an affine variant of the Smith-Waterman similarity measure. Then we present constant, variable, and function threshold algorithms that work conceptually in a divide-merge filtering fashion for detecting near duplicates as hierarchical clusters along with their corresponding representatives. The algorithms take recursive refinement approaches in the spirit of filtering, merging, and updating, cluster representatives to detect approximate duplicates at each level of the cluster tree. Experiments show a high effectiveness and accuracy of the MCF approach in detecting approximate duplicates by outperforming the seminal Monge-Elkan’s algorithm on several real-world benchmarks and generated datasets.Keywords: data mining, data cleaning, approximate duplicates, near-duplicates detection, data mining applications and discovery
Procedia PDF Downloads 38723623 Delivery Service and Online-and-Offline Purchasing for Collaborative Recommendations on Retail Cross-Channels
Authors: S. H. Liao, J. M. Huang
Abstract:
The delivery service business model is the final link in logistics for both online-and-offline businesses. The online-and-offline business model focuses on the entire customer purchasing process online and offline, placing greater emphasis on the importance of data to optimize overall retail operations. For the retail industry, it is an important task of information and management to strengthen the collection and investigation of consumers' online and offline purchasing data to better understand customers and then recommend products. This study implements two-stage data mining analytics for clustering and association rules analysis to investigate Taiwanese consumers' (n=2,209) preferences for delivery service. This process clarifies online-and-offline purchasing behaviors and preferences to find knowledge profiles/patterns/rules for cross-channel collaborative recommendations. Finally, theoretical and practical implications for methodology and enterprise are presented.Keywords: delivery service, online-and-offline purchasing, retail cross-channel, collaborative recommendations, data mining analytics
Procedia PDF Downloads 3123622 A High Reliable Space-Borne File System with Applications of Device Partition and Intra-Channel Pipeline in Nand Flash
Authors: Xin Li, Ji-Yang Yu, Yue-Hua Niu, Lu-Yuan Wang
Abstract:
As an inevitable chain of the space data acquirement system, space-borne storage system based on Nand Flash has gradually been implemented in spacecraft. In face of massive, parallel and varied data on board, efficient data management become an important issue of storage research. Face to the requirements of high-performance and reliability in Nand Flash storage system, a combination of hardware and file system design can drastically increase system dependability, even for missions with a very long duration. More sophisticated flash storage concepts with advanced operating systems have been researched to improve the reliability of Nand Flash storage system on satellites. In this paper, architecture of file system with multi-channel data acquisition and storage on board is proposed, which obtains large-capacity and high-performance with the combine of intra-channel pipeline and device partition in Nand Flash. Multi-channel data in different rate are stored as independent files with parallel-storage system in device partition, which assures the high-effective and reliable throughput of file treatments. For massive and high-speed data storage, an efficiency assessment model is established to calculate the bandwidth formula of intra-channel pipeline. Information tables designed in Magnetoresistive RAM (MRAM) hold the management of bad block in Nand Flash and the arrangement of file system address for the high-reliability of data storage. During the full-load test, the throughput of 3D PLUS Module 160Gb Nand Flash can reach 120Mbps for store and reach 120Mbps for playback, which efficiently satisfies the requirement of multi-channel data acquisition in Satellite. Compared with previous literature, the results of experiments verify the advantages of the proposed system.Keywords: device partition architecture, intra-channel pipelining, nand flash, parallel storage
Procedia PDF Downloads 28923621 A Survey in Techniques for Imbalanced Intrusion Detection System Datasets
Authors: Najmeh Abedzadeh, Matthew Jacobs
Abstract:
An intrusion detection system (IDS) is a software application that monitors malicious activities and generates alerts if any are detected. However, most network activities in IDS datasets are normal, and the relatively few numbers of attacks make the available data imbalanced. Consequently, cyber-attacks can hide inside a large number of normal activities, and machine learning algorithms have difficulty learning and classifying the data correctly. In this paper, a comprehensive literature review is conducted on different types of algorithms for both implementing the IDS and methods in correcting the imbalanced IDS dataset. The most famous algorithms are machine learning (ML), deep learning (DL), synthetic minority over-sampling technique (SMOTE), and reinforcement learning (RL). Most of the research use the CSE-CIC-IDS2017, CSE-CIC-IDS2018, and NSL-KDD datasets for evaluating their algorithms.Keywords: IDS, imbalanced datasets, sampling algorithms, big data
Procedia PDF Downloads 32823620 LLM-Powered User-Centric Knowledge Graphs for Unified Enterprise Intelligence
Authors: Rajeev Kumar, Harishankar Kumar
Abstract:
Fragmented data silos within enterprises impede the extraction of meaningful insights and hinder efficiency in tasks such as product development, client understanding, and meeting preparation. To address this, we propose a system-agnostic framework that leverages large language models (LLMs) to unify diverse data sources into a cohesive, user-centered knowledge graph. By automating entity extraction, relationship inference, and semantic enrichment, the framework maps interactions, behaviors, and data around the user, enabling intelligent querying and reasoning across various data types, including emails, calendars, chats, documents, and logs. Its domain adaptability supports applications in contextual search, task prioritization, expertise identification, and personalized recommendations, all rooted in user-centric insights. Experimental results demonstrate its effectiveness in generating actionable insights, enhancing workflows such as trip planning, meeting preparation, and daily task management. This work advances the integration of knowledge graphs and LLMs, bridging the gap between fragmented data systems and intelligent, unified enterprise solutions focused on user interactions.Keywords: knowledge graph, entity extraction, relation extraction, LLM, activity graph, enterprise intelligence
Procedia PDF Downloads 5