Search results for: scenario-based queries
76 Developing Primary Care Datasets for a National Asthma Audit
Authors: Rachael Andrews, Viktoria McMillan, Shuaib Nasser, Christopher M. Roberts
Abstract:
Background and objective: The National Review of Asthma Deaths (NRAD) found that asthma management and care was inadequate in 26% of cases reviewed. Major shortfalls identified were adherence to national guidelines and standards and, particularly, the organisation of care, including supervision and monitoring in primary care, with 70% of cases reviewed having at least one avoidable factor in this area. 5.4 million people in the UK are diagnosed with and actively treated for asthma, and approximately 60,000 are admitted to hospital with acute exacerbations each year. The majority of people with asthma receive management and treatment solely in primary care. This has therefore created concern that many people within the UK are receiving sub-optimal asthma care resulting in unnecessary morbidity and risk of adverse outcome. NRAD concluded that a national asthma audit programme should be established to measure and improve processes, organisation, and outcomes of asthma care. Objective: To develop a primary care dataset enabling extraction of information from GP practices in Wales and providing robust data by which results and lessons could be drawn and drive service development and improvement. Methods: A multidisciplinary group of experts, including general practitioners, primary care organisation representatives, and asthma patients was formed and used as a source of governance and guidance. A review of asthma literature, guidance, and standards took place and was used to identify areas of asthma care which, if improved, would lead to better patient outcomes. Modified Delphi methodology was used to gain consensus from the expert group on which of the areas identified were to be prioritised, and an asthma patient and carer focus group held to seek views and feedback on areas of asthma care that were important to them. Areas of asthma care identified by both groups were mapped to asthma guidelines and standards to inform and develop primary and secondary care datasets covering both adult and pediatric care. Dataset development consisted of expert review and a targeted consultation process in order to seek broad stakeholder views and feedback. Results: Areas of asthma care identified as requiring prioritisation by the National Asthma Audit were: (i) Prescribing, (ii) Asthma diagnosis (iii) Asthma Reviews (iv) Personalised Asthma Action Plans (PAAPs) (v) Primary care follow-up after discharge from hospital (vi) Methodologies and primary care queries were developed to cover each of the areas of poor and variable asthma care identified and the queries designed to extract information directly from electronic patients’ records. Conclusion: This paper describes the methodological approach followed to develop primary care datasets for a National Asthma Audit. It sets out the principles behind the establishment of a National Asthma Audit programme in response to a national asthma mortality review and describes the development activities undertaken. Key process elements included: (i) mapping identified areas of poor and variable asthma care to national guidelines and standards, (ii) early engagement of experts, including clinicians and patients in the process, and (iii) targeted consultation of the queries to provide further insight into measures that were collectable, reproducible and relevant.Keywords: asthma, primary care, general practice, dataset development
Procedia PDF Downloads 17675 Peer Support Groups as a Tool to Increase Chances of Passing General Practice UK Qualification Exams
Authors: Thomas Abraham, Garcia de la Vega Felipe, Lubna Nishath, Nzekwe Nduka, Powell Anne-Marie
Abstract:
Introduction: The purpose of this paper is to discuss the effectiveness of a peer support network created to provide medical education, pastoral support, and reliable resources to registrars to help them pass the MRCGP exams. This paper will include a description of the network and its purpose, discuss how it has been used by trainees since its creation, and explain how this methodology can be applied to other areas of medical education and primary care. Background: The peer support network was created in February 2021, using Facebook, Telegram, and WhatsApp platforms to facilitate discussion of cases and answer queries about the exams, share resources, and offer peer support from qualified GPs and specialists. The network was created and is maintained by the authors of this paper and is open to anyone who is registered with the General Medical Council (GMC) and is studying for the MRCGP exams. Purpose: The purpose of the network is to provide medical education, pastoral support, and reliable resources to registrars to help them pass the exams. The network is free to use and is designed to take the onus away from a single medical educator and collate a vast amount of information from multiple medical educators/trainers; thereby creating a digital library of information for all trainees - exam related or otherwise. Methodology The network is managed by a team of moderators who respond to queries and facilitate discussion. Smaller study groups are created from the main group and provide a platform for trainees to work together, share resources, and provide peer support. The network has had thousands of trainees using it since February 2021, with positive feedback from all trainees. Results: The feedback from trainees has been overwhelmingly positive. Word of mouth has spread rapidly, growing the groups exponentially. Trainees add colleagues to the groups and often stay after they pass their exams to 'give back' to their fellow trainees. To date, thousands of trainees have passed the MRCGP exams using the resources and support provided by the network. Conclusion The success of this peer support network demonstrates the effectiveness of creating a network of thousands of doctors to provide medical education and support.Keywords: peer support, medical education, pastoral support, MRCGP exams
Procedia PDF Downloads 13774 A Survey of Semantic Integration Approaches in Bioinformatics
Authors: Chaimaa Messaoudi, Rachida Fissoune, Hassan Badir
Abstract:
Technological advances of computer science and data analysis are helping to provide continuously huge volumes of biological data, which are available on the web. Such advances involve and require powerful techniques for data integration to extract pertinent knowledge and information for a specific question. Biomedical exploration of these big data often requires the use of complex queries across multiple autonomous, heterogeneous and distributed data sources. Semantic integration is an active area of research in several disciplines, such as databases, information-integration, and ontology. We provide a survey of some approaches and techniques for integrating biological data, we focus on those developed in the ontology community.Keywords: biological ontology, linked data, semantic data integration, semantic web
Procedia PDF Downloads 44973 Social Information Seeking: Studying the Effect of Question Type on Responses in Social Q&A Sites
Authors: Arshia Ayoub, Zahid Ashraf Wani
Abstract:
With the introduction of online social Q&A sites, people are able to reach each other efficiently for information seeking and simultaneously creating social bonds. There prevails an issue of low or no response for some questions posed by an information seeker on these sites. So this study tries to understand the effect of question type on responses in Social Q & A sites. The study found that among the answered queries, majority of them were answered within 24 hours of posting the questions and surprisingly most replies were received within one hour of posting. It was observed that questions of general information type were most likely to be answered followed by verification type.Keywords: community‐based services, information seeking, social search, social Q&A site
Procedia PDF Downloads 17672 Interactive, Topic-Oriented Search Support by a Centroid-Based Text Categorisation
Authors: Mario Kubek, Herwig Unger
Abstract:
Centroid terms are single words that semantically and topically characterise text documents and so may serve as their very compact representation in automatic text processing. In the present paper, centroids are used to measure the relevance of text documents with respect to a given search query. Thus, a new graphbased paradigm for searching texts in large corpora is proposed and evaluated against keyword-based methods. The first, promising experimental results demonstrate the usefulness of the centroid-based search procedure. It is shown that especially the routing of search queries in interactive and decentralised search systems can be greatly improved by applying this approach. A detailed discussion on further fields of its application completes this contribution.Keywords: search algorithm, centroid, query, keyword, co-occurrence, categorisation
Procedia PDF Downloads 28371 Evaluating Alternative Structures for Prefix Trees
Authors: Feras Hanandeh, Izzat Alsmadi, Muhammad M. Kwafha
Abstract:
Prefix trees or tries are data structures that are used to store data or index of data. The goal is to be able to store and retrieve data by executing queries in quick and reliable manners. In principle, the structure of the trie depends on having letters in nodes at the different levels to point to the actual words in the leafs. However, the exact structure of the trie may vary based on several aspects. In this paper, we evaluated different structures for building tries. Using datasets of words of different sizes, we evaluated the different forms of trie structures. Results showed that some characteristics may impact significantly, positively or negatively, the size and the performance of the trie. We investigated different forms and structures for the trie. Results showed that using an array of pointers in each level to represent the different alphabet letters is the best choice.Keywords: data structures, indexing, tree structure, trie, information retrieval
Procedia PDF Downloads 45270 The Emergence of Information and Communication Technologies Acting as a Challenge for Media Literacy
Authors: Geetu Gahlawat, Manisha Singh
Abstract:
In the recent years, the concept of media literacy is being extended from its traditional focus on print and audio-visual media to encompass the internet and other new media within academic and policy discourses. This article throws revolves around three significant queries which are to be dealt by the academia, general public and the policy-makers: What is media literacy? How is it changing? And what is the significance of media literacy? At the beginning of the article, the definition 'media literacy' is the ability to access, analyse, evaluate and create messages across a variety of contexts are given and then this is further being tested in connection with the internet and other information and communication technologies.Having advocated this skills-based approach to media literacy in relation to the internet, the article identifies some outstanding issues for new media literacy crucial to any policy of promoting media literacy among the population. The outcome is better understanding of media literacy and also the impact of ICT on media literacy by the public as well as media literate people.Keywords: media literacy, ICT, internet, education
Procedia PDF Downloads 61169 Framework for Assessment of Non-financial Concentration Risk
Authors: Anchal Gupta
Abstract:
Amid the escalating digitalization and deployment of cross-border technological solutions, a significant portion of the industry and regulatory bodies have begun to pose queries concerning the formulation, computation, and contemplation of concentration risk. In the financial sector, well-established parameters exist for gauging the concentration of a portfolio and similar elements. However, a unified framework appears to be absent, which could guide industry and regulators pertaining to non-financial concentration risk. This paper introduces a framework, constructed on the foundation of multiple regulations where regulators are advocating for licensed corporations to evaluate their concentration risk. The lacuna lies in the fact that, while regulators delineate what constitutes concentration risk, unlike other domains, no guidelines are provided that could assist firms. This frequently results in ambiguity and individual corporate interpretation, which, from a risk management standpoint, is less than ideal.Keywords: concentration risk, non-financial risk, government regulation, financial regulation, non-market risk, MAS, DORA, EDSP, SFC
Procedia PDF Downloads 4268 Developing a Rational Database Management System (RDBMS) Supporting Product Life Cycle Appications
Authors: Yusri Yusof, Chen Wong Keong
Abstract:
This paper presents the implementation details of a Relational Database Management System of a STEP-technology product model repository. It is able support the implementation of any EXPRESS language schema, although it has been primarily implemented to support mechanical product life cycle applications. This database support the input of STEP part 21 file format from CAD in geometrical and topological data format and support a range of queries for mechanical product life cycle applications. This proposed relational database management system uses entity-to-table method (R1) rather than type-to-table method (R4). The two mapping methods have their own strengths and drawbacks.Keywords: RDBMS, CAD, ISO 10303, part-21 file
Procedia PDF Downloads 53767 On the Interactive Search with Web Documents
Authors: Mario Kubek, Herwig Unger
Abstract:
Due to the large amount of information in the World Wide Web (WWW, web) and the lengthy and usually linearly ordered result lists of web search engines that do not indicate semantic relationships between their entries, the search for topically similar and related documents can become a tedious task. Especially, the process of formulating queries with proper terms representing specific information needs requires much effort from the user. This problem gets even bigger when the user's knowledge on a subject and its technical terms is not sufficient enough to do so. This article presents the new and interactive search application DocAnalyser that addresses this problem by enabling users to find similar and related web documents based on automatic query formulation and state-of-the-art search word extraction. Additionally, this tool can be used to track topics across semantically connected web documentsKeywords: DocAnalyser, interactive web search, search word extraction, query formulation, source topic detection, topic tracking
Procedia PDF Downloads 39466 Financial Crises in the Context of Behavioral Finance
Authors: Nousheen Tariq Bhutta, Syed Zulfiqar Ali Shah
Abstract:
Financial crises become a key impediment towards the development of countries especially in emerging economies. Based on standard finance, many researchers investigated the financial crises in different countries in order to find the underlying reason regarding occurrence these event; however they were unable to provide it. In this essence behavioral finance may be helpful in providing answers to some queries regarding occurrence and prevention of financial crises. In this paper, we explore the some psychological factors comprises of our inspiration, emotion, cognition and culture along with their reflection companies, financial markets and governments that present some supportive arguments. Moreover, we compared the views of Keynes and Minsky in order to validate the underling justification towards occurrence of financial crises and their prevention in future. This study helps the practitioners and policy makers through providing valuable recommendation in order to protect the economies.Keywords: financial crises, behavioral finance, financial markets, emerging economies
Procedia PDF Downloads 49965 Analysis of Jenni: Essay Writing Artificial Intelligence
Authors: Joud Tayeb, Dunia Moussa, Rafal Al-Khawlani, Huda Elyas
Abstract:
This research delves into the intricate AI features of Jenni, an AI-powered chatbot designed to offer personalized and engaging conversations. We explore the fundamental technologies driving Jenni's capabilities, including natural language processing (NLP), machine learning, and deep learning. Through a meticulous analysis of these technologies, we aim to unravel how Jenni effectively processes and understands user queries, generates contextually relevant responses, and continuously learns from interactions. To gain deeper insights into user experiences and satisfaction, a comprehensive survey was conducted. By analyzing the collected data, we determine that consumers mostly like Jenni AI and reported that it has improved their essay writing process, yet the algorithm needs to improve certain aspects, such as accuracy.Keywords: natural language processing, machine learning, deep learning, artificial intelligence, Jenni
Procedia PDF Downloads 364 HD-WSComp: Hypergraph Decomposition for Web Services Composition Based on QoS
Authors: Samah Benmerbi, Kamal Amroun, Abdelkamel Tari
Abstract:
The increasing number of Web service (WS)providers throughout the globe, have produced numerous Web services providing the same or similar functionality. Therefore, there is a need of tools developing the best answer of queries by selecting and composing services with total transparency. This paper reviews various QoS based Web service selection mechanisms and architectures which facilitate qualitatively optimal selection, in other fact Web service composition is required when a request cannot be fulfilled by a single web service. In such cases, it is preferable to integrate existing web services to satisfy user’s request. We introduce an automatic Web service composition method based on hypergraph decomposition using hypertree decomposition method. The problem of selection and the composition of the web services is transformed into a resolution in a hypertree by exploring the relations of dependency between web services to get composite web service via employing an execution order of WS satisfying global request.Keywords: web service, web service selection, web service composition, QoS, hypergraph decomposition, BE hypergraph decomposition, hypertree resolution
Procedia PDF Downloads 51063 Validation of Mapping Historical Linked Data to International Committee for Documentation (CIDOC) Conceptual Reference Model Using Shapes Constraint Language
Authors: Ghazal Faraj, András Micsik
Abstract:
Shapes Constraint Language (SHACL), a World Wide Web Consortium (W3C) language, provides well-defined shapes and RDF graphs, named "shape graphs". These shape graphs validate other resource description framework (RDF) graphs which are called "data graphs". The structural features of SHACL permit generating a variety of conditions to evaluate string matching patterns, value type, and other constraints. Moreover, the framework of SHACL supports high-level validation by expressing more complex conditions in languages such as SPARQL protocol and RDF Query Language (SPARQL). SHACL includes two parts: SHACL Core and SHACL-SPARQL. SHACL Core includes all shapes that cover the most frequent constraint components. While SHACL-SPARQL is an extension that allows SHACL to express more complex customized constraints. Validating the efficacy of dataset mapping is an essential component of reconciled data mechanisms, as the enhancement of different datasets linking is a sustainable process. The conventional validation methods are the semantic reasoner and SPARQL queries. The former checks formalization errors and data type inconsistency, while the latter validates the data contradiction. After executing SPARQL queries, the retrieved information needs to be checked manually by an expert. However, this methodology is time-consuming and inaccurate as it does not test the mapping model comprehensively. Therefore, there is a serious need to expose a new methodology that covers the entire validation aspects for linking and mapping diverse datasets. Our goal is to conduct a new approach to achieve optimal validation outcomes. The first step towards this goal is implementing SHACL to validate the mapping between the International Committee for Documentation (CIDOC) conceptual reference model (CRM) and one of its ontologies. To initiate this project successfully, a thorough understanding of both source and target ontologies was required. Subsequently, the proper environment to run SHACL and its shape graphs were determined. As a case study, we performed SHACL over a CIDOC-CRM dataset after running a Pellet reasoner via the Protégé program. The applied validation falls under multiple categories: a) data type validation which constrains whether the source data is mapped to the correct data type. For instance, checking whether a birthdate is assigned to xsd:datetime and linked to Person entity via crm:P82a_begin_of_the_begin property. b) Data integrity validation which detects inconsistent data. For instance, inspecting whether a person's birthdate occurred before any of the linked event creation dates. The expected results of our work are: 1) highlighting validation techniques and categories, 2) selecting the most suitable techniques for those various categories of validation tasks. The next plan is to establish a comprehensive validation model and generate SHACL shapes automatically.Keywords: SHACL, CIDOC-CRM, SPARQL, validation of ontology mapping
Procedia PDF Downloads 25462 How to Perform Proper Indexing?
Authors: Watheq Mansour, Waleed Bin Owais, Mohammad Basheer Kotit, Khaled Khan
Abstract:
Efficient query processing is one of the utmost requisites in any business environment to satisfy consumer needs. This paper investigates the various types of indexing models, viz. primary, secondary, and multi-level. The investigation is done under the ambit of various types of queries to which each indexing model performs with efficacy. This study also discusses the inherent advantages and disadvantages of each indexing model and how indexing models can be chosen based on a particular environment. This paper also draws parallels between various indexing models and provides recommendations that would help a Database administrator to zero-in on a particular indexing model attributed to the needs and requirements of the production environment. In addition, to satisfy industry and consumer needs attributed to the colossal data generation nowadays, this study has proposed two novel indexing techniques that can be used to index highly unstructured and structured Big Data with efficacy. The study also briefly discusses some best practices that the industry should follow in order to choose an indexing model that is apposite to their prerequisites and requirements.Keywords: indexing, hashing, latent semantic indexing, B-tree
Procedia PDF Downloads 15961 Metric Suite for Schema Evolution of a Relational Database
Authors: S. Ravichandra, D. V. L. N. Somayajulu
Abstract:
Requirement of stakeholders for adding more details to the database is the main cause of the schema evolution in the relational database. Further, this schema evolution causes the instability to the database. Hence, it is aimed to define a metric suite for schema evolution of a relational database. The metric suite will calculate the metrics based on the features of the database, analyse the queries on the database and measures the coupling, cohesion and component dependencies of the schema for existing and evolved versions of the database. This metric suite will also provide an indicator for the problems related to the stability and usability of the evolved database. The degree of change in the schema of a database is presented in the forms of graphs that acts as an indicator and also provides the relations between various parameters (metrics) related to the database architecture. The acquired information is used to defend and improve the stability of database architecture. The challenges arise in incorporating these metrics with varying parameters for formulating a suitable metric suite are discussed. To validate the proposed metric suite, an experimentation has been performed on publicly available datasets.Keywords: cohesion, coupling, entropy, metric suite, schema evolution
Procedia PDF Downloads 45160 Algorithm for Information Retrieval Optimization
Authors: Kehinde K. Agbele, Kehinde Daniel Aruleba, Eniafe F. Ayetiran
Abstract:
When using Information Retrieval Systems (IRS), users often present search queries made of ad-hoc keywords. It is then up to the IRS to obtain a precise representation of the user’s information need and the context of the information. This paper investigates optimization of IRS to individual information needs in order of relevance. The study addressed development of algorithms that optimize the ranking of documents retrieved from IRS. This study discusses and describes a Document Ranking Optimization (DROPT) algorithm for information retrieval (IR) in an Internet-based or designated databases environment. Conversely, as the volume of information available online and in designated databases is growing continuously, ranking algorithms can play a major role in the context of search results. In this paper, a DROPT technique for documents retrieved from a corpus is developed with respect to document index keywords and the query vectors. This is based on calculating the weight (Keywords: information retrieval, document relevance, performance measures, personalization
Procedia PDF Downloads 24259 Flow Conservation Framework for Monitoring Software Defined Networks
Authors: Jesús Antonio Puente Fernández, Luis Javier Garcia Villalba
Abstract:
New trends on streaming videos such as series or films require a high demand of network resources. This fact results in a huge problem within traditional IP networks due to the rigidity of its architecture. In this way, Software Defined Networks (SDN) is a new concept of network architecture that intends to be more flexible and it simplifies the management in networks with respect to the existing ones. These aspects are possible due to the separation of control plane (controller) and data plane (switches). Taking the advantage of this separated control, it is easy to deploy a monitoring tool independent of device vendors since the existing ones are dependent on the installation of specialized and expensive hardware. In this paper, we propose a framework that optimizes the traffic monitoring in SDN networks that decreases the number of monitoring queries to improve the network traffic and also reduces the overload. The performed experiments (with and without the optimization) using a video streaming delivery between two hosts demonstrate the feasibility of our monitoring proposal.Keywords: optimization, monitoring, software defined networking, statistics, query
Procedia PDF Downloads 33358 User Guidance for Effective Query Interpretation in Natural Language Interfaces to Ontologies
Authors: Aliyu Isah Agaie, Masrah Azrifah Azmi Murad, Nurfadhlina Mohd Sharef, Aida Mustapha
Abstract:
Natural Language Interfaces typically support a restricted language and also have scopes and limitations that naïve users are unaware of, resulting in errors when the users attempt to retrieve information from ontologies. To overcome this challenge, an auto-suggest feature is introduced into the querying process where users are guided through the querying process using interactive query construction system. Guiding users to formulate their queries, while providing them with an unconstrained (or almost unconstrained) way to query the ontology results in better interpretation of the query and ultimately lead to an effective search. The approach described in this paper is unobtrusive and subtly guides the users, so that they have a choice of either selecting from the suggestion list or typing in full. The user is not coerced into accepting system suggestions and can express himself using fragments or full sentences.Keywords: auto-suggest, expressiveness, habitability, natural language interface, query interpretation, user guidance
Procedia PDF Downloads 47457 Annotation Ontology for Semantic Web Development
Authors: Hadeel Al Obaidy, Amani Al Heela
Abstract:
The main purpose of this paper is to examine the concept of semantic web and the role that ontology and semantic annotation plays in the development of semantic web services. The paper focuses on semantic web infrastructure illustrating how ontology and annotation work to provide the learning capabilities for building content semantically. To improve productivity and quality of software, the paper applies approaches, notations and techniques offered by software engineering. It proposes a conceptual model to develop semantic web services for the infrastructure of web information retrieval system of digital libraries. The developed system uses ontology and annotation to build a knowledge based system to define and link the meaning of a web content to retrieve information for users’ queries. The results are more relevant through keywords and ontology rule expansion that will be more accurate to satisfy the requested information. The level of results accuracy would be enhanced since the query semantically analyzed work with the conceptual architecture of the proposed system.Keywords: semantic web services, software engineering, semantic library, knowledge representation, ontology
Procedia PDF Downloads 17456 Design and Development of a Platform for Analyzing Spatio-Temporal Data from Wireless Sensor Networks
Authors: Walid Fantazi
Abstract:
The development of sensor technology (such as microelectromechanical systems (MEMS), wireless communications, embedded systems, distributed processing and wireless sensor applications) has contributed to a broad range of WSN applications which are capable of collecting a large amount of spatiotemporal data in real time. These systems require real-time data processing to manage storage in real time and query the data they process. In order to cover these needs, we propose in this paper a Snapshot spatiotemporal data model based on object-oriented concepts. This model allows saving storing and reducing data redundancy which makes it easier to execute spatiotemporal queries and save analyzes time. Further, to ensure the robustness of the system as well as the elimination of congestion from the main access memory we propose a spatiotemporal indexing technique in RAM called Captree *. As a result, we offer an RIA (Rich Internet Application) -based SOA application architecture which allows the remote monitoring and control.Keywords: WSN, indexing data, SOA, RIA, geographic information system
Procedia PDF Downloads 25655 Digitalization, Supply Chain Integration and Financial Performance: Case of Tunisian Agro-industrial Sector
Authors: Rym Ghariani, Younes Boujelbene
Abstract:
In contemporary times, global technological advancements, particularly those in the realm of digital technology, have emerged as pivotal instruments for enterprises in fostering viable partnerships and forging meaningful alliances with other firms. The advent of these digital innovations is poised to revolutionize nearly every facet and operation within corporate entities. The primary objective of this study is to explore the correlation between digitization, integration of supply chains, and the financial efficacy of the agro-industrial sector in Tunisia. To accomplish this, data collection employed a questionnaire as the primary research instrument. Subsequently, the research queries were addressed, and hypotheses were examined by subjecting the gathered data to principal component analysis and linear regression modeling, facilitated by the utilization of SPSS26 software. The findings revealed that digitalization within the supply chain, along with external supply chain integration, exerted discernible impacts on the financial performance of the organization.Keywords: digitalization, supply chain integration, financial performance, Tunisian agro-industrial sector
Procedia PDF Downloads 5054 Use of Generative Adversarial Networks (GANs) in Neuroimaging and Clinical Neuroscience Applications
Authors: Niloufar Yadgari
Abstract:
GANs are a potent form of deep learning models that have found success in various fields. They are part of the larger group of generative techniques, which aim to produce authentic data using a probabilistic model that learns distributions from actual samples. In clinical settings, GANs have demonstrated improved abilities in capturing spatially intricate, nonlinear, and possibly subtle disease impacts in contrast to conventional generative techniques. This review critically evaluates the current research on how GANs are being used in imaging studies of different neurological conditions like Alzheimer's disease, brain tumors, aging of the brain, and multiple sclerosis. We offer a clear explanation of different GAN techniques for each use case in neuroimaging and delve into the key hurdles, unanswered queries, and potential advancements in utilizing GANs in this field. Our goal is to connect advanced deep learning techniques with neurology studies, showcasing how GANs can assist in clinical decision-making and enhance our comprehension of the structural and functional aspects of brain disorders.Keywords: GAN, pathology, generative adversarial network, neuro imaging
Procedia PDF Downloads 3453 Dynamics of a Reaction-Diffusion Problems Modeling Two Predators Competing for a Prey
Authors: Owolabi Kolade Matthew
Abstract:
In this work, we investigate both the analytical and numerical studies of the dynamical model comprising of three species system. We analyze the linear stability of stationary solutions in the one-dimensional multi-system modeling the interactions of two predators and one prey species. The stability analysis has a lot of implications for understanding the various spatiotemporal and chaotic behaviors of the species in the spatial domain. The analysis results presented have established the possibility of the three interacting species to coexist harmoniously, this feat is achieved by combining the local and global analyzes to determine the global dynamics of the system. In the presence of diffusion, a viable exponential time differencing method is applied to multi-species nonlinear time-dependent partial differential equation to address the points and queries that may naturally arise. The scheme is described in detail, and justified by a number of computational experiments.Keywords: asymptotically stable, coexistence, exponential time differencing method, global and local stability, predator-prey model, nonlinear, reaction-diffusion system
Procedia PDF Downloads 41352 Influential Factors Affecting the Creativity Scientific Problem Finding Ability of Social Science Ph.D. Students
Authors: Yuanyuan Song
Abstract:
For Ph.D. students, the skill of formulating incisive inquiries holds immense importance, as adept questioning can significantly unravel research complexities. Social Science Ph.D. students should possess specific abilities to formulate creative research questions, and identifying the most influential factors is essential. To respond to these questions, in this study, we engaged with Ph.D. candidates with social sciences backgrounds through interviews and questionnaires. Our objective was to identify the predominant determinants influencing their capacity to formulate inventive research queries, ultimately aiming to enhance the academic journey of social science doctoral candidates. Insights gleaned from semi-structured interviews and questionnaires with 15 doctoral scholars from different universities around the world highlighted that mentorship and scholarly exchanges, prior knowledge, positive mindset, and personal interests played pivotal roles in catalyzing these students' contemplation of research inquiries.Keywords: Ph.D. education, higher education, creativity cultivation, creativity scientific problem finding ability
Procedia PDF Downloads 6851 SPARK: An Open-Source Knowledge Discovery Platform That Leverages Non-Relational Databases and Massively Parallel Computational Power for Heterogeneous Genomic Datasets
Authors: Thilina Ranaweera, Enes Makalic, John L. Hopper, Adrian Bickerstaffe
Abstract:
Data are the primary asset of biomedical researchers, and the engine for both discovery and research translation. As the volume and complexity of research datasets increase, especially with new technologies such as large single nucleotide polymorphism (SNP) chips, so too does the requirement for software to manage, process and analyze the data. Researchers often need to execute complicated queries and conduct complex analyzes of large-scale datasets. Existing tools to analyze such data, and other types of high-dimensional data, unfortunately suffer from one or more major problems. They typically require a high level of computing expertise, are too simplistic (i.e., do not fit realistic models that allow for complex interactions), are limited by computing power, do not exploit the computing power of large-scale parallel architectures (e.g. supercomputers, GPU clusters etc.), or are limited in the types of analysis available, compounded by the fact that integrating new analysis methods is not straightforward. Solutions to these problems, such as those developed and implemented on parallel architectures, are currently available to only a relatively small portion of medical researchers with access and know-how. The past decade has seen a rapid expansion of data management systems for the medical domain. Much attention has been given to systems that manage phenotype datasets generated by medical studies. The introduction of heterogeneous genomic data for research subjects that reside in these systems has highlighted the need for substantial improvements in software architecture. To address this problem, we have developed SPARK, an enabling and translational system for medical research, leveraging existing high performance computing resources, and analysis techniques currently available or being developed. It builds these into The Ark, an open-source web-based system designed to manage medical data. SPARK provides a next-generation biomedical data management solution that is based upon a novel Micro-Service architecture and Big Data technologies. The system serves to demonstrate the applicability of Micro-Service architectures for the development of high performance computing applications. When applied to high-dimensional medical datasets such as genomic data, relational data management approaches with normalized data structures suffer from unfeasibly high execution times for basic operations such as insert (i.e. importing a GWAS dataset) and the queries that are typical of the genomics research domain. SPARK resolves these problems by incorporating non-relational NoSQL databases that have been driven by the emergence of Big Data. SPARK provides researchers across the world with user-friendly access to state-of-the-art data management and analysis tools while eliminating the need for high-level informatics and programming skills. The system will benefit health and medical research by eliminating the burden of large-scale data management, querying, cleaning, and analysis. SPARK represents a major advancement in genome research technologies, vastly reducing the burden of working with genomic datasets, and enabling cutting edge analysis approaches that have previously been out of reach for many medical researchers.Keywords: biomedical research, genomics, information systems, software
Procedia PDF Downloads 27050 MindFlow: A Collective Intelligence-Based System for Helping Stress Pattern Diagnosis
Authors: Andres Frederic
Abstract:
We present the MindFlow system supporting the detection and the diagnosis of stresses. The heart of the system is a knowledge synthesis engine allowing occupational health stakeholders (psychologists, occupational therapists and human resource managers) to formulate queries related to stress and responding to users requests by recommending a pattern of stress if one exists. The stress pattern diagnosis is based on expert knowledge stored in the MindFlow stress ontology including stress feature vector. The query processing may involve direct access to the MindFlow system by occupational health stakeholders, online communication between the MindFlow system and the MindFlow domain experts, or direct dialog between a occupational health stakeholder and a MindFlow domain expert. The MindFlow knowledge model is generic in the sense that it supports the needs of psychologists, occupational therapists and human resource managers. The system presented in this paper is currently under development as part of a Dutch-Japanese project and aims to assist organisation in the quick diagnosis of stress patterns.Keywords: occupational stress, stress management, physiological measurement, accident prevention
Procedia PDF Downloads 43349 A Framework for Analyzing Public Interaction of Saudi Universities on Twitter
Authors: Sahar Al-Qahtani, Rabeeh Ayaz Abbasi, Naif Radi Aljohani
Abstract:
Many universities use social media platforms as new communication channels to disseminate information and promptly communicate with their audience. As Twitter is one of the widely used social media platforms, this research aims to explore the adaption and utilization of Twitter by universities. We propose a framework called 'Social Network Analysis for Universities on Twitter' (SNAUT) to analyze the usage of Twitter by universities and to measure their interaction with public. The study includes a sample of around 110,000 tweets from 36 Saudi universities, including both public and private universities. Using SNAUT, we can (1) investigate the purpose of using Twitter by universities, (2) determine the broad topics discussed by them, and (3) identify the groups closely associated with the universities. The results show that most of the Saudi universities (whether public or private) actively use Twitter. Results also reveal that public universities respond to public queries more frequently, but private universities stand out more in terms of information dissemination using retweets and diverse hashtags. Finally, we develop a ranking mechanism in SNAUT for ranking universities based on their social interaction with the public on Twitter.Keywords: social media, twitter, social network analysis, universities, higher education, Saudi Arabia
Procedia PDF Downloads 13848 Finding the Optimal Meeting Point Based on Travel Plans in Road Networks
Authors: Mohammad H. Ahmadi, Vahid Haghighatdoost
Abstract:
Given a set of source locations for a group of friends, and a set of trip plans for each group member as a sequence of Categories-of-Interests (COIs) (e.g., restaurant), and finally a specific COI as a common destination that all group members will gather together, in Meeting Point Based on Trip Plans (MPTPs) queries our goal is to find a Point-of-Interest (POI) from different COIs, such that the aggregate travel distance for the group is minimized. In this work, we considered two cases for aggregate function as Sum and Max. For solving this query, we propose an efficient pruning technique for shrinking the search space. Our approach contains three steps. In the first step, it prunes the search space around the source locations. In the second step, it prunes the search space around the centroid of source locations. Finally, we compute the intersection of all pruned areas as the final refined search space. We prove that the POIs beyond the refined area cannot be part of optimal answer set. The paper also covers an extensive performance study of the proposed technique.Keywords: meeting point, trip plans, road networks, spatial databases
Procedia PDF Downloads 18647 An Approach to Integrate Ontologies of Open Educational Resources in Knowledge Base Management Systems
Authors: Firas A. Al Laban, Mohamed Chabi, Sammani Danwawu Abdullahi
Abstract:
There are a real needs to integrate types of Open Educational Resources (OER) with an intelligent system to extract information and knowledge in the semantic searching level. Those needs raised because most of current learning standard adopted web based learning and the e-learning systems does not always serve all educational goals. Semantic Web systems provide educators, students, and researchers with intelligent queries based on a semantic knowledge management learning system. An ontology-based learning system is an advanced system, where ontology plays the core of the semantic web in a smart learning environment. The objective of this paper is to discuss the potentials of ontologies and mapping different kinds of ontologies; heterogeneous or homogenous to manage and control different types of Open Educational Resources. The important contribution of this research is to approach a methodology uses logical rules and conceptual relations to map between ontologies of different educational resources. We expect from this methodology to establish for an intelligent educational system supporting student tutoring, self and lifelong learning system.Keywords: knowledge management systems, ontologies, semantic web, open educational resources
Procedia PDF Downloads 500