Search results for: hybrid searches
577 A Survey of Recognizing of Daily Living Activities in Multi-User Smart Home Environments
Authors: Kulsoom S. Bughio, Naeem K. Janjua, Gordana Dermody, Leslie F. Sikos, Shamsul Islam
Abstract:
The advancement in information and communication technologies (ICT) and wireless sensor networks have played a pivotal role in the design and development of real-time healthcare solutions, mainly targeting the elderly living in health-assistive smart homes. Such smart homes are equipped with sensor technologies to detect and record activities of daily living (ADL). This survey reviews and evaluates existing approaches and techniques based on real-time sensor-based modeling and reasoning in single-user and multi-user environments. It classifies the approaches into three main categories: learning-based, knowledge-based, and hybrid, and evaluates how they handle temporal relations, granularity, and uncertainty. The survey also highlights open challenges across various disciplines (including computer and information sciences and health sciences) to encourage interdisciplinary research for the detection and recognition of ADLs and discusses future directions.Keywords: daily living activities, smart homes, single-user environment, multi-user environment
Procedia PDF Downloads 141576 Deep Learning Approaches for Accurate Detection of Epileptic Seizures from Electroencephalogram Data
Authors: Ramzi Rihane, Yassine Benayed
Abstract:
Epilepsy is a chronic neurological disorder characterized by recurrent, unprovoked seizures resulting from abnormal electrical activity in the brain. Timely and accurate detection of these seizures is essential for improving patient care. In this study, we leverage the UK Bonn University open-source EEG dataset and employ advanced deep-learning techniques to automate the detection of epileptic seizures. By extracting key features from both time and frequency domains, as well as Spectrogram features, we enhance the performance of various deep learning models. Our investigation includes architectures such as Long Short-Term Memory (LSTM), Bidirectional LSTM (Bi-LSTM), 1D Convolutional Neural Networks (1D-CNN), and hybrid CNN-LSTM and CNN-BiLSTM models. The models achieved impressive accuracies: LSTM (98.52%), Bi-LSTM (98.61%), CNN-LSTM (98.91%), CNN-BiLSTM (98.83%), and CNN (98.73%). Additionally, we utilized a data augmentation technique called SMOTE, which yielded the following results: CNN (97.36%), LSTM (97.01%), Bi-LSTM (97.23%), CNN-LSTM (97.45%), and CNN-BiLSTM (97.34%). These findings demonstrate the effectiveness of deep learning in capturing complex patterns in EEG signals, providing a reliable and scalable solution for real-time seizure detection in clinical environments.Keywords: electroencephalogram, epileptic seizure, deep learning, LSTM, CNN, BI-LSTM, seizure detection
Procedia PDF Downloads 14575 Application of a Hybrid QFD-FEA Methodology for Nigerian Garment Designs
Authors: Adepeju A. Opaleye, Adekunle Kolawole, Muyiwa A. Opaleye
Abstract:
Consumers’ perceived quality of imported product has been an impediment to business in the Nigeria garment industry. To improve patronage of made- in-Nigeria designs, the first step is to understand what the consumer expects, then proffer ways to meet this expectation through product redesign or improvement of the garment production process. The purpose of this study is to investigate drivers of consumers’ value for typical Nigerian garment design (NGD). An integrated quality function deployment (QFD) and functional, expressive and aesthetic (FEA) Consumer Needs methodology helps to minimize incorrect understanding of potential consumer’s requirements in mass customized garments. Six themes emerged as drivers of consumer’s satisfaction: (1) Style variety (2) Dimensions (3) Finishing (4) Fabric quality (5) Garment Durability and (6) Aesthetics. Existing designs found to lead foreign designs in terms of its acceptance for informal events, style variety and fit. The latter may be linked to its mode of acquisition. A conceptual model of NGD acceptance in the context of consumer’s inherent characteristics, social and the business environment is proposed.Keywords: Perceived quality, Garment design, Quality function deployment, FEA Model , Mass customisation
Procedia PDF Downloads 137574 A Fresh Approach to Learn Evidence-Based Practice, a Prospective Interventional Study
Authors: Ebtehal Qulisy, Geoffrey Dougherty, Kholoud Hothan, Mylene Dandavino
Abstract:
Background: For more than 200 years, journal clubs (JCs) have been used to teach the fundamentals of critical appraisal and evidence-based practice (EBP). However, JCs curricula face important challenges, including poor sustainability, insufficient time to prepare for and conduct the activities, and lack of trainee skills and self-efficacy with critical appraisal. Andragogy principles and modern technology could help EBP be taught in more relevant, modern, and interactive ways. Method: We propose a fresh educational activity to teach EBP. Educational sessions are designed to encourage collaborative and experiential learning and do not require advanced preparation by the participants. Each session lasts 60 minutes and is adaptable to in-person, virtual, or hybrid contexts. Sessions are structured around a worksheet and include three educational objectives: “1. Identify a Clinical Conundrum”, “2. Compare and Contrast Current Guidelines”, and “3. Choose a Recent Journal Article”. Sessions begin with a short presentation by a facilitator of a clinical scenario highlighting a “grey-zone” in pediatrics. Trainees are placed in groups of two to four (based on the participants’ number) of varied training levels. The first task requires the identification of a clinical conundrum (a situation where there is no clear answer but only a reasonable solution) related to the scenario. For the second task, trainees must identify two or three clinical guidelines. The last task requires trainees to find a journal article published in the last year that reports an update regarding the scenario’s topic. Participants are allowed to use their electronic devices throughout the session. Our university provides full-text access to major journals, which facilitated this exercise. Results: Participants were a convenience sample of trainees in the inpatient services at the Montréal Children’s Hospital, McGill University. Sessions were conducted as a part of an existing weekly academic activity and facilitated by pediatricians with experience in critical appraisal. There were 28 participants in 4 sessions held during Spring 2022. Time was allocated at the end of each session to collect participants’ feedback via a self-administered online survey. There were 22 responses, were 41%(n=9) pediatric residents, 22.7%(n=5) family medicine residents, 31.8%(n=7) medical students, and 4.5%(n=1) nurse practitioner. Four respondents participated in more than one session. The “Satisfied” rates were 94.7% for session format, 100% for topic selection, 89.5% for time allocation, and 84.3% for worksheet structure. 60% of participants felt that including the sessions during the clinical ward rotation was “Feasible.” As per self-efficacy, participants reported being “Confident” for the tasks as follows: 89.5% for the ability to identify a relevant conundrum, 94.8% for the compare and contrast task, and 84.2% for the identification of a published update. The perceived effectiveness to learn EBP was reported as “Agreed” by all participants. All participants would recommend this session for further teaching. Conclusion: We developed a modern approach to teach EBP, enjoyed by all levels of participants, who also felt it was a useful learning experience. Our approach addresses known JCs challenges by being relevant to clinical care, fostering active engagement but not requiring any preparation, using available technology, and being adaptable to hybrid contexts.Keywords: medical education, journal clubs, post-graduate teaching, andragogy, experiential learning, evidence-based practice
Procedia PDF Downloads 116573 Effect of Aging Time on CeO2 Nanoparticle Size Distribution Synthesized via Sol-Gel Method
Authors: Navid Zanganeh, Hafez Balavi, Farbod Sharif, Mahla Zabet, Marzieh Bakhtiary Noodeh
Abstract:
Cerium oxide (CeO2) also known as cerium dioxide or ceria is a pale yellow-white powder with various applications in the industry from wood coating to cosmetics, filtration, fuel cell electrolytes, gas sensors, hybrid solar cells and catalysts. In this research, attempts were made to synthesize and characterization of CeO2 nano-particles via sol-gel method. In addition, the effect of aging time on the size of particles was investigated. For this purpose, the aging times adjusted 48, 56, 64, and 72 min. The obtained particles were characterized by x-ray diffraction spectroscopy (XRD), scanning electron microscopy (SEM), transmitted electron microscopy (TEM), and Brunauer–Emmett–Teller (BET). As a result, XRD patterns confirmed the formation of CeO2 nanoparticles. SEM and TEM images illustrated the nano-particles with cluster shape, spherical and a nano-size range which was in agreement with XRD results. The finest particles (7.3 nm) was obtained at the optimum condition which was aging time of 48 min, calcination temperature at 400 ⁰C, and cerium concentration of 0.004 mol. Average specific surface area of the particles at optimum condition was measured by BET analysis and recorded as 47.57 m2/g.Keywords: aging time, CeO2 nanoparticles, size distribution, sol-gel
Procedia PDF Downloads 456572 Design and Performance Evaluation of Hybrid Corrugated-GFRP Infill Panels
Authors: Woo Young Jung, Sung Min Park, Ho Young Son, Viriyavudh Sim
Abstract:
This study presents a way to reduce earthquake damage and emergency rehabilitation of critical structures such as schools, high-tech factories, and hospitals due to strong ground motions associated with climate changes. Regarding recent trend, a strong earthquake causes serious damage to critical structures and then the critical structure might be influenced by sequence aftershocks (or tsunami) due to fault plane adjustments. Therefore, in order to improve seismic performance of critical structures, retrofitted or strengthening study of the structures under aftershocks sequence after emergency rehabilitation of the structures subjected to strong earthquakes is widely carried out. Consequently, this study used composite material for emergency rehabilitation of the structure rather than concrete and steel materials because of high strength and stiffness, lightweight, rapid manufacturing, and dynamic performance. Also, this study was to develop or improve the seismic performance or seismic retrofit of critical structures subjected to strong ground motions and earthquake aftershocks, by utilizing GFRP-Corrugated Infill Panels (GCIP).Keywords: aftershock, composite material, GFRP, infill panel
Procedia PDF Downloads 334571 Developing Fault Tolerance Metrics of Web and Mobile Applications
Authors: Ahmad Mohsin, Irfan Raza Naqvi, Syda Fatima Usamn
Abstract:
Applications with higher fault tolerance index are considered more reliable and trustworthy to drive quality. In recent years application development has been shifted from traditional desktop and web to native and hybrid application(s) for the web and mobile platforms. With the emergence of Internet of things IOTs, cloud and big data trends, the need for measuring Fault Tolerance for these complex nature applications has increased to evaluate their performance. There is a phenomenal gap between fault tolerance metrics development and measurement. Classic quality metric models focused on metrics for traditional systems ignoring the essence of today’s applications software, hardware & deployment characteristics. In this paper, we have proposed simple metrics to measure fault tolerance considering general requirements for Web and Mobile Applications. We have aligned factors – subfactors, using GQM for metrics development considering the nature of mobile we apps. Systematic Mathematical formulation is done to measure metrics quantitatively. Three web mobile applications are selected to measure Fault Tolerance factors using formulated metrics. Applications are then analysed on the basis of results from observations in a controlled environment on different mobile devices. Quantitative results are presented depicting Fault tolerance in respective applications.Keywords: web and mobile applications, reliability, fault tolerance metric, quality metrics, GQM based metrics
Procedia PDF Downloads 344570 Hybridized Simulated Annealing with Chemical Reaction Optimization for Solving to Sequence Alignment Problem
Authors: Ernesto Linan, Linda Cruz, Lucero Becerra
Abstract:
In this paper, a new hybridized algorithm based on Chemical Reaction Optimization and Simulated Annealing is proposed to solve the alignment sequence Problem. The Chemical Reaction Optimization is a population-based meta-heuristic algorithm based on the principles of a chemical reaction. Simulated Annealing is applied to solve a large number of combinatorial optimization problems of general-purpose. In this paper, we propose hybridization between Chemical Reaction Optimization algorithm and Simulated Annealing in order to solve the Sequence Alignment Problem. An initial population of molecules is defined at beginning of the proposed algorithm, where each molecule represents a sequence alignment problem. In order to simulate inter-molecule collisions, the process of Chemical Reaction is placed inside the Metropolis Cycle at certain values of temperature. Inside this cycle, change of molecules is done due to collisions; some molecules are accepted by applying Boltzmann probability. The results with the hybrid scheme are better than the results obtained separately.Keywords: chemical reaction optimization, sequence alignment problem, simulated annealing algorithm, metaheuristics
Procedia PDF Downloads 211569 Human Performance Evaluating of Advanced Cardiac Life Support Procedure Using Fault Tree and Bayesian Network
Authors: Shokoufeh Abrisham, Seyed Mahmoud Hossieni, Elham Pishbin
Abstract:
In this paper, a hybrid method based on the fault tree analysis (FTA) and Bayesian networks (BNs) are employed to evaluate the team performance quality of advanced cardiac life support (ACLS) procedures in emergency department. According to American Heart Association (AHA) guidelines, a category relying on staff action leading to clinical incidents and also some discussions with emergency medicine experts, a fault tree model for ACLS procedure is obtained based on the human performance. The obtained FTA model is converted into BNs, and some different scenarios are defined to demonstrate the efficiency and flexibility of the presented model of BNs. Also, a sensitivity analysis is conducted to indicate the effects of team leader presence and uncertainty knowledge of experts on the quality of ACLS. The proposed model based on BNs shows that how the results of risk analysis can be closed to reality comparing to the obtained results based on only FTA in medical procedures.Keywords: advanced cardiac life support, fault tree analysis, Bayesian belief networks, numan performance, healthcare systems
Procedia PDF Downloads 147568 Structured Cross System Planning and Control in Modular Production Systems by Using Agent-Based Control Loops
Authors: Simon Komesker, Achim Wagner, Martin Ruskowski
Abstract:
In times of volatile markets with fluctuating demand and the uncertainty of global supply chains, flexible production systems are the key to an efficient implementation of a desired production program. In this publication, the authors present a holistic information concept taking into account various influencing factors for operating towards the global optimum. Therefore, a strategy for the implementation of multi-level planning for a flexible, reconfigurable production system with an alternative production concept in the automotive industry is developed. The main contribution of this work is a system structure mixing central and decentral planning and control evaluated in a simulation framework. The information system structure in current production systems in the automotive industry is rigidly hierarchically organized in monolithic systems. The production program is created rule-based with the premise of achieving uniform cycle time. This program then provides the information basis for execution in subsystems at the station and process execution level. In today's era of mixed-(car-)model factories, complex conditions and conflicts arise in achieving logistics, quality, and production goals. There is no provision for feedback loops of results from the process execution level (resources) and process supporting (quality and logistics) systems and reconsideration in the planning systems. To enable a robust production flow, the complexity of production system control is artificially reduced by the line structure and results, for example in material-intensive processes (buffers and safety stocks - two container principle also for different variants). The limited degrees of freedom of line production have produced the principle of progress figure control, which results in one-time sequencing, sequential order release, and relatively inflexible capacity control. As a result, modularly structured production systems such as modular production according to known approaches with more degrees of freedom are currently difficult to represent in terms of information technology. The remedy is an information concept that supports cross-system and cross-level information processing for centralized and decentralized decision-making. Through an architecture of hierarchically organized but decoupled subsystems, the paradigm of hybrid control is used, and a holonic manufacturing system is offered, which enables flexible information provisioning and processing support. In this way, the influences from quality, logistics, and production processes can be linked holistically with the advantages of mixed centralized and decentralized planning and control. Modular production systems also require modularly networked information systems with semi-autonomous optimization for a robust production flow. Dynamic prioritization of different key figures between subsystems should lead the production system to an overall optimum. The tasks and goals of quality, logistics, process, resource, and product areas in a cyber-physical production system are designed as an interconnected multi-agent-system. The result is an alternative system structure that executes centralized process planning and decentralized processing. An agent-based manufacturing control is used to enable different flexibility and reconfigurability states and manufacturing strategies in order to find optimal partial solutions of subsystems, that lead to a near global optimum for hybrid planning. This allows a robust near to plan execution with integrated quality control and intralogistics.Keywords: holonic manufacturing system, modular production system, planning, and control, system structure
Procedia PDF Downloads 169567 Daylightophil Approach towards High-Performance Architecture for Hybrid-Optimization of Visual Comfort and Daylight Factor in BSk
Authors: Mohammadjavad Mahdavinejad, Hadi Yazdi
Abstract:
The greatest influence we have from the world is shaped through the visual form, thus light is an inseparable element in human life. The use of daylight in visual perception and environment readability is an important issue for users. With regard to the hazards of greenhouse gas emissions from fossil fuels, and in line with the attitudes on the reduction of energy consumption, the correct use of daylight results in lower levels of energy consumed by artificial lighting, heating and cooling systems. Windows are usually the starting points for analysis and simulations to achieve visual comfort and energy optimization; therefore, attention should be paid to the orientation of buildings to minimize electrical energy and maximize the use of daylight. In this paper, by using the Design Builder Software, the effect of the orientation of an 18m2(3m*6m) room with 3m height in city of Tehran has been investigated considering the design constraint limitations. In these simulations, the dimensions of the building have been changed with one degree and the window is located on the smaller face (3m*3m) of the building with 80% ratio. The results indicate that the orientation of building has a lot to do with energy efficiency to meet high-performance architecture and planning goals and objectives.Keywords: daylight, window, orientation, energy consumption, design builder
Procedia PDF Downloads 233566 Geographic Information Systems and a Breath of Opportunities for Supply Chain Management: Results from a Systematic Literature Review
Authors: Anastasia Tsakiridi
Abstract:
Geographic information systems (GIS) have been utilized in numerous spatial problems, such as site research, land suitability, and demographic analysis. Besides, GIS has been applied in scientific fields like geography, health, and economics. In business studies, GIS has been used to provide insights and spatial perspectives in demographic trends, spending indicators, and network analysis. To date, the information regarding the available usages of GIS in supply chain management (SCM) and how these analyses can benefit businesses is limited. A systematic literature review (SLR) of the last 5-year peer-reviewed academic literature was conducted, aiming to explore the existing usages of GIS in SCM. The searches were performed in 3 databases (Web of Science, ProQuest, and Business Source Premier) and reported using the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) methodology. The analysis resulted in 79 papers. The results indicate that the existing GIS applications used in SCM were in the following domains: a) network/ transportation analysis (in 53 of the papers), b) location – allocation site search/ selection (multiple-criteria decision analysis) (in 45 papers), c) spatial analysis (demographic or physical) (in 34 papers), d) combination of GIS and supply chain/network optimization tools (in 32 papers), and e) visualization/ monitoring or building information modeling applications (in 8 papers). An additional categorization of the literature was conducted by examining the usage of GIS in the supply chain (SC) by the business sectors, as indicated by the volume of the papers. The results showed that GIS is mainly being applied in the SC of the biomass biofuel/wood industry (33 papers). Other industries that are currently utilizing GIS in their SC were the logistics industry (22 papers), the humanitarian/emergency/health care sector (10 papers), the food/agro-industry sector (5 papers), the petroleum/ coal/ shale gas sector (3 papers), the faecal sludge sector (2 papers), the recycle and product footprint industry (2 papers), and the construction sector (2 papers). The results were also presented by the geography of the included studies and the GIS software used to provide critical business insights and suggestions for future research. The results showed that research case studies of GIS in SCM were conducted in 26 countries (mainly in the USA) and that the most prominent GIS software provider was the Environmental Systems Research Institute’s ArcGIS (in 51 of the papers). This study is a systematic literature review of the usage of GIS in SCM. The results showed that the GIS capabilities could offer substantial benefits in SCM decision-making by providing key insights to cost minimization, supplier selection, facility location, SC network configuration, and asset management. However, as presented in the results, only eight industries/sectors are currently using GIS in their SCM activities. These findings may offer essential tools to SC managers who seek to optimize the SC activities and/or minimize logistic costs and to consultants and business owners that want to make strategic SC decisions. Furthermore, the findings may be of interest to researchers aiming to investigate unexplored research areas where GIS may improve SCM.Keywords: supply chain management, logistics, systematic literature review, GIS
Procedia PDF Downloads 142565 Designing Sustainable Building Based on Iranian's Windmills
Authors: Negar Sartipzadeh
Abstract:
Energy-conscious design, which coordinates with the Earth ecological systems during its life cycle, has the least negative impact on the environment with the least waste of resources. Due to the increasing in world population as well as the consumption of fossil fuels that cause the production of greenhouse gasses and environmental pollution, mankind is looking for renewable and also sustainable energies. The Iranian native construction is a clear evidence of energy-aware designing. Our predecessors were forced to rely on the natural resources and sustainable energies as well as environmental issues which have been being considered in the recent world. One of these endless energies is wind energy. Iranian traditional architecture foundations is a appropriate model in solving the environmental crisis and the contemporary energy. What will come in this paper is an effort to recognition and introduction of the unique characteristics of the Iranian architecture in the application of aerodynamic and hydraulic energies derived from the wind, which are the most common and major type of using sustainable energies in the traditional architecture of Iran. Therefore, the recent research attempts to offer a hybrid system suggestions for application in new constructions designing in a region such as Nashtifan, which has potential through reviewing windmills and how they deal with sustainable energy sources, as a model of Iranian native construction.Keywords: renewable energy, sustainable building, windmill, Iranian architecture
Procedia PDF Downloads 422564 Analyzing the Factors that Cause Parallel Performance Degradation in Parallel Graph-Based Computations Using Graph500
Authors: Mustafa Elfituri, Jonathan Cook
Abstract:
Recently, graph-based computations have become more important in large-scale scientific computing as they can provide a methodology to model many types of relations between independent objects. They are being actively used in fields as varied as biology, social networks, cybersecurity, and computer networks. At the same time, graph problems have some properties such as irregularity and poor locality that make their performance different than regular applications performance. Therefore, parallelizing graph algorithms is a hard and challenging task. Initial evidence is that standard computer architectures do not perform very well on graph algorithms. Little is known exactly what causes this. The Graph500 benchmark is a representative application for parallel graph-based computations, which have highly irregular data access and are driven more by traversing connected data than by computation. In this paper, we present results from analyzing the performance of various example implementations of Graph500, including a shared memory (OpenMP) version, a distributed (MPI) version, and a hybrid version. We measured and analyzed all the factors that affect its performance in order to identify possible changes that would improve its performance. Results are discussed in relation to what factors contribute to performance degradation.Keywords: graph computation, graph500 benchmark, parallel architectures, parallel programming, workload characterization.
Procedia PDF Downloads 147563 Studies on Mechanical Behavior of Kevlar/Kenaf/Graphene Reinforced Polymer Based Hybrid Composites
Authors: H. K. Shivanand, Ranjith R. Hombal, Paraveej Shirahatti, Gujjalla Anil Babu, S. ShivaPrakash
Abstract:
When it comes to the selection of materials the knowledge of materials science plays a vital role in selection and enhancements of materials properties. In the world of material science a composite material has the significant role based on its application. The composite materials are those in which two or more components having different physical and chemical properties are combined to create a new enhanced property substance. In this study three different materials (Kenaf, Kevlar and Graphene) been chosen based on their properties and a composite material is developed with help of vacuum bagging process. The fibers (Kenaf and Kevlar) and Resin(vinyl ester) ratio was maintained at 70:30 during the process and 0.5% 1% and 1.5% of Graphene was added during fabrication process. The material was machined to thedimension ofASTM standards(300×300mm and thickness 3mm)with help of water jet cutting machine. The composite materials were tested for Mechanical properties such as Interlaminar shear strength(ILSS) and Flexural strength. It is found that there is significant increase in material properties in the developed composite material.Keywords: Kevlar, Kenaf, graphene, vacuum bagging process, Interlaminar shear strength test, flexural test
Procedia PDF Downloads 93562 A New Fuzzy Fractional Order Model of Transmission of Covid-19 With Quarantine Class
Authors: Asma Hanif, A. I. K. Butt, Shabir Ahmad, Rahim Ud Din, Mustafa Inc
Abstract:
This paper is devoted to a study of the fuzzy fractional mathematical model reviewing the transmission dynamics of the infectious disease Covid-19. The proposed dynamical model consists of susceptible, exposed, symptomatic, asymptomatic, quarantine, hospitalized and recovered compartments. In this study, we deal with the fuzzy fractional model defined in Caputo’s sense. We show the positivity of state variables that all the state variables that represent different compartments of the model are positive. Using Gronwall inequality, we show that the solution of the model is bounded. Using the notion of the next-generation matrix, we find the basic reproduction number of the model. We demonstrate the local and global stability of the equilibrium point by using the concept of Castillo-Chavez and Lyapunov theory with the Lasalle invariant principle, respectively. We present the results that reveal the existence and uniqueness of the solution of the considered model through the fixed point theorem of Schauder and Banach. Using the fuzzy hybrid Laplace method, we acquire the approximate solution of the proposed model. The results are graphically presented via MATLAB-17.Keywords: Caputo fractional derivative, existence and uniqueness, gronwall inequality, Lyapunov theory
Procedia PDF Downloads 106561 Tea Club (Singapore)-Learning to Navigate the Social World without Fear: Adapted from PEERS® for Young Adults
Authors: Janice Cheong, Tan Seying
Abstract:
The growing years in adolescence are often a tumultuous time for both the individual and family; this is especially so for individuals with Autism Spectrum Disorder (ASD) and Social Communication Disorder (SCD). Tea Club, which is adapted from the PEERS® for Young Adults, seeks to address some of the social challenges faced by Singaporean adolescents with ASD/SCD while navigating social situations. Tea club (hybrid) consists of face-to-face sessions and virtual sessions. These sessions work with both the adolescent and their parents to tackle the individual's difficulties with social skills, empathy, and loneliness. Prior to the group intervention, both participants and their parents scored on the Test of Adolescent Social Skills Knowledge (TASSK) and Autism Spectrum Quotient (AQ), respectively. The session was spread across four months. At the end of the group based intervention, participants’ and parents’ scores were collected again and compared. Inputs on the programme and participant’s confidence in socialization were also gathered from both participants and their parents and looked at thematically. The findings highlight some of the challenges faced by teens with ASD in Singapore and the benefits of the intervention. Parental sentiments are also examined and discussed.Keywords: adolescence autism, group intervention, social communication disorder, social skills
Procedia PDF Downloads 145560 Channel Estimation Using Deep Learning for Reconfigurable Intelligent Surfaces-Assisted Millimeter Wave Systems
Authors: Ting Gao, Mingyue He
Abstract:
Reconfigurable intelligent surfaces (RISs) are expected to be an important part of next-generation wireless communication networks due to their potential to reduce the hardware cost and energy consumption of millimeter Wave (mmWave) massive multiple-input multiple-output (MIMO) technology. However, owing to the lack of signal processing abilities of the RIS, the perfect channel state information (CSI) in RIS-assisted communication systems is difficult to acquire. In this paper, the uplink channel estimation for mmWave systems with a hybrid active/passive RIS architecture is studied. Specifically, a deep learning-based estimation scheme is proposed to estimate the channel between the RIS and the user. In particular, the sparse structure of the mmWave channel is exploited to formulate the channel estimation as a sparse reconstruction problem. To this end, the proposed approach is derived to obtain the distribution of non-zero entries in a sparse channel. After that, the channel is reconstructed by utilizing the least-squares (LS) algorithm and compressed sensing (CS) theory. The simulation results demonstrate that the proposed channel estimation scheme is superior to existing solutions even in low signal-to-noise ratio (SNR) environments.Keywords: channel estimation, reconfigurable intelligent surface, wireless communication, deep learning
Procedia PDF Downloads 151559 Assessing of Social Comfort of the Russian Population with Big Data
Authors: Marina Shakleina, Konstantin Shaklein, Stanislav Yakiro
Abstract:
The digitalization of modern human life over the last decade has facilitated the acquisition, storage, and processing of data, which are used to detect changes in consumer preferences and to improve the internal efficiency of the production process. This emerging trend has attracted academic interest in the use of big data in research. The study focuses on modeling the social comfort of the Russian population for the period 2010-2021 using big data. Big data provides enormous opportunities for understanding human interactions at the scale of society with plenty of space and time dynamics. One of the most popular big data sources is Google Trends. The methodology for assessing social comfort using big data involves several steps: 1. 574 words were selected based on the Harvard IV-4 Dictionary adjusted to fit the reality of everyday Russian life. The set of keywords was further cleansed by excluding queries consisting of verbs and words with several lexical meanings. 2. Search queries were processed to ensure comparability of results: the transformation of data to a 10-point scale, elimination of popularity peaks, detrending, and deseasoning. The proposed methodology for keyword search and Google Trends processing was implemented in the form of a script in the Python programming language. 3. Block and summary integral indicators of social comfort were constructed using the first modified principal component resulting in weighting coefficients values of block components. According to the study, social comfort is described by 12 blocks: ‘health’, ‘education’, ‘social support’, ‘financial situation’, ‘employment’, ‘housing’, ‘ethical norms’, ‘security’, ‘political stability’, ‘leisure’, ‘environment’, ‘infrastructure’. According to the model, the summary integral indicator increased by 54% and was 4.631 points; the average annual rate was 3.6%, which is higher than the rate of economic growth by 2.7 p.p. The value of the indicator describing social comfort in Russia is determined by 26% by ‘social support’, 24% by ‘education’, 12% by ‘infrastructure’, 10% by ‘leisure’, and the remaining 28% by others. Among 25% of the most popular searches, 85% are of negative nature and are mainly related to the blocks ‘security’, ‘political stability’, ‘health’, for example, ‘crime rate’, ‘vulnerability’. Among the 25% most unpopular queries, 99% of the queries were positive and mostly related to the blocks ‘ethical norms’, ‘education’, ‘employment’, for example, ‘social package’, ‘recycling’. In conclusion, the introduction of the latent category ‘social comfort’ into the scientific vocabulary deepens the theory of the quality of life of the population in terms of the study of the involvement of an individual in the society and expanding the subjective aspect of the measurements of various indicators. Integral assessment of social comfort demonstrates the overall picture of the development of the phenomenon over time and space and quantitatively evaluates ongoing socio-economic policy. The application of big data in the assessment of latent categories gives stable results, which opens up possibilities for their practical implementation.Keywords: big data, Google trends, integral indicator, social comfort
Procedia PDF Downloads 200558 Wireless Sensor Network Energy Efficient and QoS-Aware MAC Protocols: A Survey
Authors: Bashir Abdu Muzakkari, Mohamad Afendee Mohamad, Mohd Fadzil Abdul Kadir
Abstract:
Wireless Sensor Networks (WSNs) is an aggregation of several tiny, low-cost sensor nodes, spatially distributed to monitor physical or environmental status. WSN is constantly changing because of the rapid technological advancements in sensor elements such as radio, battery and operating systems. The Medium Access Control (MAC) protocols remain very vital in the WSN because of its role in coordinating communication amongst the sensors. Other than battery consumption, packet collision, network lifetime and latency are factors that largely depend on WSN MAC protocol and these factors have been widely treated in recent days. In this paper, we survey some latest proposed WSN Contention-based, Scheduling-based and Hybrid MAC protocols while presenting an examination, correlation of advantages and limitations of each protocol. Concentration is directed towards investigating the treatment of Quality of Service (QoS) performance metrics within these particular protocols. The result shows that majority of the protocols leaned towards energy conservation. We, therefore, believe that other performance metrics of guaranteed QoS such as latency, throughput, packet loss, network and bandwidth availability may play a critical role in the design of future MAC protocols for WSNs.Keywords: WSN, QoS, energy consumption, MAC protocol
Procedia PDF Downloads 400557 Climate Adaptive Building Shells for Plus-Energy-Buildings, Designed on Bionic Principles
Authors: Andreas Hammer
Abstract:
Six peculiar architecture designs from the Frankfurt University will be discussed within this paper and their future potential of the adaptable and solar thin-film sheets implemented facades will be shown acting and reacting on climate/solar changes of their specific sites. The different aspects, as well as limitations with regard to technical and functional restrictions, will be named. The design process for a “multi-purpose building”, a “high-rise building refurbishment” and a “biker’s lodge” on the river Rheine valley, has been critically outlined and developed step by step from an international studentship towards an overall energy strategy, that firstly had to push the design to a plus-energy building and secondly had to incorporate bionic aspects into the building skins design. Both main parameters needed to be reviewed and refined during the whole design process. Various basic bionic approaches have been given [e.g. solar ivyᵀᴹ, flectofinᵀᴹ or hygroskinᵀᴹ, which were to experiment with, regarding the use of bendable photovoltaic thin film elements being parts of a hybrid, kinetic façade system.Keywords: bionic and bioclimatic design, climate adaptive building shells [CABS], energy-strategy, harvesting façade, high-efficiency building skin, photovoltaic in building skins, plus-energy-buildings, solar gain, sustainable building concept
Procedia PDF Downloads 430556 Cloning of Strawberry’s Malonyltransferase Genes and Characterisation of Their Enzymes
Authors: Xiran Wang, Johanna Trinkl, Thomas Hoffmann, Wilfried Schwab
Abstract:
Malonyltransferases (MATs) are enzymes that play a key role in the biosynthesis of secondary metabolites in plants, such as flavonoids and anthocyanins. As a kind of flavonoid-rich fruit, strawberries are an ideal model to study MATs. From Goodberry metabolome data, in the hybrid generation of 2 strawberries various, Fragaria × ananassa cv. 'Senga Sengana' and 'Candonga', we found the malonylated flavonoid concentration is significantly higher in 'Senga Sengana' compared with 'Candonga'. Therefore, we aimed to identify and characterize the malonyltransferases responsible for the different malonylated flavonoid concentrations in two different strawberry cultivars. In this study, we have found 6 MATs via genome mapping, metabolome analysis, gene cloning, and enzyme assay from strawberries, which catalyzed the malonylation of flavonoid substrates: quercetin-3-glucoside, kaempferol-3-glucoside, pelargonidin-3-glucoside, and cyanidin-3-glucoside. All four compounds reacted with FaMATs to varying degrees. These MATs have important implication into strawberries’ flavonoid biosynthesis, and also provide insights into insights into flavonoid biosynthesis, potential applications in agriculture, plant science, and pharmacy, and information on the regulation of secondary metabolism in plants.Keywords: malonyltransferase, strawberry, flavonoid biosynthesis, enzyme assay
Procedia PDF Downloads 135555 High Order Block Implicit Multi-Step (Hobim) Methods for the Solution of Stiff Ordinary Differential Equations
Authors: J. P. Chollom, G. M. Kumleng, S. Longwap
Abstract:
The search for higher order A-stable linear multi-step methods has been the interest of many numerical analysts and has been realized through either higher derivatives of the solution or by inserting additional off step points, supper future points and the likes. These methods are suitable for the solution of stiff differential equations which exhibit characteristics that place a severe restriction on the choice of step size. It becomes necessary that only methods with large regions of absolute stability remain suitable for such equations. In this paper, high order block implicit multi-step methods of the hybrid form up to order twelve have been constructed using the multi-step collocation approach by inserting one or more off step points in the multi-step method. The accuracy and stability properties of the new methods are investigated and are shown to yield A-stable methods, a property desirable of methods suitable for the solution of stiff ODE’s. The new High Order Block Implicit Multistep methods used as block integrators are tested on stiff differential systems and the results reveal that the new methods are efficient and compete favourably with the state of the art Matlab ode23 code.Keywords: block linear multistep methods, high order, implicit, stiff differential equations
Procedia PDF Downloads 358554 Development of Biosensor Chip for Detection of Specific Antibodies to HSV-1
Authors: Zatovska T. V., Nesterova N. V., Baranova G. V., Zagorodnya S. D.
Abstract:
In recent years, biosensor technologies based on the phenomenon of surface plasmon resonance (SPR) are becoming increasingly used in biology and medicine. Their application facilitates exploration in real time progress of binding of biomolecules and identification of agents that specifically interact with biologically active substances immobilized on the biosensor surface (biochips). Special attention is paid to the use of Biosensor analysis in determining the antibody-antigen interaction in the diagnostics of diseases caused by viruses and bacteria. According to WHO, the diseases that are caused by the herpes simplex virus (HSV), take second place (15.8%) after influenza as a cause of death from viral infections. Current diagnostics of HSV infection include PCR and ELISA assays. The latter allows determination the degree of immune response to viral infection and respective stages of its progress. In this regard, the searches for new and available diagnostic methods are very important. This work was aimed to develop Biosensor chip for detection of specific antibodies to HSV-1 in the human blood serum. The proteins of HSV1 (strain US) were used as antigens. The viral particles were accumulated in cell culture MDBK and purified by differential centrifugation in cesium chloride density gradient. Analysis of the HSV1 proteins was performed by polyacrylamide gel electrophoresis and ELISA. The protein concentration was measured using De Novix DS-11 spectrophotometer. The device for detection of antigen-antibody interactions was an optoelectronic two-channel spectrometer ‘Plasmon-6’, using the SPR phenomenon in the Krechman optical configuration. It was developed at the Lashkarev Institute of Semiconductor Physics of NASU. The used carrier was a glass plate covered with 45 nm gold film. Screening of human blood serums was performed using the test system ‘HSV-1 IgG ELISA’ (GenWay, USA). Development of Biosensor chip included optimization of conditions of viral antigen sorption and analysis steps. For immobilization of viral proteins 0.2% solution of Dextran 17, 200 (Sigma, USA) was used. Sorption of antigen took place at 4-8°C within 18-24 hours. After washing of chip, three times with citrate buffer (pH 5,0) 1% solution of BSA was applied to block the sites not occupied by viral antigen. It was found direct dependence between the amount of immobilized HSV1 antigen and SPR response. Using obtained biochips, panels of 25 positive and 10 negative for the content of antibodies to HSV-1 human sera were analyzed. The average value of SPR response was 185 a.s. for negative sera and from 312 to. 1264 a.s. for positive sera. It was shown that SPR data were agreed with ELISA results in 96% of samples proving the great potential of SPR in such researches. It was investigated the possibility of biochip regeneration and it was shown that application of 10 mM NaOH solution leads to rupture of intermolecular bonds. This allows reuse the chip several times. Thus, in this study biosensor chip for detection of specific antibodies to HSV1 was successfully developed expanding a range of diagnostic methods for this pathogen.Keywords: biochip, herpes virus, SPR
Procedia PDF Downloads 417553 The Methods of Immobilization of Laccase for Direct Transfer in an Enzymatic Fuel Cell
Authors: Afshin Farahbakhsh, Hoda Khodadadi
Abstract:
In this paper, we compare five methods of biological fuel cell fabrication by combining a Shewanella oneidensis microbial anode and a laccase-modified air-breathing cathode. As a result of biofuel cell laccase with graphite nanofibers, carbon surface (PAMAN) on the pt/hpg electrode, graphite sheets MWCNT and with (PG) and (MWCNT) showed, respectively. Describes methods for creating controllable and reproducible bio-anodes and demonstrates the versatility of hybrid biological fuel cells. The laccase-based biocathodes prepared either with the crude extract or with the purified enzyme can provide electrochemically active and stable biomaterials. The laccase-based biocathodes prepared either with the crude extract or with the purified enzyme can provide electrochemically active and stable biomaterials. When the device was fed with transdermal extracts, containing only 30μM of glucose, the average peak power was proportionally lower (0.004mW). The result of biofuel cell with graphite nanofibers showed the enzymatic fuel cell reaches 0.5 V at open circuit voltage with both, ethanol and methanol and the maximum current density observed for E2electrode was 228.94mAcm.Keywords: enzymatic electrode, fuel cell, immobilization, laccase
Procedia PDF Downloads 262552 The Motivating and Limiting Factors of Learners’ Engagement in an Online Discussion Forum
Authors: K. Durairaj, I. N. Umar
Abstract:
Lately, asynchronous discussion forum is integrated in higher educational institutions as it may increase learning process, learners’ understanding, achievement and knowledge construction. Asynchronous discussion forum is used to complement the traditional, face-to-face learning session in hybrid learning courses. However, studies have proven that students’ engagement in online forum are still unconvincing. Thus, the aim of this study is to investigate the motivating factors and obstacles that affect the learners’ engagement in asynchronous discussion forum. This study is carried out in one of the public higher educational institutions in Malaysia with 18 postgraduate students as samples. The authors have developed a 40-items questionnaire based on literature review. The results indicate several factors that have encouraged or limited students’ engagement in asynchronous discussion forum: (a) the practices or behaviors of peers, or instructors, (b) the needs for the discussions, (c) the learners’ personalities, (d) constraints in continuing the discussion forum, (e) lack of ideas, (f) the level of thoughts, (g) the level of knowledge construction, (h) technical problems, (i) time constraints and (j) misunderstanding. This study suggests some recommendations to increase the students’ engagement in online forums. Finally, based upon the findings, some implications are proposed for further research.Keywords: asynchronous discussion forum, engagement, factors, motivating, limiting
Procedia PDF Downloads 328551 Mondoc: Informal Lightweight Ontology for Faceted Semantic Classification of Hypernymy
Authors: M. Regina Carreira-Lopez
Abstract:
Lightweight ontologies seek to concrete union relationships between a parent node, and a secondary node, also called "child node". This logic relation (L) can be formally defined as a triple ontological relation (LO) equivalent to LO in ⟨LN, LE, LC⟩, and where LN represents a finite set of nodes (N); LE is a set of entities (E), each of which represents a relationship between nodes to form a rooted tree of ⟨LN, LE⟩; and LC is a finite set of concepts (C), encoded in a formal language (FL). Mondoc enables more refined searches on semantic and classified facets for retrieving specialized knowledge about Atlantic migrations, from the Declaration of Independence of the United States of America (1776) and to the end of the Spanish Civil War (1939). The model looks forward to increasing documentary relevance by applying an inverse frequency of co-ocurrent hypernymy phenomena for a concrete dataset of textual corpora, with RMySQL package. Mondoc profiles archival utilities implementing SQL programming code, and allows data export to XML schemas, for achieving semantic and faceted analysis of speech by analyzing keywords in context (KWIC). The methodology applies random and unrestricted sampling techniques with RMySQL to verify the resonance phenomena of inverse documentary relevance between the number of co-occurrences of the same term (t) in more than two documents of a set of texts (D). Secondly, the research also evidences co-associations between (t) and their corresponding synonyms and antonyms (synsets) are also inverse. The results from grouping facets or polysemic words with synsets in more than two textual corpora within their syntagmatic context (nouns, verbs, adjectives, etc.) state how to proceed with semantic indexing of hypernymy phenomena for subject-heading lists and for authority lists for documentary and archival purposes. Mondoc contributes to the development of web directories and seems to achieve a proper and more selective search of e-documents (classification ontology). It can also foster on-line catalogs production for semantic authorities, or concepts, through XML schemas, because its applications could be used for implementing data models, by a prior adaptation of the based-ontology to structured meta-languages, such as OWL, RDF (descriptive ontology). Mondoc serves to the classification of concepts and applies a semantic indexing approach of facets. It enables information retrieval, as well as quantitative and qualitative data interpretation. The model reproduces a triple tuple ⟨LN, LE, LT, LCF L, BKF⟩ where LN is a set of entities that connect with other nodes to concrete a rooted tree in ⟨LN, LE⟩. LT specifies a set of terms, and LCF acts as a finite set of concepts, encoded in a formal language, L. Mondoc only resolves partial problems of linguistic ambiguity (in case of synonymy and antonymy), but neither the pragmatic dimension of natural language nor the cognitive perspective is addressed. To achieve this goal, forthcoming programming developments should target at oriented meta-languages with structured documents in XML.Keywords: hypernymy, information retrieval, lightweight ontology, resonance
Procedia PDF Downloads 125550 A Distinct Method Based on Mamba-Unet for Brain Tumor Image Segmentation
Authors: Djallel Bouamama, Yasser R. Haddadi
Abstract:
Accurate brain tumor segmentation is crucial for diagnosis and treatment planning, yet it remains a challenging task due to the variability in tumor shapes and intensities. This paper introduces a distinct approach to brain tumor image segmentation by leveraging an advanced architecture known as Mamba-Unet. Building on the well-established U-Net framework, Mamba-Unet incorporates distinct design enhancements to improve segmentation performance. Our proposed method integrates a multi-scale attention mechanism and a hybrid loss function to effectively capture fine-grained details and contextual information in brain MRI scans. We demonstrate that Mamba-Unet significantly enhances segmentation accuracy compared to conventional U-Net models by utilizing a comprehensive dataset of annotated brain MRI scans. Quantitative evaluations reveal that Mamba-Unet surpasses traditional U-Net architectures and other contemporary segmentation models regarding Dice coefficient, sensitivity, and specificity. The improvements are attributed to the method's ability to manage class imbalance better and resolve complex tumor boundaries. This work advances the state-of-the-art in brain tumor segmentation and holds promise for improving clinical workflows and patient outcomes through more precise and reliable tumor detection.Keywords: brain tumor classification, image segmentation, CNN, U-NET
Procedia PDF Downloads 34549 Neuroanatomical Specificity in Reporting & Diagnosing Neurolinguistic Disorders: A Functional & Ethical Primer
Authors: Ruairi J. McMillan
Abstract:
Introduction: This critical analysis aims to ascertain how well neuroanatomical aetiologies are communicated within 20 case reports of aphasia. Neuroanatomical visualisations based on dissected brain specimens were produced and combined with white matter tract and vascular taxonomies of function in order to address the most consistently underreported features found within the aphasic case study reports. Together, these approaches are intended to integrate aphasiological knowledge from the past 20 years with aphasiological diagnostics, and to act as prototypal resources for both researchers and clinical professionals. The medico-legal precedent for aphasia diagnostics under Canadian, US and UK case law and the neuroimaging/neurological diagnostics relative to the functional capacity of aphasic patients are discussed in relation to the major findings of the literary analysis, neuroimaging protocols in clinical use today, and the neuroanatomical aetiologies of different aphasias. Basic Methodology: Literature searches of relevant scientific databases (e.g, OVID medline) were carried out using search terms such as aphasia case study (year) & stroke induced aphasia case study. A series of 7 diagnostic reporting criteria were formulated, and the resulting case studies were scored / 7 alongside clinical stroke criteria. In order to focus on the diagnostic assessment of the patient’s condition, only the case report proper (not the discussion) was used to quantify results. Statistical testing established if specific reporting criteria were associated with higher overall scores and potentially inferable increases in quality of reporting. Statistical testing of whether criteria scores were associated with an unclear/adjusted diagnosis were also tested, as well as the probability of a given criterion deviating from an expected estimate. Major Findings: The quantitative analysis of neuroanatomically driven diagnostics in case studies of aphasia revealed particularly low scores in the connection of neuroanatomical functions to aphasiological assessment (10%), and in the inclusion of white matter tracts within neuroimaging or assessment diagnostics (30%). Case studies which included clinical mention of white matter tracts within the report itself were distributed among higher scoring cases, as were case studies which (as clinically indicated) related the affected vascular region to the brain parenchyma of the language network. Concluding Statement: These findings indicate that certain neuroanatomical functions are integrated less often within the patient report than others, despite a precedent for well-integrated neuroanatomical aphasiology also being found among the case studies sampled, and despite these functions being clinically essential in diagnostic neuroimaging and aphasiological assessment. Therefore, ultimately the integration and specificity of aetiological neuroanatomy may contribute positively to the capacity and autonomy of aphasic patients as well as their clinicians. The integration of a full aetiological neuroanatomy within the reporting of aphasias may improve patient outcomes and sustain autonomy in the event of medico-ethical investigation.Keywords: aphasia, language network, functional neuroanatomy, aphasiological diagnostics, medico-legal ethics
Procedia PDF Downloads 67548 Imp_hist-Si: Improved Hybrid Image Segmentation Technique for Satellite Imagery to Decrease the Segmentation Error Rate
Authors: Neetu Manocha
Abstract:
Image segmentation is a technique where a picture is parted into distinct parts having similar features which have a place with similar items. Various segmentation strategies have been proposed as of late by prominent analysts. But, after ultimate thorough research, the novelists have analyzed that generally, the old methods do not decrease the segmentation error rate. Then author finds the technique HIST-SI to decrease the segmentation error rates. In this technique, cluster-based and threshold-based segmentation techniques are merged together. After then, to improve the result of HIST-SI, the authors added the method of filtering and linking in this technique named Imp_HIST-SI to decrease the segmentation error rates. The goal of this research is to find a new technique to decrease the segmentation error rates and produce much better results than the HIST-SI technique. For testing the proposed technique, a dataset of Bhuvan – a National Geoportal developed and hosted by ISRO (Indian Space Research Organisation) is used. Experiments are conducted using Scikit-image & OpenCV tools of Python, and performance is evaluated and compared over various existing image segmentation techniques for several matrices, i.e., Mean Square Error (MSE) and Peak Signal Noise Ratio (PSNR).Keywords: satellite image, image segmentation, edge detection, error rate, MSE, PSNR, HIST-SI, linking, filtering, imp_HIST-SI
Procedia PDF Downloads 141