Search results for: empathic accuracy
1584 Modeling of Daily Global Solar Radiation Using Ann Techniques: A Case of Study
Authors: Said Benkaciali, Mourad Haddadi, Abdallah Khellaf, Kacem Gairaa, Mawloud Guermoui
Abstract:
In this study, many experiments were carried out to assess the influence of the input parameters on the performance of multilayer perceptron which is one the configuration of the artificial neural networks. To estimate the daily global solar radiation on the horizontal surface, we have developed some models by using seven combinations of twelve meteorological and geographical input parameters collected from a radiometric station installed at Ghardaïa city (southern of Algeria). For selecting of best combination which provides a good accuracy, six statistical formulas (or statistical indicators) have been evaluated, such as the root mean square errors, mean absolute errors, correlation coefficient, and determination coefficient. We noted that multilayer perceptron techniques have the best performance, except when the sunshine duration parameter is not included in the input variables. The maximum of determination coefficient and correlation coefficient are equal to 98.20 and 99.11%. On the other hand, some empirical models were developed to compare their performances with those of multilayer perceptron neural networks. Results obtained show that the neural networks techniques give the best performance compared to the empirical models.Keywords: empirical models, multilayer perceptron neural network, solar radiation, statistical formulas
Procedia PDF Downloads 3451583 High Order Block Implicit Multi-Step (Hobim) Methods for the Solution of Stiff Ordinary Differential Equations
Authors: J. P. Chollom, G. M. Kumleng, S. Longwap
Abstract:
The search for higher order A-stable linear multi-step methods has been the interest of many numerical analysts and has been realized through either higher derivatives of the solution or by inserting additional off step points, supper future points and the likes. These methods are suitable for the solution of stiff differential equations which exhibit characteristics that place a severe restriction on the choice of step size. It becomes necessary that only methods with large regions of absolute stability remain suitable for such equations. In this paper, high order block implicit multi-step methods of the hybrid form up to order twelve have been constructed using the multi-step collocation approach by inserting one or more off step points in the multi-step method. The accuracy and stability properties of the new methods are investigated and are shown to yield A-stable methods, a property desirable of methods suitable for the solution of stiff ODE’s. The new High Order Block Implicit Multistep methods used as block integrators are tested on stiff differential systems and the results reveal that the new methods are efficient and compete favourably with the state of the art Matlab ode23 code.Keywords: block linear multistep methods, high order, implicit, stiff differential equations
Procedia PDF Downloads 3581582 A Computational Fluid Dynamics Study of Turbulence Flow and Parameterization of an Aerofoil
Authors: Mohamed Z. M. Duwahir, Shian Gao
Abstract:
The main objective of this project was to introduce and test a new scheme for parameterization of subsonic aerofoil, using a function called Shape Function. Python programming was used to create a user interactive environment for geometry generation of aerofoil using NACA and Shape Function methodologies. Two aerofoils, NACA 0012 and NACA 1412, were generated using this function. Testing the accuracy of the Shape Function scheme was done by Linear Square Fitting using Python and CFD modelling the aerofoil in Fluent. NACA 0012 (symmetrical aerofoil) was better approximated using Shape Function than NACA 1412 (cambered aerofoil). The second part of the project involved comparing two turbulent models, k-ε and Spalart-Allmaras (SA), in Fluent by modelling the aerofoils NACA 0012 and NACA 1412 in conditions of Reynolds number of 3 × 106. It was shown that SA modelling is better for aerodynamic purpose. The experimental coefficient of lift (Cl) and coefficient of drag (Cd) were compared with empirical wind tunnel data for a range of angle of attack (AOA). As a further step, this project involved drawing and meshing 3D wings in Gambit. The 3D wing flow was solved and compared with 2D aerofoil section experimental results and wind tunnel data.Keywords: CFD simulation, shape function, turbulent modelling, aerofoil
Procedia PDF Downloads 3581581 Isolation Preserving Medical Conclusion Hold Structure via C5 Algorithm
Authors: Swati Kishor Zode, Rahul Ambekar
Abstract:
Data mining is the extraction of fascinating examples on the other hand information from enormous measure of information and choice is made as indicated by the applicable information extracted. As of late, with the dangerous advancement in internet, stockpiling of information and handling procedures, privacy preservation has been one of the major (higher) concerns in data mining. Various techniques and methods have been produced for protection saving data mining. In the situation of Clinical Decision Support System, the choice is to be made on the premise of the data separated from the remote servers by means of Internet to diagnose the patient. In this paper, the fundamental thought is to build the precision of Decision Support System for multiple diseases for different maladies and in addition protect persistent information while correspondence between Clinician side (Client side) also, the Server side. A privacy preserving protocol for clinical decision support network is proposed so that patients information dependably stay scrambled amid diagnose prepare by looking after the accuracy. To enhance the precision of Decision Support System for various malady C5.0 classifiers and to save security, a Homomorphism encryption algorithm Paillier cryptosystem is being utilized.Keywords: classification, homomorphic encryption, clinical decision support, privacy
Procedia PDF Downloads 3301580 PEINS: A Generic Compression Scheme Using Probabilistic Encoding and Irrational Number Storage
Authors: P. Jayashree, S. Rajkumar
Abstract:
With social networks and smart devices generating a multitude of data, effective data management is the need of the hour for networks and cloud applications. Some applications need effective storage while some other applications need effective communication over networks and data reduction comes as a handy solution to meet out both requirements. Most of the data compression techniques are based on data statistics and may result in either lossy or lossless data reductions. Though lossy reductions produce better compression ratios compared to lossless methods, many applications require data accuracy and miniature details to be preserved. A variety of data compression algorithms does exist in the literature for different forms of data like text, image, and multimedia data. In the proposed work, a generic progressive compression algorithm, based on probabilistic encoding, called PEINS is projected as an enhancement over irrational number stored coding technique to cater to storage issues of increasing data volumes as a cost effective solution, which also offers data security as a secondary outcome to some extent. The proposed work reveals cost effectiveness in terms of better compression ratio with no deterioration in compression time.Keywords: compression ratio, generic compression, irrational number storage, probabilistic encoding
Procedia PDF Downloads 2941579 Improving the Accuracy of Oral Care Performed by ICU Nurses for Cancer Patients
Authors: Huang Wei-Yi
Abstract:
Purpose: Oral cancer patients undergoing skin flap reconstruction may have wounds in the oral cavity, leading to accumulation of blood, clots, and secretions. Inadequate oral care by nursing staff can result in oral infections and pain. Methods: An investigation revealed that ICU nurses' knowledge and adherence to oral care standards were below acceptable levels. Key issues identified included lack of hands-on training opportunities, insufficient experience, absence of oral care standards and regular audits, no in-service education programs, and a lack of oral care educational materials. Interventions: The following measures were implemented: 1) in-service education programs, 2) development of care standards, 3) creation of a monitoring plan, 4) bedside demonstration teaching, and 5) revision of educational materials. Results: The intervention demonstrated that ICU nurses' knowledge and adherence to oral care standards improved, leading to better quality oral care and reduced pain for patients. Conclusion: Through in-service education, bedside demonstrations, establishment of oral care standards, and regular audits, the oral care skills of ICU nurses were significantly enhanced, resulting in improved oral care quality and decreased patient pain.Keywords: oral care, ICU, improving, oral cancer
Procedia PDF Downloads 231578 Particle Filter State Estimation Algorithm Based on Improved Artificial Bee Colony Algorithm
Authors: Guangyuan Zhao, Nan Huang, Xuesong Han, Xu Huang
Abstract:
In order to solve the problem of sample dilution in the traditional particle filter algorithm and achieve accurate state estimation in a nonlinear system, a particle filter method based on an improved artificial bee colony (ABC) algorithm was proposed. The algorithm simulated the process of bee foraging and optimization and made the high likelihood region of the backward probability of particles moving to improve the rationality of particle distribution. The opposition-based learning (OBL) strategy is introduced to optimize the initial population of the artificial bee colony algorithm. The convergence factor is introduced into the neighborhood search strategy to limit the search range and improve the convergence speed. Finally, the crossover and mutation operations of the genetic algorithm are introduced into the search mechanism of the following bee, which makes the algorithm jump out of the local extreme value quickly and continue to search the global extreme value to improve its optimization ability. The simulation results show that the improved method can improve the estimation accuracy of particle filters, ensure the diversity of particles, and improve the rationality of particle distribution.Keywords: particle filter, impoverishment, state estimation, artificial bee colony algorithm
Procedia PDF Downloads 1511577 Detection of Cyberattacks on the Metaverse Based on First-Order Logic
Authors: Sulaiman Al Amro
Abstract:
There are currently considerable challenges concerning data security and privacy, particularly in relation to modern technologies. This includes the virtual world known as the Metaverse, which consists of a virtual space that integrates various technologies and is therefore susceptible to cyber threats such as malware, phishing, and identity theft. This has led recent studies to propose the development of Metaverse forensic frameworks and the integration of advanced technologies, including machine learning for intrusion detection and security. In this context, the application of first-order logic offers a formal and systematic approach to defining the conditions of cyberattacks, thereby contributing to the development of effective detection mechanisms. In addition, formalizing the rules and patterns of cyber threats has the potential to enhance the overall security posture of the Metaverse and, thus, the integrity and safety of this virtual environment. The current paper focuses on the primary actions employed by avatars for potential attacks, including Interval Temporal Logic (ITL) and behavior-based detection to detect an avatar’s abnormal activities within the Metaverse. The research established that the proposed framework attained an accuracy of 92.307%, resulting in the experimental results demonstrating the efficacy of ITL, including its superior performance in addressing the threats posed by avatars within the Metaverse domain.Keywords: security, privacy, metaverse, cyberattacks, detection, first-order logic
Procedia PDF Downloads 401576 Selecting Answers for Questions with Multiple Answer Choices in Arabic Question Answering Based on Textual Entailment Recognition
Authors: Anes Enakoa, Yawei Liang
Abstract:
Question Answering (QA) system is one of the most important and demanding tasks in the field of Natural Language Processing (NLP). In QA systems, the answer generation task generates a list of candidate answers to the user's question, in which only one answer is correct. Answer selection is one of the main components of the QA, which is concerned with selecting the best answer choice from the candidate answers suggested by the system. However, the selection process can be very challenging especially in Arabic due to its particularities. To address this challenge, an approach is proposed to answer questions with multiple answer choices for Arabic QA systems based on Textual Entailment (TE) recognition. The developed approach employs a Support Vector Machine that considers lexical, semantic and syntactic features in order to recognize the entailment between the generated hypotheses (H) and the text (T). A set of experiments has been conducted for performance evaluation and the overall performance of the proposed method reached an accuracy of 67.5% with C@1 score of 80.46%. The obtained results are promising and demonstrate that the proposed method is effective for TE recognition task.Keywords: information retrieval, machine learning, natural language processing, question answering, textual entailment
Procedia PDF Downloads 1451575 A Distinct Method Based on Mamba-Unet for Brain Tumor Image Segmentation
Authors: Djallel Bouamama, Yasser R. Haddadi
Abstract:
Accurate brain tumor segmentation is crucial for diagnosis and treatment planning, yet it remains a challenging task due to the variability in tumor shapes and intensities. This paper introduces a distinct approach to brain tumor image segmentation by leveraging an advanced architecture known as Mamba-Unet. Building on the well-established U-Net framework, Mamba-Unet incorporates distinct design enhancements to improve segmentation performance. Our proposed method integrates a multi-scale attention mechanism and a hybrid loss function to effectively capture fine-grained details and contextual information in brain MRI scans. We demonstrate that Mamba-Unet significantly enhances segmentation accuracy compared to conventional U-Net models by utilizing a comprehensive dataset of annotated brain MRI scans. Quantitative evaluations reveal that Mamba-Unet surpasses traditional U-Net architectures and other contemporary segmentation models regarding Dice coefficient, sensitivity, and specificity. The improvements are attributed to the method's ability to manage class imbalance better and resolve complex tumor boundaries. This work advances the state-of-the-art in brain tumor segmentation and holds promise for improving clinical workflows and patient outcomes through more precise and reliable tumor detection.Keywords: brain tumor classification, image segmentation, CNN, U-NET
Procedia PDF Downloads 341574 Reliability of Swine Estrous Detector Probe in Dairy Cattle Breeding
Authors: O. O. Leigh, L. C. Agbugba, A. O. Oyewunmi, A. E. Ibiam, A. Hassan
Abstract:
Accuracy of insemination timing is a key determinant of high pregnancy rates in livestock breeding stations. The estrous detector probes are a recent introduction into the Nigerian livestock farming sector. Many of these probes are species-labeled and they measure changes in the vaginal mucus resistivity (VMR) during the stages of the estrous cycle. With respect to size and shaft conformation, the Draminski® swine estrous detector probe (sEDP) is quite similar to the bovine estrous detector probe. We investigated the reliability of the sEDP at insemination time on two farms designated as FM A and FM B. Cows (Bunaji, n=20 per farm) were evaluated for VMR at 16th h post standard OvSynch protocol, with concurrent insemination on FM B only. The difference in the mean VMR between FM A (221 ± 24.36) Ohms and FM B (254 ± 35.59) Ohms was not significant (p > 0.05). Sixteen cows (80%) at FM B were later (day 70) confirmed pregnant via rectal palpation and calved at term. These findings suggest consistency in VMR evaluated with sEDP at insemination as well as a high predictability for VMR associated with good pregnancy rates in dairy cattle. We conclude that Draminski® swine estrous detector probe is reliable in determining time of insemination in cattle breeding stations.Keywords: dairy cattle, insemination, swine estrous probe, vaginal mucus resistivity
Procedia PDF Downloads 1241573 Analysis of Vocal Fold Vibrations from High-Speed Digital Images Based on Dynamic Time Warping
Authors: A. I. A. Rahman, Sh-Hussain Salleh, K. Ahmad, K. Anuar
Abstract:
Analysis of vocal fold vibration is essential for understanding the mechanism of voice production and for improving clinical assessment of voice disorders. This paper presents a Dynamic Time Warping (DTW) based approach to analyze and objectively classify vocal fold vibration patterns. The proposed technique was designed and implemented on a Glottal Area Waveform (GAW) extracted from high-speed laryngeal images by delineating the glottal edges for each image frame. Feature extraction from the GAW was performed using Linear Predictive Coding (LPC). Several types of voice reference templates from simulations of clear, breathy, fry, pressed and hyperfunctional voice productions were used. The patterns of the reference templates were first verified using the analytical signal generated through Hilbert transformation of the GAW. Samples from normal speakers’ voice recordings were then used to evaluate and test the effectiveness of this approach. The classification of the voice patterns using the technique of LPC and DTW gave the accuracy of 81%.Keywords: dynamic time warping, glottal area waveform, linear predictive coding, high-speed laryngeal images, Hilbert transform
Procedia PDF Downloads 2391572 Influence of Environmental Temperature on Dairy Herd Performance and Behaviour
Authors: L. Krpalkova, N. O' Mahony, A. Carvalho, S. Campbell, S. Harapanahalli, J. Walsh
Abstract:
The objective of this study was to determine the effects of environmental stressors on the performance of lactating dairy cows and discuss some future trends. There exists a relationship between the meteorological data and milk yield prediction accuracy in pasture-based dairy systems. New precision technologies are available and are being developed to improve the sustainability of the dairy industry. Some of these technologies focus on welfare of individual animals on dairy farms. These technologies allow the automatic identification of animal behaviour and health events, greatly increasing overall herd health and yield while reducing animal health inspection demands and long-term animal healthcare costs. The data set consisted of records from 489 dairy cows at two dairy farms and temperature measured from the nearest meteorological weather station in 2018. The effects of temperature on milk production and behaviour of animals were analyzed. The statistical results indicate different effects of temperature on milk yield and behaviour. The “comfort zone” for animals is in the range 10 °C to 20 °C. Dairy cows out of this zone had to decrease or increase their metabolic heat production, and it affected their milk production and behaviour.Keywords: behavior, milk yield, temperature, precision technologies
Procedia PDF Downloads 1091571 Evaluation of Spatial Correlation Length and Karhunen-Loeve Expansion Terms for Predicting Reliability Level of Long-Term Settlement in Soft Soils
Authors: Mehrnaz Alibeikloo, Hadi Khabbaz, Behzad Fatahi
Abstract:
The spectral random field method is one of the widely used methods to obtain more reliable and accurate results in geotechnical problems involving material variability. Karhunen-Loeve (K-L) expansion method was applied to perform random field discretization of cross-correlated creep parameters. Karhunen-Loeve expansion method is based on eigenfunctions and eigenvalues of covariance function adopting Kernel integral solution. In this paper, the accuracy of Karhunen-Loeve expansion was investigated to predict long-term settlement of soft soils adopting elastic visco-plastic creep model. For this purpose, a parametric study was carried to evaluate the effect of K-L expansion terms and spatial correlation length on the reliability of results. The results indicate that small values of spatial correlation length require more K-L expansion terms. Moreover, by increasing spatial correlation length, the coefficient of variation (COV) of creep settlement increases, confirming more conservative and safer prediction.Keywords: Karhunen-Loeve expansion, long-term settlement, reliability analysis, spatial correlation length
Procedia PDF Downloads 1591570 Identifying the Structural Components of Old Buildings from Floor Plans
Authors: Shi-Yu Xu
Abstract:
The top three risk factors that have contributed to building collapses during past earthquake events in Taiwan are: "irregular floor plans or elevations," "insufficient columns in single-bay buildings," and the "weak-story problem." Fortunately, these unsound structural characteristics can be directly identified from the floor plans. However, due to the vast number of old buildings, conducting manual inspections to identify these compromised structural features in all existing structures would be time-consuming and prone to human errors. This study aims to develop an algorithm that utilizes artificial intelligence techniques to automatically pinpoint the structural components within a building's floor plans. The obtained spatial information will be utilized to construct a digital structural model of the building. This information, particularly regarding the distribution of columns in the floor plan, can then be used to conduct preliminary seismic assessments of the building. The study employs various image processing and pattern recognition techniques to enhance detection efficiency and accuracy. The study enables a large-scale evaluation of structural vulnerability for numerous old buildings, providing ample time to arrange for structural retrofitting in those buildings that are at risk of significant damage or collapse during earthquakes.Keywords: structural vulnerability detection, object recognition, seismic capacity assessment, old buildings, artificial intelligence
Procedia PDF Downloads 891569 The Rendering of Sex-Related Expressions by Court Interpreters in Hong Kong: A Corpus-Based Approach
Authors: Yee Yan Crystal Kwong
Abstract:
The essence of rape is the absence of consent to sexual intercourse. Yet, the definition of consent is not absolute and allows for subjectivity. In this case, the accuracy of oral interpretation becomes very important as the narratives of events and situation, as well as the register and style of speakers would influence the juror decision making. This paper first adopts a corpus-based approach to investigate how court interpreters in Hong Kong handle expressions that refer to sexual activities. The data of this study will be based on online corpus :From legislation to translation, from translation to interpretation: The narrative of sexual offences. The corpus comprises the transcription of five separate rape trials and all of these trials were heard with the presence of an interpreter. Since there are plenty of sex-related expressions used by witnesses and defendants in the five cases, emphasis will be put on those which have an impact on the definition of rape. With an in-depth analysis of the interpreted utterances, different interpreting approaches will be identified to observe how interpreters retain the intended meanings. Interviews with experienced court interpreters will also be conducted to revisit the validity of the traditional verbatim standard. At the end of this research, various interpreting approaches will be compared and evaluated. A redefinition of interpreters' institutional role, as well as recommendations for interpreting learners will be provided.Keywords: court interpreting, interpreters, legal translation, slangs
Procedia PDF Downloads 2621568 Analysing Techniques for Fusing Multimodal Data in Predictive Scenarios Using Convolutional Neural Networks
Authors: Philipp Ruf, Massiwa Chabbi, Christoph Reich, Djaffar Ould-Abdeslam
Abstract:
In recent years, convolutional neural networks (CNN) have demonstrated high performance in image analysis, but oftentimes, there is only structured data available regarding a specific problem. By interpreting structured data as images, CNNs can effectively learn and extract valuable insights from tabular data, leading to improved predictive accuracy and uncovering hidden patterns that may not be apparent in traditional structured data analysis. In applying a single neural network for analyzing multimodal data, e.g., both structured and unstructured information, significant advantages in terms of time complexity and energy efficiency can be achieved. Converting structured data into images and merging them with existing visual material offers a promising solution for applying CNN in multimodal datasets, as they often occur in a medical context. By employing suitable preprocessing techniques, structured data is transformed into image representations, where the respective features are expressed as different formations of colors and shapes. In an additional step, these representations are fused with existing images to incorporate both types of information. This final image is finally analyzed using a CNN.Keywords: CNN, image processing, tabular data, mixed dataset, data transformation, multimodal fusion
Procedia PDF Downloads 1231567 Design of a Real Time Heart Sounds Recognition System
Authors: Omer Abdalla Ishag, Magdi Baker Amien
Abstract:
Physicians used the stethoscope for listening patient heart sounds in order to make a diagnosis. However, the determination of heart conditions by acoustic stethoscope is a difficult task so it requires special training of medical staff. This study developed an accurate model for analyzing the phonocardiograph signal based on PC and DSP processor. The system has been realized into two phases; offline and real time phase. In offline phase, 30 cases of heart sounds files were collected from medical students and doctor's world website. For experimental phase (real time), an electronic stethoscope has been designed, implemented and recorded signals from 30 volunteers, 17 were normal cases and 13 were various pathologies cases, these acquired 30 signals were preprocessed using an adaptive filter to remove lung sounds. The background noise has been removed from both offline and real data, using wavelet transform, then graphical and statistics features vector elements were extracted, finally a look-up table was used for classification heart sounds cases. The obtained results of the implemented system showed accuracy of 90%, 80% and sensitivity of 87.5%, 82.4% for offline data, and real data respectively. The whole system has been designed on TMS320VC5509a DSP Platform.Keywords: code composer studio, heart sounds, phonocardiograph, wavelet transform
Procedia PDF Downloads 4461566 Forecasting 24-Hour Ahead Electricity Load Using Time Series Models
Authors: Ramin Vafadary, Maryam Khanbaghi
Abstract:
Forecasting electricity load is important for various purposes like planning, operation, and control. Forecasts can save operating and maintenance costs, increase the reliability of power supply and delivery systems, and correct decisions for future development. This paper compares various time series methods to forecast 24 hours ahead of electricity load. The methods considered are the Holt-Winters smoothing, SARIMA Modeling, LSTM Network, Fbprophet, and Tensorflow probability. The performance of each method is evaluated by using the forecasting accuracy criteria, namely, the mean absolute error and root mean square error. The National Renewable Energy Laboratory (NREL) residential energy consumption data is used to train the models. The results of this study show that the SARIMA model is superior to the others for 24 hours ahead forecasts. Furthermore, a Bagging technique is used to make the predictions more robust. The obtained results show that by Bagging multiple time-series forecasts, we can improve the robustness of the models for 24 hours ahead of electricity load forecasting.Keywords: bagging, Fbprophet, Holt-Winters, LSTM, load forecast, SARIMA, TensorFlow probability, time series
Procedia PDF Downloads 951565 Identity Management in Virtual Worlds Based on Biometrics Watermarking
Authors: S. Bader, N. Essoukri Ben Amara
Abstract:
With the technological development and rise of virtual worlds, these spaces are becoming more and more attractive for cybercriminals, hidden behind avatars and fictitious identities. Since access to these spaces is not restricted or controlled, some impostors take advantage of gaining unauthorized access and practicing cyber criminality. This paper proposes an identity management approach for securing access to virtual worlds. The major purpose of the suggested solution is to install a strong security mechanism to protect virtual identities represented by avatars. Thus, only legitimate users, through their corresponding avatars, are allowed to access the platform resources. Access is controlled by integrating an authentication process based on biometrics. In the request process for registration, a user fingerprint is enrolled and then encrypted into a watermark utilizing a cancelable and non-invertible algorithm for its protection. After a user personalizes their representative character, the biometric mark is embedded into the avatar through a watermarking procedure. The authenticity of the avatar identity is verified when it requests authorization for access. We have evaluated the proposed approach on a dataset of avatars from various virtual worlds, and we have registered promising performance results in terms of authentication accuracy, acceptation and rejection rates.Keywords: identity management, security, biometrics authentication and authorization, avatar, virtual world
Procedia PDF Downloads 2651564 Decision-Making using Fuzzy Linguistic Hypersoft Set Topology
Authors: Muhammad Saqlain, Poom Kumam
Abstract:
Language being an abstract system and creative act, is quite complicated as its meaning varies depending on the context. The context is determined by the empirical knowledge of a person, which is derived from observation and experience. About further subdivided attributes, the decision-making challenges may entail quantitative and qualitative factors. However, because there is no norm for putting a numerical value on language, existing approaches cannot carry out the operations of linguistic knowledge. The assigning of mathematical values (fuzzy, intuitionistic, and neutrosophic) to any decision-making problem; without considering any rule of linguistic knowledge is ambiguous and inaccurate. Thus, this paper aims to provide a generic model for these issues. This paper provides the linguistic set structure of the fuzzy hypersoft set (FLHSS) to solve decision-making issues. We have proposed the definition some basic operations like AND, NOT, OR, AND, compliment, negation, etc., along with Topology and examples, and properties. Secondly, the operational laws for the fuzzy linguistic hypersoft set have been proposed to deal with the decision-making issues. Implementing proposed aggregate operators and operational laws can be used to convert linguistic quantifiers into numerical values. This will increase the accuracy and precision of the fuzzy hypersoft set structure to deal with decision-making issues.Keywords: linguistic quantifiers, aggregate operators, multi-criteria decision making (mcdm)., fuzzy topology
Procedia PDF Downloads 971563 Virtualization and Visualization Based Driver Configuration in Operating System
Authors: Pavan Shah
Abstract:
In an Embedded system, Virtualization and visualization technology can provide us an effective response and measurable work in a software development environment. In addition to work of virtualization and virtualization can be easily deserved to provide the best resource sharing between real-time hardware applications and a healthy environment. However, the virtualization is noticeable work to minimize the I/O work and utilize virtualization & virtualization technology for either a software development environment (SDE) or a runtime environment of real-time embedded systems (RTMES) or real-time operating system (RTOS) eras. In this Paper, we particularly focus on virtualization and visualization overheads data of network which generates the I/O and implementation of standardized I/O (i.e., Virto), which can work as front-end network driver in a real-time operating system (RTOS) hardware module. Even there have been several work studies are available based on the virtualization operating system environment, but for the Virto on a general-purpose OS, my implementation is on the open-source Virto for a real-time operating system (RTOS). In this paper, the measurement results show that implementation which can improve the bandwidth and latency of memory management of the real-time operating system environment (RTMES) for getting more accuracy of the trained model.Keywords: virtualization, visualization, network driver, operating system
Procedia PDF Downloads 1331562 Sensitivity and Specificity of Clinical Testing for Digital Nerve Injury
Authors: Guy Rubin, Ravit Shay, Nimrod Rozen
Abstract:
The accuracy of a diagnostic test used to classify a patient as having disease or being disease-free is a valuable piece of information to be used by the physician when making treatment decisions. Finger laceration, suspected to have nerve injury is a challenging decision for the treating surgeon. The purpose of this study was to evaluate the sensitivity, specificity and predictive values of six clinical tests in the diagnosis of digital nerve injury. The six clinical tests included light touch, pin prick, static and dynamic 2-point discrimination, Semmes Weinstein monofilament and wrinkle test. Data comparing pre-surgery examination with post-surgery results of 42 patients with 52 digital nerve injury was evaluated. The subjective examinations, light touch, pin prick, static and dynamic 2-point discrimination and Semmes-Weinstein monofilament were not sensitive (57.6, 69.7, 42.4, 40 and 66.8% respectively) and specific (36.8, 36.8, 47.4, 42.1 and 31.6% respectively). Wrinkle test, the only objective examination, was the most sensitive (78.1%) and specific (55.6%). This result gives no pre-operative examination the ability to predict the result of explorative surgery.Keywords: digital nerve, injury, nerve examination, Semmes-Weinstein monofilamen, sensitivity, specificity, two point discrimination, wrinkle test
Procedia PDF Downloads 3441561 Deep Supervision Based-Unet to Detect Buildings Changes from VHR Aerial Imagery
Authors: Shimaa Holail, Tamer Saleh, Xiongwu Xiao
Abstract:
Building change detection (BCD) from satellite imagery is an essential topic in urbanization monitoring, agricultural land management, and updating geospatial databases. Recently, methods for detecting changes based on deep learning have made significant progress and impressive results. However, it has the problem of being insensitive to changes in buildings with complex spectral differences, and the features being extracted are not discriminatory enough, resulting in incomplete buildings and irregular boundaries. To overcome these problems, we propose a dual Siamese network based on the Unet model with the addition of a deep supervision strategy (DS) in this paper. This network consists of a backbone (encoder) based on ImageNet pre-training, a fusion block, and feature pyramid networks (FPN) to enhance the step-by-step information of the changing regions and obtain a more accurate BCD map. To train the proposed method, we created a new dataset (EGY-BCD) of high-resolution and multi-temporal aerial images captured over New Cairo in Egypt to detect building changes for this purpose. The experimental results showed that the proposed method is effective and performs well with the EGY-BCD dataset regarding the overall accuracy, F1-score, and mIoU, which were 91.6 %, 80.1 %, and 73.5 %, respectively.Keywords: building change detection, deep supervision, semantic segmentation, EGY-BCD dataset
Procedia PDF Downloads 1201560 A Study on the Assessment of Prosthetic Infection after Total Knee Replacement Surgery
Authors: Chun-Lang Chang, Chun-Kai Liu
Abstract:
In this study, the patients that have undergone total knee replacement surgery from the 2010 National Health Insurance database were adopted as the study participants. The important factors were screened and selected through literature collection and interviews with physicians. Through the Cross Entropy Method (CE), Genetic Algorithm Logistic Regression (GALR), and Particle Swarm Optimization (PSO), the weights of the factors were obtained. In addition, the weights of the respective algorithms, coupled with the Excel VBA were adopted to construct the Case Based Reasoning (CBR) system. The results through statistical tests show that the GALR and PSO produced no significant differences, and the accuracy of both models were above 97%. Moreover, the area under the curve of ROC for these two models also exceeded 0.87. This study shall serve as a reference for medical staff as an assistance for clinical assessment of infections in order to effectively enhance medical service quality and efficiency, avoid unnecessary medical waste, and substantially contribute to resource allocations in medical institutions.Keywords: Case Based Reasoning, Cross Entropy Method, Genetic Algorithm Logistic Regression, Particle Swarm Optimization, Total Knee Replacement Surgery
Procedia PDF Downloads 3221559 Forecasting Age-Specific Mortality Rates and Life Expectancy at Births for Malaysian Sub-Populations
Authors: Syazreen N. Shair, Saiful A. Ishak, Aida Y. Yusof, Azizah Murad
Abstract:
In this paper, we forecast age-specific Malaysian mortality rates and life expectancy at births by gender and ethnic groups including Malay, Chinese and Indian. Two mortality forecasting models are adopted the original Lee-Carter model and its recent modified version, the product ratio coherent model. While the first forecasts the mortality rates for each subpopulation independently, the latter accounts for the relationship between sub-populations. The evaluation of both models is performed using the out-of-sample forecast errors which are mean absolute percentage errors (MAPE) for mortality rates and mean forecast errors (MFE) for life expectancy at births. The best model is then used to perform the long-term forecasts up to the year 2030, the year when Malaysia is expected to become an aged nation. Results suggest that in terms of overall accuracy, the product ratio model performs better than the original Lee-Carter model. The association of lower mortality group (Chinese) in the subpopulation model can improve the forecasts of high mortality groups (Malay and Indian).Keywords: coherent forecasts, life expectancy at births, Lee-Carter model, product-ratio model, mortality rates
Procedia PDF Downloads 2191558 Multi-Criteria Inventory Classification Process Based on Logical Analysis of Data
Authors: Diana López-Soto, Soumaya Yacout, Francisco Ángel-Bello
Abstract:
Although inventories are considered as stocks of money sitting on shelve, they are needed in order to secure a constant and continuous production. Therefore, companies need to have control over the amount of inventory in order to find the balance between excessive and shortage of inventory. The classification of items according to certain criteria such as the price, the usage rate and the lead time before arrival allows any company to concentrate its investment in inventory according to certain ranking or priority of items. This makes the decision making process for inventory management easier and more justifiable. The purpose of this paper is to present a new approach for the classification of new items based on the already existing criteria. This approach is called the Logical Analysis of Data (LAD). It is used in this paper to assist the process of ABC items classification based on multiple criteria. LAD is a data mining technique based on Boolean theory that is used for pattern recognition. This technique has been tested in medicine, industry, credit risk analysis, and engineering with remarkable results. An application on ABC inventory classification is presented for the first time, and the results are compared with those obtained when using the well-known AHP technique and the ANN technique. The results show that LAD presented very good classification accuracy.Keywords: ABC multi-criteria inventory classification, inventory management, multi-class LAD model, multi-criteria classification
Procedia PDF Downloads 8811557 The Impact of Artificial Intelligence on Spare Parts Technology
Authors: Amir Andria Gad Shehata
Abstract:
Minimizing the inventory cost, optimizing the inventory quantities, and increasing system operational availability are the main motivations to enhance forecasting demand of spare parts in a major power utility company in Medina. This paper reports in an effort made to optimize the orders quantities of spare parts by improving the method of forecasting the demand. The study focuses on equipment that has frequent spare parts purchase orders with uncertain demand. The pattern of the demand considers a lumpy pattern which makes conventional forecasting methods less effective. A comparison was made by benchmarking various methods of forecasting based on experts’ criteria to select the most suitable method for the case study. Three actual data sets were used to make the forecast in this case study. Two neural networks (NN) approaches were utilized and compared, namely long short-term memory (LSTM) and multilayer perceptron (MLP). The results as expected, showed that the NN models gave better results than traditional forecasting method (judgmental method). In addition, the LSTM model had a higher predictive accuracy than the MLP model.Keywords: spare part, spare part inventory, inventory model, optimization, maintenanceneural network, LSTM, MLP, forecasting demand, inventory management
Procedia PDF Downloads 631556 Detection of Paenibacillus larvae (American Foulbrood Disease) by the PCR and Culture in the Remains of the Hive Collected at the Bottom of the Colony
Authors: N. Adjlane, N. Haddad
Abstract:
The American foulbrood is one of the most serious diseases that may affect brood of larvae and pupae stages. The causative organism is a gram positive bacterium Paaenibacillus larvae. American foulbrood infected apiaries suffer from severe economic losses, resulting from significant decreases in honeybee populations and honey production. The aim of this study was to detect Paenibacillus larvae in the remains collected at the bottom of the hive from the suspected hives by direct PCR and culture growth. A total of 56 suspected beehive wax debris samples collected in 40 different apiaries located in the central region of Algeria. MYPGP the culture medium is used during all the identifications of the bacterium. After positive results on samples, biochemical confirmation tests (test of catalase, presence hydrolysis of casein) and microscopic (gram stain) are used in order to verify the accuracy of the initial results. The QIAamp DNA Mini Kit is used to identify the DNA of Paaenibacillus larvae. Paaenibacillus larvae were identified in 14 samples out of 16 by the PCR. A suspected culture-negative sample was found positive through evaluation with PCR. This research is for the bacterium Paaenibacillus larvae in the debris of the colony is an effective method for diagnosis of the pathology of American foulbrood.Keywords: Paenibacillus larvae, honeybee, PCR, microbiological method
Procedia PDF Downloads 4111555 Predictive Analytics in Traffic Flow Management: Integrating Temporal Dynamics and Traffic Characteristics to Estimate Travel Time
Authors: Maria Ezziani, Rabie Zine, Amine Amar, Ilhame Kissani
Abstract:
This paper introduces a predictive model for urban transportation engineering, which is vital for efficient traffic management. Utilizing comprehensive datasets and advanced statistical techniques, the model accurately forecasts travel times by considering temporal variations and traffic dynamics. Machine learning algorithms, including regression trees and neural networks, are employed to capture sequential dependencies. Results indicate significant improvements in predictive accuracy, particularly during peak hours and holidays, with the incorporation of traffic flow and speed variables. Future enhancements may integrate weather conditions and traffic incidents. The model's applications range from adaptive traffic management systems to route optimization algorithms, facilitating congestion reduction and enhancing journey reliability. Overall, this research extends beyond travel time estimation, offering insights into broader transportation planning and policy-making realms, empowering stakeholders to optimize infrastructure utilization and improve network efficiency.Keywords: predictive analytics, traffic flow, travel time estimation, urban transportation, machine learning, traffic management
Procedia PDF Downloads 84