Search results for: software vulnerability detection
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 8530

Search results for: software vulnerability detection

8110 Cognitive Methods for Detecting Deception During the Criminal Investigation Process

Authors: Laid Fekih

Abstract:

Background: It is difficult to detect lying, deception, and misrepresentation just by looking at verbal or non-verbal expression during the criminal investigation process, as there is a common belief that it is possible to tell whether a person is lying or telling the truth just by looking at the way they act or behave. The process of detecting lies and deception during the criminal investigation process needs more studies and research to overcome the difficulties facing the investigators. Method: The present study aimed to identify the effectiveness of cognitive methods and techniques in detecting deception during the criminal investigation. It adopted the quasi-experimental method and covered a sample of (20) defendants distributed randomly into two homogeneous groups, an experimental group of (10) defendants be subject to criminal investigation by applying cognitive techniques to detect deception and a second experimental group of (10) defendants be subject to the direct investigation method. The tool that used is a guided interview based on models of investigative questions according to the cognitive deception detection approach, which consists of three techniques of Vrij: imposing the cognitive burden, encouragement to provide more information, and ask unexpected questions, and the Direct Investigation Method. Results: Results revealed a significant difference between the two groups in term of lie detection accuracy in favour of defendants be subject to criminal investigation by applying cognitive techniques, the cognitive deception detection approach produced superior total accuracy rates both with human observers and through an analysis of objective criteria. The cognitive deception detection approach produced superior accuracy results in truth detection: 71%, deception detection: 70% compared to a direct investigation method truth detection: 52%; deception detection: 49%. Conclusion: The study recommended if practitioners use a cognitive deception detection technique, they will correctly classify more individuals than when they use a direct investigation method.

Keywords: the cognitive lie detection approach, deception, criminal investigation, mental health

Procedia PDF Downloads 63
8109 Advancing in Cricket Analytics: Novel Approaches for Pitch and Ball Detection Employing OpenCV and YOLOV8

Authors: Pratham Madnur, Prathamkumar Shetty, Sneha Varur, Gouri Parashetti

Abstract:

In order to overcome conventional obstacles, this research paper investigates novel approaches for cricket pitch and ball detection that make use of cutting-edge technologies. The research integrates OpenCV for pitch inspection and modifies the YOLOv8 model for cricket ball detection in order to overcome the shortcomings of manual pitch assessment and traditional ball detection techniques. To ensure flexibility in a range of pitch environments, the pitch detection method leverages OpenCV’s color space transformation, contour extraction, and accurate color range defining features. Regarding ball detection, the YOLOv8 model emphasizes the preservation of minor object details to improve accuracy and is specifically trained to the unique properties of cricket balls. The methods are more reliable because of the careful preparation of the datasets, which include novel ball and pitch information. These cutting-edge methods not only improve cricket analytics but also set the stage for flexible methods in more general sports technology applications.

Keywords: OpenCV, YOLOv8, cricket, custom dataset, computer vision, sports

Procedia PDF Downloads 63
8108 Simplifying Seismic Vulnerability Analysis for Existing Reinforced Concrete Buildings

Authors: Maryam Solgi, Behzad Shahmohammadi, Morteza Raissi Dehkordi

Abstract:

One of the main steps for seismic retrofitting of buildings is to determine the vulnerability of structures. While current procedures for evaluating existing buildings are complicated, and there is no limitation between short, middle-high, and tall buildings. This research utilizes a simplified method for assessing structures, which is adequate for existing reinforced concrete buildings. To approach this aim, Simple Lateral Mechanisms Analysis (SLaMA) procedure proposed by NZSEE (New Zealand Society for Earthquake Engineering) has been carried out. In this study, three RC moment-resisting frame buildings are determined. First, these buildings have been evaluated by inelastic static procedure (Pushover) based on acceptance criteria. Then, Park-Ang Damage Index is determined for the whole members of each building by Inelastic Time History Analysis. Next, the Simple Lateral Mechanisms Analysis procedure, a hand method, is carried out to define the capacity of structures. Ultimately, existing procedures are compared with Peak Ground Acceleration caused to fail (PGAfail). The results of this comparison emphasize that the Pushover procedure and SLaMA method define a greater value of PGAfail than the Park-Ang Damage model.

Keywords: peak ground acceleration caused to fail, reinforced concrete moment-frame buildings, seismic vulnerability analysis, simple lateral mechanisms analysis

Procedia PDF Downloads 89
8107 The Use of Building Energy Simulation Software in Case Studies: A Literature Review

Authors: Arman Ameen, Mathias Cehlin

Abstract:

The use of Building Energy Simulation (BES) software has increased in the last two decades, parallel to the development of increased computing power and easy to use software applications. This type of software is primarily used to simulate the energy use and the indoor environment for a building. The rapid development of these types of software has raised their level of user-friendliness, better parameter input options and the increased possibility of analysis, both for a single building component or an entire building. This, in turn, has led to many researchers utilizing BES software in their research in various degrees. The aim of this paper is to carry out a literature review concerning the use of the BES software IDA Indoor Climate and Energy (IDA ICE) in the scientific community. The focus of this paper will be specifically the use of the software for whole building energy simulation, number and types of articles and publications dates, the area of application, types of parameters used, the location of the studied building, type of building, type of analysis and solution methodology. Another aspect that is examined, which is of great interest, is the method of validations regarding the simulation results. The results show that there is an upgoing trend in the use of IDA ICE and that researchers use the software in their research in various degrees depending on case and aim of their research. The satisfactory level of validation of the simulations carried out in these articles varies depending on the type of article and type of analysis.

Keywords: building simulation, IDA ICE, literature review, validation

Procedia PDF Downloads 128
8106 Development of Cost-effective Sensitive Methods for Pathogen Detection in Community Wastewater for Disease Surveillance

Authors: Jesmin Akter, Chang Hyuk Ahn, Ilho Kim, Jaiyeop Lee

Abstract:

Global pandemic coronavirus disease (COVID-19) caused by Severe acute respiratory syndrome SARS-CoV-2, to control the spread of the COVID-19 pandemic, wastewater surveillance has been used to monitor SARS-CoV2 prevalence in the community. The challenging part is establishing wastewater surveillance; there is a need for a well-equipped laboratory for wastewater sample analysis. According to many previous studies, reverse transcription-polymerase chain reaction (RT-PCR) based molecular tests are the most widely used and popular detection method worldwide. However, the RT-qPCR based approaches for the detection or quantification of SARS-CoV-2 genetic fragments ribonucleic acid (RNA) from wastewater require a specialized laboratory, skilled personnel, expensive instruments, and a workflow that typically requires 6 to 8 hours to provide results for just minimum samples. Rapid and reliable alternative detection methods are needed to enable less-well-qualified practitioners to set up and provide sensitive detection of SARS-CoV-2 within wastewater at less-specialized regional laboratories. Therefore, scientists and researchers are conducting experiments for rapid detection methods of COVID-19; in some cases, the structural and molecular characteristics of SARS-CoV-2 are unknown, and various strategies for the correct diagnosis of COVID-19 have been proposed by research laboratories, which are presented in the present study. The ongoing research and development of these highly sensitive and rapid technologies, namely RT-LAMP, ELISA, Biosensors, GeneXpert, allows a wide range of potential options not only for SARS-CoV-2 detection but also for other viruses as well. The effort of this study is to discuss the above effective and regional rapid detection and quantification methods in community wastewater as an essential step in advancing scientific goals.

Keywords: rapid detection, SARS-CoV-2, sensitive detection, wastewater surveillance

Procedia PDF Downloads 79
8105 Design and Fabrication of Optical Nanobiosensors for Detection of MicroRNAs Involved in Neurodegenerative Diseases

Authors: Mahdi Rahaie

Abstract:

MicroRNAs are a novel class of small RNAs which regulate gene expression by translational repression or degradation of messenger RNAs. To produce sensitive, simple and cost-effective assays for microRNAs, detection is in urgent demand due to important role of these biomolecules in progression of human disease such as Alzheimer’s, Multiple sclerosis, and some other neurodegenerative diseases. Herein, we report several novel, sensitive and specific microRNA nanobiosensors which were designed based on colorimetric and fluorescence detection of nanoparticles and hybridization chain reaction amplification as an enzyme-free amplification. These new strategies eliminate the need for enzymatic reactions, chemical changes, separation processes and sophisticated equipment whereas less limit of detection with most specify are acceptable. The important features of these methods are high sensitivity and specificity to differentiate between perfectly matched, mismatched and non-complementary target microRNAs and also decent response in the real sample analysis with blood plasma. These nanobiosensors can clinically be used not only for the early detection of neuro diseases but also for every sickness related to miRNAs by direct detection of the plasma microRNAs in real clinical samples, without a need for sample preparation, RNA extraction and/or amplification.

Keywords: hybridization chain reaction, microRNA, nanobiosensor, neurodegenerative diseases

Procedia PDF Downloads 145
8104 Evaluation of the Efficiency of French Language Educational Software for Learners in Semnan Province, Iran

Authors: Alireza Hashemi

Abstract:

In recent decades, language teaching methodology has undergone significant changes due to the advent of computers and the growth of educational software. French language education has also benefited from these developments, and various software has been produced to facilitate the learning of this language. However, the question arises whether these software programs meet the educational needs of Iranian learners, particularly in Semnan Province. The aim of this study is to evaluate the efficiency and effectiveness of French language educational software for learners in Semnan Province, considering educational, cultural, and technical criteria. In this study, content analysis and performance evaluation methods were used to examine the educational software ‘Français Facile’. This software was evaluated based on criteria such as teaching methods, cultural compatibility, and technical features. To collect data, standardized questionnaires and semi-structured interviews with learners in Semnan Province were used. Additionally, the SPSS statistical software was employed for quantitative data analysis, and the thematic analysis method was used for qualitative data. The results indicated that the ‘Français Facile’ software has strengths such as providing diverse educational content and an interactive learning environment. However, some weaknesses include the lack of alignment of educational content with the learning culture of learners in Semnan Province and technical issues in software execution. Statistical data showed that 65% of learners were satisfied with the educational content, but 55% reported issues related to cultural alignment with their needs. This study indicates that to enhance the efficiency of French language educational software, there is a need to localize educational content and improve technical infrastructure. Producing locally adapted educational software can improve the quality of language learning and increase the motivation of learners in Semnan Province. This research emphasizes the importance of understanding the cultural and educational needs of learners in the development of educational software and recommends that developers of educational software pay special attention to these aspects.

Keywords: educational software, French language, Iran, learners in Semnan province

Procedia PDF Downloads 29
8103 Detection of Keypoint in Press-Fit Curve Based on Convolutional Neural Network

Authors: Shoujia Fang, Guoqing Ding, Xin Chen

Abstract:

The quality of press-fit assembly is closely related to reliability and safety of product. The paper proposed a keypoint detection method based on convolutional neural network to improve the accuracy of keypoint detection in press-fit curve. It would provide an auxiliary basis for judging quality of press-fit assembly. The press-fit curve is a curve of press-fit force and displacement. Both force data and distance data are time-series data. Therefore, one-dimensional convolutional neural network is used to process the press-fit curve. After the obtained press-fit data is filtered, the multi-layer one-dimensional convolutional neural network is used to perform the automatic learning of press-fit curve features, and then sent to the multi-layer perceptron to finally output keypoint of the curve. We used the data of press-fit assembly equipment in the actual production process to train CNN model, and we used different data from the same equipment to evaluate the performance of detection. Compared with the existing research result, the performance of detection was significantly improved. This method can provide a reliable basis for the judgment of press-fit quality.

Keywords: keypoint detection, curve feature, convolutional neural network, press-fit assembly

Procedia PDF Downloads 221
8102 Artificially Intelligent Context Aware Personal Computer Assistant (ACPCA)

Authors: Abdul Mannan Akhtar

Abstract:

In this paper a novel concept of a self learning smart personalized computer assistant (ACPCA) is established which is a context aware system. Based on user habits, moods, and other routines/situational reactions the system will manage various services and suggestions at appropriate times including what schedule to follow, what to watch, what software to be used, what should be deleted etc. This system will utilize a hybrid fuzzyNeural model to predict what the user will do next and support his actions. This will be done by establishing fuzzy sets of user activities, choices, preferences etc. and utilizing their combinations to predict his moods and immediate preferences. Various application of context aware systems exist separately e.g. on certain websites for music or multimedia suggestions but a personalized autonomous system that could adapt to user’s personality does not exist at present. Due to the novelty and massiveness of this concept, this paper will primarily focus on the problem establishment, product features and its functionality; however a small mini case is also implemented on MATLAB to demonstrate some of the aspects of ACPCA. The mini case involves prediction of user moods, activity, routine and food preference using a hybrid fuzzy-Neural soft computing technique.

Keywords: context aware systems, APCPCA, soft computing techniques, artificial intelligence, fuzzy logic, neural network, mood detection, face detection, activity detection

Procedia PDF Downloads 460
8101 Agile Software Development Implementation in Developing a Diet Tracker Mobile Application

Authors: Dwi Puspita Sari, Gulnur Baltabayeva, Nadia Salman, Maxut Toleuov, Vijay Kanabar

Abstract:

Technology era drives people to use mobile phone to support their daily life activities. Technology development has a rapid phase which pushes the IT company to adjust any technology changes in order to fulfill customer’s satisfaction. As a result of that, many companies in the USA emerged from systematics software development approach to agile software development approach in developing systems and applications to develop many mobile phone applications in a short phase to fulfill user’s needs. As a systematic approach is considered as time consuming, costly, and too risky, agile software development has become a more popular approach to use for developing software including mobile applications. This paper reflects a short-term project to develop a diet tracker mobile application using agile software development that focused on applying scrum framework in the development process.

Keywords: agile software development, scrum, diet tracker, mobile application

Procedia PDF Downloads 247
8100 A Character Detection Method for Ancient Yi Books Based on Connected Components and Regressive Character Segmentation

Authors: Xu Han, Shanxiong Chen, Shiyu Zhu, Xiaoyu Lin, Fujia Zhao, Dingwang Wang

Abstract:

Character detection is an important issue for character recognition of ancient Yi books. The accuracy of detection directly affects the recognition effect of ancient Yi books. Considering the complex layout, the lack of standard typesetting and the mixed arrangement between images and texts, we propose a character detection method for ancient Yi books based on connected components and regressive character segmentation. First, the scanned images of ancient Yi books are preprocessed with nonlocal mean filtering, and then a modified local adaptive threshold binarization algorithm is used to obtain the binary images to segment the foreground and background for the images. Second, the non-text areas are removed by the method based on connected components. Finally, the single character in the ancient Yi books is segmented by our method. The experimental results show that the method can effectively separate the text areas and non-text areas for ancient Yi books and achieve higher accuracy and recall rate in the experiment of character detection, and effectively solve the problem of character detection and segmentation in character recognition of ancient books.

Keywords: CCS concepts, computing methodologies, interest point, salient region detections, image segmentation

Procedia PDF Downloads 124
8099 Motion-Based Detection and Tracking of Multiple Pedestrians

Authors: A. Harras, A. Tsuji, K. Terada

Abstract:

Tracking of moving people has gained a matter of great importance due to rapid technological advancements in the field of computer vision. The objective of this study is to design a motion based detection and tracking multiple walking pedestrians randomly in different directions. In our proposed method, Gaussian mixture model (GMM) is used to determine moving persons in image sequences. It reacts to changes that take place in the scene like different illumination; moving objects start and stop often, etc. Background noise in the scene is eliminated through applying morphological operations and the motions of tracked people which is determined by using the Kalman filter. The Kalman filter is applied to predict the tracked location in each frame and to determine the likelihood of each detection. We used a benchmark data set for the evaluation based on a side wall stationary camera. The actual scenes from the data set are taken on a street including up to eight people in front of the camera in different two scenes, the duration is 53 and 35 seconds, respectively. In the case of walking pedestrians in close proximity, the proposed method has achieved the detection ratio of 87%, and the tracking ratio is 77 % successfully. When they are deferred from each other, the detection ratio is increased to 90% and the tracking ratio is also increased to 79%.

Keywords: automatic detection, tracking, pedestrians, counting

Procedia PDF Downloads 251
8098 Plastic Pipe Defect Detection Using Nonlinear Acoustic Modulation

Authors: Gigih Priyandoko, Mohd Fairusham Ghazali, Tan Siew Fun

Abstract:

This paper discusses about the defect detection of plastic pipe by using nonlinear acoustic wave modulation method. It is a sensitive method for damage detection and it is based on the propagation of high frequency acoustic waves in plastic pipe with low frequency excitation. The plastic pipe is excited simultaneously with a slow amplitude modulated vibration pumping wave and a constant amplitude probing wave. The frequency of both the excitation signals coincides with the resonances of the plastic pipe. A PVP pipe is used as the specimen as it is commonly used for the conveyance of liquid in many fields. The results obtained are being observed and the difference between uncracked specimen and cracked specimen can be distinguished clearly.

Keywords: plastic pipe, defect detection, nonlinear acoustic modulation, excitation

Procedia PDF Downloads 444
8097 Aspects and Studies of Fractal Geometry in Automatic Breast Cancer Detection

Authors: Mrinal Kanti Bhowmik, Kakali Das Jr., Barin Kumar De, Debotosh Bhattacharjee

Abstract:

Breast cancer is the most common cancer and a leading cause of death for women in the 35 to 55 age group. Early detection of breast cancer can decrease the mortality rate of breast cancer. Mammography is considered as a ‘Gold Standard’ for breast cancer detection and a very popular modality, presently used for breast cancer screening and detection. The screening of digital mammograms often leads to over diagnosis and a consequence to unnecessary traumatic & painful biopsies. For that reason recent studies involving the use of thermal imaging as a screening technique have generated a growing interest especially in cases where the mammography is limited, as in young patients who have dense breast tissue. Tumor is a significant sign of breast cancer in both mammography and thermography. The tumors are complex in structure and they also exhibit a different statistical and textural features compared to the breast background tissue. Fractal geometry is a geometry which is used to describe this type of complex structure as per their main characteristic, where traditional Euclidean geometry fails. Over the last few years, fractal geometrics have been applied mostly in many medical image (1D, 2D, or 3D) analysis applications. In breast cancer detection using digital mammogram images, also it plays a significant role. Fractal is also used in thermography for early detection of the masses using the thermal texture. This paper presents an overview of the recent aspects and initiatives of fractals in breast cancer detection in both mammography and thermography. The scope of fractal geometry in automatic breast cancer detection using digital mammogram and thermogram images are analysed, which forms a foundation for further study on application of fractal geometry in medical imaging for improving the efficiency of automatic detection.

Keywords: fractal, tumor, thermography, mammography

Procedia PDF Downloads 379
8096 Italian Speech Vowels Landmark Detection through the Legacy Tool 'xkl' with Integration of Combined CNNs and RNNs

Authors: Kaleem Kashif, Tayyaba Anam, Yizhi Wu

Abstract:

This paper introduces a methodology for advancing Italian speech vowels landmark detection within the distinctive feature-based speech recognition domain. Leveraging the legacy tool 'xkl' by integrating combined convolutional neural networks (CNNs) and recurrent neural networks (RNNs), the study presents a comprehensive enhancement to the 'xkl' legacy software. This integration incorporates re-assigned spectrogram methodologies, enabling meticulous acoustic analysis. Simultaneously, our proposed model, integrating combined CNNs and RNNs, demonstrates unprecedented precision and robustness in landmark detection. The augmentation of re-assigned spectrogram fusion within the 'xkl' software signifies a meticulous advancement, particularly enhancing precision related to vowel formant estimation. This augmentation catalyzes unparalleled accuracy in landmark detection, resulting in a substantial performance leap compared to conventional methods. The proposed model emerges as a state-of-the-art solution in the distinctive feature-based speech recognition systems domain. In the realm of deep learning, a synergistic integration of combined CNNs and RNNs is introduced, endowed with specialized temporal embeddings, harnessing self-attention mechanisms, and positional embeddings. The proposed model allows it to excel in capturing intricate dependencies within Italian speech vowels, rendering it highly adaptable and sophisticated in the distinctive feature domain. Furthermore, our advanced temporal modeling approach employs Bayesian temporal encoding, refining the measurement of inter-landmark intervals. Comparative analysis against state-of-the-art models reveals a substantial improvement in accuracy, highlighting the robustness and efficacy of the proposed methodology. Upon rigorous testing on a database (LaMIT) speech recorded in a silent room by four Italian native speakers, the landmark detector demonstrates exceptional performance, achieving a 95% true detection rate and a 10% false detection rate. A majority of missed landmarks were observed in proximity to reduced vowels. These promising results underscore the robust identifiability of landmarks within the speech waveform, establishing the feasibility of employing a landmark detector as a front end in a speech recognition system. The synergistic integration of re-assigned spectrogram fusion, CNNs, RNNs, and Bayesian temporal encoding not only signifies a significant advancement in Italian speech vowels landmark detection but also positions the proposed model as a leader in the field. The model offers distinct advantages, including unparalleled accuracy, adaptability, and sophistication, marking a milestone in the intersection of deep learning and distinctive feature-based speech recognition. This work contributes to the broader scientific community by presenting a methodologically rigorous framework for enhancing landmark detection accuracy in Italian speech vowels. The integration of cutting-edge techniques establishes a foundation for future advancements in speech signal processing, emphasizing the potential of the proposed model in practical applications across various domains requiring robust speech recognition systems.

Keywords: landmark detection, acoustic analysis, convolutional neural network, recurrent neural network

Procedia PDF Downloads 51
8095 Urban Flood Risk Mapping–a Review

Authors: Sherly M. A., Subhankar Karmakar, Terence Chan, Christian Rau

Abstract:

Floods are one of the most frequent natural disasters, causing widespread devastation, economic damage and threat to human lives. Hydrologic impacts of climate change and intensification of urbanization are two root causes of increased flood occurrences, and recent research trends are oriented towards understanding these aspects. Due to rapid urbanization, population of cities across the world has increased exponentially leading to improperly planned developments. Climate change due to natural and anthropogenic activities on our environment has resulted in spatiotemporal changes in rainfall patterns. The combined effect of both aggravates the vulnerability of urban populations to floods. In this context, an efficient and effective flood risk management with its core component as flood risk mapping is essential in prevention and mitigation of flood disasters. Urban flood risk mapping involves zoning of an urban region based on its flood risk, which depicts the spatiotemporal pattern of frequency and severity of hazards, exposure to hazards, and degree of vulnerability of the population in terms of socio-economic, environmental and infrastructural aspects. Although vulnerability is a key component of risk, its assessment and mapping is often less advanced than hazard mapping and quantification. A synergic effort from technical experts and social scientists is vital for the effectiveness of flood risk management programs. Despite an increasing volume of quality research conducted on urban flood risk, a comprehensive multidisciplinary approach towards flood risk mapping still remains neglected due to which many of the input parameters and definitions of flood risk concepts are imprecise. Thus, the objectives of this review are to introduce and precisely define the relevant input parameters, concepts and terms in urban flood risk mapping, along with its methodology, current status and limitations. The review also aims at providing thought-provoking insights to potential future researchers and flood management professionals.

Keywords: flood risk, flood hazard, flood vulnerability, flood modeling, urban flooding, urban flood risk mapping

Procedia PDF Downloads 581
8094 Developing the P1-P7 Management and Analysis Software for Thai Child Evaluation (TCE) of Food and Nutrition Status

Authors: S. Damapong, C. Kingkeow, W. Kongnoo, P. Pattapokin, S. Pruenglamphu

Abstract:

As the presence of Thai children double burden malnutrition, we conducted a project to promote holistic age-appropriate nutrition for Thai children. Researchers developed P1-P7 computer software for managing and analyzing diverse types of collected data. The study objectives were: i) to use software to manage and analyze the collected data, ii) to evaluate the children nutritional status and their caretakers’ nutrition practice to create regulations for improving nutrition. Data were collected by means of questionnaires, called P1-P7. P1, P2 and P5 were for children and caretakers, and others were for institutions. The children nutritional status, height-for-age, weight-for-age, and weight-for-height standards were calculated using Thai child z-score references. Institution evaluations consisted of various standard regulations including the use of our software. The results showed that the software was used in 44 out of 118 communities (37.3%), 57 out of 240 child development centers and nurseries (23.8%), and 105 out of 152 schools (69.1%). No major problems have been reported with the software, although user efficiency can be increased further through additional training. As the result, the P1-P7 software was used to manage and analyze nutritional status, nutrition behavior, and environmental conditions, in order to conduct Thai Child Evaluation (TCE). The software was most widely used in schools. Some aspects of P1-P7’s questionnaires could be modified to increase ease of use and efficiency.

Keywords: P1-P7 software, Thai child evaluation, nutritional status, malnutrition

Procedia PDF Downloads 347
8093 Identifying Mitigation Plans in Reducing Usability Risk Using Delphi Method

Authors: Jayaletchumi T. Sambantha Moorthy, Suhaimi bin Ibrahim, Mohd Naz’ri Mahrin

Abstract:

Most quality models have defined usability as a significant factor that leads to improving product acceptability, increasing user satisfaction, improving product reliability, and also financially benefiting companies. Usability is also the best factor that acts as a balance for both the technical and human aspects of a software product, which is an important aspect in defining quality during software development process. A usability risk can be defined as a potential usability risk factor that a chosen action or activity may lead to a possible loss or an undesirable outcome. This could impact the usability of a software product thereby contributing to negative user experiences and causing a possible software product failure. Hence, it is important to mitigate and reduce usability risks in the software development process itself. By managing possible involved usability risks in software development process, failure of software product could be reduced. Therefore, this research uses the Delphi method to identify mitigation plans to reduce potential usability risks. The Delphi method is conducted with seven experts from the field of risk management and software development.

Keywords: usability, usability risk, risk management, risk mitigation, delphi study

Procedia PDF Downloads 459
8092 Short-Path Near-Infrared Laser Detection of Environmental Gases by Wavelength-Modulation Spectroscopy

Authors: Isao Tomita

Abstract:

The detection of environmental gases, 12CO_2, 13CO_2, and CH_4, using near-infrared semiconductor lasers with a short laser path length is studied by means of wavelength-modulation spectroscopy. The developed system is compact and has high sensitivity enough to detect the absorption peaks of isotopic 13CO_2 of a 3-% CO_2 gas at 2 um with a path length of 2.4 m, where its peak size is two orders of magnitude smaller than that of the ordinary 12CO_2 peaks. In addition, the detection of 12CO_2 peaks of a 385-ppm (0.0385-%) CO_2 gas in the air is made at 2 um with a path length of 1.4 m. Furthermore, in pursuing the detection of an ancient environmental CH_4 gas confined to a bubble in ice at the polar regions, measurements of the absorption spectrum for a trace gas of CH_4 in a small area are attempted. For a 100-% CH_4 gas trapped in a 1 mm^3 glass container, the absorption peaks of CH_4 are obtained at 1.65 um with a path length of 3 mm, and also the gas pressure is extrapolated from the measured data.

Keywords: environmental gases, Near-Infrared Laser Detection, Wavelength-Modulation Spectroscopy, gas pressure

Procedia PDF Downloads 416
8091 Investigating the Viability of Ultra-Low Parameter Count Networks for Real-Time Football Detection

Authors: Tim Farrelly

Abstract:

In recent years, AI-powered object detection systems have opened the doors for innovative new applications and products, especially those operating in the real world or ‘on edge’ – namely, in sport. This paper investigates the viability of an ultra-low parameter convolutional neural network specially designed for the detection of footballs on ‘on the edge’ devices. The main contribution of this paper is the exploration of integrating new design features (depth-wise separable convolutional blocks and squeezed and excitation modules) into an ultra-low parameter network and demonstrating subsequent improvements in performance. The results show that tracking the ball from Full HD images with negligibly high accu-racy is possible in real-time.

Keywords: deep learning, object detection, machine vision applications, sport, network design

Procedia PDF Downloads 140
8090 Approximate-Based Estimation of Single Event Upset Effect on Statistic Random-Access Memory-Based Field-Programmable Gate Arrays

Authors: Mahsa Mousavi, Hamid Reza Pourshaghaghi, Mohammad Tahghighi, Henk Corporaal

Abstract:

Recently, Statistic Random-Access Memory-based (SRAM-based) Field-Programmable Gate Arrays (FPGAs) are widely used in aeronautics and space systems where high dependability is demanded and considered as a mandatory requirement. Since design’s circuit is stored in configuration memory in SRAM-based FPGAs; they are very sensitive to Single Event Upsets (SEUs). In addition, the adverse effects of SEUs on the electronics used in space are much higher than in the Earth. Thus, developing fault tolerant techniques play crucial roles for the use of SRAM-based FPGAs in space. However, fault tolerance techniques introduce additional penalties in system parameters, e.g., area, power, performance and design time. In this paper, an accurate estimation of configuration memory vulnerability to SEUs is proposed for approximate-tolerant applications. This vulnerability estimation is highly required for compromising between the overhead introduced by fault tolerance techniques and system robustness. In this paper, we study applications in which the exact final output value is not necessarily always a concern meaning that some of the SEU-induced changes in output values are negligible. We therefore define and propose Approximate-based Configuration Memory Vulnerability Factor (ACMVF) estimation to avoid overestimating configuration memory vulnerability to SEUs. In this paper, we assess the vulnerability of configuration memory by injecting SEUs in configuration memory bits and comparing the output values of a given circuit in presence of SEUs with expected correct output. In spite of conventional vulnerability factor calculation methods, which accounts any deviations from the expected value as failures, in our proposed method a threshold margin is considered depending on user-case applications. Given the proposed threshold margin in our model, a failure occurs only when the difference between the erroneous output value and the expected output value is more than this margin. The ACMVF is subsequently calculated by acquiring the ratio of failures with respect to the total number of SEU injections. In our paper, a test-bench for emulating SEUs and calculating ACMVF is implemented on Zynq-7000 FPGA platform. This system makes use of the Single Event Mitigation (SEM) IP core to inject SEUs into configuration memory bits of the target design implemented in Zynq-7000 FPGA. Experimental results for 32-bit adder show that, when 1% to 10% deviation from correct output is considered, the counted failures number is reduced 41% to 59% compared with the failures number counted by conventional vulnerability factor calculation. It means that estimation accuracy of the configuration memory vulnerability to SEUs is improved up to 58% in the case that 10% deviation is acceptable in output results. Note that less than 10% deviation in addition result is reasonably tolerable for many applications in approximate computing domain such as Convolutional Neural Network (CNN).

Keywords: fault tolerance, FPGA, single event upset, approximate computing

Procedia PDF Downloads 192
8089 HRV Analysis Based Arrhythmic Beat Detection Using kNN Classifier

Authors: Onder Yakut, Oguzhan Timus, Emine Dogru Bolat

Abstract:

Health diseases have a vital significance affecting human being's life and life quality. Sudden death events can be prevented owing to early diagnosis and treatment methods. Electrical signals, taken from the human being's body using non-invasive methods and showing the heart activity is called Electrocardiogram (ECG). The ECG signal is used for following daily activity of the heart by clinicians. Heart Rate Variability (HRV) is a physiological parameter giving the variation between the heart beats. ECG data taken from MITBIH Arrhythmia Database is used in the model employed in this study. The detection of arrhythmic heart beats is aimed utilizing the features extracted from the HRV time domain parameters. The developed model provides a satisfactory performance with ~89% accuracy, 91.7 % sensitivity and 85% specificity rates for the detection of arrhythmic beats.

Keywords: arrhythmic beat detection, ECG, HRV, kNN classifier

Procedia PDF Downloads 349
8088 Design of an Improved Distributed Framework for Intrusion Detection System Based on Artificial Immune System and Neural Network

Authors: Yulin Rao, Zhixuan Li, Burra Venkata Durga Kumar

Abstract:

Intrusion detection refers to monitoring the actions of internal and external intruders on the system and detecting the behaviours that violate security policies in real-time. In intrusion detection, there has been much discussion about the application of neural network technology and artificial immune system (AIS). However, many solutions use static methods (signature-based and stateful protocol analysis) or centralized intrusion detection systems (CIDS), which are unsuitable for real-time intrusion detection systems that need to process large amounts of data and detect unknown intrusions. This article proposes a framework for a distributed intrusion detection system (DIDS) with multi-agents based on the concept of AIS and neural network technology to detect anomalies and intrusions. In this framework, multiple agents are assigned to each host and work together, improving the system's detection efficiency and robustness. The trainer agent in the central server of the framework uses the artificial neural network (ANN) rather than the negative selection algorithm of AIS to generate mature detectors. Mature detectors can distinguish between self-files and non-self-files after learning. Our analyzer agents use genetic algorithms to generate memory cell detectors. This kind of detector will effectively reduce false positive and false negative errors and act quickly on known intrusions.

Keywords: artificial immune system, distributed artificial intelligence, multi-agent, intrusion detection system, neural network

Procedia PDF Downloads 103
8087 Intrusion Detection System Based on Peer to Peer

Authors: Alireza Pour Ebrahimi, Vahid Abasi

Abstract:

Recently by the extension of internet usage, Research on the intrusion detection system takes a significant importance. Many of improvement systems prevent internal and external network attacks by providing security through firewalls and antivirus. In recently years, intrusion detection systems gradually turn from host-based systems and depend on O.S to the distributed systems which are running on multiple O.S. In this work, by considering the diversity of computer networks whit respect to structure, architecture, resource, services, users and also security goals requirement a fully distributed collaborative intrusion detection system based on peer to peer architecture is suggested. in this platform each partner device (matched device) considered as a peer-to-peer network. All transmitted information to network are visible only for device that use security scanning of a source. Experimental results show that the distributed architecture is significantly upgradeable in respect to centralized approach.

Keywords: network, intrusion detection system, peer to peer, internal and external network

Procedia PDF Downloads 540
8086 Rapid and Culture-Independent Detection of Staphylococcus Aureus by PCR Based Protocols

Authors: V. Verma, Syed Riyaz-ul-Hassan

Abstract:

Staphylococcus aureus is one of the most commonly found pathogenic bacteria and is hard to eliminate from the human environment. It is responsible for many nosocomial infections, besides being the main causative agent of food intoxication by virtue of its variety of enterotoxins. Routine detection of S. aureus in food is usually carried out by traditional methods based on morphological and biochemical characterization. These methods are time-consuming and tedious. In addition, misclassifications with automated susceptibility testing systems or commercially available latex agglutination kits have been reported by several workers. Consequently, there is a need for methods to specifically discriminate S. aureus from other staphylococci as quickly as possible. Data on protocols developed using molecular means like PCR technology will be presented for rapid and specific detection of this pathogen in food, clinical and environmental samples, especially milk.

Keywords: food Pathogens, PCR technology, rapid and specific detection, staphylococcus aureus

Procedia PDF Downloads 511
8085 Component Based Testing Using Clustering and Support Vector Machine

Authors: Iqbaldeep Kaur, Amarjeet Kaur

Abstract:

Software Reusability is important part of software development. So component based software development in case of software testing has gained a lot of practical importance in the field of software engineering from academic researcher and also from software development industry perspective. Finding test cases for efficient reuse of test cases is one of the important problems aimed by researcher. Clustering reduce the search space, reuse test cases by grouping similar entities according to requirements ensuring reduced time complexity as it reduce the search time for retrieval the test cases. In this research paper we proposed approach for re-usability of test cases by unsupervised approach. In unsupervised learning we proposed k-mean and Support Vector Machine. We have designed the algorithm for requirement and test case document clustering according to its tf-idf vector space and the output is set of highly cohesive pattern groups.

Keywords: software testing, reusability, clustering, k-mean, SVM

Procedia PDF Downloads 424
8084 Detecting Heartbeat Architectural Tactic in Source Code Using Program Analysis

Authors: Ananta Kumar Das, Sujit Kumar Chakrabarti

Abstract:

Architectural tactics such as heartbeat, ping-echo, encapsulate, encrypt data are techniques that are used to achieve quality attributes of a system. Detecting architectural tactics has several benefits: it can aid system comprehension (e.g., legacy systems) and in the estimation of quality attributes such as safety, security, maintainability, etc. Architectural tactics are typically spread over the source code and are implicit. For large codebases, manual detection is often not feasible. Therefore, there is a need for automated methods of detection of architectural tactics. This paper presents a formalization of the heartbeat architectural tactic and a program analytic approach to detect this tactic in source code. The experiment of the proposed method is done on a set of Java applications. The outcome of the experiment strongly suggests that the method compares well with a manual approach in terms of its sensitivity and specificity, and far supersedes a manual exercise in terms of its scalability.

Keywords: software architecture, architectural tactics, detecting architectural tactics, program analysis, AST, alias analysis

Procedia PDF Downloads 148
8083 Fabrication of Immune-Affinity Monolithic Array for Detection of α-Fetoprotein and Carcinoembryonic Antigen

Authors: Li Li, Li-Ru Xia, He-Ye Wang, Xiao-Dong Bi

Abstract:

In this paper, we presented a highly sensitive immune-affinity monolithic array for detection of α-fetoprotein (AFP) and carcinoembryonic antigen (CEA). Firstly, the epoxy functionalized monolith arrays were fabricated using UV initiated copolymerization method. Scanning electron microscopy (SEM) image showed that the poly(BABEA-co-GMA) monolith exhibited a well-controlled skeletal and well-distributed porous structure. Then, AFP and CEA immune-affinity monolithic arrays were prepared by immobilization of AFP and CEA antibodies on epoxy functionalized monolith arrays. With a non-competitive immune response format, the presented AFP and CEA immune-affinity arrays were demonstrated as an inexpensive, flexible, homogeneous and stable array for detection of AFP and CEA.

Keywords: chemiluminescent detection, immune-affinity, monolithic copolymer array, UV-initiated copolymerization

Procedia PDF Downloads 331
8082 Coping Strategies and Characterization of Vulnerability in the Perspective of Climate Change

Authors: Muhammad Umer Mehmood, Muhammad Luqman, Muhammad Yaseen, Imtiaz Hussain

Abstract:

Climate change is an arduous fact, which could not be unheeded easily. It is a phenomenon which has brought a collection of challenges for the mankind. Scientists have found many of its negative impacts on the life of human being and the resources on which the life of humanity is dependent. There are many issues which are associated with the factor of prime importance in this study, 'climate change'. Whenever changes happen in nature, they strike the whole globe. Effects of these changes vary from region to region. Climate of every region of this globe is different from the other. Even within a state, country or the province has different climatic conditions. So it is mandatory that the response in that specific region and the coping strategy of this specific region should be according to the prevailing risk. In the present study, the objective was to assess the coping strategies and vulnerability of small landholders. So that a professional suggestion could be made to cope with the vulnerability factor of small farmers. The cross-sectional research design was used with the intervention of quantitative approach. The study was conducted in the Khanewal district, of Punjab, Pakistan. 120 small farmers were interviewed after randomized sampling from the population of respective area. All respondents were above the age of 15 years. A questionnaire was developed after keen observation of facts in the respective area. Content and face validity of the instrument was assessed with SPSS and experts in the field. Data were analyzed through SPSS using descriptive statistics. From the sample of 120, 81.67% of the respondents claimed that the environment is getting warmer and not fit for their present agricultural practices. 84.17% of the sample expressed serious concern that they are disturbed due to change in rainfall pattern and vulnerability towards the climatic effects. On the other hand, they expressed that they are not good at tackling the effects of climate change. Adaptation of coping strategies like change in cropping pattern, use of resistant varieties, varieties with minimum water requirement, intercropping and tree planting was low by more than half of the sample. From the sample 63.33% small farmers said that the coping strategies they adopt are not effective enough. The present study showed that subsistence farming, lack of marketing and overall infrastructure, lack of access to social security networks, limited access to agriculture extension services, inappropriate access to agrometeorological system, unawareness and access to scientific development and low crop yield are the prominent factors which are responsible for the vulnerability of small farmers. A comprehensive study should be conducted at national level so that a national policy could be formulated to cope with the dilemma in future with relevance to climate change. Mainstreaming and collaboration among the researchers and academicians could prove beneficiary in this regard the interest of national leaders’ does matter. Proper policies to avoid the vulnerability factors should be the top priority. The world is taking up this issue with full responsibility as should we, keeping in view the local situation.

Keywords: adaptation, coping strategies, climate change, Pakistan, small farmers, vulnerability

Procedia PDF Downloads 135
8081 An Optimal Matching Design Method of Space-Based Optical Payload for Typical Aerial Target Detection

Authors: Yin Zhang, Kai Qiao, Xiyang Zhi, Jinnan Gong, Jianming Hu

Abstract:

In order to effectively detect aerial targets over long distances, an optimal matching design method of space-based optical payload is proposed. Firstly, main factors affecting optical detectability of small targets under complex environment are analyzed based on the full link of a detection system, including band center, band width and spatial resolution. Then a performance characterization model representing the relationship between image signal-to-noise ratio (SCR) and the above influencing factors is established to describe a detection system. Finally, an optimal matching design example is demonstrated for a typical aerial target by simulating and analyzing its SCR under different scene clutter coupling with multi-scale characteristics, and the optimized detection band and spatial resolution are presented. The method can provide theoretical basis and scientific guidance for space-based detection system design, payload specification demonstration and information processing algorithm optimization.

Keywords: space-based detection, aerial targets, optical system design, detectability characterization

Procedia PDF Downloads 163