Search results for: Extended Park´s vector approach
15760 Moving Object Detection Using Histogram of Uniformly Oriented Gradient
Authors: Wei-Jong Yang, Yu-Siang Su, Pau-Choo Chung, Jar-Ferr Yang
Abstract:
Moving object detection (MOD) is an important issue in advanced driver assistance systems (ADAS). There are two important moving objects, pedestrians and scooters in ADAS. In real-world systems, there exist two important challenges for MOD, including the computational complexity and the detection accuracy. The histogram of oriented gradient (HOG) features can easily detect the edge of object without invariance to changes in illumination and shadowing. However, to reduce the execution time for real-time systems, the image size should be down sampled which would lead the outlier influence to increase. For this reason, we propose the histogram of uniformly-oriented gradient (HUG) features to get better accurate description of the contour of human body. In the testing phase, the support vector machine (SVM) with linear kernel function is involved. Experimental results show the correctness and effectiveness of the proposed method. With SVM classifiers, the real testing results show the proposed HUG features achieve better than classification performance than the HOG ones.Keywords: moving object detection, histogram of oriented gradient, histogram of uniformly-oriented gradient, linear support vector machine
Procedia PDF Downloads 59515759 A Machine Learning Approach for Assessment of Tremor: A Neurological Movement Disorder
Authors: Rajesh Ranjan, Marimuthu Palaniswami, A. A. Hashmi
Abstract:
With the changing lifestyle and environment around us, the prevalence of the critical and incurable disease has proliferated. One such condition is the neurological disorder which is rampant among the old age population and is increasing at an unstoppable rate. Most of the neurological disorder patients suffer from some movement disorder affecting the movement of their body parts. Tremor is the most common movement disorder which is prevalent in such patients that infect the upper or lower limbs or both extremities. The tremor symptoms are commonly visible in Parkinson’s disease patient, and it can also be a pure tremor (essential tremor). The patients suffering from tremor face enormous trouble in performing the daily activity, and they always need a caretaker for assistance. In the clinics, the assessment of tremor is done through a manual clinical rating task such as Unified Parkinson’s disease rating scale which is time taking and cumbersome. Neurologists have also affirmed a challenge in differentiating a Parkinsonian tremor with the pure tremor which is essential in providing an accurate diagnosis. Therefore, there is a need to develop a monitoring and assistive tool for the tremor patient that keep on checking their health condition by coordinating them with the clinicians and caretakers for early diagnosis and assistance in performing the daily activity. In our research, we focus on developing a system for automatic classification of tremor which can accurately differentiate the pure tremor from the Parkinsonian tremor using a wearable accelerometer-based device, so that adequate diagnosis can be provided to the correct patient. In this research, a study was conducted in the neuro-clinic to assess the upper wrist movement of the patient suffering from Pure (Essential) tremor and Parkinsonian tremor using a wearable accelerometer-based device. Four tasks were designed in accordance with Unified Parkinson’s disease motor rating scale which is used to assess the rest, postural, intentional and action tremor in such patient. Various features such as time-frequency domain, wavelet-based and fast-Fourier transform based cross-correlation were extracted from the tri-axial signal which was used as input feature vector space for the different supervised and unsupervised learning tools for quantification of severity of tremor. A minimum covariance maximum correlation energy comparison index was also developed which was used as the input feature for various classification tools for distinguishing the PT and ET tremor types. An automatic system for efficient classification of tremor was developed using feature extraction methods, and superior performance was achieved using K-nearest neighbors and Support Vector Machine classifiers respectively.Keywords: machine learning approach for neurological disorder assessment, automatic classification of tremor types, feature extraction method for tremor classification, neurological movement disorder, parkinsonian tremor, essential tremor
Procedia PDF Downloads 15415758 A Study of Evolving Cloud Computing Data Security: A Machine Learning Perspective
Authors: Shinoy Vengaramkode Bhaskaran
Abstract:
The advancement of cloud computing led to a variety of security issues for both consumers and industries. Whereas machine learning (ML) is one approach to securing Cloud-based systems. Various methods have been employed to prevent or detect attacks and security vulnerabilities on the Cloud using ML techniques. In this paper, we present an ML perspective on the methodologies and techniques of cloud security. Initially, an investigative study on cloud computing is conducted with a primary emphasis on the gaps with two research questions that are impeding the adoption of cloud technology, as well as the challenges associated with threat solutions. Next, some ideas are generated based on machine learning methods to mitigate certain types of attacks that are frequently discussed through the application of ML techniques. Finally, we review different machine learning algorithms and their adoption in cloud computing.Keywords: artificial intelligence, machine learning, cloud computing infrastructure as a service, support vector machine, platform as a service
Procedia PDF Downloads 1015757 The Fiscal-Monetary Policy and Economic Growth in Algeria: VECM Approach
Authors: K. Bokreta, D. Benanaya
Abstract:
The objective of this study is to examine the relative effectiveness of monetary and fiscal policy in Algeria using the econometric modelling techniques of cointegration and vector error correction modelling to analyse and draw policy inferences. The chosen variables of fiscal policy are government expenditure and net taxes on products, while the effect of monetary policy is presented by the inflation rate and the official exchange rate. From the results, we find that in the long-run, the impact of government expenditures is positive, while the effect of taxes is negative on growth. Additionally, we find that the inflation rate is found to have little effect on GDP per capita but the impact of the exchange rate is insignificant. We conclude that fiscal policy is more powerful then monetary policy in promoting economic growth in Algeria.Keywords: economic growth, monetary policy, fiscal policy, VECM
Procedia PDF Downloads 31115756 A General Framework for Knowledge Discovery from Echocardiographic and Natural Images
Authors: S. Nandagopalan, N. Pradeep
Abstract:
The aim of this paper is to propose a general framework for storing, analyzing, and extracting knowledge from two-dimensional echocardiographic images, color Doppler images, non-medical images, and general data sets. A number of high performance data mining algorithms have been used to carry out this task. Our framework encompasses four layers namely physical storage, object identification, knowledge discovery, user level. Techniques such as active contour model to identify the cardiac chambers, pixel classification to segment the color Doppler echo image, universal model for image retrieval, Bayesian method for classification, parallel algorithms for image segmentation, etc., were employed. Using the feature vector database that have been efficiently constructed, one can perform various data mining tasks like clustering, classification, etc. with efficient algorithms along with image mining given a query image. All these facilities are included in the framework that is supported by state-of-the-art user interface (UI). The algorithms were tested with actual patient data and Coral image database and the results show that their performance is better than the results reported already.Keywords: active contour, Bayesian, echocardiographic image, feature vector
Procedia PDF Downloads 44615755 Dynamic Modeling of the Exchange Rate in Tunisia: Theoretical and Empirical Study
Authors: Chokri Slim
Abstract:
The relative failure of simultaneous equation models in the seventies has led researchers to turn to other approaches that take into account the dynamics of economic and financial systems. In this paper, we use an approach based on vector autoregressive model that is widely used in recent years. Their popularity is due to their flexible nature and ease of use to produce models with useful descriptive characteristics. It is also easy to use them to test economic hypotheses. The standard econometric techniques assume that the series studied are stable over time (stationary hypothesis). Most economic series do not verify this hypothesis, which assumes, when one wishes to study the relationships that bind them to implement specific techniques. This is cointegration which characterizes non-stationary series (integrated) with a linear combination is stationary, will also be presented in this paper. Since the work of Johansen, this approach is generally presented as part of a multivariate analysis and to specify long-term stable relationships while at the same time analyzing the short-term dynamics of the variables considered. In the empirical part, we have applied these concepts to study the dynamics of of the exchange rate in Tunisia, which is one of the most important economic policy of a country open to the outside. According to the results of the empirical study by the cointegration method, there is a cointegration relationship between the exchange rate and its determinants. This relationship shows that the variables have a significant influence in determining the exchange rate in Tunisia.Keywords: stationarity, cointegration, dynamic models, causality, VECM models
Procedia PDF Downloads 36615754 A Framework for Defining Innovation Districts: A Case Study of 22@ Barcelona
Authors: Arnault Morisson
Abstract:
Innovation districts are being implemented as urban regeneration strategies in cities as diverse as Barcelona (Spain), Boston (Massachusetts), Chattanooga (Tennessee), Detroit (Michigan), Medellin (Colombia), and Montréal (Canada). Little, however, is known about the concept. This paper aims to provide a framework to define innovation districts. The research methodology is based on a qualitative approach using 22@ Barcelona as a case study. 22@ Barcelona was the first innovation district ever created and has been a model for the innovation districts of Medellin (Colombia) and Boston (Massachusetts) among others. Innovation districts based on the 22@ Barcelona’s model can be defined as top-down urban innovation ecosystems designed around four multilayered and multidimensional models of innovation: urban planning, productive, collaborative, and creative, all coordinated under strong leadership, with the ultimate objectives to accelerate the innovation process and competitiveness of a locality. Innovation districts aim to respond to a new economic paradigm in which economic production flows back to cities.Keywords: innovation ecosystem, governance, technology park, urban planning, urban policy, urban regeneration
Procedia PDF Downloads 37415753 The Design Optimization for Sound Absorption Material of Multi-Layer Structure
Authors: Un-Hwan Park, Jun-Hyeok Heo, In-Sung Lee, Tae-Hyeon Oh, Dae-Kyu Park
Abstract:
Sound absorbing material is used as automotive interior material. Sound absorption coefficient should be predicted to design it. But it is difficult to predict sound absorbing coefficient because it is comprised of several material layers. So, its targets are achieved through many experimental tunings. It causes a lot of cost and time. In this paper, we propose the process to estimate the sound absorption coefficient with multi-layer structure. In order to estimate the coefficient, physical properties of each material are used. These properties also use predicted values by Foam-X software using the sound absorption coefficient data measured by impedance tube. Since there are many physical properties and the measurement equipment is expensive, the values predicted by software are used. Through the measurement of the sound absorption coefficient of each material, its physical properties are calculated inversely. The properties of each material are used to calculate the sound absorption coefficient of the multi-layer material. Since the absorption coefficient of multi-layer can be calculated, optimization design is possible through simulation. Then, we will compare and analyze the calculated sound absorption coefficient with the data measured by scaled reverberation chamber and impedance tubes for a prototype. If this method is used when developing automotive interior materials with multi-layer structure, the development effort can be reduced because it can be optimized by simulation. So, cost and time can be saved.Keywords: sound absorption material, sound impedance tube, sound absorption coefficient, optimization design
Procedia PDF Downloads 29115752 An Approach Based on Statistics and Multi-Resolution Representation to Classify Mammograms
Authors: Nebi Gedik
Abstract:
One of the significant and continual public health problems in the world is breast cancer. Early detection is very important to fight the disease, and mammography has been one of the most common and reliable methods to detect the disease in the early stages. However, it is a difficult task, and computer-aided diagnosis (CAD) systems are needed to assist radiologists in providing both accurate and uniform evaluation for mass in mammograms. In this study, a multiresolution statistical method to classify mammograms as normal and abnormal in digitized mammograms is used to construct a CAD system. The mammogram images are represented by wave atom transform, and this representation is made by certain groups of coefficients, independently. The CAD system is designed by calculating some statistical features using each group of coefficients. The classification is performed by using support vector machine (SVM).Keywords: wave atom transform, statistical features, multi-resolution representation, mammogram
Procedia PDF Downloads 22315751 Disaster Adaptation Mechanism and Disaster Prevention Adaptation Planning Strategies for Industrial Parks in Response to Climate Change and Different Socio-Economic Disasters
Authors: Jen-Te Pai, Jao-Heng Liu, Shin-En Pai
Abstract:
The impact of climate change has intensified in recent years, causing Taiwan to face higher frequency and serious natural disasters. Therefore, it is imperative for industrial parks manufacturers to promote adaptation policies in response to climate change. On the other hand, with the rise of the international anti-terrorism situation, once a terrorist attack occurs, it will attract domestic and international media attention, especially the strategic and economic status of the science park. Thus, it is necessary to formulate adaptation and mitigation strategies under climate change and social economic disasters. After reviewed the literature about climate change, urban disaster prevention, vulnerability assessment, and risk communication, the study selected 62 industrial parks compiled by the Industrial Bureau of the Ministry of Economic Affairs of Taiwan as the research object. This study explored the vulnerability and disaster prevention and disaster relief functional assessment of these industrial parks facing of natural and socio-economic disasters. Furthermore, this study explored planned adaptation of industrial parks management section and autonomous adaptation of corporate institutions in the park. The conclusion of this study is that Taiwan industrial parks with a higher vulnerability to natural and socio-economic disasters should employ positive adaptive behaviours.Keywords: adaptive behaviours, analytic network process, vulnerability, industrial parks
Procedia PDF Downloads 14815750 Bifid Ureters: Arising Directly from the Separate Calyces and Renal Pelvis of the Kidney: A Case Report
Authors: Yuri Seu, Hyun Jin Park, Jin Seo Park, Yong-Suk Moon, HongtaeKim, Mi-Sun Hur
Abstract:
The present case report describes bifid ureters arising directly from the separate calyces and renal pelvis of the kidney. It was a single common ureter leading away from the bladder, which was separated into incompletely duplicated ureters near the level of the anterior superior iliac supine. These two branches then entered the left kidney through their own courses. Each ureter traveled anterior and posterior to the renal vein, respectively. These two ureters formed a Y-shaped pattern. One ureter coursed anterior to the renal vein with shorter length, and it terminated at the renal pelvis that was divided into major calices in approximately lower two thirds of the kidney. The other ureter coursed posterior to the renal vein with longer length, terminating at approximately the upper third of the kidney. The renal calices in the upper third of the kidney were directly connected to the posterior ureter, whereas the other major calices in the lower two thirds of the kidney formed the renal pelvis connecting to the anterior ureter. Thus, convergence of the major calices was separated according to the terminations of two ureters. These anomalous ureters were traced to the calices of the kidney, thereby providing a reference of a rare variation of the ureter. The bifid ureters arising from the separate calyces and renal pelvis should be considered by radiologists when evaluating images and diagnosing possible complications of these anomalies.Keywords: bifid ureters, kidney, major calices, renal pelvis
Procedia PDF Downloads 8715749 Analytical Design of IMC-PID Controller for Ideal Decoupling Embedded in Multivariable Smith Predictor Control System
Authors: Le Hieu Giang, Truong Nguyen Luan Vu, Le Linh
Abstract:
In this paper, the analytical tuning rules of IMC-PID controller are presented for the multivariable Smith predictor that involved the ideal decoupling. Accordingly, the decoupler is first introduced into the multivariable Smith predictor control system by a well-known approach of ideal decoupling, which is compactly extended for general nxn multivariable processes and the multivariable Smith predictor controller is then obtained in terms of the multiple single-loop Smith predictor controllers. The tuning rules of PID controller in series with filter are found by using Maclaurin approximation. Many multivariable industrial processes are employed to demonstrate the simplicity and effectiveness of the presented method. The simulation results show the superior performances of presented method in compared with the other methods.Keywords: ideal decoupler, IMC-PID controller, multivariable smith predictor, Padé approximation
Procedia PDF Downloads 42115748 From Linear to Circular Model: An Artificial Intelligence-Powered Approach in Fosso Imperatore
Authors: Carlotta D’Alessandro, Giuseppe Ioppolo, Katarzyna Szopik-Depczyńska
Abstract:
— The growing scarcity of resources and the mounting pressures of climate change, water pollution, and chemical contamination have prompted societies, governments, and businesses to seek ways to minimize their environmental impact. To combat climate change, and foster sustainability, Industrial Symbiosis (IS) offers a powerful approach, facilitating the shift toward a circular economic model. IS has gained prominence in the European Union's policy framework as crucial enabler of resource efficiency and circular economy practices. The essence of IS lies in the collaborative sharing of resources such as energy, material by-products, waste, and water, thanks to geographic proximity. It can be exemplified by eco-industrial parks (EIPs), which are natural environments for boosting cooperation and resource sharing between businesses. EIPs are characterized by group of businesses situated in proximity, connected by a network of both cooperative and competitive interactions. They represent a sustainable industrial model aimed at reducing resource use, waste, and environmental impact while fostering economic and social wellbeing. IS, combined with Artificial Intelligence (AI)-driven technologies, can further optimize resource sharing and efficiency within EIPs. This research, supported by the “CE_IPs” project, aims to analyze the potential for IS and AI, in advancing circularity and sustainability at Fosso Imperatore. The Fosso Imperatore Industrial Park in Nocera Inferiore, Italy, specializes in agriculture and the industrial transformation of agricultural products, particularly tomatoes, tobacco, and textile fibers. This unique industrial cluster, centered around tomato cultivation and processing, also includes mechanical engineering enterprises and agricultural packaging firms. To stimulate the shift from a traditional to a circular economic model, an AI-powered Local Development Plan (LDP) is developed for Fosso Imperatore. It can leverage data analytics, predictive modeling, and stakeholder engagement to optimize resource utilization, reduce waste, and promote sustainable industrial practices. A comprehensive SWOT analysis of the AI-powered LDP revealed several key factors influencing its potential success and challenges. Among the notable strengths and opportunities arising from AI implementation are reduced processing times, fewer human errors, and increased revenue generation. Furthermore, predictive analytics minimize downtime, bolster productivity, and elevate quality while mitigating workplace hazards. However, the integration of AI also presents potential weaknesses and threats, including significant financial investment, since implementing and maintaining AI systems can be costly. The widespread adoption of AI could lead to job losses in certain sectors. Lastly, AI systems are susceptible to cyberattacks, posing risks to data security and operational continuity. Moreover, an Analytic Hierarchy Process (AHP) analysis was employed to yield a prioritized ranking of the outlined AI-driven LDP practices based on the stakeholder input, ensuring a more comprehensive and representative understanding of their relative significance for achieving sustainability in Fosso Imperatore Industrial Park. While this study provides valuable insights into the potential of AIpowered LDP at the Fosso Imperatore, it is important to note that the findings may not be directly applicable to all industrial parks, particularly those with different sizes, geographic locations, or industry compositions. Additional study is necessary to scrutinize the generalizability of these results and to identify best practices for implementing AI-driven LDP in diverse contexts.Keywords: artificial intelligence, climate change, Fosso Imperatore, industrial park, industrial symbiosis
Procedia PDF Downloads 3015747 Developing High-Definition Flood Inundation Maps (HD-Fims) Using Raster Adjustment with Scenario Profiles (RASPTM)
Authors: Robert Jacobsen
Abstract:
Flood inundation maps (FIMs) are an essential tool in communicating flood threat scenarios to the public as well as in floodplain governance. With an increasing demand for online raster FIMs, the FIM State-of-the-Practice (SOP) is rapidly advancing to meet the dual requirements for high-resolution and high-accuracy—or High-Definition. Importantly, today’s technology also enables the resolution of problems of local—neighborhood-scale—bias errors that often occur in FIMs, even with the use of SOP two-dimensional flood modeling. To facilitate the development of HD-FIMs, a new GIS method--Raster Adjustment with Scenario Profiles, RASPTM—is described for adjusting kernel raster FIMs to match refined scenario profiles. With RASPTM, flood professionals can prepare HD-FIMs for a wide range of scenarios with available kernel rasters, including kernel rasters prepared from vector FIMs. The paper provides detailed procedures for RASPTM, along with an example of applying RASPTM to prepare an HD-FIM for the August 2016 Flood in Louisiana using both an SOP kernel raster and a kernel raster derived from an older vector-based flood insurance rate map. The accuracy of the HD-FIMs achieved with the application of RASPTM to the two kernel rasters is evaluated.Keywords: hydrology, mapping, high-definition, inundation
Procedia PDF Downloads 8215746 Simulation-Based Investigation of Ferroresonance in Different Transformer Configurations
Authors: George Eduful, Yuanyuan Fan, Ahmed Abu-Siada
Abstract:
Ferroresonance poses a substantial threat to the quality and reliability of power distribution systems due to its inherent characteristics of sustained overvoltages and currents. This paper aims to enhance the understanding and reduce the ferroresonance threat by investigating the susceptibility of different transformer configurations using MATLAB/Simulink simulations. To achieve this, four 200 kVA transformers with different vector groups (D-Yn, Yg-Yg, Yn-Yn, and Y-D11) and core types (3-limb, 5-limb, single-phase) were systematically exposed to controlled ferroresonance conditions. The impact of varying the length of the 11 kV cable connected to the transformers was also examined. Through comprehensive voltage, current, and total harmonic distortion analyses, the performance of each configuration was evaluated and compared. The results of the study indicate that transformers with Y-D11 and Yg-Yg configurations exhibited lower susceptibility to ferroresonance, in comparison to those with D-Y11 and Yg-Yg configurations. This implies that the Y-D11 and Yg-Yg transformers are better suited for applications with high risks of ferroresonance. The insights provided by this study are of significant value for the strategic selection and deployment of transformers in power systems, particularly in settings prone to ferroresonance. By identifying and recommending transformer configurations that demonstrate better resilience, this paper contributes to enhancing the overall robustness and reliability of power grid infrastructure.Keywords: about cable-connected, core type, ferroresonance, over voltages, power transformer, vector group
Procedia PDF Downloads 4415745 Analytical Study Of Holographic Polymer Dispersed Liquid Crystals Using Finite Difference Time Domain Method
Authors: N. R. Mohamad, H. Ono, H. Haroon, A. Salleh, N. M. Z. Hashim
Abstract:
In this research, we have studied and analyzed the modulation of light and liquid crystal in HPDLCs using Finite Domain Time Difference (FDTD) method. HPDLCs are modeled as a mixture of polymer and liquid crystals (LCs) that categorized as an anisotropic medium. FDTD method is directly solves Maxwell’s equation with less approximation, so this method can analyze more flexible and general approach for the arbitrary anisotropic media. As the results from FDTD simulation, the highest diffraction efficiency occurred at ±19 degrees (Bragg angle) using p polarization incident beam to Bragg grating, Q > 10 when the pitch is 1µm. Therefore, the liquid crystal is assumed to be aligned parallel to the grating constant vector during these parameters.Keywords: birefringence, diffraction efficiency, finite domain time difference, nematic liquid crystals
Procedia PDF Downloads 46215744 Visitor Management in the National Parks: Recreational Carrying Capacity Assessment of Çıralı Coast, Turkey
Authors: Tendü H. Göktuğ, Gönül T. İçemer, Bülent Deniz
Abstract:
National parks, which are rich in natural and cultural resources values are protected in the context of the idea to develop sustainability, are among the most important recreated areas demanding with each passing day. Increasing recreational use or unplanned use forms negatively affect the resource values and visitor satisfaction. The intent of national parks management is to protect the natural and cultural resource values and to provide the visitors with a quality of recreational experience, as well. In this context, the current studies to improve the appropriate tourism and recreation planning and visitor management, approach have focused on recreational carrying capacity analysis. The aim of this study is to analyze recreational carrying capacity of Çıralı Coast in the Bey Mountains Coastal National Park to compare the analyze results with the current usage format and to develop alternative management strategies. In the first phase of the study, the annual and daily visitations, geographic, bio-physical, and managerial characteristics of the park and the type of recreational usage and the recreational areas were analyzed. In addition to these, ecological observations were carried out in order to determine recreational-based pressures on the ecosystems. On-site questionnaires were administrated to a sample of 284 respondents in the August 2015 - 2016 to collect data concerning the demographics and visit characteristics. The second phase of the study, the coastal area separated into four different usage zones and the methodology proposed by Cifuentes (1992) was used for capacity analyses. This method supplies the calculation of physical, real and effective carrying capacities by using environmental, ecological, climatic and managerial parameters in a formulation. Expected numbers which estimated three levels of carrying capacities were compared to current numbers of national parks’ visitors. In the study, it was determined that the current recreational uses in the north of the beach were caused by ecological pressures, and the current numbers in the south of beach much more than estimated numbers of visitors. Based on these results management strategies were defined and the appropriate management tools were developed in accordance with these strategies. The authors are grateful for the financial support of this project by The Scientific and Technological Research Council of Turkey (No: 114O344)Keywords: Çıralı Coast, national parks, recreational carrying capacity, visitor management
Procedia PDF Downloads 27615743 A Method for False Alarm Recognition Based on Multi-Classification Support Vector Machine
Authors: Weiwei Cui, Dejian Lin, Leigang Zhang, Yao Wang, Zheng Sun, Lianfeng Li
Abstract:
Built-in test (BIT) is an important technology in testability field, and it is widely used in state monitoring and fault diagnosis. With the improvement of modern equipment performance and complexity, the scope of BIT becomes larger, and it leads to the emergence of false alarm problem. The false alarm makes the health assessment unstable, and it reduces the effectiveness of BIT. The conventional false alarm suppression methods such as repeated test and majority voting cannot meet the requirement for a complicated system, and the intelligence algorithms such as artificial neural networks (ANN) are widely studied and used. However, false alarm has a very low frequency and small sample, yet a method based on ANN requires a large size of training sample. To recognize the false alarm, we propose a method based on multi-classification support vector machine (SVM) in this paper. Firstly, we divide the state of a system into three states: healthy, false-alarm, and faulty. Then we use multi-classification with '1 vs 1' policy to train and recognize the state of a system. Finally, an example of fault injection system is taken to verify the effectiveness of the proposed method by comparing ANN. The result shows that the method is reasonable and effective.Keywords: false alarm, fault diagnosis, SVM, k-means, BIT
Procedia PDF Downloads 15715742 An Approach for the Assessment of Semi-Elliptical Surface Crack
Authors: Muhammad Naweed, Usman Tariq Murtaza, Waseem Siddique
Abstract:
A pallet body approach is a finite element-based computational approach used for the modeling and assessment of a three-dimensional surface crack. The approach is capable of inserting the crack in an engineering structure and generating high-quality hexahedral mesh in the cracked region of the structure. The approach is capable of computing the stress intensity factors along a semi-elliptical surface crack numerically. The objective of this work is to present that the stress intensity factors produced by the approach can be used with confidence for capturing the parameters during the fatigue crack growth.Keywords: pallet body approach, semi-elliptical surface crack, stress intensity factors, fatigue crack growth
Procedia PDF Downloads 10315741 A General Framework for Knowledge Discovery Using High Performance Machine Learning Algorithms
Authors: S. Nandagopalan, N. Pradeep
Abstract:
The aim of this paper is to propose a general framework for storing, analyzing, and extracting knowledge from two-dimensional echocardiographic images, color Doppler images, non-medical images, and general data sets. A number of high performance data mining algorithms have been used to carry out this task. Our framework encompasses four layers namely physical storage, object identification, knowledge discovery, user level. Techniques such as active contour model to identify the cardiac chambers, pixel classification to segment the color Doppler echo image, universal model for image retrieval, Bayesian method for classification, parallel algorithms for image segmentation, etc., were employed. Using the feature vector database that have been efficiently constructed, one can perform various data mining tasks like clustering, classification, etc. with efficient algorithms along with image mining given a query image. All these facilities are included in the framework that is supported by state-of-the-art user interface (UI). The algorithms were tested with actual patient data and Coral image database and the results show that their performance is better than the results reported already.Keywords: active contour, bayesian, echocardiographic image, feature vector
Procedia PDF Downloads 42115740 Interdisciplinary Collaborative Innovation Mechanism for Sustainability Challenges
Authors: C. Park, H. Lee, Y-J. Lee
Abstract:
Aim: This study presents Interdisciplinary Collaborative Innovation Mechanism as a medium to enable the effective generation of innovations for sustainability challenges facing humanities. Background: Interdisciplinary approach of fusing disparate knowledge and perspectives from diverse expertise and subject areas is one of the key requirements to address the intricate nature of sustainability issues. There is a lack of rigorous empirical study of the systematic structure of interdisciplinary collaborative innovation for sustainability to date. Method: To address this research gap, the action research approach is adopted to develop the Interdisciplinary Collaborative Innovation Mechanism (ICIM) framework based on an empirical study of a total of 28 open innovation competitions in the format of MAKEathons between 2016 to 2023. First, the conceptual framework was formulated based on the literature findings, and the framework was subsequently tested and iterated. Outcomes: The findings provide the ICIM framework composed of five elements: Discipline Diversity Quadruple; Systematic Structure; Inspirational Stimuli; Supportive Collaboration Environment; and Hardware and Intellectual Support. The framework offers a discussion of the key elements when attempting to facilitate interdisciplinary collaboration for sustainability innovation. Contributions: This study contributes to two burgeoning areas of sustainable development and open innovation studies by articulating the concrete structure to bridge the gap. In practice, the framework helps facilitate effective innovation processes and positive social and environmental impact created for real-world sustainability challenges.Keywords: action research, interdisciplinary collaboration, open innovation, problem-solving, sustainable development, sustainability challenges
Procedia PDF Downloads 24815739 Pharmaceutical Applications of Newton's Second Law and Disc Inertia
Authors: Nicholas Jensen
Abstract:
As the effort to create new drugs to treat rare conditions cost-effectively intensifies, there is a need to ensure maximum efficiency in the manufacturing process. This includes the creation of ultracompact treatment forms, which can best be achieved via applications of fundamental laws of physics. This paper reports an experiment exploring the relationship between the forms of Newton's 2ⁿᵈ Law appropriate to linear motion and to transversal architraves. The moment of inertia of three discs was determined by experiments and compared with previous data derived from a theoretical relationship. The method used was to attach the discs to a moment arm. Comparing the results with those obtained from previous experiments, it is found to be consistent with the first law of thermodynamics. It was further found that Newton's 2ⁿᵈ law violates the second law of thermodynamics. The purpose of this experiment was to explore the relationship between the forms of Newton's 2nd Law appropriate to linear motion and to apply torque to a twisting force, which is determined by position vector r and force vector F. Substituting equation alpha in place of beta; angular acceleration is a linear acceleration divided by radius r of the moment arm. The nevrological analogy of Newton's 2nd Law states that these findings can contribute to a fuller understanding of thermodynamics in relation to viscosity. Implications for the pharmaceutical industry will be seen to be fruitful from these findings.Keywords: Newtonian physics, inertia, viscosity, pharmaceutical applications
Procedia PDF Downloads 11815738 Temporal Change in Bonding Strength and Antimicrobial Effect of a Zirconia after Nonthermal Atmospheric Pressure Plasma Treatment
Authors: Chan Park, Sang-Won Park, Kwi-Dug Yun, Hyun-Pil Lim
Abstract:
Purpose: Plasma treatment under various conditions has been studied to increase the bonding strength and surface sterilization of dental ceramic materials. We assessed the evolution of the shear bond strength (SBS) and antimicrobial effect of nonthermal atmospheric pressure plasma (NTAPP) treatment over time. Methods: Presintered zirconia specimens were manufactured as discs (diameter: 15 mm, height: 2 mm) after final sintering. The specimens then received a 30-min treatment with argon gas (Ar², 99.999%; 10 L/min) using an NTAPP device. Five post-treatment intervals were evaluated: control (no treatment), P0 (within 1 h), P1 (24 h), P2 (48 h), and P3 (72 h). This study investigated the surface characteristics, SBS of two different resin cement (RelyXTM U200 self-adhesive resin cement, Panavia F2.0 methacryloyloxydecyl dihydrogen phosphate (MDP)-based resin cement), and Streptococcus mutans biofilm formation. Results: The SBS of RelyXTM U200 increased significantly (p < 0.05) within 2 days following plasma treatment (P0, P1, P2). For Panavia F 2.0, a significant decrease (p < 0.05) was detected only in the group that had undergone cementation immediately after plasma treatment (P0). S. mutans adhesion decreased significantly (p < 0.05) within 2 days of plasma treatment (P0, P1, P2) compared to the control group. The P0 group displayed a lower biofilm thickness than the P1 and P2 groups (p < 0.05). Conclusions: After NTAPP treatment of zirconia, the effects on bonding strength and antimicrobial growth persist for a limited duration. The effect of NTAPP treatment on bonding strength depends on the resin cement.Keywords: NTAPP, SBS, antimicrobial effect, zirconia
Procedia PDF Downloads 24415737 Dengue Virus Infection Rate in Mosquitoes Collected in Thailand Related to Environmental Factors
Authors: Chanya Jetsukontorn
Abstract:
Dengue hemorrhagic fever is the most important Mosquito-borne disease and the major public health problem in Thailand. The most important vector is Aedes aegypti. Environmental factors such as temperature, relative humidity, and biting rate affect dengue virus infection. The most effective measure for prevention is controlling of vector mosquitoes. In addition, surveillance of field-caught mosquitoes is imperative for determining the natural vector and can provide an early warning sign at risk of transmission in an area. In this study, Aedes aegypti mosquitoes were collected in Amphur Muang, Phetchabun Province, Thailand. The mosquitoes were collected in the rainy season and the dry season both indoor and outdoor. During mosquito’s collection, the data of environmental factors such as temperature, humidity and breeding sites were observed and recorded. After identified to species, mosquitoes were pooled according to genus/species, and sampling location. Pools consisted of a maximum of 10 Aedes mosquitoes. 70 pools of 675 Aedes aegypti were screened with RT-PCR for flaviviruses. To confirm individual infection for determining True infection rate, individual mosquitoes which gave positive results of flavivirus detection were tested for dengue virus by RT-PCR. The infection rate was 5.93% (4 positive individuals from 675 mosquitoes). The probability to detect dengue virus in mosquitoes at the neighbour’s houses was 1.25 times, especially where distances between neighboring houses and patient’s houses were less than 50 meters. The relative humidity in dengue-infected villages with dengue-infected mosquitoes was significantly higher than villages that free from dengue-infected mosquitoes. Indoor biting rate of Aedes aegypti was 14.87 times higher than outdoor, and biting times of 09.00-10.00, 10.00-11.00, 11.00-12.00 yielded 1.77, 1.46, 0.68mosquitoes/man-hour, respectively. These findings confirm environmental factors were related to Dengue infection in Thailand. Data obtained from this study will be useful for the prevention and control of the diseases.Keywords: Aedes aegypti, Dengue virus, environmental factors, one health, PCR
Procedia PDF Downloads 14615736 Analysing Time Series for a Forecasting Model to the Dynamics of Aedes Aegypti Population Size
Authors: Flavia Cordeiro, Fabio Silva, Alvaro Eiras, Jose Luiz Acebal
Abstract:
Aedes aegypti is present in the tropical and subtropical regions of the world and is a vector of several diseases such as dengue fever, yellow fever, chikungunya, zika etc. The growth in the number of arboviruses cases in the last decades became a matter of great concern worldwide. Meteorological factors like mean temperature and precipitation are known to influence the infestation by the species through effects on physiology and ecology, altering the fecundity, mortality, lifespan, dispersion behaviour and abundance of the vector. Models able to describe the dynamics of the vector population size should then take into account the meteorological variables. The relationship between meteorological factors and the population dynamics of Ae. aegypti adult females are studied to provide a good set of predictors to model the dynamics of the mosquito population size. The time-series data of capture of adult females of a public health surveillance program from the city of Lavras, MG, Brazil had its association with precipitation, humidity and temperature analysed through a set of statistical methods for time series analysis commonly adopted in Signal Processing, Information Theory and Neuroscience. Cross-correlation, multicollinearity test and whitened cross-correlation were applied to determine in which time lags would occur the influence of meteorological variables on the dynamics of the mosquito abundance. Among the findings, the studied case indicated strong collinearity between humidity and precipitation, and precipitation was selected to form a pair of descriptors together with temperature. In the techniques used, there were observed significant associations between infestation indicators and both temperature and precipitation in short, mid and long terms, evincing that those variables should be considered in entomological models and as public health indicators. A descriptive model used to test the results exhibits a strong correlation to data.Keywords: Aedes aegypti, cross-correlation, multicollinearity, meteorological variables
Procedia PDF Downloads 18115735 Comparison of Crossover Types to Obtain Optimal Queries Using Adaptive Genetic Algorithm
Authors: Wafa’ Alma'Aitah, Khaled Almakadmeh
Abstract:
this study presents an information retrieval system of using genetic algorithm to increase information retrieval efficiency. Using vector space model, information retrieval is based on the similarity measurement between query and documents. Documents with high similarity to query are judge more relevant to the query and should be retrieved first. Using genetic algorithms, each query is represented by a chromosome; these chromosomes are fed into genetic operator process: selection, crossover, and mutation until an optimized query chromosome is obtained for document retrieval. Results show that information retrieval with adaptive crossover probability and single point type crossover and roulette wheel as selection type give the highest recall. The proposed approach is verified using (242) proceedings abstracts collected from the Saudi Arabian national conference.Keywords: genetic algorithm, information retrieval, optimal queries, crossover
Procedia PDF Downloads 29415734 Identifying Protein-Coding and Non-Coding Regions in Transcriptomes
Authors: Angela U. Makolo
Abstract:
Protein-coding and Non-coding regions determine the biology of a sequenced transcriptome. Research advances have shown that Non-coding regions are important in disease progression and clinical diagnosis. Existing bioinformatics tools have been targeted towards Protein-coding regions alone. Therefore, there are challenges associated with gaining biological insights from transcriptome sequence data. These tools are also limited to computationally intensive sequence alignment, which is inadequate and less accurate to identify both Protein-coding and Non-coding regions. Alignment-free techniques can overcome the limitation of identifying both regions. Therefore, this study was designed to develop an efficient sequence alignment-free model for identifying both Protein-coding and Non-coding regions in sequenced transcriptomes. Feature grouping and randomization procedures were applied to the input transcriptomes (37,503 data points). Successive iterations were carried out to compute the gradient vector that converged the developed Protein-coding and Non-coding Region Identifier (PNRI) model to the approximate coefficient vector. The logistic regression algorithm was used with a sigmoid activation function. A parameter vector was estimated for every sample in 37,503 data points in a bid to reduce the generalization error and cost. Maximum Likelihood Estimation (MLE) was used for parameter estimation by taking the log-likelihood of six features and combining them into a summation function. Dynamic thresholding was used to classify the Protein-coding and Non-coding regions, and the Receiver Operating Characteristic (ROC) curve was determined. The generalization performance of PNRI was determined in terms of F1 score, accuracy, sensitivity, and specificity. The average generalization performance of PNRI was determined using a benchmark of multi-species organisms. The generalization error for identifying Protein-coding and Non-coding regions decreased from 0.514 to 0.508 and to 0.378, respectively, after three iterations. The cost (difference between the predicted and the actual outcome) also decreased from 1.446 to 0.842 and to 0.718, respectively, for the first, second and third iterations. The iterations terminated at the 390th epoch, having an error of 0.036 and a cost of 0.316. The computed elements of the parameter vector that maximized the objective function were 0.043, 0.519, 0.715, 0.878, 1.157, and 2.575. The PNRI gave an ROC of 0.97, indicating an improved predictive ability. The PNRI identified both Protein-coding and Non-coding regions with an F1 score of 0.970, accuracy (0.969), sensitivity (0.966), and specificity of 0.973. Using 13 non-human multi-species model organisms, the average generalization performance of the traditional method was 74.4%, while that of the developed model was 85.2%, thereby making the developed model better in the identification of Protein-coding and Non-coding regions in transcriptomes. The developed Protein-coding and Non-coding region identifier model efficiently identified the Protein-coding and Non-coding transcriptomic regions. It could be used in genome annotation and in the analysis of transcriptomes.Keywords: sequence alignment-free model, dynamic thresholding classification, input randomization, genome annotation
Procedia PDF Downloads 6815733 Direct and Indirect Impacts of Predator Conflict in Kanha National Park, India
Authors: Diane H. Dotson, Shari L. Rodriguez
Abstract:
Habitat for predators is on the decline worldwide, which often brings humans and predators into conflict over remaining shared space and common resources. While the direct impacts of human predator conflict on humans (i.e., attacks on livestock or humans resulting in injury or death) are well documented, the indirect impacts of conflict on humans (i.e., downstream effects such as fear, stress, opportunity costs, PTSD) have not been addressed. We interviewed 437 people living in 54 villages on the periphery of Kanha National Park, India, to assess the amount and severity of direct and indirect impacts of predator conflict. While 58% of livestock owners believed that predator attacks on livestock guards occurred frequently and 62% of those who collect forest products believed that predator attacks on those collecting occurred frequently, less than 20% of all participants knew of someone who had experienced an attack. Data related to indirect impacts suggest that such impacts are common; 76% of participants indicated they were afraid a predator will physically injure them. Livestock owners reported that livestock guarding took time away from their primary job (61%) and getting enough sleep (73%), and believed that it increased their vulnerability to illnesses (80%). These results suggest that the perceptions of risk of predator attack are likely inflated, yet the costs of human predator impacts may be substantially higher than previously estimated, particularly related to human well-being, making the implementation of appropriate and effective conservation and conflict mitigation strategies and policies increasingly urgent.Keywords: direct impacts, indirect impacts, human-predator conflict, India
Procedia PDF Downloads 15615732 Machine Learning for Aiding Meningitis Diagnosis in Pediatric Patients
Authors: Karina Zaccari, Ernesto Cordeiro Marujo
Abstract:
This paper presents a Machine Learning (ML) approach to support Meningitis diagnosis in patients at a children’s hospital in Sao Paulo, Brazil. The aim is to use ML techniques to reduce the use of invasive procedures, such as cerebrospinal fluid (CSF) collection, as much as possible. In this study, we focus on predicting the probability of Meningitis given the results of a blood and urine laboratory tests, together with the analysis of pain or other complaints from the patient. We tested a number of different ML algorithms, including: Adaptative Boosting (AdaBoost), Decision Tree, Gradient Boosting, K-Nearest Neighbors (KNN), Logistic Regression, Random Forest and Support Vector Machines (SVM). Decision Tree algorithm performed best, with 94.56% and 96.18% accuracy for training and testing data, respectively. These results represent a significant aid to doctors in diagnosing Meningitis as early as possible and in preventing expensive and painful procedures on some children.Keywords: machine learning, medical diagnosis, meningitis detection, pediatric research
Procedia PDF Downloads 15115731 Normal and Peaberry Coffee Beans Classification from Green Coffee Bean Images Using Convolutional Neural Networks and Support Vector Machine
Authors: Hira Lal Gope, Hidekazu Fukai
Abstract:
The aim of this study is to develop a system which can identify and sort peaberries automatically at low cost for coffee producers in developing countries. In this paper, the focus is on the classification of peaberries and normal coffee beans using image processing and machine learning techniques. The peaberry is not bad and not a normal bean. The peaberry is born in an only single seed, relatively round seed from a coffee cherry instead of the usual flat-sided pair of beans. It has another value and flavor. To make the taste of the coffee better, it is necessary to separate the peaberry and normal bean before green coffee beans roasting. Otherwise, the taste of total beans will be mixed, and it will be bad. In roaster procedure time, all the beans shape, size, and weight must be unique; otherwise, the larger bean will take more time for roasting inside. The peaberry has a different size and different shape even though they have the same weight as normal beans. The peaberry roasts slower than other normal beans. Therefore, neither technique provides a good option to select the peaberries. Defect beans, e.g., sour, broken, black, and fade bean, are easy to check and pick up manually by hand. On the other hand, the peaberry pick up is very difficult even for trained specialists because the shape and color of the peaberry are similar to normal beans. In this study, we use image processing and machine learning techniques to discriminate the normal and peaberry bean as a part of the sorting system. As the first step, we applied Deep Convolutional Neural Networks (CNN) and Support Vector Machine (SVM) as machine learning techniques to discriminate the peaberry and normal bean. As a result, better performance was obtained with CNN than with SVM for the discrimination of the peaberry. The trained artificial neural network with high performance CPU and GPU in this work will be simply installed into the inexpensive and low in calculation Raspberry Pi system. We assume that this system will be used in under developed countries. The study evaluates and compares the feasibility of the methods in terms of accuracy of classification and processing speed.Keywords: convolutional neural networks, coffee bean, peaberry, sorting, support vector machine
Procedia PDF Downloads 145