Search results for: zernike moment feature descriptor
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 2441

Search results for: zernike moment feature descriptor

881 Visualization of Corrosion at Plate-Like Structures Based on Ultrasonic Wave Propagation Images

Authors: Aoqi Zhang, Changgil Lee Lee, Seunghee Park

Abstract:

A non-contact nondestructive technique using laser-induced ultrasonic wave generation method was applied to visualize corrosion damage at aluminum alloy plate structures. The ultrasonic waves were generated by a Nd:YAG pulse laser, and a galvanometer-based laser scanner was used to scan specific area at a target structure. At the same time, wave responses were measured at a piezoelectric sensor which was attached on the target structure. The visualization of structural damage was achieved by calculating logarithmic values of root mean square (RMS). Damage-sensitive feature was defined as the scattering characteristics of the waves that encounter corrosion damage. The corroded damage was artificially formed by hydrochloric acid. To observe the effect of the location where the corrosion was formed, the both sides of the plate were scanned with same scanning area. Also, the effect on the depth of the corrosion was considered as well as the effect on the size of the corrosion. The results indicated that the damages were successfully visualized for almost cases, whether the damages were formed at the front or back side. However, the damage could not be clearly detected because the depth of the corrosion was shallow. In the future works, it needs to develop signal processing algorithm to more clearly visualize the damage by improving signal-to-noise ratio.

Keywords: non-destructive testing, corrosion, pulsed laser scanning, ultrasonic waves, plate structure

Procedia PDF Downloads 300
880 PointNetLK-OBB: A Point Cloud Registration Algorithm with High Accuracy

Authors: Wenhao Lan, Ning Li, Qiang Tong

Abstract:

To improve the registration accuracy of a source point cloud and template point cloud when the initial relative deflection angle is too large, a PointNetLK algorithm combined with an oriented bounding box (PointNetLK-OBB) is proposed. In this algorithm, the OBB of a 3D point cloud is used to represent the macro feature of source and template point clouds. Under the guidance of the iterative closest point algorithm, the OBB of the source and template point clouds is aligned, and a mirror symmetry effect is produced between them. According to the fitting degree of the source and template point clouds, the mirror symmetry plane is detected, and the optimal rotation and translation of the source point cloud is obtained to complete the 3D point cloud registration task. To verify the effectiveness of the proposed algorithm, a comparative experiment was performed using the publicly available ModelNet40 dataset. The experimental results demonstrate that, compared with PointNetLK, PointNetLK-OBB improves the registration accuracy of the source and template point clouds when the initial relative deflection angle is too large, and the sensitivity of the initial relative position between the source point cloud and template point cloud is reduced. The primary contribution of this paper is the use of PointNetLK to avoid the non-convex problem of traditional point cloud registration and leveraging the regularity of the OBB to avoid the local optimization problem in the PointNetLK context.

Keywords: mirror symmetry, oriented bounding box, point cloud registration, PointNetLK-OBB

Procedia PDF Downloads 150
879 A Group Setting of IED in Microgrid Protection Management System

Authors: Jyh-Cherng Gu, Ming-Ta Yang, Chao-Fong Yan, Hsin-Yung Chung, Yung-Ruei Chang, Yih-Der Lee, Chen-Min Chan, Chia-Hao Hsu

Abstract:

There are a number of distributed generations (DGs) installed in microgrid, which may have diverse path and direction of power flow or fault current. The overcurrent protection scheme for the traditional radial type distribution system will no longer meet the needs of microgrid protection. Integrating the intelligent electronic device (IED) and a supervisory control and data acquisition (SCADA) with IEC 61850 communication protocol, the paper proposes a microgrid protection management system (MPMS) to protect power system from the fault. In the proposed method, the MPMS performs logic programming of each IED to coordinate their tripping sequence. The GOOSE message defined in IEC 61850 is used as the transmission information medium among IEDs. Moreover, to cope with the difference in fault current of microgrid between grid-connected mode and islanded mode, the proposed MPMS applies the group setting feature of IED to protect system and robust adaptability. Once the microgrid topology varies, the MPMS will recalculate the fault current and update the group setting of IED. Provided there is a fault, IEDs will isolate the fault at once. Finally, the Matlab/Simulink and Elipse Power Studio software are used to simulate and demonstrate the feasibility of the proposed method.

Keywords: IEC 61850, IED, group Setting, microgrid

Procedia PDF Downloads 463
878 Geology, Geomorphology and Genesis of Andarokh Karstic Cave, North-East Iran

Authors: Mojtaba Heydarizad

Abstract:

Andarokh basin is one of the main karstic regions in Khorasan Razavi province NE Iran. This basin is part of Kopeh-Dagh mega zone extending from Caspian Sea in the east to northern Afghanistan in the west. This basin is covered by Mozdooran Formation, Ngr evaporative formation and quaternary alluvium deposits in descending order of age. Mozdooran carbonate formation is notably karstified. The main surface karstic features in Mozdooran formation are Groove karren, Cleft karren, Rain pit, Rill karren, Tritt karren, Kamintza, Domes, and Table karren. In addition to surface features, deep karstic feature Andarokh Cave also exists in the region. Studying Ca, Mg, Mn, Sr, Fe concentration and Sr/Mn ratio in Mozdooran formation samples with distance to main faults and joints system using PCA analyses demonstrates intense meteoric digenesis role in controlling carbonate rock geochemistry. The karst evaluation in Andarokh basin varies from early stages 'deep seated karst' in Mesozoic to mature karstic system 'Exhumed karst' in quaternary period. Andarokh cave (the main cave in Andarokh basin) is rudimentary branch work consists of three passages of A, B and C and two entrances Andarokh and Sky.

Keywords: Andarokh basin, Andarokh cave, geochemical analyses, karst evaluation

Procedia PDF Downloads 154
877 Reed: An Approach Towards Quickly Bootstrapping Multilingual Acoustic Models

Authors: Bipasha Sen, Aditya Agarwal

Abstract:

Multilingual automatic speech recognition (ASR) system is a single entity capable of transcribing multiple languages sharing a common phone space. Performance of such a system is highly dependent on the compatibility of the languages. State of the art speech recognition systems are built using sequential architectures based on recurrent neural networks (RNN) limiting the computational parallelization in training. This poses a significant challenge in terms of time taken to bootstrap and validate the compatibility of multiple languages for building a robust multilingual system. Complex architectural choices based on self-attention networks are made to improve the parallelization thereby reducing the training time. In this work, we propose Reed, a simple system based on 1D convolutions which uses very short context to improve the training time. To improve the performance of our system, we use raw time-domain speech signals directly as input. This enables the convolutional layers to learn feature representations rather than relying on handcrafted features such as MFCC. We report improvement on training and inference times by atleast a factor of 4x and 7.4x respectively with comparable WERs against standard RNN based baseline systems on SpeechOcean's multilingual low resource dataset.

Keywords: convolutional neural networks, language compatibility, low resource languages, multilingual automatic speech recognition

Procedia PDF Downloads 123
876 Mechanism of Action of Troxerutin in Reducing Oxidative Stress

Authors: Nasrin Hosseinzad

Abstract:

Troxerutin, a trihydroxyethylated derived of rutin, is a flavonoid existing in tea, coffee, cereal grains, various fruits and vegetables have been conveyed to display radioprotective, antithrombotic, nephron-protective and hepato-protective possessions. Troxerutin, has been well-proved to utilize hepatoprotective assets. Troxerutin could upturn the resistance of hippocampal neurons alongside apoptosis by lessening the action of AChE and oxidative stress. Consequently, troxerutin may have advantageous properties in the administration of Alzheimer's disease and cancer. Troxerutin has been testified to have several welfares and medicinal stuffs. It could shelter the mouse kidney against d-gal-induced damage by refining renal utility, decreasing histopathologic changes, dropping ROS construction, reintroducing the activities of antioxidant enzymes and reducing DNA oxidative destruction. The DNA cleavage study clarifies that troxerutin showed DNA protection against hydroxyl radical persuaded DNA mutilation. Troxerutin uses anti-cancer effect in HuH-7 hepatocarcinoma cells conceivably through synchronized regulation of the molecular signalling pathways, Nrf2 and NF-κB. DNA binding at slight channel by troxerutin may have donated to feature breaks leading to improved radiation brought cell death. Furthermore, the mechanism principal the observed variance in the antioxidant activities of troxerutin and its esters was qualified to equally their free radical scavenging capabilities and dissemination on the cell membrane outward.

Keywords: troxerutin, DNA, oxidative stress, antioxidant, free radical

Procedia PDF Downloads 160
875 Modeling Optimal Lipophilicity and Drug Performance in Ligand-Receptor Interactions: A Machine Learning Approach to Drug Discovery

Authors: Jay Ananth

Abstract:

The drug discovery process currently requires numerous years of clinical testing as well as money just for a single drug to earn FDA approval. For drugs that even make it this far in the process, there is a very slim chance of receiving FDA approval, resulting in detrimental hurdles to drug accessibility. To minimize these inefficiencies, numerous studies have implemented computational methods, although few computational investigations have focused on a crucial feature of drugs: lipophilicity. Lipophilicity is a physical attribute of a compound that measures its solubility in lipids and is a determinant of drug efficacy. This project leverages Artificial Intelligence to predict the impact of a drug’s lipophilicity on its performance by accounting for factors such as binding affinity and toxicity. The model predicted lipophilicity and binding affinity in the validation set with very high R² scores of 0.921 and 0.788, respectively, while also being applicable to a variety of target receptors. The results expressed a strong positive correlation between lipophilicity and both binding affinity and toxicity. The model helps in both drug development and discovery, providing every pharmaceutical company with recommended lipophilicity levels for drug candidates as well as a rapid assessment of early-stage drugs prior to any testing, eliminating significant amounts of time and resources currently restricting drug accessibility.

Keywords: drug discovery, lipophilicity, ligand-receptor interactions, machine learning, drug development

Procedia PDF Downloads 111
874 Taylor’s Law and Relationship between Life Expectancy at Birth and Variance in Age at Death in Period Life Table

Authors: David A. Swanson, Lucky M. Tedrow

Abstract:

Taylor’s Law is a widely observed empirical pattern that relates variances to means in sets of non-negative measurements via an approximate power function, which has found application to human mortality. This study adds to this research by showing that Taylor’s Law leads to a model that reasonably describes the relationship between life expectancy at birth (e0, which also is equal to mean age at death in a life table) and variance at age of death in seven World Bank regional life tables measured at two points in time, 1970 and 2000. Using as a benchmark a non-random sample of four Japanese female life tables covering the period from 1950 to 2004, the study finds that the simple linear model provides reasonably accurate estimates of variance in age at death in a life table from e0, where the latter range from 60.9 to 85.59 years. Employing 2017 life tables from the Human Mortality Database, the simple linear model is used to provide estimates of variance at age in death for six countries, three of which have high e0 values and three of which have lower e0 values. The paper provides a substantive interpretation of Taylor’s Law relative to e0 and concludes by arguing that reasonably accurate estimates of variance in age at death in a period life table can be calculated using this approach, which also can be used where e0 itself is estimated rather than generated through the construction of a life table, a useful feature of the model.

Keywords: empirical pattern, mean age at death in a life table, mean age of a stationary population, stationary population

Procedia PDF Downloads 330
873 Rational Allocation of Resources in Water Infrastructure Development Projects

Authors: M. Macchiaroli, V. Pellecchia, L. Dolores

Abstract:

Within any European and world model of management of the integrated water service (in Italy only since 2012 is regulated by a national Authority, that is ARERA), a significant part is covered by the development of assets in terms of hydraulic networks and wastewater collection networks, including all their relative building works. The process of selecting the investments to be made starts from the preventive analysis of critical issues (water losses, unserved areas, low service standards, etc.) who occur in the managed territory of the Operator. Through the Program of Interventions (Provision by ARERA n. 580/2019/R/idr), the Operator provides to program the projects that can meet the emerged needs to determine the improvement of the water service levels. This phase (analyzed and solved by the author with a work published in 2019) involves the use of evaluation techniques (cost-benefit analysis, multi-criteria, and multi-objective techniques, neural networks, etc.) useful in selecting the most appropriate design answers to the different criticalities. However, at this point, the problem of establishing the time priorities between the various works deemed necessary remains open. That is, it is necessary to hierarchize the investments. In this decision-making moment, the interests of the private Operator are often opposed, which favors investments capable of generating high profitability, compared to those of the public controller (ARERA), which favors investments in greater social impact. In support of the concertation between these two actors, the protocol set out in the research has been developed, based on the AHP and capable of borrowing from the programmatic documents an orientation path for the settlement of the conflict. The protocol is applied to a case study of the Campania Region in Italy and has been professionally applied in the shared decision process between the manager and the local Authority.

Keywords: analytic hierarchy process, decision making, economic evaluation of projects, integrated water service

Procedia PDF Downloads 123
872 Determination of Crustal Structure and Moho Depth within the Jammu and Kashmir Region, Northwest Himalaya through Receiver Function

Authors: Shiv Jyoti Pandey, Shveta Puri, G. M. Bhat, Neha Raina

Abstract:

The Jammu and Kashmir (J&K) region of Northwest Himalaya has a long history of earthquake activity which falls within Seismic Zones IV and V. To know the crustal structure beneath this region, we utilized teleseismic receiver function method. This paper presents the results of the analyses of the teleseismic earthquake waves recorded by 10 seismic observatories installed in the vicinity of major thrusts and faults. The teleseismic waves at epicentral distance between 30o and 90o with moment magnitudes greater than or equal to 5.5 that contains large amount of information about the crust and upper mantle structure directly beneath a receiver has been used. The receiver function (RF) technique has been widely applied to investigate crustal structures using P-to-S converted (Ps) phases from velocity discontinuities. The arrival time of the Ps, PpPs and PpSs+ PsPs converted and reverberated phases from the Moho can be combined to constrain the mean crustal thickness and Vp/Vs ratio. Over 500 receiver functions from 10 broadband stations located in the Jammu & Kashmir region of Northwest Himalaya were analyzed. With the help of H-K stacking method, we determined the crustal thickness (H) and average crustal Vp/Vs ratio (K) in this region. We also used Neighbourhood algorithm technique to verify our results. The receiver function results for these stations show that the crustal thickness under Jammu & Kashmir ranges from 45.0 to 53.6 km with an average value of 50.01 km. The Vp/Vs ratio varies from 1.63 to 1.99 with an average value of 1.784 which corresponds to an average Poisson’s ratio of 0.266 with a range from 0.198 to 0.331. High Poisson’s ratios under some stations may be related to partial melting in the crust near the uppermost mantle. The crustal structure model developed from this study can be used to refine the velocity model used in the precise epicenter location in the region, thereby increasing the knowledge to understand current seismicity in the region.

Keywords: H-K stacking, Poisson’s ratios, receiver function, teleseismic

Procedia PDF Downloads 248
871 Optoelectronic Hardware Architecture for Recurrent Learning Algorithm in Image Processing

Authors: Abdullah Bal, Sevdenur Bal

Abstract:

This paper purposes a new type of hardware application for training of cellular neural networks (CNN) using optical joint transform correlation (JTC) architecture for image feature extraction. CNNs require much more computation during the training stage compare to test process. Since optoelectronic hardware applications offer possibility of parallel high speed processing capability for 2D data processing applications, CNN training algorithm can be realized using Fourier optics technique. JTC employs lens and CCD cameras with laser beam that realize 2D matrix multiplication and summation in the light speed. Therefore, in the each iteration of training, JTC carries more computation burden inherently and the rest of mathematical computation realized digitally. The bipolar data is encoded by phase and summation of correlation operations is realized using multi-object input joint images. Overlapping properties of JTC are then utilized for summation of two cross-correlations which provide less computation possibility for training stage. Phase-only JTC does not require data rearrangement, electronic pre-calculation and strict system alignment. The proposed system can be incorporated simultaneously with various optical image processing or optical pattern recognition techniques just in the same optical system.

Keywords: CNN training, image processing, joint transform correlation, optoelectronic hardware

Procedia PDF Downloads 506
870 Effect of Duration and Frequency on Ground Motion: Case Study of Guwahati City

Authors: Amar F. Siddique

Abstract:

The Guwahati city is one of the fastest growing cities of the north-eastern region of India, situated on the South Bank of the Brahmaputra River falls in the highest seismic zone level V. The city has witnessed many high magnitude earthquakes in the past decades. The Assam earthquake occurred on August 15, 1950, of moment magnitude 8.7 epicentered near Rima, Tibet was one of the major earthquakes which caused a serious structural damage and widespread soil liquefaction in and around the region. Hence the study of ground motion characteristics of Guwahati city is very essential. In this present work 1D equivalent linear ground response analysis (GRA) has been adopted using Deep soil software. The analysis has been done for two typical sites namely, Panbazar and Azara comprising total four boreholes location in Guwahati city of India. GRA of the sites is carried out by using an input motion recorded at Nongpoh station (recorded PGA 0.048g) and Nongstoin station (recorded PGA 0.047g) of 1997 Indo-Burma earthquake. In comparison to motion recorded at Nongpoh, different amplifications of bedrock peak ground acceleration (PGA) are obtained for all the boreholes by the motion recorded at Nongstoin station; although, the Fourier amplitude ratios (FAR) and fundamental frequencies remain almost same. The difference in recorded duration and frequency content of the two motions mainly influence the amplification of motions thus getting different surface PGA and amplification factor keeping a constant bedrock PGA. From the results of response spectra, it is found that at the period of less than 0.2 sec the ground motion recorded at Nongpoh station will give a high spectral acceleration (SA) on the structures than at Nongstoin station. Again for a period greater than 0.2 sec the ground motion recorded at Nongstoin station will give a high SA on the structures than at Nongpoh station.

Keywords: fourier amplitude ratio, ground response analysis, peak ground acceleration, spectral acceleration

Procedia PDF Downloads 179
869 Characteristics and Key Exploration Directions of Gold Deposits in China

Authors: Bin Wang, Yong Xu, Honggang Qu, Rongmei Liu, Zhenji Gao

Abstract:

Based on the geodynamic environment, basic geological characteristics of minerals and so on, gold deposits in China are divided into 11 categories, of which tectonic fracture altered rock, mid-intrudes and contact zone, micro-fine disseminated and continental volcanic types are the main prospecting kinds. The metallogenic age of gold deposits in China is dominated by the Mesozoic and Cenozoic. According to the geotectonic units, geological evolution, geological conditions, spatial distribution, gold deposits types, metallogenic factors etc., 42 gold concentration areas are initially determined and have a concentrated distribution feature. On the basis of the gold exploration density, gold concentration areas are divided into high, medium and low level areas. High ones are mainly distributed in the central and eastern regions. 93.04% of the gold exploration drillings are within 500 meters, but there are some problems, such as less and shallower of drilling verification etc.. The paper discusses the resource potentials of gold deposits and proposes the future prospecting directions and suggestions. The deep and periphery of old mines in the central and eastern regions and western area, especially in Xinjiang and Qinghai, will be the future key prospecting one and have huge potential gold reserves. If the exploration depth is extended to 2,000 meters shallow, the gold resources will double.

Keywords: gold deposits, gold deposits types, gold concentration areas, prospecting, resource potentiality

Procedia PDF Downloads 77
868 Gender Difference and Conflict Management Strategy Preference among Managers in Public Organizations in South-Western Nigeria

Authors: D. I. Akintayo, C. O. Aje

Abstract:

This study investigated the moderating influence of gender difference and conflict resolution strategy preference on managers` efficiency in managing industrial conflict in work organizations in South-Western Nigeria. This was for the purpose of ascertaining the relevance of gender difference and conflict resolution strategy preference to managerial efficiency towards ensuring sustainable industrial peace and harmonious labour-management relations at workplaces in Nigeria. Descriptive ex-post-facto research design was adopted for the study. A total of 185 respondents were selected for the study using purposive stratified sampling technique. A set of questionnaire titled ‘Rahim Organizational Conflict Inventory’ (ROCI) and Managerial Conflict Efficiency Scale (MCES) were adopted for the study. The three generated hypotheses were tested using Pearson Product Moment Correlation and t-test statistical methods. The findings of the study revealed that: A significant relationship exists between gender difference and conflict management preference of the managers(r = 0.644; P < 0.05). I t was also found that there was no significant difference between male and female managers’ conflict management strategy preference (t (181) = 11.08; P > 0.05).The finding reveals that there is no significant difference between female and male managers’ conflict management efficiency on the basis of conflict management preference of the managers (t (181) = 10.23; P > 0.05). Based on the findings of the study, it is recommended that collective bargaining strategy should be encouraged as conflict resolution strategy in order to guarantee effective management of industrial conflict and harmonious labour-management relations. Also, both male and female managers should be empowered to be appointed to managerial positions and should avoid the use of coercion, competition, aggressiveness and pro-task in the course of managing industrial conflict. Rather, persuasion, compromising, relational, lobbying and participatory approaches should be employed during collective bargaining process in order to foster effective management of conflict at workplaces.

Keywords: conflict management, gender difference, managerial studies, public organization and managers, strategy preference

Procedia PDF Downloads 459
867 Effective Parameter Selection for Audio-Based Music Mood Classification for Christian Kokborok Song: A Regression-Based Approach

Authors: Sanchali Das, Swapan Debbarma

Abstract:

Music mood classification is developing in both the areas of music information retrieval (MIR) and natural language processing (NLP). Some of the Indian languages like Hindi English etc. have considerable exposure in MIR. But research in mood classification in regional language is very less. In this paper, powerful audio based feature for Kokborok Christian song is identified and mood classification task has been performed. Kokborok is an Indo-Burman language especially spoken in the northeastern part of India and also some other countries like Bangladesh, Myanmar etc. For performing audio-based classification task, useful audio features are taken out by jMIR software. There are some standard audio parameters are there for the audio-based task but as known to all that every language has its unique characteristics. So here, the most significant features which are the best fit for the database of Kokborok song is analysed. The regression-based model is used to find out the independent parameters that act as a predictor and predicts the dependencies of parameters and shows how it will impact on overall classification result. For classification WEKA 3.5 is used, and selected parameters create a classification model. And another model is developed by using all the standard audio features that are used by most of the researcher. In this experiment, the essential parameters that are responsible for effective audio based mood classification and parameters that do not significantly change for each of the Christian Kokborok songs are analysed, and a comparison is also shown between the two above model.

Keywords: Christian Kokborok song, mood classification, music information retrieval, regression

Procedia PDF Downloads 222
866 Design and Development of 5-DOF Color Sorting Manipulator for Industrial Applications

Authors: Atef A. Ata, Sohair F. Rezeka, Ahmed El-Shenawy, Mohammed Diab

Abstract:

Image processing in today’s world grabs massive attentions as it leads to possibilities of broaden application in many fields of high technology. The real challenge is how to improve existing sorting system applications which consists of two integrated stations of processing and handling with a new image processing feature. Existing color sorting techniques use a set of inductive, capacitive, and optical sensors to differentiate object color. This research presents a mechatronics color sorting system solution with the application of image processing. A 5-DOF robot arm is designed and developed with pick and place operation to be main part of the color sorting system. Image processing procedure senses the circular objects in an image captured in real time by a webcam attached at the end-effector then extracts color and position information out of it. This information is passed as a sequence of sorting commands to the manipulator that has pick-and-place mechanism. Performance analysis proves that this color based object sorting system works very accurate under ideal condition in term of adequate illumination, circular objects shape and color. The circular objects tested for sorting are red, green and blue. For non-ideal condition, such as unspecified color the accuracy reduces to 80%.

Keywords: robotics manipulator, 5-DOF manipulator, image processing, color sorting, pick-and-place

Procedia PDF Downloads 374
865 Establishment of a Classifier Model for Early Prediction of Acute Delirium in Adult Intensive Care Unit Using Machine Learning

Authors: Pei Yi Lin

Abstract:

Objective: The objective of this study is to use machine learning methods to build an early prediction classifier model for acute delirium to improve the quality of medical care for intensive care patients. Background: Delirium is a common acute and sudden disturbance of consciousness in critically ill patients. After the occurrence, it is easy to prolong the length of hospital stay and increase medical costs and mortality. In 2021, the incidence of delirium in the intensive care unit of internal medicine was as high as 59.78%, which indirectly prolonged the average length of hospital stay by 8.28 days, and the mortality rate is about 2.22% in the past three years. Therefore, it is expected to build a delirium prediction classifier through big data analysis and machine learning methods to detect delirium early. Method: This study is a retrospective study, using the artificial intelligence big data database to extract the characteristic factors related to delirium in intensive care unit patients and let the machine learn. The study included patients aged over 20 years old who were admitted to the intensive care unit between May 1, 2022, and December 31, 2022, excluding GCS assessment <4 points, admission to ICU for less than 24 hours, and CAM-ICU evaluation. The CAMICU delirium assessment results every 8 hours within 30 days of hospitalization are regarded as an event, and the cumulative data from ICU admission to the prediction time point are extracted to predict the possibility of delirium occurring in the next 8 hours, and collect a total of 63,754 research case data, extract 12 feature selections to train the model, including age, sex, average ICU stay hours, visual and auditory abnormalities, RASS assessment score, APACHE-II Score score, number of invasive catheters indwelling, restraint and sedative and hypnotic drugs. Through feature data cleaning, processing and KNN interpolation method supplementation, a total of 54595 research case events were extracted to provide machine learning model analysis, using the research events from May 01 to November 30, 2022, as the model training data, 80% of which is the training set for model training, and 20% for the internal verification of the verification set, and then from December 01 to December 2022 The CU research event on the 31st is an external verification set data, and finally the model inference and performance evaluation are performed, and then the model has trained again by adjusting the model parameters. Results: In this study, XG Boost, Random Forest, Logistic Regression, and Decision Tree were used to analyze and compare four machine learning models. The average accuracy rate of internal verification was highest in Random Forest (AUC=0.86), and the average accuracy rate of external verification was in Random Forest and XG Boost was the highest, AUC was 0.86, and the average accuracy of cross-validation was the highest in Random Forest (ACC=0.77). Conclusion: Clinically, medical staff usually conduct CAM-ICU assessments at the bedside of critically ill patients in clinical practice, but there is a lack of machine learning classification methods to assist ICU patients in real-time assessment, resulting in the inability to provide more objective and continuous monitoring data to assist Clinical staff can more accurately identify and predict the occurrence of delirium in patients. It is hoped that the development and construction of predictive models through machine learning can predict delirium early and immediately, make clinical decisions at the best time, and cooperate with PADIS delirium care measures to provide individualized non-drug interventional care measures to maintain patient safety, and then Improve the quality of care.

Keywords: critically ill patients, machine learning methods, delirium prediction, classifier model

Procedia PDF Downloads 75
864 Awareness and Utilization of Social Network Tools among Agricultural Science Students in Colleges of Education in Ogun State, Nigeria

Authors: Adebowale Olukayode Efunnowo

Abstract:

This study was carried out to assess the awareness and utilization of Social Network Tools (SNTs) among agricultural science students in Colleges of Education in Ogun State, Nigeria. Simple random sampling techniques were used to select 280 respondents from the study area. Descriptive statistics was used to describe the objectives while Pearson Product Moment Correlation was used to test the hypothesis. The result showed that the majority (71.8%) of the respondents were single, with a mean age of 20 years. Almost all (95.7%) the respondents were aware of Facebook and 2go as a Social Network Tools (SNTs) while 85.0% of the respondents were not aware of Blackplanet, LinkedIn, MyHeritage and Bebo. Many (41.1%) of the respondents had views that using SNTs can enhance extensive literature survey, increase internet browsing potential, promote teaching proficiency, and update on outcomes of researches. However, 51.4% of the respondents perceived that SNTs usage as what is meant for the lecturers/adults only while 16.1% considered it as mainly used by internet fraudsters. Findings revealed that about 50.0% of the respondents browsed Facebook and 2go daily while more than 80% of the respondents used Blackplanet, MyHeritage, Skyrock, Bebo, LinkedIn and My YearBook as the need arise. Major constraints to the awareness and utilization of SNTs were high cost and poor quality of ICTs facilities (77.1%), epileptic power supply (75.0%), inadequate telecommunication infrastructure (71.1%), low technical know-how (62.9%) and inadequate computer knowledge (61.1%). The result of PPMC analysis showed that there was an inverse relationship between constraints and utilization of SNTs at p < 0.05. It can be concluded that constraints affect efficient and effective utilization of SNTs in the study area. It is hereby recommended that management of colleges of education and agricultural institutes should provide good internet connectivity, computer facilities, and alternative power supply in order to increase the awareness and utilization of SNTs among students.

Keywords: awareness, utilization, social network tools, constraints, students

Procedia PDF Downloads 352
863 Design and Control of a Knee Rehabilitation Device Using an MR-Fluid Brake

Authors: Mina Beheshti, Vida Shams, Mojtaba Esfandiari, Farzaneh Abdollahi, Abdolreza Ohadi

Abstract:

Most of the people who survive a stroke need rehabilitation tools to regain their mobility. The core function of these devices is a brake actuator. The goal of this study is to design and control a magnetorheological brake which can be used as a rehabilitation tool. In fact, the fluid used in this brake is called magnetorheological fluid or MR that properties can change by variation of the magnetic field. The braking properties can be set as control by using this feature of the fluid. In this research, different MR brake designs are first introduced in each design, and the dimensions of the brake have been determined based on the required torque for foot movement. To calculate the brake dimensions, it is assumed that the shear stress distribution in the fluid is uniform and the fluid is in its saturated state. After designing the rehabilitation brake, the mathematical model of the healthy movement of a healthy person is extracted. Due to the nonlinear nature of the system and its variability, various adaptive controllers, neural networks, and robust have been implemented to estimate the parameters and control the system. After calculating torque and control current, the best type of controller in terms of error and control current has been selected. Finally, this controller is implemented on the experimental data of the patient's movements, and the control current is calculated to achieve the desired torque and motion.

Keywords: rehabilitation, magnetorheological fluid, knee, brake, adaptive control, robust control, neural network control, torque control

Procedia PDF Downloads 151
862 Analysis of Bridge-Pile Foundation System in Multi-layered Non-Linear Soil Strata Using Energy-Based Method

Authors: Arvan Prakash Ankitha, Madasamy Arockiasamy

Abstract:

The increasing demand for adopting pile foundations in bridgeshas pointed towardsthe need to constantly improve the existing analytical techniques for better understanding of the behavior of such foundation systems. This study presents a simplistic approach using the energy-based method to assess the displacement responses of piles subjected to general loading conditions: Axial Load, Lateral Load, and a Bending Moment. The governing differential equations and the boundary conditions for a bridge pile embedded in multi-layered soil strata subjected to the general loading conditions are obtained using the Hamilton’s principle employing variational principles and minimization of energies. The soil non-linearity has been incorporated through simple constitutive relationships that account for degradation of soil moduli with increasing strain values.A simple power law based on published literature is used where the soil is assumed to be nonlinear-elastic and perfectly plastic. A Tresca yield surface is assumed to develop the soil stiffness variation with different strain levels that defines the non-linearity of the soil strata. This numerical technique has been applied to a pile foundation in a two - layered soil strata for a pier supporting the bridge and solved using the software MATLAB R2019a. The analysis yields the bridge pile displacements at any depth along the length of the pile. The results of the analysis are in good agreement with the published field data and the three-dimensional finite element analysis results performed using the software ANSYS 2019R3. The methodology can be extended to study the response of the multi-strata soil supporting group piles underneath the bridge piers.

Keywords: pile foundations, deep foundations, multilayer soil strata, energy based method

Procedia PDF Downloads 140
861 Anomaly Detection in a Data Center with a Reconstruction Method Using a Multi-Autoencoders Model

Authors: Victor Breux, Jérôme Boutet, Alain Goret, Viviane Cattin

Abstract:

Early detection of anomalies in data centers is important to reduce downtimes and the costs of periodic maintenance. However, there is little research on this topic and even fewer on the fusion of sensor data for the detection of abnormal events. The goal of this paper is to propose a method for anomaly detection in data centers by combining sensor data (temperature, humidity, power) and deep learning models. The model described in the paper uses one autoencoder per sensor to reconstruct the inputs. The auto-encoders contain Long-Short Term Memory (LSTM) layers and are trained using the normal samples of the relevant sensors selected by correlation analysis. The difference signal between the input and its reconstruction is then used to classify the samples using feature extraction and a random forest classifier. The data measured by the sensors of a data center between January 2019 and May 2020 are used to train the model, while the data between June 2020 and May 2021 are used to assess it. Performances of the model are assessed a posteriori through F1-score by comparing detected anomalies with the data center’s history. The proposed model outperforms the state-of-the-art reconstruction method, which uses only one autoencoder taking multivariate sequences and detects an anomaly with a threshold on the reconstruction error, with an F1-score of 83.60% compared to 24.16%.

Keywords: anomaly detection, autoencoder, data centers, deep learning

Procedia PDF Downloads 194
860 Population Stereotype Production, User Factors, and Icon Design for Underserved Communities of Rural India

Authors: Avijit Sengupta, Klarissa Ting Ting Cheng, Maffee Peng-Hui Wan

Abstract:

This study investigates the influence of user factors and referent characteristics on representation types generated using the stereotype production method for designing icons. Sixty-eight participants of farming communities were asked to draw images based on sixteen feature referents. Significant statistical differences were found between the types of representations generated for contextual and context-independent referents. Strong correlations were observed between years of formal education and total number of abstract representations produced for both contextual and context-independent referents. However, representation characteristics were not influenced by other user factors such as participants’ experience with mobile phone and years of farming experience. A statistically significant tendency of making concrete representations was observed for both contextual and context-independent referents. These findings provide insights on community members’ involvement in icon design and suggest a consolidated icon design strategy based on population stereotype, particularly for under-served rural communities of India.

Keywords: abstract representation, concrete representation, participatory design, population stereotype

Procedia PDF Downloads 375
859 Ab Initio Study of Electronic Structure and Transport of Graphyne and Graphdiyne

Authors: Zeljko Crljen, Predrag Lazic

Abstract:

Graphene has attracted a tremendous interest in the field of nanoelectronics and spintronics due to its exceptional electronic properties. However, pristine graphene has no band gap, a feature needed in building some of the electronic elements. Recently, a growing attention has been given to a class of carbon allotropes of graphene with honeycomb structures, in particular to graphyne and graphdiyne. They are characterized with a single and double acetylene bonding chains respectively, connecting the nearest-neighbor hexagonal rings. With an electron density comparable to that of graphene and a prominent gap in electronic band structures they appear as promising materials for nanoelectronic components. We studied the electronic structure and transport of infinite sheets of graphyne and graphdiyne and compared them with graphene. The method based on the non-equilibrium Green functions and density functional theory has been used in order to obtain a full ab initio self-consistent description of the transport current with different electrochemical bias potentials. The current/voltage (I/V) characteristics show a semi-conducting behavior with prominent nonlinearities at higher voltages. The calculated band gaps are 0.52V and 0.59V, respectively, and the effective masses are considerably smaller compared to typical semiconductors. We analyzed the results in terms of transmission eigenchannels and showed that the difference in conductance is directly related to the difference of the internal structure of the allotropes.

Keywords: electronic transport, graphene-like structures, nanoelectronics, two-dimensional materials

Procedia PDF Downloads 189
858 Insufficiency Fracture of Femoral Head in Patients Treated With Intramedullary Nailing for Proximal Femur Fracture

Authors: Jai Hyung Park, Eugene Kim, Jin Hun Park, Min Joon Oh

Abstract:

Introduction: Subchondral insufficiency fracture of the femoral head (SIF) is a rare complication; however, it has been recognized to cause femoral head collapse. Subchondral insufficiency fracture (SIF) is caused by normal or physiological stress without any trauma. It has been reported in osteoporotic patients after the fixation of the proximal femur with an Intramedullary nail. Case presentation: We reported 5 cases with SIF of the femoral head after proximal femur fracture fixation with Intra-medullary nail. All patients had osteoporosis as an underlying disease. Good reduction was achieved in all 5 patients. SIF was found from about 3 months to 4 years after the initial operation, and all the fractures were solidly united at the final diagnosis. We investigated retrospectively the feature of those cases and several factors that affected the occurrence of SIF. Discussion: There are a few discussions regarding the SIF of the femoral head. These discussions may include the predisposing risk factors, how to diagnose the SIF in osteoporotic patients, and the peri-operative factors to prevent SIF. Conclusion: Subchondral insufficiency fracture of the femoral head is a considerable complication after the internal fixation of the proximal femur. There are several factors that can be modified. If they could be controlled in the peri-operative period, SIF could be prevented or handled in advance. Other options related to arthroplasty can be considered in old osteoporotic patients.

Keywords: insufficiency fracture of femoral head, intra-medullary nail, osteoporosis, proximal femur fracture

Procedia PDF Downloads 128
857 Constructing the Joint Mean-Variance Regions for Univariate and Bivariate Normal Distributions: Approach Based on the Measure of Cumulative Distribution Functions

Authors: Valerii Dashuk

Abstract:

The usage of the confidence intervals in economics and econometrics is widespread. To be able to investigate a random variable more thoroughly, joint tests are applied. One of such examples is joint mean-variance test. A new approach for testing such hypotheses and constructing confidence sets is introduced. Exploring both the value of the random variable and its deviation with the help of this technique allows checking simultaneously the shift and the probability of that shift (i.e., portfolio risks). Another application is based on the normal distribution, which is fully defined by mean and variance, therefore could be tested using the introduced approach. This method is based on the difference of probability density functions. The starting point is two sets of normal distribution parameters that should be compared (whether they may be considered as identical with given significance level). Then the absolute difference in probabilities at each 'point' of the domain of these distributions is calculated. This measure is transformed to a function of cumulative distribution functions and compared to the critical values. Critical values table was designed from the simulations. The approach was compared with the other techniques for the univariate case. It differs qualitatively and quantitatively in easiness of implementation, computation speed, accuracy of the critical region (theoretical vs. real significance level). Stable results when working with outliers and non-normal distributions, as well as scaling possibilities, are also strong sides of the method. The main advantage of this approach is the possibility to extend it to infinite-dimension case, which was not possible in the most of the previous works. At the moment expansion to 2-dimensional state is done and it allows to test jointly up to 5 parameters. Therefore the derived technique is equivalent to classic tests in standard situations but gives more efficient alternatives in nonstandard problems and on big amounts of data.

Keywords: confidence set, cumulative distribution function, hypotheses testing, normal distribution, probability density function

Procedia PDF Downloads 174
856 Friend or Foe: Decoding the Legal Challenges Posed by Artificial Intellegence in the Era of Intellectual Property

Authors: Latika Choudhary

Abstract:

“The potential benefits of Artificial Intelligence are huge, So are the dangers.” - Dave Water. Artificial intelligence is one of the facet of Information technology domain which despite several attempts does not have a clear definition or ambit. However it can be understood as technology to solve problems via automated decisions and predictions. Artificial intelligence is essentially an algorithm based technology which analyses the large amounts of data and then solves problems by detecting useful patterns. Owing to its automated feature it will not be wrong to say that humans & AI have more utility than humans alone or computers alone.1 For many decades AI experienced enthusiasm as well as setbacks, yet it has today become part and parcel of our everyday life, making it convenient or at times problematic. AI and related technology encompass Intellectual Property in multiple ways, the most important being AI technology for management of Intellectual Property, IP for protecting AI and IP as a hindrance to the transparency of AI systems. Thus the relationship between the two is of reciprocity as IP influences AI and vice versa. While AI is a recent concept, the IP laws for protection or even dealing with its challenges are relatively older, raising the need for revision to keep up with the pace of technological advancements. This paper will analyze the relationship between AI and IP to determine how beneficial or conflictual the same is, address how the old concepts of IP are being stretched to its maximum limits so as to accommodate the unwanted consequences of the Artificial Intelligence and propose ways to mitigate the situation so that AI becomes the friend it is and not turn into a potential foe it appears to be.

Keywords: intellectual property rights, information technology, algorithm, artificial intelligence

Procedia PDF Downloads 87
855 Compilation of Islamic Law as Law Applied Religious Courts in Indonesia (Responding to Changes in Religious Courts Authority)

Authors: Hamdan Arief Hanif, Rahmat Sidiq

Abstract:

Indonesia is a country of law, the legal system adopted by Indonesia is a civil law system. A major feature of the civil law is the codified legislation. Meanwhile the majority of society Indonesia are Muslims, whilst Islamic law itself having the sources written in Qur'an, Sunnah and the opinion of Muslim scholars, generally not codified in book form of legislation that is easy on the set as a reference. in Indonesia, many scholars have different opinions in decisions so that there is no legal certainty in Muslim civil cases, so the need for legal codification, which, as the source of the judges in deciding a case, especially a case in religious courts. This paper raised the topic of discussion which offers a solution to the application of the codification of the Islamic Law which became the core resources in delivering a verdict against Islamic civil related issue; codification usually called a compilation of Islamic Law. Compilation of Islamic Law is highly recommended as a core reference for the judges in religious courts in Indonesia. This compilation which includes a collection of large number of opinions scholars (book of fiqh) that existed previously and are ripened in deduce in order to unify the existing differences. This paper also discusses how the early formation of the compilation and as the right solution in order to create legal certainty and justice especially for the muslim community in Indonesia.

Keywords: Islamic law, compilation, law applied core, religious court

Procedia PDF Downloads 355
854 Multi-Layer Multi-Feature Background Subtraction Using Codebook Model Framework

Authors: Yun-Tao Zhang, Jong-Yeop Bae, Whoi-Yul Kim

Abstract:

Background modeling and subtraction in video analysis has been widely proved to be an effective method for moving objects detection in many computer vision applications. Over the past years, a large number of approaches have been developed to tackle different types of challenges in this field. However, the dynamic background and illumination variations are two of the most frequently occurring issues in the practical situation. This paper presents a new two-layer model based on codebook algorithm incorporated with local binary pattern (LBP) texture measure, targeted for handling dynamic background and illumination variation problems. More specifically, the first layer is designed by block-based codebook combining with LBP histogram and mean values of RGB color channels. Because of the invariance of the LBP features with respect to monotonic gray-scale changes, this layer can produce block-wise detection results with considerable tolerance of illumination variations. The pixel-based codebook is employed to reinforce the precision from the outputs of the first layer which is to eliminate false positives further. As a result, the proposed approach can greatly promote the accuracy under the circumstances of dynamic background and illumination changes. Experimental results on several popular background subtraction datasets demonstrate a very competitive performance compared to previous models.

Keywords: background subtraction, codebook model, local binary pattern, dynamic background, illumination change

Procedia PDF Downloads 217
853 Signal Integrity Performance Analysis in Capacitive and Inductively Coupled Very Large Scale Integration Interconnect Models

Authors: Mudavath Raju, Bhaskar Gugulothu, B. Rajendra Naik

Abstract:

The rapid advances in Very Large Scale Integration (VLSI) technology has resulted in the reduction of minimum feature size to sub-quarter microns and switching time in tens of picoseconds or even less. As a result, the degradation of high-speed digital circuits due to signal integrity issues such as coupling effects, clock feedthrough, crosstalk noise and delay uncertainty noise. Crosstalk noise in VLSI interconnects is a major concern and reduction in VLSI interconnect has become more important for high-speed digital circuits. It is the most effectively considered in Deep Sub Micron (DSM) and Ultra Deep Sub Micron (UDSM) technology. Increasing spacing in-between aggressor and victim line is one of the technique to reduce the crosstalk. Guard trace or shield insertion in-between aggressor and victim is also one of the prominent options for the minimization of crosstalk. In this paper, far end crosstalk noise is estimated with mutual inductance and capacitance RLC interconnect model. Also investigated the extent of crosstalk in capacitive and inductively coupled interconnects to minimizes the same through shield insertion technique.

Keywords: VLSI, interconnects, signal integrity, crosstalk, shield insertion, guard trace, deep sub micron

Procedia PDF Downloads 186
852 Analysis Model for the Relationship of Users, Products, and Stores on Online Marketplace Based on Distributed Representation

Authors: Ke He, Wumaier Parezhati, Haruka Yamashita

Abstract:

Recently, online marketplaces in the e-commerce industry, such as Rakuten and Alibaba, have become some of the most popular online marketplaces in Asia. In these shopping websites, consumers can select purchase products from a large number of stores. Additionally, consumers of the e-commerce site have to register their name, age, gender, and other information in advance, to access their registered account. Therefore, establishing a method for analyzing consumer preferences from both the store and the product side is required. This study uses the Doc2Vec method, which has been studied in the field of natural language processing. Doc2Vec has been used in many cases to analyze the extraction of semantic relationships between documents (represented as consumers) and words (represented as products) in the field of document classification. This concept is applicable to represent the relationship between users and items; however, the problem is that one more factor (i.e., shops) needs to be considered in Doc2Vec. More precisely, a method for analyzing the relationship between consumers, stores, and products is required. The purpose of our study is to combine the analysis of the Doc2vec model for users and shops, and for users and items in the same feature space. This method enables the calculation of similar shops and items for each user. In this study, we derive the real data analysis accumulated in the online marketplace and demonstrate the efficiency of the proposal.

Keywords: Doc2Vec, online marketplace, marketing, recommendation systems

Procedia PDF Downloads 112