Search results for: loan loss recognition
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 5028

Search results for: loan loss recognition

4398 Binarization and Recognition of Characters from Historical Degraded Documents

Authors: Bency Jacob, S.B. Waykar

Abstract:

Degradations in historical document images appear due to aging of the documents. It is very difficult to understand and retrieve text from badly degraded documents as there is variation between the document foreground and background. Thresholding of such document images either result in broken characters or detection of false texts. Numerous algorithms exist that can separate text and background efficiently in the textual regions of the document; but portions of background are mistaken as text in areas that hardly contain any text. This paper presents a way to overcome these problems by a robust binarization technique that recovers the text from a severely degraded document images and thereby increases the accuracy of optical character recognition systems. The proposed document recovery algorithm efficiently removes degradations from document images. Here we are using the ostus method ,local thresholding and global thresholding and after the binarization training and recognizing the characters in the degraded documents.

Keywords: binarization, denoising, global thresholding, local thresholding, thresholding

Procedia PDF Downloads 336
4397 The Effects of Some Organic Amendments on Sediment Yield, Splash Loss, and Runoff of Soils of Selected Parent Materials in Southeastern Nigeria

Authors: Leonard Chimaobi Agim, Charles Arinzechukwu Igwe, Emmanuel Uzoma Onweremadu, Gabreil Osuji

Abstract:

Soil erosion has been linked to stream sedimentation, ecosystem degradation, and loss of soil nutrients. A study was conducted to evaluate the effect of some organic amendment on sediment yield, splash loss, and runoff of soils of selected parent materials in southeastern Nigeria. A total of 20 locations, five from each of four parent materials namely: Asu River Group (ARG), Bende Ameki Group (BAG), Coastal Plain Sand (CPS) and Falsebedded Sandstone (FBS) were used for the study. Collected soil samples were analyzed with standard methods for the initial soil properties. Rainfall simulation at an intensity of 190 mm hr-1was conducted for 30 minutes on the soil samples at both the initial stage and after amendment to obtain erosion parameters. The influence of parent material on sediment yield, splash loss and runoff based on rainfall simulation was tested for using one way analyses of variance, while the influence of organic material and their combinations were a factorially fitted in a randomized complete block design. The organic amendments include; goat dropping (GD), poultry dropping (PD), municipal solid waste (MSW) and their combinations (COA) applied at four rates of 0, 10, 20 and 30 t ha-1 respectively. Data were analyzed using analyses of variance suitable for a factorial experiment. Significant means were separated using LSD at 5 % probability levels. Result showed significant (p ≤ 0.05) lower values of sediment yield, splash loss and runoff following amendment. For instance, organic amendment reduced sediment yield under wet and dry runs by 12.91 % and 26.16% in Ishiagu, 40.76% and 45.67%, in Bende, 16.17% and 50% in Obinze and 22.80% and 42.35% in Umulolo respectively. Goat dropping and combination of amendment gave the best results in reducing sediment yield.

Keywords: organic amendment, parent material, rainfall simulation, soil erosion

Procedia PDF Downloads 334
4396 Analytic Hierarchy Process and Multi-Criteria Decision-Making Approach for Selecting the Most Effective Soil Erosion Zone in Gomati River Basin

Authors: Rajesh Chakraborty, Dibyendu Das, Rabindra Nath Barman, Uttam Kumar Mandal

Abstract:

In the present study, the objective is to find out the most effective zone causing soil erosion in the Gumati river basin located in the state of Tripura, a north eastern state of India using analytical hierarchy process (AHP) and multi-objective optimization on the basis of ratio analysis (MOORA).The watershed is segmented into 20 zones based on Area. The watershed is considered by pointing the maximum elevation from sea lever from Google earth. The soil erosion is determined using the universal soil loss equation. The different independent variables of soil loss equation bear different weightage for different soil zones. And therefore, to find the weightage factor for all the variables of soil loss equation like rainfall runoff erosivity index, soil erodibility factor etc, analytical hierarchy process (AHP) is used. And thereafter, multi-objective optimization on the basis of ratio analysis (MOORA) approach is used to select the most effective zone causing soil erosion. The MCDM technique concludes that the maximum soil erosion is occurring in the zone 14.

Keywords: soil erosion, analytic hierarchy process (AHP), multi criteria decision making (MCDM), universal soil loss equation (USLE), multi-objective optimization on the basis of ratio analysis (MOORA)

Procedia PDF Downloads 524
4395 Drastic Improvement in Vision Following Surgical Excision of Juvenile Nasopharyngeal Angiofibroma with Compressive Optic Neuropathy

Authors: Sweta Das

Abstract:

This case report is a 15-year-old male who presented with painless unilateral vision loss from left optic nerve compression due to juvenile nasopharyngeal angiofibroma. JNA is a rare, benign neoplasm that causes intracranial and intraorbital bone destruction and extends aggressively into surrounding soft tissues. It accounts for <1% of all head and neck tumors, is predominantly found in pediatric males and tends to affect indigenous population disproportionately. The most common presenting symptom for JNA is epistaxis and nasal obstruction. However, it can invade orbit, chiasm and pituitary gland, causing loss of vision and field. Visual acuity and function near normalized following surgical excision. Optometry plays an important role in the diagnosis and co-management of JNA with optic nerve compression by closely monitoring afferent optic nerve function and structure, and extraocular motility. Visual function and acuity in patients with short-term compressive neuropathy may drastically improve following surgical resection as this case demonstrates.

Keywords: orbital mass, painless monocular vision loss, compressive optic neuropathy, pediatric tumor

Procedia PDF Downloads 50
4394 Molecular Characterization and Identification of C-Type Lectin in Red Palm Weevil, Rhynchophorus ferrugineus Oliver

Authors: Hafiza Javaria Ashraf, Xinghong Wang, Zhanghong Shi, Youming Hou

Abstract:

Insect’s innate immunity depends on a variety of defense responses for the recognition of invading pathogens. Pathogen recognition involves particular proteins known as pattern recognition receptors (PRRs). These PRRs interact with pathogen-associated molecular patterns (PAMPs) present on the surface of pathogens to distinguish between self and non-self. C-type lectins (CTLs) belong to a superfamily of PPRs which involved in insect immunity and defense mechanism. Rhynchophorus ferrugineus Olivier is a devastating pest of Palm cultivations in China. Although studies on R. ferrugineus immune mechanism and host defense have conducted, however, the role of CTL in immune responses of R. ferrugineus remains elusive. Here, we report RfCTL, which is a secreted protein containing a single-CRD domain. The open reading frame (ORF) of CTL is 226 bp, which encodes a putative protein of 168 amino acids. Transcript expression analysis revealed that RfCTL highly expressed in immune-related tissues, i.e., hemolymph and fat body. The abundance of RfCTL in the gut and fat body dramatically increased upon Staphylococcus aureus and Escherichia coli bacterial challenges, suggesting a role in defense against gram-positive and gram-negative bacterial infection. Taken together, we inferred that RfCTL might be involved in the immune defense of R. ferrugineus and established a solid foundation for future studies on R. ferrugineus CTL domain proteins for better understanding of insect immunity.

Keywords: biological invasion, c-type lectin, insect immunity, Rhynchophorus ferrugineus Oliver

Procedia PDF Downloads 140
4393 Managing Incomplete PSA Observations in Prostate Cancer Data: Key Strategies and Best Practices for Handling Loss to Follow-Up and Missing Data

Authors: Madiha Liaqat, Rehan Ahmed Khan, Shahid Kamal

Abstract:

Multiple imputation with delta adjustment is a versatile and transparent technique for addressing univariate missing data in the presence of various missing mechanisms. This approach allows for the exploration of sensitivity to the missing-at-random (MAR) assumption. In this review, we outline the delta-adjustment procedure and illustrate its application for assessing the sensitivity to deviations from the MAR assumption. By examining diverse missingness scenarios and conducting sensitivity analyses, we gain valuable insights into the implications of missing data on our analyses, enhancing the reliability of our study's conclusions. In our study, we focused on assessing logPSA, a continuous biomarker in incomplete prostate cancer data, to examine the robustness of conclusions against plausible departures from the MAR assumption. We introduced several approaches for conducting sensitivity analyses, illustrating their application within the pattern mixture model (PMM) under the delta adjustment framework. This proposed approach effectively handles missing data, particularly loss to follow-up.

Keywords: loss to follow-up, incomplete response, multiple imputation, sensitivity analysis, prostate cancer

Procedia PDF Downloads 75
4392 Enhancing Metaverse Security: A Multi-Factor Authentication Scheme

Authors: R. Chinnaiyaprabhu, S. Bharanidharan, V. Dharsana, Rajalavanya

Abstract:

The concept of the Metaverse represents a potential evolution in the realm of cyberspace. In the early stages of Web 2.0, we observed a proliferation of online pseudonyms or 'nyms,' which increased the prevalence of fake accounts and made it challenging to establish unique online identities for various roles. However, in the era of Web 3.0, particularly in the context of the Metaverse, an individual's digital identity is intrinsically linked to their real-world identity. Consequently, actions taken in the Metaverse can carry significant consequences in the physical world. In light of these considerations, we propose the development of an innovative authentication system known as 'Metasec.' This system is designed to enhance security for digital assets, online identities, avatars, and user accounts within the Metaverse. Notably, Metasec operates as a password less authentication solution, relying on a multifaceted approach to security, encompassing device attestation, facial recognition, and pattern-based security keys.

Keywords: metaverse, multifactor authentication, security, facial recognition, patten password

Procedia PDF Downloads 54
4391 Criterion-Referenced Test Reliability through Threshold Loss Agreement: Fuzzy Logic Analysis Approach

Authors: Mohammad Ali Alavidoost, Hossein Bozorgian

Abstract:

Criterion-referenced tests (CRTs) are designed to measure student performance against a fixed set of predetermined criteria or learning standards. The reliability of such tests cannot be based on internal reliability. Threshold loss agreement is one way to calculate the reliability of CRTs. However, the selection of master and non-master in such agreement is determined by the threshold point. The problem is if the threshold point witnesses a minute change, the selection of master and non-master may have a drastic change, leading to the change in reliability results. Therefore, in this study, the Fuzzy logic approach is employed as a remedial procedure for data analysis to obviate the threshold point problem. Forty-one Iranian students were selected; the participants were all between 20 and 30 years old. A quantitative approach was used to address the research questions. In doing so, a quasi-experimental design was utilized since the selection of the participants was not randomized. Based on the Fuzzy logic approach, the threshold point would be more stable during the analysis, resulting in rather constant reliability results and more precise assessment.

Keywords: criterion-referenced tests, threshold loss agreement, threshold point, fuzzy logic approach

Procedia PDF Downloads 357
4390 A CD40 Variant is Associated with Systemic Bone Loss Among Patients with Rheumatoid Arthritis

Authors: Rim Sghiri, Samia Al Shouli, Hana Benhassine, Nejla Elamri, Zahid Shakoor, Foued Slama, Adel Almogren, Hala Zeglaoui, Elyes Bouajina, Ramzi Zemni

Abstract:

Objectives: Little is known about genes predisposing to systemic bone loss (SBL) in rheumatoid arthritis (RA). Therefore, we examined the association between SBL and a variant of CD40 gene, which is known to play a critical role in both immune response and bone homeostasis among patients with RA. Methods: CD40 rs48104850 was genotyped in 176 adult RA patients. Bone mineral density (BMD) was measured using dual-energy X-ray absorptiometry (DXA). Results: Low BMD was observed in 116 (65.9%) patients. Among them, 60 (34.1%) had low femoral neck (FN) Z score, 72 (40.9%) had low total femur (TF) Z score, and 105 (59.6%) had low lumbar spine (LS) Z score. CD40 rs4810485 was found to be associated with reduced TF Z score with the CD40 rs4810485 T allele protecting against reduced TF Z score (OR = 0.40, 95% CI = 0.23-0.68, p = 0.0005). This association was confirmed in the multivariate logistic regression analysis (OR=0.31, 95% CI= 0.16-0.59, p=3.84 x 10₋₄). Moreover, median FN BMD was reduced among RA patients with CD40 rs4810485 GG genotype compared to RA patients harbouring CD40 rs4810485 TT and GT genotypes (0.788± 0.136 versus 0.826± 0.146g/cm², p=0.001). Conclusion: This study, for the first time ever, demonstrated an association between a CD40 genetic variant and SBL among patients with RA.

Keywords: rheumatoid arthritis, CD40 gene, bone mineral density, systemic bone loss, rs48104850

Procedia PDF Downloads 447
4389 A High Performance Piano Note Recognition Scheme via Precise Onset Detection and Segmented Short-Time Fourier Transform

Authors: Sonali Banrjee, Swarup Kumar Mitra, Aritra Acharyya

Abstract:

A piano note recognition method has been proposed by the authors in this paper. The authors have used a comprehensive method for onset detection of each note present in a piano piece followed by segmented short-time Fourier transform (STFT) for the identification of piano notes. The performance evaluation of the proposed method has been carried out in different harsh noisy environments by adding different levels of additive white Gaussian noise (AWGN) having different signal-to-noise ratio (SNR) in the original signal and evaluating the note detection error rate (NDER) of different piano pieces consisting of different number of notes at different SNR levels. The NDER is found to be remained within 15% for all piano pieces under consideration when the SNR is kept above 8 dB.

Keywords: AWGN, onset detection, piano note, STFT

Procedia PDF Downloads 153
4388 Understanding Factors that Affect the Prior Knowledge of Deaf and Hard of Hearing Students and their Relation to Reading Comprehension

Authors: Khalid Alasim

Abstract:

The reading comprehension levels of students who are deaf or hard of hearing (DHH) are low compared to those of their hearing peers. One possible reason for this low reading levels is related to the students’ prior knowledge. This study investigated the potential factors that might affected DHH students’ prior knowledge, including their degree of hearing loss, the presence or absence of family members with a hearing loss, and educational stage (elementary–middle school). The study also examined the contribution of prior knowledge in predicting DHH students’ reading comprehension levels, and investigated the differences in the students’ scores based on the type of questions, including text-explicit (TE), text-implicit (TI), and script-implicit (SI) questions. Thirty-one elementary and middle-school students completed a demographic form and assessment, and descriptive statistics and multiple and simple linear regressions were used to answer the research questions. The findings indicated that the independent variables—degree of hearing loss, presence or absence of family members with hearing loss, and educational stage—explained little of the variance in DHH students’ prior knowledge. Further, the results showed that the DHH students’ prior knowledge affected their reading comprehension. Finally, the result demonstrated that the participants were able to answer more of the TI questions correctly than the TE and SI questions. The study concluded that prior knowledge is important in these students’ reading comprehension, and it is also important for teachers and parents of DHH children to use effective ways to increase their students’ and children’s prior knowledge.

Keywords: reading comprehension, prior knowledge, metacognition, elementary, self-contained classrooms

Procedia PDF Downloads 90
4387 Thick Data Analytics for Learning Cataract Severity: A Triplet Loss Siamese Neural Network Model

Authors: Jinan Fiaidhi, Sabah Mohammed

Abstract:

Diagnosing cataract severity is an important factor in deciding to undertake surgery. It is usually conducted by an ophthalmologist or through taking a variety of fundus photography that needs to be examined by the ophthalmologist. This paper carries out an investigation using a Siamese neural net that can be trained with small anchor samples to score cataract severity. The model used in this paper is based on a triplet loss function that takes the ophthalmologist best experience in rating positive and negative anchors to a specific cataract scaling system. This approach that takes the heuristics of the ophthalmologist is generally called the thick data approach, which is a kind of machine learning approach that learn from a few shots. Clinical Relevance: The lens of the eye is mostly made up of water and proteins. A cataract occurs when these proteins at the eye lens start to clump together and block lights causing impair vision. This research aims at employing thick data machine learning techniques to rate the severity of the cataract using Siamese neural network.

Keywords: thick data analytics, siamese neural network, triplet-loss model, few shot learning

Procedia PDF Downloads 99
4386 Efficacy of Agrobacterium Tumefaciens as a Possible Entomopathogenic Agent

Authors: Fouzia Qamar, Shahida Hasnain

Abstract:

The objective of the present study was to evaluate the possible role of Agrobacterium tumefaciens as a possible insect biocontrol agent. Pests selected for the present challenge were adult males of Periplaneta americana and last instar larvae of Pieris brassicae and Spodoptera litura. Different ranges of bacterial doses were selected and tested to score the mortalities of the insects after 24 hours, for the lethal dose estimation studies. Mode of application for the inoculation of the bacteria, was the microinjection technique. The evaluation of the possible entomopathogenic carrying attribute of bacterial Ti plasmid, led to the conclusion that the loss of plasmid was associated with the loss of virulence against target insects.

Keywords: agrobacterium tumefaciens, toxicity assessment, biopesticidal attribute, entomopathogenic agent

Procedia PDF Downloads 366
4385 The Effect of the Combination of Mouthwash and Saliva Substitutes on Tooth Erosion: An in Vitro Study

Authors: Young-Eun Jang, Mi-Hye Ma, Yemi Kim

Abstract:

As the elderly population increases, the number of patients complaining of dry mouth is also increasing. Elderly people often use mouthwash to prevent periodontal disease. Mouthwash and saliva substitutes with low pH were reported to be able to cause enamel erosion. To the best of our knowledge, there have been no studies showing the effect of mouthwash on patients using saliva substitutes. Therefore, the purpose of this study was to evaluate the effect of the use of mouthwash in combination with saliva substitutes on tooth erosion using a quantitative light-induced fluorescence-digital (QLF-D) system. A total of 96 bovine specimens were embedded in putty blocks and randomly allocated to the following groups with n = 12 each: Group 1, application of mouthwash; Group 2, application of saliva substitutes; Group 3, application of saliva substitutes in combination with mouthwash; and control group, application of saline. The bovine samples were eroded using a demineralization solution and then saliva substitutes and mouthwash were applied according to the groups for 2 weeks. For saliva substitutes, three different products were used; Oasis (Oasis Consumer Health, Cleveland, OH, USA), Xeromia solution (Osstem Pharma Co., Seoul, Korea), and Drymund gel (Dong-A Pharma Co., Seoul, Korea). The pH values of the saliva substitutes were determined using a pH meter. Loss of enamel and root dentin was measured using the QLF-D system immediately after demineralization on the 3rd, 7th, and 14th days. The data were analyzed using repeated measures ANOVA followed by Tukey’s post hoc tests (p < 0.05). Mineral loss in enamel and root dentin was detected when mouthwash and saliva substitutes were used alone, respectively (p < 0.05). Also, when mouthwash was used with saliva substitutes, the mineral loss was observed in enamel and root dentin (p < 0.05). The use of Xeromia and Drymund gel increased mineral loss of enamel significantly compared to the use of Oasis (p < 0.05). However, when Drymund gel and Xeromia were used in combination with mouthwash, mineral loss of enamel was significantly reduced compared to when they were used alone (p < 0.05). The pH values of Drymund gel, Xeromia, Oasis, and mouthwash were 5.5, 5.52, 6.2, and 6.37, respectively. Based on these results, it can be concluded that the use of mouthwash with a higher pH value than that of saliva substitutes could help patients suffering from xerostomia avoid the risk of dental erosion.

Keywords: saliva substitute, mouthwash, tooth erosion, dry mouth

Procedia PDF Downloads 76
4384 Enhanced Multi-Scale Feature Extraction Using a DCNN by Proposing Dynamic Soft Margin SoftMax for Face Emotion Detection

Authors: Armin Nabaei, M. Omair Ahmad, M. N. S. Swamy

Abstract:

Many facial expression and emotion recognition methods in the traditional approaches of using LDA, PCA, and EBGM have been proposed. In recent years deep learning models have provided a unique platform addressing by automatically extracting the features for the detection of facial expression and emotions. However, deep networks require large training datasets to extract automatic features effectively. In this work, we propose an efficient emotion detection algorithm using face images when only small datasets are available for training. We design a deep network whose feature extraction capability is enhanced by utilizing several parallel modules between the input and output of the network, each focusing on the extraction of different types of coarse features with fined grained details to break the symmetry of produced information. In fact, we leverage long range dependencies, which is one of the main drawback of CNNs. We develop this work by introducing a Dynamic Soft-Margin SoftMax.The conventional SoftMax suffers from reaching to gold labels very soon, which take the model to over-fitting. Because it’s not able to determine adequately discriminant feature vectors for some variant class labels. We reduced the risk of over-fitting by using a dynamic shape of input tensor instead of static in SoftMax layer with specifying a desired Soft- Margin. In fact, it acts as a controller to how hard the model should work to push dissimilar embedding vectors apart. For the proposed Categorical Loss, by the objective of compacting the same class labels and separating different class labels in the normalized log domain.We select penalty for those predictions with high divergence from ground-truth labels.So, we shorten correct feature vectors and enlarge false prediction tensors, it means we assign more weights for those classes with conjunction to each other (namely, “hard labels to learn”). By doing this work, we constrain the model to generate more discriminate feature vectors for variant class labels. Finally, for the proposed optimizer, our focus is on solving weak convergence of Adam optimizer for a non-convex problem. Our noteworthy optimizer is working by an alternative updating gradient procedure with an exponential weighted moving average function for faster convergence and exploiting a weight decay method to help drastically reducing the learning rate near optima to reach the dominant local minimum. We demonstrate the superiority of our proposed work by surpassing the first rank of three widely used Facial Expression Recognition datasets with 93.30% on FER-2013, and 16% improvement compare to the first rank after 10 years, reaching to 90.73% on RAF-DB, and 100% k-fold average accuracy for CK+ dataset, and shown to provide a top performance to that provided by other networks, which require much larger training datasets.

Keywords: computer vision, facial expression recognition, machine learning, algorithms, depp learning, neural networks

Procedia PDF Downloads 62
4383 Tip60 Histone Acetyltransferase Activators as Neuroepigenetic Therapeutic Modulators for Alzheimer’s Disease

Authors: Akanksha Bhatnagar, Sandhya Kortegare, Felice Elefant

Abstract:

Context: Alzheimer's disease (AD) is a neurodegenerative disorder that is characterized by progressive cognitive decline and memory loss. The cause of AD is not fully understood, but it is thought to be caused by a combination of genetic, environmental, and lifestyle factors. One of the hallmarks of AD is the loss of neurons in the hippocampus, a brain region that is important for memory and learning. This loss of neurons is thought to be caused by a decrease in histone acetylation, which is a process that regulates gene expression. Research Aim: The research aim of the study was to develop mall molecule compounds that can enhance the activity of Tip60, a histone acetyltransferase that is important for memory and learning. Methodology/Analysis: The researchers used in silico structural modeling and a pharmacophore-based virtual screening approach to design and synthesize small molecule compounds strongly predicted to target and enhance Tip60’s HAT activity. The compounds were then tested in vitro and in vivo to assess their ability to enhance Tip60 activity and rescue cognitive deficits in AD models. Findings: The researchers found that several of the compounds were able to enhance Tip60 activity and rescue cognitive deficits in AD models. The compounds were also developed to cross the blood-brain barrier, which is an important factor for the development of potential AD therapeutics. Theoretical Importance: The findings of this study suggest that Tip60 HAT activators have the potential to be developed as therapeutic agents for AD. The compounds are specific to Tip60, which suggests that they may have fewer side effects than other HDAC inhibitors. Additionally, the compounds are able to cross the blood-brain barrier, which is a major hurdle for the development of AD therapeutics. Data Collection: The study collected data from a variety of sources, including in vitro assays and animal models. The in vitro assays assessed the ability of compounds to enhance Tip60 activity using histone acetyltransferase (HAT) enzyme assays and chromatin immunoprecipitation assays. Animal models were used to assess the ability of the compounds to rescue cognitive deficits in AD models using a variety of behavioral tests, including locomotor ability, sensory learning, and recognition tasks. The human clinical trials will be used to assess the safety and efficacy of the compounds in humans. Questions: The question addressed by this study was whether Tip60 HAT activators could be developed as therapeutic agents for AD. Conclusions: The findings of this study suggest that Tip60 HAT activators have the potential to be developed as therapeutic agents for AD. The compounds are specific to Tip60, which suggests that they may have fewer side effects than other HDAC inhibitors. Additionally, the compounds are able to cross the blood-brain barrier, which is a major hurdle for the development of AD therapeutics. Further research is needed to confirm the safety and efficacy of these compounds in humans.

Keywords: Alzheimer's disease, cognition, neuroepigenetics, drug discovery

Procedia PDF Downloads 63
4382 Loss Analysis by Loading Conditions of Distribution Transformers

Authors: A. Bozkurt, C. Kocatepe, R. Yumurtaci, İ. C. Tastan, G. Tulun

Abstract:

Efficient use of energy, with the increase in demand of energy and also with the reduction of natural energy sources, has improved its importance in recent years. Most of the losses in the system from electricity produced until the point of consumption is mostly composed by the energy distribution system. In this study, analysis of the resulting loss in power distribution transformer and distribution power cable is realized which are most of the losses in the distribution system. Transformer losses in the real distribution system were analyzed by CYME Power Engineering Software program. These losses are disclosed for different voltage levels and different loading conditions.

Keywords: distribution system, distribution transformer, power cable, technical losses

Procedia PDF Downloads 635
4381 Predicting the Human Impact of Natural Onset Disasters Using Pattern Recognition Techniques and Rule Based Clustering

Authors: Sara Hasani

Abstract:

This research focuses on natural sudden onset disasters characterised as ‘occurring with little or no warning and often cause excessive injuries far surpassing the national response capacities’. Based on the panel analysis of the historic record of 4,252 natural onset disasters between 1980 to 2015, a predictive method was developed to predict the human impact of the disaster (fatality, injured, homeless) with less than 3% of errors. The geographical dispersion of the disasters includes every country where the data were available and cross-examined from various humanitarian sources. The records were then filtered into 4252 records of the disasters where the five predictive variables (disaster type, HDI, DRI, population, and population density) were clearly stated. The procedure was designed based on a combination of pattern recognition techniques and rule-based clustering for prediction and discrimination analysis to validate the results further. The result indicates that there is a relationship between the disaster human impact and the five socio-economic characteristics of the affected country mentioned above. As a result, a framework was put forward, which could predict the disaster’s human impact based on their severity rank in the early hours of disaster strike. The predictions in this model were outlined in two worst and best-case scenarios, which respectively inform the lower range and higher range of the prediction. A necessity to develop the predictive framework can be highlighted by noticing that despite the existing research in literature, a framework for predicting the human impact and estimating the needs at the time of the disaster is yet to be developed. This can further be used to allocate the resources at the response phase of the disaster where the data is scarce.

Keywords: disaster management, natural disaster, pattern recognition, prediction

Procedia PDF Downloads 144
4380 Cultural Disposition and Implicit Dehumanization of Sexualized Females by Women

Authors: Hong Im Shin

Abstract:

Previous research demonstrated that self-objectification (women view themselves as objects for use) is related to system-justification. Three studies investigated whether cultural disposition as its system-justifying function could have an impact on self-objectification and dehumanization of sexualized women and men. Study 1 (N = 91) employed a survey methodology to examine the relationship between cultural disposition (collectivism vs. individualism), trait of system-justification, and self-objectification. The results showed that the higher tendency of collectivism was related to stronger system-justification and self-objectification. Study 2 (N = 60 females) introduced a single category implicit association task (SC-IAT) to assess the extent to which sexually objectified women were associated with uniquely human attributes (i.e., culture) compared to animal-related attributes (i.e., nature). According to results, female participants associated sexually objectified female targets less with human attributes compared to animal-related attributes. Study 3 (N = 46) investigated whether priming to individualism or collectivism was associated to system justification and sexual objectification of men and women with the use of a recognition task involving upright and inverted pictures of sexualized women and men. The results indicated that the female participants primed to individualism showed an inversion effect for sexualized women and men (person-like recognition), whereas there was no inversion effect for sexualized women in the priming condition of collectivism (object-like recognition). This implies that cultural disposition plays a mediating role for rationalizing the gender status, implicit dehumanization of sexualized females and self-objectification. Future research directions are discussed.

Keywords: cultural disposition, dehumanization, implicit test, self-objectification

Procedia PDF Downloads 225
4379 Towards a Complete Automation Feature Recognition System for Sheet Metal Manufacturing

Authors: Bahaa Eltahawy, Mikko Ylihärsilä, Reino Virrankoski, Esko Petäjä

Abstract:

Sheet metal processing is automated, but the step from product models to the production machine control still requires human intervention. This may cause time consuming bottlenecks in the production process and increase the risk of human errors. In this paper we present a system, which automatically recognizes features from the CAD-model of the sheet metal product. By using these features, the system produces a complete model of the particular sheet metal product. Then the model is used as an input for the sheet metal processing machine. Currently the system is implemented, capable to recognize more than 11 of the most common sheet metal structural features, and the procedure is fully automated. This provides remarkable savings in the production time, and protects against the human errors. This paper presents the developed system architecture, applied algorithms and system software implementation and testing.

Keywords: feature recognition, automation, sheet metal manufacturing, CAD, CAM

Procedia PDF Downloads 343
4378 Biosignal Recognition for Personal Identification

Authors: Hadri Hussain, M.Nasir Ibrahim, Chee-Ming Ting, Mariani Idroas, Fuad Numan, Alias Mohd Noor

Abstract:

A biometric security system has become an important application in client identification and verification system. A conventional biometric system is normally based on unimodal biometric that depends on either behavioural or physiological information for authentication purposes. The behavioural biometric depends on human body biometric signal (such as speech) and biosignal biometric (such as electrocardiogram (ECG) and phonocardiogram or heart sound (HS)). The speech signal is commonly used in a recognition system in biometric, while the ECG and the HS have been used to identify a person’s diseases uniquely related to its cluster. However, the conventional biometric system is liable to spoof attack that will affect the performance of the system. Therefore, a multimodal biometric security system is developed, which is based on biometric signal of ECG, HS, and speech. The biosignal data involved in the biometric system is initially segmented, with each segment Mel Frequency Cepstral Coefficients (MFCC) method is exploited for extracting the feature. The Hidden Markov Model (HMM) is used to model the client and to classify the unknown input with respect to the modal. The recognition system involved training and testing session that is known as client identification (CID). In this project, twenty clients are tested with the developed system. The best overall performance at 44 kHz was 93.92% for ECG and the worst overall performance was ECG at 88.47%. The results were compared to the best overall performance at 44 kHz for (20clients) to increment of clients, which was 90.00% for HS and the worst overall performance falls at ECG at 79.91%. It can be concluded that the difference multimodal biometric has a substantial effect on performance of the biometric system and with the increment of data, even with higher frequency sampling, the performance still decreased slightly as predicted.

Keywords: electrocardiogram, phonocardiogram, hidden markov model, mel frequency cepstral coeffiecients, client identification

Procedia PDF Downloads 270
4377 Composite Kernels for Public Emotion Recognition from Twitter

Authors: Chien-Hung Chen, Yan-Chun Hsing, Yung-Chun Chang

Abstract:

The Internet has grown into a powerful medium for information dispersion and social interaction that leads to a rapid growth of social media which allows users to easily post their emotions and perspectives regarding certain topics online. Our research aims at using natural language processing and text mining techniques to explore the public emotions expressed on Twitter by analyzing the sentiment behind tweets. In this paper, we propose a composite kernel method that integrates tree kernel with the linear kernel to simultaneously exploit both the tree representation and the distributed emotion keyword representation to analyze the syntactic and content information in tweets. The experiment results demonstrate that our method can effectively detect public emotion of tweets while outperforming the other compared methods.

Keywords: emotion recognition, natural language processing, composite kernel, sentiment analysis, text mining

Procedia PDF Downloads 209
4376 Small Target Recognition Based on Trajectory Information

Authors: Saad Alkentar, Abdulkareem Assalem

Abstract:

Recognizing small targets has always posed a significant challenge in image analysis. Over long distances, the image signal-to-noise ratio tends to be low, limiting the amount of useful information available to detection systems. Consequently, visual target recognition becomes an intricate task to tackle. In this study, we introduce a Track Before Detect (TBD) approach that leverages target trajectory information (coordinates) to effectively distinguish between noise and potential targets. By reframing the problem as a multivariate time series classification, we have achieved remarkable results. Specifically, our TBD method achieves an impressive 97% accuracy in separating target signals from noise within a mere half-second time span (consisting of 10 data points). Furthermore, when classifying the identified targets into our predefined categories—airplane, drone, and bird—we achieve an outstanding classification accuracy of 96% over a more extended period of 1.5 seconds (comprising 30 data points).

Keywords: small targets, drones, trajectory information, TBD, multivariate time series

Procedia PDF Downloads 32
4375 A Laundry Algorithm for Colored Textiles

Authors: H. E. Budak, B. Arslan-Ilkiz, N. Cakmakci, I. Gocek, U. K. Sahin, H. Acikgoz-Tufan, M. H. Arslan

Abstract:

The aim of this study is to design a novel laundry algorithm for colored textiles which have significant decoloring problem. During the experimental work, bleached knitted single jersey fabric made of 100% cotton and dyed with reactive dyestuff was utilized, since according to a conducted survey textiles made of cotton are the most demanded textile products in the textile market by the textile consumers and for coloration of textiles reactive dyestuffs are the ones that are the most commonly used in the textile industry for dyeing cotton-made products. Therefore, the fabric used in this study was selected and purchased in accordance with the survey results. The fabric samples cut out of this fabric were dyed with different dyeing parameters by using Remazol Brilliant Red 3BS dyestuff in Gyrowash machine at laboratory conditions. From the alternative reactive-dyed cotton fabric samples, the ones that have high tendency to color loss were determined and examined. Accordingly, the parameters of the dyeing process used for these fabric samples were evaluated and the dyeing process which was chosen to be used for causing high tendency to color loss for the cotton fabrics was determined in order to reveal the level of improvement in color loss during this study clearly. Afterwards, all of the untreated fabric samples cut out of the fabric purchased were dyed with the dyeing process selected. When dyeing process was completed, an experimental design was created for the laundering process by using Minitab® program considering temperature, time and mechanical action as parameters. All of the washing experiments were performed in domestic washing machine. 16 washing experiments were performed with 8 different experimental conditions and 2 repeats for each condition. After each of the washing experiments, water samples of the main wash of the laundering process were measured with UV spectrophotometer. The values obtained were compared with the calibration curve of the materials used for the dyeing process. The results of the washing experiments were statistically analyzed with Minitab® program. According to the results, the most suitable washing algorithm to be used in terms of the parameters temperature, time and mechanical action for domestic washing machines for minimizing fabric color loss was chosen. The laundry algorithm proposed in this study have the ability of minimalizing the problem of color loss of colored textiles in washing machines by eliminating the negative effects of the parameters of laundering process on color of textiles without compromising the fundamental effects of basic cleaning action being performed properly. Therefore, since fabric color loss is minimized with this washing algorithm, dyestuff residuals will definitely be lower in the grey water released from the laundering process. In addition to this, with this laundry algorithm it is possible to wash and clean other types of textile products with proper cleaning effect and minimized color loss.

Keywords: color loss, laundry algorithm, textiles, domestic washing process

Procedia PDF Downloads 337
4374 Identity Verification Based on Multimodal Machine Learning on Red Green Blue (RGB) Red Green Blue-Depth (RGB-D) Voice Data

Authors: LuoJiaoyang, Yu Hongyang

Abstract:

In this paper, we experimented with a new approach to multimodal identification using RGB, RGB-D and voice data. The multimodal combination of RGB and voice data has been applied in tasks such as emotion recognition and has shown good results and stability, and it is also the same in identity recognition tasks. We believe that the data of different modalities can enhance the effect of the model through mutual reinforcement. We try to increase the three modalities on the basis of the dual modalities and try to improve the effectiveness of the network by increasing the number of modalities. We also implemented the single-modal identification system separately, tested the data of these different modalities under clean and noisy conditions, and compared the performance with the multimodal model. In the process of designing the multimodal model, we tried a variety of different fusion strategies and finally chose the fusion method with the best performance. The experimental results show that the performance of the multimodal system is better than that of the single modality, especially in dealing with noise, and the multimodal system can achieve an average improvement of 5%.

Keywords: multimodal, three modalities, RGB-D, identity verification

Procedia PDF Downloads 60
4373 Modal Approach for Decoupling Damage Cost Dependencies in Building Stories

Authors: Haj Najafi Leila, Tehranizadeh Mohsen

Abstract:

Dependencies between diverse factors involved in probabilistic seismic loss evaluation are recognized to be an imperative issue in acquiring accurate loss estimates. Dependencies among component damage costs could be taken into account considering two partial distinct states of independent or perfectly-dependent for component damage states; however, in our best knowledge, there is no available procedure to take account of loss dependencies in story level. This paper attempts to present a method called "modal cost superposition method" for decoupling story damage costs subjected to earthquake ground motions dealt with closed form differential equations between damage cost and engineering demand parameters which should be solved in complex system considering all stories' cost equations by the means of the introduced "substituted matrixes of mass and stiffness". Costs are treated as probabilistic variables with definite statistic factors of median and standard deviation amounts and a presumed probability distribution. To supplement the proposed procedure and also to display straightforwardness of its application, one benchmark study has been conducted. Acceptable compatibility has been proven for the estimated damage costs evaluated by the new proposed modal and also frequently used stochastic approaches for entire building; however, in story level, insufficiency of employing modification factor for incorporating occurrence probability dependencies between stories has been revealed due to discrepant amounts of dependency between damage costs of different stories. Also, more dependency contribution in occurrence probability of loss could be concluded regarding more compatibility of loss results in higher stories than the lower ones, whereas reduction in incorporation portion of cost modes provides acceptable level of accuracy and gets away from time consuming calculations including some limited number of cost modes in high mode situation.

Keywords: dependency, story-cost, cost modes, engineering demand parameter

Procedia PDF Downloads 168
4372 Time Pressure and Its Effect at Tactical Level of Disaster Management

Authors: Agoston Restas

Abstract:

Introduction: In case of managing disasters decision makers can face many times such a special situation where any pre-sign of the drastically change is missing therefore the improvised decision making can be required. The complexity, ambiguity, uncertainty or the volatility of the situation can require many times the improvisation as decision making. It can be taken at any level of the management (strategic, operational and tactical) but at tactical level the main reason of the improvisation is surely time pressure. It is certainly the biggest problem during the management. Methods: The author used different tools and methods to achieve his goals; one of them was the study of the relevant literature, the other one was his own experience as a firefighting manager. Other results come from two surveys that are referred to; one of them was an essay analysis, the second one was a word association test, specially created for the research. Results and discussion: This article proves that, in certain situations, the multi-criteria, evaluating decision-making processes simply cannot be used or only in a limited manner. However, it can be seen that managers, directors or commanders are many times in situations that simply cannot be ignored when making decisions which should be made in a short time. The functional background of decisions made in a short time, their mechanism, which is different from the conventional, was studied lately and this special decision procedure was given the name recognition-primed decision. In the article, author illustrates the limits of the possibilities of analytical decision-making, presents the general operating mechanism of recognition-primed decision-making, elaborates on its special model relevant to managers at tactical level, as well as explore and systemize the factors that facilitate (catalyze) the processes with an example with fire managers.

Keywords: decision making, disaster managers, recognition primed decision, model for making decisions in emergencies

Procedia PDF Downloads 249
4371 Online Pose Estimation and Tracking Approach with Siamese Region Proposal Network

Authors: Cheng Fang, Lingwei Quan, Cunyue Lu

Abstract:

Human pose estimation and tracking are to accurately identify and locate the positions of human joints in the video. It is a computer vision task which is of great significance for human motion recognition, behavior understanding and scene analysis. There has been remarkable progress on human pose estimation in recent years. However, more researches are needed for human pose tracking especially for online tracking. In this paper, a framework, called PoseSRPN, is proposed for online single-person pose estimation and tracking. We use Siamese network attaching a pose estimation branch to incorporate Single-person Pose Tracking (SPT) and Visual Object Tracking (VOT) into one framework. The pose estimation branch has a simple network structure that replaces the complex upsampling and convolution network structure with deconvolution. By augmenting the loss of fully convolutional Siamese network with the pose estimation task, pose estimation and tracking can be trained in one stage. Once trained, PoseSRPN only relies on a single bounding box initialization and producing human joints location. The experimental results show that while maintaining the good accuracy of pose estimation on COCO and PoseTrack datasets, the proposed method achieves a speed of 59 frame/s, which is superior to other pose tracking frameworks.

Keywords: computer vision, pose estimation, pose tracking, Siamese network

Procedia PDF Downloads 142
4370 Fight against Money Laundering with Optical Character Recognition

Authors: Saikiran Subbagari, Avinash Malladhi

Abstract:

Anti Money Laundering (AML) regulations are designed to prevent money laundering and terrorist financing activities worldwide. Financial institutions around the world are legally obligated to identify, assess and mitigate the risks associated with money laundering and report any suspicious transactions to governing authorities. With increasing volumes of data to analyze, financial institutions seek to automate their AML processes. In the rise of financial crimes, optical character recognition (OCR), in combination with machine learning (ML) algorithms, serves as a crucial tool for automating AML processes by extracting the data from documents and identifying suspicious transactions. In this paper, we examine the utilization of OCR for AML and delve into various OCR techniques employed in AML processes. These techniques encompass template-based, feature-based, neural network-based, natural language processing (NLP), hidden markov models (HMMs), conditional random fields (CRFs), binarizations, pattern matching and stroke width transform (SWT). We evaluate each technique, discussing their strengths and constraints. Also, we emphasize on how OCR can improve the accuracy of customer identity verification by comparing the extracted text with the office of foreign assets control (OFAC) watchlist. We will also discuss how OCR helps to overcome language barriers in AML compliance. We also address the implementation challenges that OCR-based AML systems may face and offer recommendations for financial institutions based on the data from previous research studies, which illustrate the effectiveness of OCR-based AML.

Keywords: anti-money laundering, compliance, financial crimes, fraud detection, machine learning, optical character recognition

Procedia PDF Downloads 128
4369 Analysis of 3 dB Directional Coupler Based On Silicon-On-Insulator (SOI) Large Cross-Section Rib Waveguide

Authors: Nurdiani Zamhari, Abang Annuar Ehsan

Abstract:

The 3 dB directional coupler is designed by using silicon-on-insulator (SOI) large cross-section and simulate by Beam Propagation Method at the communication wavelength of 1.55 µm and 1.48 µm. The geometry is shaped with rib height (H) of 6 µm and varied in step factor (r) which is 0.5, 0.6, 0.7 and 0.8. The wave guide spacing is also fixed to 5 µm and the slab width is symmetrical. In general, the 3 dB coupling lengths for four different cross-sections are several millimetre long. The 1.48 of wavelength give the longer coupling length if compare to 1.55 at the same step factor (r). Besides, the low loss propagation is achieved with less than 2 % of propagation loss.

Keywords: 3 dB directional couplers, silicon-on-insulator, symmetrical rib waveguide, OptiBPM 9

Procedia PDF Downloads 505