Search results for: binary vector quantization (BVQ)
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 1769

Search results for: binary vector quantization (BVQ)

209 Analysis of the Savings Behaviour of Rice Farmers in Tiaong, Quezon, Philippines

Authors: Angelika Kris D. Dalangin, Cesar B. Quicoy

Abstract:

Rice farming is a major source of livelihood and employment in the Philippines, but it requires a substantial amount of capital. Capital may come from income (farm, non-farm, and off-farm), savings and credit. However, rice farmers suffer from lack of capital due to high costs of inputs and low productivity. Capital insufficiency, coupled with low productivity, hindered them to meet their basic household and production needs. Hence, they resorted to borrowing money, mostly from informal lenders who charge very high interest rates. As another source of capital, savings can help rice farmers meet their basic needs for both the household and the farm. However, information is inadequate whether the farmers save or not, as well as, why they do not depend on savings to augment their lack of capital. Thus, it is worth analyzing how rice farmers saved. The study revealed, using the actual savings which is the difference between the household income and expenditure, that about three-fourths (72%) of the total number of farmers interviewed are savers. However, when they were asked whether they are savers or not, more than half of them considered themselves as non-savers. This gap shows that there are many farmers who think that they do not have savings at all; hence they continue to borrow money and do not depend on savings to augment their lack of capital. The study also identified the forms of savings, saving motives, and savings utilization among rice farmers. Results revealed that, for the past 12 months, most of the farmers saved cash at home for liquidity purposes while others deposited cash in banks and/or saved their money in the form of livestock. Among the most important reasons of farmers for saving are for daily household expenses, for building a house, for emergency purposes, for retirement, and for their next production. Furthermore, the study assessed the factors affecting the rice farmers’ savings behaviour using logistic regression. Results showed that the factors found to be significant were presence of non-farm income, per capita net farm income, and per capita household expense. The presence of non-farm income and per capita net farm income positively affects the farmers’ savings behaviour. On the other hand, per capita household expenses have negative effect. The effect, however, of per capita net farm income and household expenses is very negligible because of the very small chance that the farmer is a saver. Generally, income and expenditure were proved to be significant factors that affect the savings behaviour of the rice farmers. However, most farmers could not save regularly due to low farm income and high household and farm expenditures. Thus, it is highly recommended that government should develop programs or implement policies that will create more jobs for the farmers and their family members. In addition, programs and policies should be implemented to increase farm productivity and income.

Keywords: agricultural economics, agricultural finance, binary logistic regression, logit, Philippines, Quezon, rice farmers, savings, savings behaviour

Procedia PDF Downloads 227
208 Production and Purification of Salmonella Typhimurium MisL Autotransporter Protein in Escherichia coli

Authors: Neslihan Taskale Karatug, Mustafa Akcelik

Abstract:

Some literature data show that misL protein play a role on host immune response formed against Salmonella Typhimurium. The aim of the present study is to learn the role of the protein in S. Typhimurium pathogenicity. To describe certain functions of the protein, primarily recombinant misL protein was produced and purified. PCR was performed using a primer set targeted to passenger domain of the misL gene on S. Typhimurium LT2 genome. Amplicon and pet28a vector were enzymatically cleaved with EcoRI and NheI. The digested DNA materials were purified with High Pure PCR Product Purification Kit. The ligation reaction was achieved with the pure products. After preparation of competent Escherichia coli Dh5α, ligation mix was transformed into the cell by electroporation. To confirm the existence of insert gene, recombinant plasmid DNA of Dh5α was isolated with high pure plasmid DNA kit. Proved the correctness of recombinant plasmid was electroporated to BL21. The cell was induced by IPTG. After induction, the presence of recombinant protein was checked by SDS-PAGE. The recombinant misL protein was purified using HisPur Ni-NTA spin colon. The pure protein was shown by SDS-PAGE and western blot immünoassay. The concentration of the protein was measured BCA Protein Assay kit. In the wake of ligation with digested products (2 kb misL and 5.4 kb pet28a) visualised on gel size of the band was about 7.4 kb and was named as pNT01. The pNT01 recombinant plasmid was transformed into Dh5α and colonies were chosen in selective medium. Plasmid DNA isolation from them was carried out. PCR was achieved on the pNT01 to check misL and 2 kb band was observed on the agarose gel. After electroporation of the plasmid and induction of the cell, 68 kDa misL protein was seen. Subsequent to the purification of the protein, only a band was observed on SDS-PAGE. Association of the pure protein with anti-his antibody was verified by the western blot assay. The concentration of the pure misL protein was determined as 345 μg/mL. Production of polyclonal antibody will be achieved by using the obtained pure recombinant misL protein as next step. The role of the protein will come out on the immune system together some assays.

Keywords: cloning, Escherichia coli, recombinant protein purification, Salmonella Typhimurium

Procedia PDF Downloads 389
207 Preschoolers’ Selective Trust in Moral Promises

Authors: Yuanxia Zheng, Min Zhong, Cong Xin, Guoxiong Liu, Liqi Zhu

Abstract:

Trust is a critical foundation of social interaction and development, playing a significant role in the physical and mental well-being of children, as well as their social participation. Previous research has demonstrated that young children do not blindly trust others but make selective trust judgments based on available information. The characteristics of speakers can influence children’s trust judgments. According to Mayer et al.’s model of trust, these characteristics of speakers, including ability, benevolence, and integrity, can influence children’s trust judgments. While previous research has focused primarily on the effects of ability and benevolence, there has been relatively little attention paid to integrity, which refers to individuals’ adherence to promises, fairness, and justice. This study focuses specifically on how keeping/breaking promises affects young children’s trust judgments. The paradigm of selective trust was employed in two experiments. A sample size of 100 children was required for an effect size of w = 0.30,α = 0.05,1-β = 0.85, using G*Power 3.1. This study employed a 2×2 within-subjects design to investigate the effects of moral valence of promises (within-subjects factor: moral vs. immoral promises), and fulfilment of promises (within-subjects factor: kept vs. broken promises) on children’s trust judgments (divided into declarative and promising contexts). In Experiment 1 adapted binary choice paradigms, presenting 118 preschoolers (62 girls, Mean age = 4.99 years, SD = 0.78) with four conflict scenarios involving the keeping or breaking moral/immoral promises, in order to investigate children’s trust judgments. Experiment 2 utilized single choice paradigms, in which 112 preschoolers (57 girls, Mean age = 4.94 years, SD = 0.80) were presented four stories to examine their level of trust. The results of Experiment 1 showed that preschoolers selectively trusted both promisors who kept moral promises and those who broke immoral promises, as well as their assertions and new promises. Additionally, the 5.5-6.5-year-old children are more likely to trust both promisors who keep moral promises and those who break immoral promises more than the 3.5- 4.5-year-old children. Moreover, preschoolers are more likely to make accurate trust judgments towards promisor who kept moral promise compared to those who broke immoral promises. The results of Experiment 2 showed significant differences of preschoolers’ trust degree: kept moral promise > broke immoral promise > broke moral promise ≈ kept immoral promise. This study is the first to investigate the development of trust judgement in moral promise among preschoolers aged 3.5-6.5. The results show that preschoolers can consider both valence and fulfilment of promises when making trust judgments. Furthermore, as preschoolers mature, they become more inclined to trust promisors who keep moral promises and those who break immoral promises. Additionally, the study reveals that preschoolers have the highest level of trust in promisors who kept moral promises, followed by those who broke immoral promises. Promisors who broke moral promises and those who kept immoral promises are trusted the least. These findings contribute valuable insights to our understanding of moral promises and trust judgment.

Keywords: promise, trust, moral judgement, preschoolers

Procedia PDF Downloads 53
206 Quantifying Multivariate Spatiotemporal Dynamics of Malaria Risk Using Graph-Based Optimization in Southern Ethiopia

Authors: Yonas Shuke Kitawa

Abstract:

Background: Although malaria incidence has substantially fallen sharply over the past few years, the rate of decline varies by district, time, and malaria type. Despite this turn-down, malaria remains a major public health threat in various districts of Ethiopia. Consequently, the present study is aimed at developing a predictive model that helps to identify the spatio-temporal variation in malaria risk by multiple plasmodium species. Methods: We propose a multivariate spatio-temporal Bayesian model to obtain a more coherent picture of the temporally varying spatial variation in disease risk. The spatial autocorrelation in such a data set is typically modeled by a set of random effects that assign a conditional autoregressive prior distribution. However, the autocorrelation considered in such cases depends on a binary neighborhood matrix specified through the border-sharing rule. Over here, we propose a graph-based optimization algorithm for estimating the neighborhood matrix that merely represents the spatial correlation by exploring the areal units as the vertices of a graph and the neighbor relations as the series of edges. Furthermore, we used aggregated malaria count in southern Ethiopia from August 2013 to May 2019. Results: We recognized that precipitation, temperature, and humidity are positively associated with the malaria threat in the area. On the other hand, enhanced vegetation index, nighttime light (NTL), and distance from coastal areas are negatively associated. Moreover, nonlinear relationships were observed between malaria incidence and precipitation, temperature, and NTL. Additionally, lagged effects of temperature and humidity have a significant effect on malaria risk by either species. More elevated risk of P. falciparum was observed following the rainy season, and unstable transmission of P. vivax was observed in the area. Finally, P. vivax risks are less sensitive to environmental factors than those of P. falciparum. Conclusion: The improved inference was gained by employing the proposed approach in comparison to the commonly used border-sharing rule. Additionally, different covariates are identified, including delayed effects, and elevated risks of either of the cases were observed in districts found in the central and western regions. As malaria transmission operates in a spatially continuous manner, a spatially continuous model should be employed when it is computationally feasible.

Keywords: disease mapping, MSTCAR, graph-based optimization algorithm, P. falciparum, P. vivax, waiting matrix

Procedia PDF Downloads 76
205 Kinematic Analysis of Human Gait for Typical Postures of Walking, Running and Cart Pulling

Authors: Nupur Karmaker, Hasin Aupama Azhari, Abdul Al Mortuza, Abhijit Chanda, Golam Abu Zakaria

Abstract:

Purpose: The purpose of gait analysis is to determine the biomechanics of the joint, phases of gait cycle, graphical and analytical analysis of degree of rotation, analysis of the electrical activity of muscles and force exerted on the hip joint at different locomotion during walking, running and cart pulling. Methods and Materials: Visual gait analysis and electromyography method has been used to detect the degree of rotation of joints and electrical activity of muscles. In cinematography method an object is observed from different sides and takes its video. Cart pulling length has been divided into frames with respect to time by using video splitter software. Phases of gait cycle, degree of rotation of joints, EMG profile and force analysis during walking and running has been taken from different papers. Gait cycle and degree of rotation of joints during cart pulling has been prepared by using video camera, stop watch, video splitter software and Microsoft Excel. Results and Discussion: During the cart pulling the force exerted on hip is the resultant of various forces. The force on hip is the vector sum of the force Fg= mg, due the body of weight of the person and Fa= ma, due to the velocity. Maximum stance phase shows during cart pulling and minimum shows during running. During cart pulling shows maximum degree of rotation of hip joint, knee: running, and ankle: cart pulling. During walking, it has been observed minimum degree of rotation of hip, ankle: during running. During cart pulling, dynamic force depends on the walking velocity, body weight and load weight. Conclusions: 80% people suffer gait related disease with increasing their age. Proper care should take during cart pulling. It will be better to establish the gait laboratory to determine the gait related diseases. If the way of cart pulling is changed i.e the design of cart pulling machine, load bearing system is changed then it would possible to reduce the risk of limb loss, flat foot syndrome and varicose vein in lower limb.

Keywords: kinematic, gait, gait lab, phase, force analysis

Procedia PDF Downloads 575
204 Standard Essential Patents for Artificial Intelligence Hardware and the Implications For Intellectual Property Rights

Authors: Wendy de Gomez

Abstract:

Standardization is a critical element in the ability of a society to reduce uncertainty, subjectivity, misrepresentation, and interpretation while simultaneously contributing to innovation. Technological standardization is critical to codify specific operationalization through legal instruments that provide rules of development, expectation, and use. In the current emerging technology landscape Artificial Intelligence (AI) hardware as a general use technology has seen incredible growth as evidenced from AI technology patents between 2012 and 2018 in the United States Patent Trademark Office (USPTO) AI dataset. However, as outlined in the 2023 United States Government National Standards Strategy for Critical and Emerging Technology the codification through standardization of emerging technologies such as AI has not kept pace with its actual technological proliferation. This gap has the potential to cause significant divergent possibilities for the downstream outcomes of AI in both the short and long term. This original empirical research provides an overview of the standardization efforts around AI in different geographies and provides a background to standardization law. It quantifies the longitudinal trend of Artificial Intelligence hardware patents through the USPTO AI dataset. It seeks evidence of existing Standard Essential Patents from these AI hardware patents through a text analysis of the Statement of patent history and the Field of the invention of these patents in Patent Vector and examines their determination as a Standard Essential Patent and their inclusion in existing AI technology standards across the four main AI standards bodies- European Telecommunications Standards Institute (ETSI); International Telecommunication Union (ITU)/ Telecommunication Standardization Sector (-T); Institute of Electrical and Electronics Engineers (IEEE); and the International Organization for Standardization (ISO). Once the analysis is complete the paper will discuss both the theoretical and operational implications of F/Rand Licensing Agreements for the owners of these Standard Essential Patents in the United States Court and Administrative system. It will conclude with an evaluation of how Standard Setting Organizations (SSOs) can work with SEP owners more effectively through various forms of Intellectual Property mechanisms such as patent pools.

Keywords: patents, artifical intelligence, standards, F/Rand agreements

Procedia PDF Downloads 85
203 Unsupervised Detection of Burned Area from Remote Sensing Images Using Spatial Correlation and Fuzzy Clustering

Authors: Tauqir A. Moughal, Fusheng Yu, Abeer Mazher

Abstract:

Land-cover and land-use change information are important because of their practical uses in various applications, including deforestation, damage assessment, disasters monitoring, urban expansion, planning, and land management. Therefore, developing change detection methods for remote sensing images is an important ongoing research agenda. However, detection of change through optical remote sensing images is not a trivial task due to many factors including the vagueness between the boundaries of changed and unchanged regions and spatial dependence of the pixels to its neighborhood. In this paper, we propose a binary change detection technique for bi-temporal optical remote sensing images. As in most of the optical remote sensing images, the transition between the two clusters (change and no change) is overlapping and the existing methods are incapable of providing the accurate cluster boundaries. In this regard, a methodology has been proposed which uses the fuzzy c-means clustering to tackle the problem of vagueness in the changed and unchanged class by formulating the soft boundaries between them. Furthermore, in order to exploit the neighborhood information of the pixels, the input patterns are generated corresponding to each pixel from bi-temporal images using 3×3, 5×5 and 7×7 window. The between images and within image spatial dependence of the pixels to its neighborhood is quantified by using Pearson product moment correlation and Moran’s I statistics, respectively. The proposed technique consists of two phases. At first, between images and within image spatial correlation is calculated to utilize the information that the pixels at different locations may not be independent. Second, fuzzy c-means technique is used to produce two clusters from input feature by not only taking care of vagueness between the changed and unchanged class but also by exploiting the spatial correlation of the pixels. To show the effectiveness of the proposed technique, experiments are conducted on multispectral and bi-temporal remote sensing images. A subset (2100×1212 pixels) of a pan-sharpened, bi-temporal Landsat 5 thematic mapper optical image of Los Angeles, California, is used in this study which shows a long period of the forest fire continued from July until October 2009. Early forest fire and later forest fire optical remote sensing images were acquired on July 5, 2009 and October 25, 2009, respectively. The proposed technique is used to detect the fire (which causes change on earth’s surface) and compared with the existing K-means clustering technique. Experimental results showed that proposed technique performs better than the already existing technique. The proposed technique can be easily extendable for optical hyperspectral images and is suitable for many practical applications.

Keywords: burned area, change detection, correlation, fuzzy clustering, optical remote sensing

Procedia PDF Downloads 168
202 God, The Master Programmer: The Relationship Between God and Computers

Authors: Mohammad Sabbagh

Abstract:

Anyone who reads the Torah or the Quran learns that GOD created everything that is around us, seen and unseen, in six days. Within HIS plan of creation, HE placed for us a key proof of HIS existence which is essentially computers and the ability to program them. Digital computer programming began with binary instructions, which eventually evolved to what is known as high-level programming languages. Any programmer in our modern time can attest that you are essentially giving the computer commands by words and when the program is compiled, whatever is processed as output is limited to what the computer was given as an ability and furthermore as an instruction. So one can deduce that GOD created everything around us with HIS words, programming everything around in six days, just like how we can program a virtual world on the computer. GOD did mention in the Quran that one day where GOD’s throne is, is 1000 years of what we count; therefore, one might understand that GOD spoke non-stop for 6000 years of what we count, and gave everything it’s the function, attributes, class, methods and interactions. Similar to what we do in object-oriented programming. Of course, GOD has the higher example, and what HE created is much more than OOP. So when GOD said that everything is already predetermined, it is because any input, whether physical, spiritual or by thought, is outputted by any of HIS creatures, the answer has already been programmed. Any path, any thought, any idea has already been laid out with a reaction to any decision an inputter makes. Exalted is GOD!. GOD refers to HIMSELF as The Fastest Accountant in The Quran; the Arabic word that was used is close to processor or calculator. If you create a 3D simulation of a supernova explosion to understand how GOD produces certain elements and fuses protons together to spread more of HIS blessings around HIS skies; in 2022 you are going to require one of the strongest, fastest, most capable supercomputers of the world that has a theoretical speed of 50 petaFLOPS to accomplish that. In other words, the ability to perform one quadrillion (1015) floating-point operations per second. A number a human cannot even fathom. To put in more of a perspective, GOD is calculating when the computer is going through those 50 petaFLOPS calculations per second and HE is also calculating all the physics of every atom and what is smaller than that in all the actual explosion, and it’s all in truth. When GOD said HE created the world in truth, one of the meanings a person can understand is that when certain things occur around you, whether how a car crashes or how a tree grows; there is a science and a way to understand it, and whatever programming or science you deduce from whatever event you observed, it can relate to other similar events. That is why GOD might have said in The Quran that it is the people of knowledge, scholars, or scientist that fears GOD the most! One thing that is essential for us to keep up with what the computer is doing and for us to track our progress along with any errors is we incorporate logging mechanisms and backups. GOD in The Quran said that ‘WE used to copy what you used to do’. Essentially as the world is running, think of it as an interactive movie that is being played out in front of you, in a full-immersive non-virtual reality setting. GOD is recording it, from every angle to every thought, to every action. This brings the idea of how scary the Day of Judgment will be when one might realize that it’s going to be a fully immersive video when we would be getting and reading our book.

Keywords: programming, the Quran, object orientation, computers and humans, GOD

Procedia PDF Downloads 106
201 Oligoalkylamine Modified Poly(Amidoamine) Generation 4.5 Dendrimer for the Delivery of Small Interfering RNA

Authors: Endris Yibru Hanurry, Wei-Hsin Hsu, Hsieh-Chih Tsai

Abstract:

In recent years, the discovery of small interfering RNAs (siRNAs) has got great attention for the treatment of cancer and other diseases. However, the therapeutic efficacy of siRNAs has been faced with many drawbacks because of short half-life in blood circulation, poor membrane penetration, weak endosomal escape and inadequate release into the cytosol. To overcome these drawbacks, we designed a non-viral vector by conjugating polyamidoamine generation 4.5 dendrimer (PDG4.5) with diethylenetriamine (DETA)- and tetraethylenepentamine (TEPA) followed by binding with siRNA to form polyplexes through electrostatic interaction. The result of 1H nuclear magnetic resonance (NMR), 13C NMR, correlation spectroscopy, heteronuclear single–quantum correlation spectroscopy, and Fourier transform infrared spectroscopy confirmed the successful conjugation of DETA and TEPA with PDG4.5. Then, the size, surface charge, morphology, binding ability, stability, release assay, toxicity and cellular internalization were analyzed to explore the physicochemical and biological properties of PDG4.5-DETA and PDG4.5-TEPA polyplexes at specific N/P ratios. The polyplexes (N/P = 8) exhibited spherical nanosized (125 and 85 nm) particles with optimum surface charge (13 and 26 mV), showed strong siRNA binding ability, protected the siRNA against enzyme digestion and accepted biocompatibility to the HeLa cells. Qualitatively, the fluorescence microscopy image revealed the delocalization (Manders’ coefficient 0.63 and 0.53 for PDG4.5-DETA and PDG4.5-TEPA, respectively) of polyplexes and the translocation of the siRNA throughout the cytosol to show a decent cellular internalization and intracellular biodistribution of polyplexes in HeLa cells. Quantitatively, the flow cytometry result indicated that a significant (P < 0.05) amount of siRNA was internalized by cells treated with PDG4.5-DETA (68.5%) and PDG4.5-TEPA (73%) polyplexes. Generally, PDG4.5-DETA and PDG4.5-TEPA were ideal nanocarriers of siRNA in vitro and might be used as promising candidates for in vivo study and future pharmaceutical applications.

Keywords: non-viral carrier, oligoalkylamine, poly(amidoamine) dendrimer, polyplexes, siRNA

Procedia PDF Downloads 127
200 An Adaptive Oversampling Technique for Imbalanced Datasets

Authors: Shaukat Ali Shahee, Usha Ananthakumar

Abstract:

A data set exhibits class imbalance problem when one class has very few examples compared to the other class, and this is also referred to as between class imbalance. The traditional classifiers fail to classify the minority class examples correctly due to its bias towards the majority class. Apart from between-class imbalance, imbalance within classes where classes are composed of a different number of sub-clusters with these sub-clusters containing different number of examples also deteriorates the performance of the classifier. Previously, many methods have been proposed for handling imbalanced dataset problem. These methods can be classified into four categories: data preprocessing, algorithmic based, cost-based methods and ensemble of classifier. Data preprocessing techniques have shown great potential as they attempt to improve data distribution rather than the classifier. Data preprocessing technique handles class imbalance either by increasing the minority class examples or by decreasing the majority class examples. Decreasing the majority class examples lead to loss of information and also when minority class has an absolute rarity, removing the majority class examples is generally not recommended. Existing methods available for handling class imbalance do not address both between-class imbalance and within-class imbalance simultaneously. In this paper, we propose a method that handles between class imbalance and within class imbalance simultaneously for binary classification problem. Removing between class imbalance and within class imbalance simultaneously eliminates the biases of the classifier towards bigger sub-clusters by minimizing the error domination of bigger sub-clusters in total error. The proposed method uses model-based clustering to find the presence of sub-clusters or sub-concepts in the dataset. The number of examples oversampled among the sub-clusters is determined based on the complexity of sub-clusters. The method also takes into consideration the scatter of the data in the feature space and also adaptively copes up with unseen test data using Lowner-John ellipsoid for increasing the accuracy of the classifier. In this study, neural network is being used as this is one such classifier where the total error is minimized and removing the between-class imbalance and within class imbalance simultaneously help the classifier in giving equal weight to all the sub-clusters irrespective of the classes. The proposed method is validated on 9 publicly available data sets and compared with three existing oversampling techniques that rely on the spatial location of minority class examples in the euclidean feature space. The experimental results show the proposed method to be statistically significantly superior to other methods in terms of various accuracy measures. Thus the proposed method can serve as a good alternative to handle various problem domains like credit scoring, customer churn prediction, financial distress, etc., that typically involve imbalanced data sets.

Keywords: classification, imbalanced dataset, Lowner-John ellipsoid, model based clustering, oversampling

Procedia PDF Downloads 415
199 A Method for Clinical Concept Extraction from Medical Text

Authors: Moshe Wasserblat, Jonathan Mamou, Oren Pereg

Abstract:

Natural Language Processing (NLP) has made a major leap in the last few years, in practical integration into medical solutions; for example, extracting clinical concepts from medical texts such as medical condition, medication, treatment, and symptoms. However, training and deploying those models in real environments still demands a large amount of annotated data and NLP/Machine Learning (ML) expertise, which makes this process costly and time-consuming. We present a practical and efficient method for clinical concept extraction that does not require costly labeled data nor ML expertise. The method includes three steps: Step 1- the user injects a large in-domain text corpus (e.g., PubMed). Then, the system builds a contextual model containing vector representations of concepts in the corpus, in an unsupervised manner (e.g., Phrase2Vec). Step 2- the user provides a seed set of terms representing a specific medical concept (e.g., for the concept of the symptoms, the user may provide: ‘dry mouth,’ ‘itchy skin,’ and ‘blurred vision’). Then, the system matches the seed set against the contextual model and extracts the most semantically similar terms (e.g., additional symptoms). The result is a complete set of terms related to the medical concept. Step 3 –in production, there is a need to extract medical concepts from the unseen medical text. The system extracts key-phrases from the new text, then matches them against the complete set of terms from step 2, and the most semantically similar will be annotated with the same medical concept category. As an example, the seed symptom concepts would result in the following annotation: “The patient complaints on fatigue [symptom], dry skin [symptom], and Weight loss [symptom], which can be an early sign for Diabetes.” Our evaluations show promising results for extracting concepts from medical corpora. The method allows medical analysts to easily and efficiently build taxonomies (in step 2) representing their domain-specific concepts, and automatically annotate a large number of texts (in step 3) for classification/summarization of medical reports.

Keywords: clinical concepts, concept expansion, medical records annotation, medical records summarization

Procedia PDF Downloads 133
198 National Assessment for Schools in Saudi Arabia: Score Reliability and Plausible Values

Authors: Dimiter M. Dimitrov, Abdullah Sadaawi

Abstract:

The National Assessment for Schools (NAFS) in Saudi Arabia consists of standardized tests in Mathematics, Reading, and Science for school grade levels 3, 6, and 9. One main goal is to classify students into four categories of NAFS performance (minimal, basic, proficient, and advanced) by schools and the entire national sample. The NAFS scoring and equating is performed on a bounded scale (D-scale: ranging from 0 to 1) in the framework of the recently developed “D-scoring method of measurement.” The specificity of the NAFS measurement framework and data complexity presented both challenges and opportunities to (a) the estimation of score reliability for schools, (b) setting cut-scores for the classification of students into categories of performance, and (c) generating plausible values for distributions of student performance on the D-scale. The estimation of score reliability at the school level was performed in the framework of generalizability theory (GT), with students “nested” within schools and test items “nested” within test forms. The GT design was executed via a multilevel modeling syntax code in R. Cut-scores (on the D-scale) for the classification of students into performance categories was derived via a recently developed method of standard setting, referred to as “Response Vector for Mastery” (RVM) method. For each school, the classification of students into categories of NAFS performance was based on distributions of plausible values for the students’ scores on NAFS tests by grade level (3, 6, and 9) and subject (Mathematics, Reading, and Science). Plausible values (on the D-scale) for each individual student were generated via random selection from a statistical logit-normal distribution with parameters derived from the student’s D-score and its conditional standard error, SE(D). All procedures related to D-scoring, equating, generating plausible values, and classification of students into performance levels were executed via a computer program in R developed for the purpose of NAFS data analysis.

Keywords: large-scale assessment, reliability, generalizability theory, plausible values

Procedia PDF Downloads 17
197 Correlation between the Larvae Density (Diptera: Culicidae) and Physicochemical Characteristics of Habitats in Mazandaran Province, Northern Iran

Authors: Seyed Hassan Nikookar, Mahmoud Fazeli-Dinan, Seyyed Payman Ziapour, Ahmad-Ali Enayati

Abstract:

Background: Mosquitoes look for all kinds of aquatic habitats for laying eggs. Characteristics of water habitats are important factors in determining whether a mosquito can survive and successfully completed their developmental stages. Physicochemical factors can display an important role in vector control programs. This investigate determined whether physicochemical factors differ between habitats can be effective in the larvae density in Mazandaran province. Methods: Larvae were collected by the standard dipper up to 350 ml for 15-20 minutes from fixed habitats in 16 villages of 30 townships, the specimens identified by morphological key. Water samples were collected during larval collection and were evaluated for temperature (°C), acidity (pH), turbidity (NTU), electrical conductivity (μS/cm), alkalinity (mg/l), total hardness (mg/l), nitrate (mg/l), chloride (mg/l), phosphate (mg/l), sulfate (mg/l) in selected habitats using standard methods. Spearman Correlation coefficient was used for analyze data. Results: Totally 7566 mosquito larvae of three genera and 15 species were collected of fixed habitats. Cx. pipiens was the dominant species except in villages of Tileno, Zavat, Asad Abad, Shah Mansur Mahale which An. maculipennis, Cx. torrentium were as the predominant species. Turbidity in Karat Koti, Chloride in Al Tappeh, nitrate, phosphate and sulfate in Chalmardi, electrical conductivity, alkalinity, total hardness in Komishan villages were significantly higher than other villages (P < 0.05). There were a significant positive correlation between Cx. pipiens and Electrical conductivity, Alkalinity, Total hardness, Chloride, Cx. tritaeniorhynchus and Chloride, whereas a significant negative correlation observed between Sulfate and Cx. perexiguss. Conclusion: The correlations observed between physicochemical factor and larval density, possibly can confirm the effect of these parameters on the breeding activities of mosquitoes, and could probability facilitate larval control programs by the handwork of such factors.

Keywords: anopheles, culex, culiseta, physicochemical, habitats, larvae density, correlation

Procedia PDF Downloads 263
196 Chikungunya Virus Infection among Patients with Febrile Illness Attending University of Maiduguri Teaching Hospital, Nigeria

Authors: Abdul-Dahiru El-Yuguda, Saka Saheed Baba, Tawa Monilade Adisa, Mustapha Bala Abubakar

Abstract:

Background: Chikungunya (CHIK) virus, a previously anecdotally described arbovirus, is now assuming a worldwide public health burden. The CHIK virus infection is characterized by potentially life threatening and debilitating arthritis in addition to the high fever, arthralgia, myalgia, headache and rash. Method: Three hundred and seventy (370) serum samples were collected from outpatients with febrile illness attending University of Maiduguri Teaching Hospital, Nigeria, and was used to detect for Chikungunya (CHIK) virus IgG and IgM antibodies using the Enzyme Linked Immunosorbent Assays (ELISAs). Result: Out of the 370 sera tested, 39 (10.5%) were positive for presence of CHIK virus antibodies. A total of 24 (6.5%) tested positive for CHIK virus IgM only while none (0.0%) was positive for presence of CHIK virus IgG only and 15 (4.1%) of the serum samples were positive for both IgG and IgM antibodies. A significant difference (p<0.0001) was observed in the distribution of CHIK virus antibodies in relation to gender. The males had prevalence of 8.5% IgM antibodies as against 4.6% observed in females. On the other hand 4.6% of the females were positive for concurrent CHIK virus IgG and IgM antibodies when compared to a prevalence of 3.4% observed in males. Only the age groups ≤ 60 years and the undisclosed age group were positive for presence of CHIK virus IgG and/or IgM antibodies. No significant difference (p>0.05) was observed in the seasonal prevalence of CHIK virus antibodies among the study subjects Analysis of the prevalence of CHIK virus antibodies in relation to clinical presentation (as observed by Clinicians) of the patients revealed that headache and fever were the most frequently encountered ailments. Conclusion: The CHIK virus IgM and concurrent IgM and IgG antibody prevalence rates of 6.5% and 4.1% observed in this study indicates a current infection and the lack of IgG antibody alone observed shows that the infection is not endemic but sporadic. Recommendation: Further studies should be carried to establish the seasonal prevalence of CHIK virus infection vis-à-vis vector dynamics in the study area. A comprehensive study need to be carried out on the molecular characterization of the CHIK virus circulating in Nigeria with a view to developing CHIK virus vaccine.

Keywords: Chikungunya virus, IgM and IgG antibodies, febrile patients, enzyme linked immunosorbent assay

Procedia PDF Downloads 388
195 Prevalence of Fast-Food Consumption on Overweight or Obesity on Employees (Age Between 25-45 Years) in Private Sector; A Cross-Sectional Study in Colombo, Sri Lanka

Authors: Arosha Rashmi De Silva, Ananda Chandrasekara

Abstract:

This study seeks to comprehensively examine the influence of fast-food consumption and physical activity levels on the body weight of young employees within the private sector of Sri Lanka. The escalating popularity of fast food has raised concerns about its nutritional content and associated health ramifications. To investigate this phenomenon, a cohort of 100 individuals aged between 25 and 45, employed in Sri Lanka's private sector, participated in this research. These participants provided socio-demographic data through a standardized questionnaire, enabling the characterization of their backgrounds. Additionally, participants disclosed their frequency of fast-food consumption and engagement in physical activities, utilizing validated assessment tools. The collected data was meticulously compiled into an Excel spreadsheet and subjected to rigorous statistical analysis. Descriptive statistics, such as percentages and proportions, were employed to delineate the body weight status of the participants. Employing chi-square tests, our study identified significant associations between fast-food consumption, levels of physical activity, and body weight categories. Furthermore, through binary logistic regression analysis, potential risk factors contributing to overweight and obesity within the young employee cohort were elucidated. Our findings revealed a disconcerting trend, with 6% of participants classified as underweight, 32% within the normal weight range, and a substantial 62% categorized as overweight or obese. These outcomes underscore the alarming prevalence of overweight and obesity among young private-sector employees, particularly within the bustling urban landscape of Colombo, Sri Lanka. The data strongly imply a robust correlation between fast-food consumption, sedentary behaviors, and higher body weight categories, reflective of the evolving lifestyle patterns associated with the nation's economic growth. This study emphasizes the urgent need for effective interventions to counter the detrimental effects of fast-food consumption. The implementation of awareness campaigns elucidating the adverse health consequences of fast food, coupled with comprehensive nutritional education, can empower individuals to make informed dietary choices. Workplace interventions, including the provision of healthier meal alternatives and the facilitation of physical activity opportunities, are essential in fostering a healthier workforce and mitigating the escalating burden of overweight and obesity in Sri Lanka

Keywords: fast food consumption, obese, overweight, physical activity level

Procedia PDF Downloads 49
194 3D Modeling for Frequency and Time-Domain Airborne EM Systems with Topography

Authors: C. Yin, B. Zhang, Y. Liu, J. Cai

Abstract:

Airborne EM (AEM) is an effective geophysical exploration tool, especially suitable for ridged mountain areas. In these areas, topography will have serious effects on AEM system responses. However, until now little study has been reported on topographic effect on airborne EM systems. In this paper, an edge-based unstructured finite-element (FE) method is developed for 3D topographic modeling for both frequency and time-domain airborne EM systems. Starting from the frequency-domain Maxwell equations, a vector Helmholtz equation is derived to obtain a stable and accurate solution. Considering that the AEM transmitter and receiver are both located in the air, the scattered field method is used in our modeling. The Galerkin method is applied to discretize the Helmholtz equation for the final FE equations. Solving the FE equations, the frequency-domain AEM responses are obtained. To accelerate the calculation speed, the response of source in free-space is used as the primary field and the PARDISO direct solver is used to deal with the problem with multiple transmitting sources. After calculating the frequency-domain AEM responses, a Hankel’s transform is applied to obtain the time-domain AEM responses. To check the accuracy of present algorithm and to analyze the characteristic of topographic effect on airborne EM systems, both the frequency- and time-domain AEM responses for 3 model groups are simulated: 1) a flat half-space model that has a semi-analytical solution of EM response; 2) a valley or hill earth model; 3) a valley or hill earth with an abnormal body embedded. Numerical experiments show that close to the node points of the topography, AEM responses demonstrate sharp changes. Special attentions need to be paid to the topographic effects when interpreting AEM survey data over rugged topographic areas. Besides, the profile of the AEM responses presents a mirror relation with the topographic earth surface. In comparison to the topographic effect that mainly occurs at the high-frequency end and early time channels, the EM responses of underground conductors mainly occur at low frequencies and later time channels. For the signal of the same time channel, the dB/dt field reflects the change of conductivity better than the B-field. The research of this paper will serve airborne EM in the identification and correction of the topographic effects.

Keywords: 3D, Airborne EM, forward modeling, topographic effect

Procedia PDF Downloads 316
193 Using the Smith-Waterman Algorithm to Extract Features in the Classification of Obesity Status

Authors: Rosa Figueroa, Christopher Flores

Abstract:

Text categorization is the problem of assigning a new document to a set of predetermined categories, on the basis of a training set of free-text data that contains documents whose category membership is known. To train a classification model, it is necessary to extract characteristics in the form of tokens that facilitate the learning and classification process. In text categorization, the feature extraction process involves the use of word sequences also known as N-grams. In general, it is expected that documents belonging to the same category share similar features. The Smith-Waterman (SW) algorithm is a dynamic programming algorithm that performs a local sequence alignment in order to determine similar regions between two strings or protein sequences. This work explores the use of SW algorithm as an alternative to feature extraction in text categorization. The dataset used for this purpose, contains 2,610 annotated documents with the classes Obese/Non-Obese. This dataset was represented in a matrix form using the Bag of Word approach. The score selected to represent the occurrence of the tokens in each document was the term frequency-inverse document frequency (TF-IDF). In order to extract features for classification, four experiments were conducted: the first experiment used SW to extract features, the second one used unigrams (single word), the third one used bigrams (two word sequence) and the last experiment used a combination of unigrams and bigrams to extract features for classification. To test the effectiveness of the extracted feature set for the four experiments, a Support Vector Machine (SVM) classifier was tuned using 20% of the dataset. The remaining 80% of the dataset together with 5-Fold Cross Validation were used to evaluate and compare the performance of the four experiments of feature extraction. Results from the tuning process suggest that SW performs better than the N-gram based feature extraction. These results were confirmed by using the remaining 80% of the dataset, where SW performed the best (accuracy = 97.10%, weighted average F-measure = 97.07%). The second best was obtained by the combination of unigrams-bigrams (accuracy = 96.04, weighted average F-measure = 95.97) closely followed by the bigrams (accuracy = 94.56%, weighted average F-measure = 94.46%) and finally unigrams (accuracy = 92.96%, weighted average F-measure = 92.90%).

Keywords: comorbidities, machine learning, obesity, Smith-Waterman algorithm

Procedia PDF Downloads 296
192 Macroeconomic Effects and Dynamics of Natural Disaster Damages: Evidence from SETX on the Resiliency Hypothesis

Authors: Agim Kukelii, Gevorg Sargsyan

Abstract:

This study, focusing on the base regional area (county level), estimates the effect of natural disaster damages on aggregate personal income, aggregate wages, wages per worker, aggregate employment, and aggregate income transfer. The study further estimates the dynamics of personal income, employment, and wages under natural disaster shocks. Southeast Texas, located at the center of Golf Coast, is hit by meteorological and hydrological caused natural disasters yearly. On average, there are more than four natural disasters per year that cane an estimated damage average of 2.2% of real personal income. The study uses the panel data method to estimate the average effect of natural disasters on the area’s economy (personal income, wages, employment, and income transfer). It also uses Panel Vector Autoregressive (PVAR) model to study the dynamics of macroeconomic variables under natural disaster shocks. The study finds that the average effect of natural disasters is positive for personal income and income transfer and is negative for wages and employment. The PVAR and the impulse response function estimates reveal that natural disaster shocks cause a decrease in personal income, employment, and wages. However, the economy’s variables bounce back after three years. The novelty of this study rests on several aspects. First, this is the first study to investigate the effects of natural disasters on macroeconomic variables at a regional level. Second, the study uses direct measures of natural disaster damages. Third, the study estimates that the time that the local economy takes to absorb the natural disaster damages shocks is three years. This is a relatively good reaction to the local economy, therefore, adding to the “resiliency” hypothesis. The study has several implications for policymakers, businesses, and households. First, this study serves to increase the awareness of local stakeholders that natural disaster damages do worsen, macroeconomic variables, such as personal income, employment, and wages beyond the immediate damages to residential and commercial properties, physical infrastructure, and discomfort in daily lives. Second, the study estimates that these effects linger on the economy on average for three years, which would require policymakers to factor in the time area need to be on focus.

Keywords: natural disaster damages, macroeconomics effects, PVAR, panel data

Procedia PDF Downloads 86
191 Determinants of Never Users of Contraception-Results from Pakistan Demographic and Health Survey 2012-13

Authors: Arsalan Jabbar, Wajiha Javed, Nelofer Mehboob, Zahid Memon

Abstract:

Introduction: There are multiple social, individual and cultural factors that influence an individual’s decision to adopt family planning methods especially among non-users in patriarchal societies like Pakistan.Non-users, if targeted efficiently, can contribute significantly to country’s CPR. A research study showed that non-users if convinced to adopt lactational amenorrhea method can shift to long-term methods in future. Research shows that if non-users are targeted efficiently a 59% reduction in unintended pregnancies in Saharan Africa and South-Central and South-East Asia is anticipated. Methods: We did secondary data analysis on Pakistan Demographic Heath Survey (2012-13) dataset. Use of contraception (never-use/ever-use) was the outcome variable. At univariate level Chi-square/Fisher Exact test was used to assess relationship of baseline covariates with contraception use. Then variables to be incorporated in the model were checked for multi-collinearity, confounding, and interaction. Then binary logistic regression (with an urban-rural stratification) was done to find the relationship between contraception use and baseline demographic and social variables. Results: The multivariate analyses of the study showed that younger women (≤ 29 years) were more prone to be never users as compared to those who were > 30 years and this trend was seen in urban areas (AOR 1.92, CI 1.453-2.536) as well as rural areas (AOR 1.809, CI 1.421-2.303). While looking at regional variation, women from urban Sindh (AOR 1.548, CI 1.142-2.099) and urban Balochistan (AOR 2.403, CI 1.504-3.839) had more never users as compared to other urban regions. Women in the rich wealth quintile were more never users and this was seen both in urban and rural localities (urban (AOR 1.106 CI .753-1.624); rural areas (AOR 1.162, CI .887-1.524)) even though these were not statistically significant. Women idealizing more children(> 4) are more never users as compared to those idealizing less children in both urban (AOR 1.854, CI 1.275-2.697) and rural areas (AOR 2.101, CI 1.514-2.916). Women who never lost a pregnancy were more inclined to be non-users in rural areas (AOR 1.394, CI 1.127-1.723) .Women familiar with only traditional or no method had more never users in rural areas (AOR 1.717, CI 1.127-1.723) but in urban areas it wasn’t significant. Women unaware of Lady Health Worker’s presence in their area were more never users especially in rural areas (AOR 1.276, CI 1.014-1.607). Women who did not visit any care provider were more never users (urban (AOR 11.738, CI 9.112-15.121) rural areas (AOR 7.832, CI 6.243-9.826)). Discussion/Conclusion: This study concluded that government, policy makers and private sector family planning programs should focus on the untapped pool of never users (younger women from underserved provinces, in higher wealth quintiles, who desire more children.). We need to make sure to cover catchment areas where there are less LHWs and less providers as ignorance to modern methods and never been visited by an LHW are important determinants of never use. This all is in sync with previous literate from similar developing countries.

Keywords: contraception, demographic and health survey, family planning, never users

Procedia PDF Downloads 406
190 Improve Divers Tracking and Classification in Sonar Images Using Robust Diver Wake Detection Algorithm

Authors: Mohammad Tarek Al Muallim, Ozhan Duzenli, Ceyhun Ilguy

Abstract:

Harbor protection systems are so important. The need for automatic protection systems has increased over the last years. Diver detection active sonar has great significance. It used to detect underwater threats such as divers and autonomous underwater vehicle. To automatically detect such threats the sonar image is processed by algorithms. These algorithms used to detect, track and classify of underwater objects. In this work, divers tracking and classification algorithm is improved be proposing a robust wake detection method. To detect objects the sonar images is normalized then segmented based on fixed threshold. Next, the centroids of the segments are found and clustered based on distance metric. Then to track the objects linear Kalman filter is applied. To reduce effect of noise and creation of false tracks, the Kalman tracker is fine tuned. The tuning is done based on our active sonar specifications. After the tracks are initialed and updated they are subjected to a filtering stage to eliminate the noisy and unstable tracks. Also to eliminate object with a speed out of the diver speed range such as buoys and fast boats. Afterwards the result tracks are subjected to a classification stage to deiced the type of the object been tracked. Here the classification stage is to deice wither if the tracked object is an open circuit diver or a close circuit diver. At the classification stage, a small area around the object is extracted and a novel wake detection method is applied. The morphological features of the object with his wake is extracted. We used support vector machine to find the best classifier. The sonar training images and the test images are collected by ARMELSAN Defense Technologies Company using the portable diver detection sonar ARAS-2023. After applying the algorithm to the test sonar data, we get fine and stable tracks of the divers. The total classification accuracy achieved with the diver type is 97%.

Keywords: harbor protection, diver detection, active sonar, wake detection, diver classification

Procedia PDF Downloads 237
189 Effects of Magnetic Field on 4H-SiC P-N Junctions

Authors: Khimmatali Nomozovich Juraev

Abstract:

Silicon carbide is one of the promising materials with potential applications in electronic devices using high power, high frequency and high electric field. Currently, silicon carbide is used to manufacture high power and frequency diodes, transistors, radiation detectors, light emitting diodes (LEDs) and other functional devices. In this work, the effects of magnetic field on p-n junctions based on 4H-SiC were experimentally studied. As a research material, monocrystalline silicon carbide wafers (Cree Research, Inc., USA) with relatively few growth defects grown by physical vapor transport (PVT) method were used: Nd dislocations 104 cm², Nm micropipes ~ 10–10² cm-², thickness ~ 300-600 μm, surface ~ 0.25 cm², resistivity ~ 3.6–20 Ωcm, the concentration of background impurities Nd − Na ~ (0.5–1.0)×1017cm-³. The initial parameters of the samples were determined on a Hall Effect Measurement System HMS-7000 (Ecopia) measuring device. Diffusing Ni nickel atoms were covered to the silicon surface of silicon carbide in a Universal Vacuum Post device at a vacuum of 10-⁵ -10-⁶ Torr by thermal sputtering and kept at a temperature of 600-650°C for 30 minutes. Then Ni atoms were diffused into the silicon carbide 4H-SiC sample at a temperature of 1150-1300°C by low temperature diffusion method in an air atmosphere, and the effects of the magnetic field on the I-V characteristics of the samples were studied. I-V characteristics of silicon carbide 4H-SiC p-n junction sample were measured in the magnetic field and in the absence of a magnetic field. The measurements were carried out under conditions where the magnitude of the magnetic field induction vector was 0.5 T. In the state, the direction of the current flowing through the diode is perpendicular to the direction of the magnetic field. From the obtained results, it can be seen that the magnetic field significantly affects the I-V characteristics of the p-n junction in the magnetic field when it is measured in the forward direction. Under the influence of the magnetic field, the change of the magnetic resistance of the sample of silicon carbide 4H-SiC p-n junction was determined. It was found that changing the magnetic field poles increases the direct forward current of the p-n junction or decreases it when the field direction changes. These unique electrical properties of the 4H-SiC p-n junction sample of silicon carbide, that is, the change of the sample's electrical properties in a magnetic field, makes it possible to fabricate magnetic field sensing devices based on silicon carbide to use at harsh environments in future. So far, the productions of silicon carbide magnetic detectors are not available in the industry.

Keywords: 4H-SiC, diffusion Ni, effects of magnetic field, I-V characteristics

Procedia PDF Downloads 95
188 The Effect of Technology on Skin Development and Progress

Authors: Haidy Weliam Megaly Gouda

Abstract:

Dermatology is often a neglected specialty in low-resource settings despite the high morbidity associated with skin disease. This becomes even more significant when associated with HIV infection, as dermatological conditions are more common and aggressive in HIV-positive patients. African countries have the highest HIV infection rates, and skin conditions are frequently misdiagnosed and mismanaged because of a lack of dermatological training and educational material. The frequent lack of diagnostic tests in the African setting renders basic clinical skills all the more vital. This project aimed to improve the diagnosis and treatment of skin disease in the HIV population in a district hospital in Malawi. A basic dermatological clinical tool was developed and produced in collaboration with local staff and based on available literature and data collected from clinics. The aim was to improve diagnostic accuracy and provide guidance for the treatment of skin disease in HIV-positive patients. A literature search within Embassy, Medline and Google Scholar was performed and supplemented through data obtained from attending 5 Antiretroviral clinics. From the literature, conditions were selected for inclusion in the resource if they were described as specific, more prevalent, or extensive in the HIV population or have more adverse outcomes if they develop in HIV patients. Resource-appropriate treatment options were decided using Malawian Ministry of Health guidelines and textbooks specific to African dermatology. After the collection of data and discussion with local clinical and pharmacy staff, a list of 15 skin conditions was included, and a booklet was created using the simple layout of a picture, a diagnostic description of the disease and treatment options. Clinical photographs were collected from local clinics (with full consent of the patient) or from the book ‘Common Skin Diseases in Africa’ (permission granted if fully acknowledged and used in a not-for-profit capacity). This tool was evaluated by the local staff alongside an educational teaching session on skin disease. This project aimed to reduce uncertainty in diagnosis and provide guidance for appropriate treatment in HIV patients by gathering information into one practical and manageable resource. To further this project, we hope to review the effectiveness of the tool in practice.

Keywords: prevalence and pattern of skin diseases, impact on quality of life, rural Nepal, interventions, quality switched ruby laser, skin color river blindness, clinical signs, circularity index, grey level run length matrix, grey level co-occurrence matrix, local binary pattern, object detection, ring detection, shape identification

Procedia PDF Downloads 60
187 Automated Computer-Vision Analysis Pipeline of Calcium Imaging Neuronal Network Activity Data

Authors: David Oluigbo, Erik Hemberg, Nathan Shwatal, Wenqi Ding, Yin Yuan, Susanna Mierau

Abstract:

Introduction: Calcium imaging is an established technique in neuroscience research for detecting activity in neural networks. Bursts of action potentials in neurons lead to transient increases in intracellular calcium visualized with fluorescent indicators. Manual identification of cell bodies and their contours by experts typically takes 10-20 minutes per calcium imaging recording. Our aim, therefore, was to design an automated pipeline to facilitate and optimize calcium imaging data analysis. Our pipeline aims to accelerate cell body and contour identification and production of graphical representations reflecting changes in neuronal calcium-based fluorescence. Methods: We created a Python-based pipeline that uses OpenCV (a computer vision Python package) to accurately (1) detect neuron contours, (2) extract the mean fluorescence within the contour, and (3) identify transient changes in the fluorescence due to neuronal activity. The pipeline consisted of 3 Python scripts that could both be easily accessed through a Python Jupyter notebook. In total, we tested this pipeline on ten separate calcium imaging datasets from murine dissociate cortical cultures. We next compared our automated pipeline outputs with the outputs of manually labeled data for neuronal cell location and corresponding fluorescent times series generated by an expert neuroscientist. Results: Our results show that our automated pipeline efficiently pinpoints neuronal cell body location and neuronal contours and provides a graphical representation of neural network metrics accurately reflecting changes in neuronal calcium-based fluorescence. The pipeline detected the shape, area, and location of most neuronal cell body contours by using binary thresholding and grayscale image conversion to allow computer vision to better distinguish between cells and non-cells. Its results were also comparable to manually analyzed results but with significantly reduced result acquisition times of 2-5 minutes per recording versus 10-20 minutes per recording. Based on these findings, our next step is to precisely measure the specificity and sensitivity of the automated pipeline’s cell body and contour detection to extract more robust neural network metrics and dynamics. Conclusion: Our Python-based pipeline performed automated computer vision-based analysis of calcium image recordings from neuronal cell bodies in neuronal cell cultures. Our new goal is to improve cell body and contour detection to produce more robust, accurate neural network metrics and dynamic graphs.

Keywords: calcium imaging, computer vision, neural activity, neural networks

Procedia PDF Downloads 82
186 Impact of Civil Engineering and Economic Growth in the Sustainability of the Environment: Case of Albania

Authors: Rigers Dodaj

Abstract:

Nowadays, the environment is a critical goal for civil engineers, human activity, construction projects, economic growth, and whole national development. Regarding the development of Albania's economy, people's living standards are increasing, and the requirements for the living environment are also increasing. Under these circumstances, environmental protection and sustainability this is the critical issue. The rising industrialization, urbanization, and energy demand affect the environment by emission of carbon dioxide gas (CO2), a significant parameter known to impact air pollution directly. Consequently, many governments and international organizations conducted policies and regulations to address environmental degradation in the pursuit of economic development, for instance in Albania, the CO2 emission calculated in metric tons per capita has increased by 23% in the last 20 years. This paper analyzes the importance of civil engineering and economic growth in the sustainability of the environment focusing on CO2 emission. The analyzed data are time series 2001 - 2020 (with annual frequency), based on official publications of the World Bank. The statistical approach with vector error correction model and time series forecasting model are used to perform the parameter’s estimations and long-run equilibrium. The research in this paper adds a new perspective to the evaluation of a sustainable environment in the context of carbon emission reduction. Also, it provides reference and technical support for the government toward green and sustainable environmental policies. In the context of low-carbon development, effectively improving carbon emission efficiency is an inevitable requirement for achieving sustainable economic and environmental protection. Also, the study reveals that civil engineering development projects impact greatly the environment in the long run, especially in areas of flooding, noise pollution, water pollution, erosion, ecological disorder, natural hazards, etc. The potential for reducing industrial carbon emissions in recent years indicates that reduction is becoming more difficult, it needs another economic growth policy and more civil engineering development, by improving the level of industrialization and promoting technological innovation in industrial low-carbonization.

Keywords: CO₂ emission, civil engineering, economic growth, environmental sustainability

Procedia PDF Downloads 83
185 Finding the Association Rule between Nursing Interventions and Early Evaluation Results of In-Hospital Cardiac Arrest to Improve Patient Safety

Authors: Wei-Chih Huang, Pei-Lung Chung, Ching-Heng Lin, Hsuan-Chia Yang, Der-Ming Liou

Abstract:

Background: In-Hospital Cardiac Arrest (IHCA) threaten life of the inpatients, cause serious effect to patient safety, quality of inpatients care and hospital service. Health providers must identify the signs of IHCA early to avoid the occurrence of IHCA. This study will consider the potential association between early signs of IHCA and the essence of patient care provided by nurses and other professionals before an IHCA occurs. The aim of this study is to identify significant associations between nursing interventions and abnormal early evaluation results of IHCA that can assist health care providers in monitoring inpatients at risk of IHCA to increase opportunities of IHCA early detection and prevention. Materials and Methods: This study used one of the data mining techniques called association rules mining to compute associations between nursing interventions and abnormal early evaluation results of IHCA. The nursing interventions and abnormal early evaluation results of IHCA were considered to be co-occurring if nursing interventions were provided within 24 hours of last being observed in abnormal early evaluation results of IHCA. The rule based methods were utilized 23.6 million electronic medical records (EMR) from a medical center in Taipei, Taiwan. This dataset includes 733 concepts of nursing interventions that coded by clinical care classification (CCC) codes and 13 early evaluation results of IHCA with binary codes. The values of interestingness and lift were computed as Q values to measure the co-occurrence and associations’ strength between all in-hospital patient care measures and abnormal early evaluation results of IHCA. The associations were evaluated by comparing the results of Q values and verified by medical experts. Results and Conclusions: The results show that there are 4195 pairs of associations between nursing interventions and abnormal early evaluation results of IHCA with their Q values. The indication of positive association is 203 pairs with Q values greater than 5. Inpatients with high blood sugar level (hyperglycemia) have positive association with having heart rate lower than 50 beats per minute or higher than 120 beats per minute, Q value is 6.636. Inpatients with temporary pacemaker (TPM) have significant association with high risk of IHCA, Q value is 47.403. There is significant positive correlation between inpatients with hypovolemia and happened abnormal heart rhythms (arrhythmias), Q value is 127.49. The results of this study can help to prevent IHCA from occurring by making health care providers early recognition of inpatients at risk of IHCA, assist with monitoring patients for providing quality of care to patients, improve IHCA surveillance and quality of in-hospital care.

Keywords: in-hospital cardiac arrest, patient safety, nursing intervention, association rule mining

Procedia PDF Downloads 270
184 Redeeming the Self-Settling Scores with the Nazis by the Means of Poetics

Authors: Liliane Steiner

Abstract:

Beyond the testimonial act, that sheds light on the feminine experience in the Holocaust, the survivors' writing voices first and foremost the abjection of the feminine self brutally inflicted by the Nazis in the Holocaust, and in the same movement redeems the self by the means of poetics, and brings it to an existential state of being a subject. This study aims to stress the poetics of this writing in order to promote the Holocaust literature from the margins to the mainstream and to contribute to the commemoration of the Holocaust in the next generations. Methodology: The study of the survivors' redeeming of self is based on Julia Kristeva's theory of the abject: the self-throws out everything that threatens its existence and Liliane Steiner's theory of the post- abjection of hell: the belated act of vomiting the abject experiences settles cores with the author of the abject to redeem the self. The research will focus on Ruth Sender's trilogy The Cage, To Life and The Holocaust Lady as a case study. Findings: The binary mode that characterizes this writing reflects the experience of Jewish women, who were subject(s), were treated violently as object(s), debased, defeminized and, eventually turned into abject by the Nazis. In a tour de force, this writing re-enacts the postponed resistance, that vomited the abject imposed on the feminine self by the very act of narration, which denounces the real abject, the perpetrators. The post-abjection of self is acted out in constructs of abject, relating the abject experience of the Holocaust as well as the rehabilitation of the surviving self (subject). The transcription of abject surfaces in deconstructing the abject through self- characterization, and in the elusive rendering of bad memories, having recourse to literary figures. The narrative 'I' selects, obstructs, mends and tells the past events from an active standpoint, as would a subject in control of its (narrative) fate. In a compensatory movement, the narrating I tells itself by reconstructing the subject and proving time and again that I is other. Moreover, in the belated endeavor to revenge, testify and narrate the abject, the narrative I defies itself, and represents itself as a dialectical I, splitting and multiplying itself in a deconstructing way. The dialectical I is never (one) I. It voices not only the unvoiced but also and mainly the other silenced 'I's. Drawing its nature and construct from traumatic memories, the dialectical I transgresses boundaries to narrate her story, and in the same breath, the story of Jewish women doomed to silence. In this narrative feat, the dialectical I stresses its essential dialectical existence with the past, never to be (one) again. Conclusion: The pattern of I is other generates patterns of subject(s) that defy, transgress and repudiate the abject and its repercussions on the feminine I. The feminine I writes itself as a survivor that defies the abject (Nazis) and takes revenge. The paradigm of metamorphosis that accompanies the journey of the Holocaust memoirist engenders life and surviving as well as a narration that defies stagnation and death.

Keywords: abject, feminine writing, holocaust, post-abjection

Procedia PDF Downloads 102
183 Automatic Target Recognition in SAR Images Based on Sparse Representation Technique

Authors: Ahmet Karagoz, Irfan Karagoz

Abstract:

Synthetic Aperture Radar (SAR) is a radar mechanism that can be integrated into manned and unmanned aerial vehicles to create high-resolution images in all weather conditions, regardless of day and night. In this study, SAR images of military vehicles with different azimuth and descent angles are pre-processed at the first stage. The main purpose here is to reduce the high speckle noise found in SAR images. For this, the Wiener adaptive filter, the mean filter, and the median filters are used to reduce the amount of speckle noise in the images without causing loss of data. During the image segmentation phase, pixel values are ordered so that the target vehicle region is separated from other regions containing unnecessary information. The target image is parsed with the brightest 20% pixel value of 255 and the other pixel values of 0. In addition, by using appropriate parameters of statistical region merging algorithm, segmentation comparison is performed. In the step of feature extraction, the feature vectors belonging to the vehicles are obtained by using Gabor filters with different orientation, frequency and angle values. A number of Gabor filters are created by changing the orientation, frequency and angle parameters of the Gabor filters to extract important features of the images that form the distinctive parts. Finally, images are classified by sparse representation method. In the study, l₁ norm analysis of sparse representation is used. A joint database of the feature vectors generated by the target images of military vehicle types is obtained side by side and this database is transformed into the matrix form. In order to classify the vehicles in a similar way, the test images of each vehicle is converted to the vector form and l₁ norm analysis of the sparse representation method is applied through the existing database matrix form. As a result, correct recognition has been performed by matching the target images of military vehicles with the test images by means of the sparse representation method. 97% classification success of SAR images of different military vehicle types is obtained.

Keywords: automatic target recognition, sparse representation, image classification, SAR images

Procedia PDF Downloads 364
182 Meta-Magnetic Properties of LaFe₁₂B₆ Type Compounds

Authors: Baptiste Vallet-Simond, Léopold V. B. Diop, Olivier Isnard

Abstract:

The antiferromagnetic itinerant-electron compound LaFe₁₂B₆ occupies a special place among rare-earth iron-rich intermetallic; it presents exotic magnetic and physical properties. The unusual amplitude-modulated spin configuration defined by a propagation vector k = (¼, ¼, ¼), remarkably weak Fe magnetic moment (0.43 μB) in the antiferromagnetic ground state, especially low magnetic ordering temperature TN = 36 K for an Fe-rich phase, a multicritical point in the complex magnetic phase diagram, both normal and inverse magnetocaloric effects, and huge hydrostatic pressure effects can be highlighted as the most relevant. Both antiferromagnetic (AFM) and paramagnetic (PM) states can be transformed into the ferromagnetic (FM) state via a field-induced first-order metamagnetic transition. Of particular interest is the low-temperature magnetization process. This process is discontinuous and evolves unexpected huge metamagnetic transitions consisting of a succession of steep magnetization jumps separated by plateaus, giving rise to an unusual avalanche-like behavior. The metamagnetic transition is accompanied by giant magnetoresistance and large magnetostriction. In the present work, we report on the intrinsic magnetic properties of the La₁₋ₓPrₓFe₁₂B₆ series of compounds exhibiting sharp metamagnetic transitions. The study of the structural, magnetic, magneto-transport, and magnetostrictive properties of the La₁₋ₓPrₓFe₁₂B₆ system was performed by combining a wide variety of measurement techniques. Magnetic measurements were performed up to µ0H = 10 T. It was found that the proportion of Pr had a strong influence on the magnetic properties of this series of compounds. At x=0.05, the ground state at 2K is that of an antiferromagnet, but the critical transition field Hc has been lowered from Hc = 6T at x = 0 to Hc = 2.5 Tat x=0.05. And starting from x=0.10, the ground state of this series of compounds is a coexistence of AFM and FM parts. At x=0.30, the AFM order has completely vanished, and only the FM part is left. However, we still observe meta-magnetic transitions at higher temperatures (above 100 K for x=0.30) from the paramagnetic (P) state to a forced FM state. And, of course, such transitions are accompanied by strong magneto-caloric, magnetostrictive, and magnetoresistance effects. The Curie temperatures for the probed compositions going from x=0.05 to x=0.30 were spread over the temperature range of 40 K up to 100 K.

Keywords: metamagnetism, RMB intermetallic, magneto-transport effect, metamagnetic transitions

Procedia PDF Downloads 67
181 Breast Cancer Sensing and Imaging Utilized Printed Ultra Wide Band Spherical Sensor Array

Authors: Elyas Palantei, Dewiani, Farid Armin, Ardiansyah

Abstract:

High precision of printed microwave sensor utilized for sensing and monitoring the potential breast cancer existed in women breast tissue was optimally computed. The single element of UWB printed sensor that successfully modeled through several numerical optimizations was multiple fabricated and incorporated with woman bra to form the spherical sensors array. One sample of UWB microwave sensor obtained through the numerical computation and optimization was chosen to be fabricated. In overall, the spherical sensors array consists of twelve stair patch structures, and each element was individually measured to characterize its electrical properties, especially the return loss parameter. The comparison of S11 profiles of all UWB sensor elements is discussed. The constructed UWB sensor is well verified using HFSS programming, CST programming, and experimental measurement. Numerically, both HFSS and CST confirmed the potential operation bandwidth of UWB sensor is more or less 4.5 GHz. However, the measured bandwidth provided is about 1.2 GHz due to the technical difficulties existed during the manufacturing step. The configuration of UWB microwave sensing and monitoring system implemented consists of 12 element UWB printed sensors, vector network analyzer (VNA) to perform as the transceiver and signal processing part, the PC Desktop/Laptop acting as the image processing and displaying unit. In practice, all the reflected power collected from whole surface of artificial breast model are grouped into several numbers of pixel color classes positioned on the corresponding row and column (pixel number). The total number of power pixels applied in 2D-imaging process was specified to 100 pixels (or the power distribution pixels dimension 10x10). This was determined by considering the total area of breast phantom of average Asian women breast size and synchronizing with the single UWB sensor physical dimension. The interesting microwave imaging results were plotted and together with some technical problems arisen on developing the breast sensing and monitoring system are examined in the paper.

Keywords: UWB sensor, UWB microwave imaging, spherical array, breast cancer monitoring, 2D-medical imaging

Procedia PDF Downloads 191
180 Household Socioeconomic Factors Associated with Teenage Pregnancies in Kigali City, Rwanda

Authors: Dieudonne Uwizeye, Reuben Muhayiteto

Abstract:

Teenage pregnancy is a challenging problem for sustainable development due to restrictions it poses to socioeconomic opportunities for young mothers, their children and families. Being unable to take appropriate economic and social responsibilities, teen mothers get trapped into poverty and become economic burden to their family and country. Besides, teenage pregnancy is also a health problem because children born to very young mothers are vulnerable with greater risk of illnesses and deaths, and teenage mothers are more likely to be exposed to greater risk of maternal mortality and to other health and psychological problems. In Kigali city, in Rwanda, teenage pregnancy rate is currently high and its increase in recent years is worrisome. However, only individual factors influencing the teenage pregnancy tend to be the basis of interventions. It is important to understand the important socioeconomic factors at the household level that are associated with teenage pregnancy to help government, parents, and other stakeholders to appropriately address the problem with sustainable measures. This study analyzed secondary data from the Fifth Rwanda Demographic and Health Survey (RDHS-V 2014-2015) conducted by the National Institute of Statistics of Rwanda (NISR). The aim was to examine household socio-economic factors that are associated with incidence of teenage pregnancies in Kigali city. In addition to descriptive analysis, Pearson’s Chi Square and Binary Logistic Regression were used in the analysis. Findings indicate that marital status and age of household head, number of members in a household, number of rooms used for sleeping, educational level of the household head and household's wealth are significantly associated with teenage pregnancy in Rwanda ( p< 0.05). It was found that teenagers living with parents, those having parents with higher education and those from richer families are less likely to become pregnant. Age of household head was pinpointed as factor to teenage pregnancy, with teenage-headed households being more vulnerable. The findings also revealed that household composition correlates with the probability of teenage pregnancy (p < 0.05) with teenagers from households with less number of members being more vulnerable. Regarding the size of the house, the study suggested that the more rooms available in households, the less incidences of teenage pregnancy are likely to be observed (p < 0.05). However, teenage pregnancy was not significantly associated with physical violence among parents (p = 0.65) and sex of household heads (p = 0.52), except in teen-headed households of which female are predominantly heads. The study concludes that teenage pregnancy remains a serious social, economic and health problem in Rwanda. The study informs government officials, parents and other stakeholders to take interventions and preventive measures through community sex education, policies and strategies to foster effective parental guidance, care and control of young girls through meeting their necessary social and financial needs within households.

Keywords: household socio-economic factors, Rwanda, Rwanda demographic and health survey, teenage pregnancy

Procedia PDF Downloads 177