Search results for: yellow color detection
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 4490

Search results for: yellow color detection

800 A Case Study on Indian Translation Ecosystem of Point-Of-Care Solutions

Authors: Tripta Dixit, Smita Sahu, William Selvamurthy, Sadhana Srivastava

Abstract:

The translation of healthcare technologies is an expensive, complex affair, current healthcare challenges in Asian countries and their efforts to meet Millennium Development Goals (MDGs), necessitates continuous technology advancement to save countless lives, improve the quality of life and for socio-economic development. India’s consistently improving global innovation index (57) demonstrates its innovation potential, but access to health care is asymmetric and lacks priority in India. Therefore, there is utmost need of a robust translation system for point-of-care (POC) solutions, inexpensive, low-maintenance, reliable, and easy-to-use diagnostic technologies. Few cases of POC technologies viz. Elisa based diagnostic kits for regional viral disease, a device for detection of cancerous lesions were studied to understand the process and challenges involved in their translation. Accordingly, the entire translation ecosystem was summarized proposing a nexus of various actors such as technology developer, technology transferor technology receiver, funding entities, government/regulatory bodies and their effect on translation of different medical technologies. This study highlights the role and concerns pertaining to these actors for POC such as unsystematic and unvalidated research roadmap, low profit preposition, unfocused approach of up-scaling, low market acceptability and multiple window regulatory framework, etc. This provides an opportunity to devise solutions to overcome problem areas in translation path.

Keywords: healthcare technologies, point-of-care solutions, public health, translation

Procedia PDF Downloads 165
799 Prenatal Diagnosis of Beta Thalassemia Intermedia in Vietnamese Family: Case Report

Authors: Ha T. T. Ly, Truc B. Truc, Hai N. Truong, Mai P. T. Nguyen, Ngoc D. Ngo, Khanh V. Tran, Hai T. Le

Abstract:

Beta thalassemia is one of the most common inherited blood disorders, which is characterized by decreased or absent in beta globin expression. Patients with Beta thalassemia whose anemia is not so severe as to necessitate transfusions are said to have thalassemia intermedia. Objective: The goal of this study is prenatal diagnosis for pregnancy woman with Beta thalassemia intermedia and her husband with Beta thalassemia carrier at high risk of Beta thalassemia major in Northern of Vietnam. Material and method: The family has a 6 years-old compound heterozygous thalassemia major for CD71/72(+A) and Hbb:c. -78A>G/nt-28(A>G) male child. The father was heterozygous for CD71/72(+A) mutation which is Beta plus type and the mother was compound heterozygosity of two different variants, namely, Hbb: c. -78A>G/nt-28(A>G) and CD26(A-G) HbE. Prenatal Beta thalassemia mutation detection in fetal DNA was carried out using multiplex Amplification-refractory mutation system ARMS-PCR and confirmed by direct Sanger-sequencing Hbb gene. Prenatal diagnoses were perfomed by amniotic fluid sampling from pregnant woman in the 16-18th week of pregnancy after the genotypes of parents of the probands were identified. Result: When amniotic fluid sample was analyzed for Beta globin gene (Hbb), we found that the genotype is heterozygous for CD71/72(+A) and CD26(A-G) HbE. This genotype is different from the 1st child of this family. Conclusion: Prenatal diagnosis helps the parents to know the genotype and the thalassemia status of the fetus, so they can have early decision on their pregnancy. Genetic diagnosis provided a useful method in diagnosis for familial members in pedigree, genetic counseling and prenatal diagnosis.

Keywords: beta thalassemia intermedia, Hbb gene, pedigree, prenatal diagnosis

Procedia PDF Downloads 376
798 Use Multiphysics Simulations and Resistive Pulse Sensing to Study the Effect of Metal and Non-Metal Nanoparticles in Different Salt Concentration

Authors: Chun-Lin Chiang, Che-Yen Lee, Yu-Shan Yeh, Jiunn-Haur Shaw

Abstract:

Wafer fabrication is a critical part of the semiconductor process, when the finest linewidth with the improvement of technology continues to decline and the structure development from 2D towards to 3D. The nanoparticles contained in the slurry or in the ultrapure water which used for cleaning have a large influence on the manufacturing process. Therefore, semiconductor industry is hoping to find a viable method for on-line detection the nanoparticles size and concentration. The resistive pulse sensing technology is one of the methods that may cover this question. As we know that nanoparticles properties of material differ significantly from their properties at larger length scales. So, we want to clear that the metal and non-metal nanoparticles translocation dynamic when we use the resistive pulse sensing technology. In this study we try to use the finite element method that contains three governing equations to do multiphysics coupling simulations. The Navier-Stokes equation describes the laminar motion, the Nernst-Planck equation describes the ion transport, and the Poisson equation describes the potential distribution in the flow channel. To explore that the metal nanoparticles and the non-metal nanoparticles in different concentration electrolytes, through the nanochannel caused by ion current changes. Then the reliability of the simulation results was verified by resistive pulse sensing test. The existing results show that the lower ion concentration, the greater effect of nanoparticles on the ion concentration in the nanochannel. The conductive spikes are correlated with nanoparticles surface charge. Then we can be concluded that in the resistive pulse sensing technique, the ion concentration in the nanochannel and nanoparticle properties are important for the translocation dynamic, and they have the interactions.

Keywords: multiphysics simulations, resistive pulse sensing, nanoparticles, nanochannel

Procedia PDF Downloads 335
797 Local Interpretable Model-agnostic Explanations (LIME) Approach to Email Spam Detection

Authors: Rohini Hariharan, Yazhini R., Blessy Maria Mathew

Abstract:

The task of detecting email spam is a very important one in the era of digital technology that needs effective ways of curbing unwanted messages. This paper presents an approach aimed at making email spam categorization algorithms transparent, reliable and more trustworthy by incorporating Local Interpretable Model-agnostic Explanations (LIME). Our technique assists in providing interpretable explanations for specific classifications of emails to help users understand the decision-making process by the model. In this study, we developed a complete pipeline that incorporates LIME into the spam classification framework and allows creating simplified, interpretable models tailored to individual emails. LIME identifies influential terms, pointing out key elements that drive classification results, thus reducing opacity inherent in conventional machine learning models. Additionally, we suggest a visualization scheme for displaying keywords that will improve understanding of categorization decisions by users. We test our method on a diverse email dataset and compare its performance with various baseline models, such as Gaussian Naive Bayes, Multinomial Naive Bayes, Bernoulli Naive Bayes, Support Vector Classifier, K-Nearest Neighbors, Decision Tree, and Logistic Regression. Our testing results show that our model surpasses all other models, achieving an accuracy of 96.59% and a precision of 99.12%.

Keywords: text classification, LIME (local interpretable model-agnostic explanations), stemming, tokenization, logistic regression.

Procedia PDF Downloads 39
796 Chemical Characterization and Prebiotic Effect of Water-Soluble Polysaccharides from Zizyphus lotus Leaves

Authors: Zakaria Boual, Abdellah Kemassi, Toufik Chouana, Philippe Michaud, Mohammed Didi Ould El Hadj

Abstract:

In order to investigate the prebiotic potential of oligosaccharides prepared by chemical hydrolysis of water-soluble polysaccharides (WSP) from Zizyphus lotus leaves, the effect of oligosaccharides on bacterial growth was studied. The chemical composition of WSP was evaluated by colorimetric assays revealed the average values: 7.05±0.73% proteins and 86.21±0.74% carbohydrates, among them 64.81±0.42% are neutral sugar and the rest 16.25±1.62% are uronic acids. The characterization of monosaccharides was determined by high performance anion exchange chromatography with pulsed amperometric detection (HPAEC-PAD) was found to be composed of galactose (23.95%), glucose (21.30%), rhamnose (20.28%), arabinose (9.55%), and glucuronic acid (22.95%). The effects of oligosaccharides on the growth of lactic acid bacteria were compared with those of fructo-oligosaccharide (RP95). The oligosaccharides concentration was 1g/L of man rogosa sharpe broth. Bacterial growth was assessed during 2, 4.5, 6.5, 9, 12, 16 and 24 h by measuring the optical density of the cultures at 600 nm (OD600) and pH values. During fermentation, pH in broth cultures decreased from 6.7 to 5.87±0.15. The enumeration of lactic acid bacteria indicated that oligosaccharides led to a significant increase in bacteria (P≤0.05) compared to the control. The fermentative metabolism appeared to be faster on RP95 than on oligosaccharides from Zizyphus lotus leaves. Both RP95 and oligosaccharides showed clear prebiotic effects, but had differences in fermentation kinetics because of to the different degree of polymerization. This study shows the prebiotic effectiveness of oligosaccharides, and provides proof for the selection of leaves of Zizyphus lotus for use as functional food ingredients.

Keywords: Zizyphus lotus, polysaccharides, characterization, prebiotic effects

Procedia PDF Downloads 400
795 Numerical Modelling of Skin Tumor Diagnostics through Dynamic Thermography

Authors: Luiz Carlos Wrobel, Matjaz Hribersek, Jure Marn, Jurij Iljaz

Abstract:

Dynamic thermography has been clinically proven to be a valuable diagnostic technique for skin tumor detection as well as for other medical applications such as breast cancer diagnostics, diagnostics of vascular diseases, fever screening, dermatological and other applications. Thermography for medical screening can be done in two different ways, observing the temperature response under steady-state conditions (passive or static thermography), and by inducing thermal stresses by cooling or heating the observed tissue and measuring the thermal response during the recovery phase (active or dynamic thermography). The numerical modelling of heat transfer phenomena in biological tissue during dynamic thermography can aid the technique by improving process parameters or by estimating unknown tissue parameters based on measured data. This paper presents a nonlinear numerical model of multilayer skin tissue containing a skin tumor, together with the thermoregulation response of the tissue during the cooling-rewarming processes of dynamic thermography. The model is based on the Pennes bioheat equation and solved numerically by using a subdomain boundary element method which treats the problem as axisymmetric. The paper includes computational tests and numerical results for Clark II and Clark IV tumors, comparing the models using constant and temperature-dependent thermophysical properties, which showed noticeable differences and highlighted the importance of using a local thermoregulation model.

Keywords: boundary element method, dynamic thermography, static thermography, skin tumor diagnostic

Procedia PDF Downloads 96
794 Inhibition of the Activity of Polyphenol Oxidase Enzyme Present in Annona muricata and Musa acuminata by the Experimentally Identified Natural Anti-Browning Agents

Authors: Michelle Belinda S. Weerawardana, Gobika Thiripuranathar, Priyani A. Paranagama

Abstract:

Most of fresh vegetables and fruits available in the retail markets undergo a physiological disorder in its appearance and coloration, which indeed discourages consumer purchase. A loss of millions of dollars yearly to the food industry had been due to this pronounced color reaction called Enzymatic Browning which is driven due to the catalytic activity by an oxidoreductase enzyme, polyphenol oxidase (PPO). The enzyme oxidizes the phenolic compounds which are abundantly available in fruits and vegetables as substrates into quinones, which could react with proteins in its surrounding to generate black pigments, called melanins, which are highly UV-active compounds. Annona muricata (Katu anoda) and Musa acuminata (Ash plantains) is a fruit and a vegetable consumed by Sri Lankans widely due to their high nutritional values, medicinal properties and economical importance. The objective of the present study was to evaluate and determine the effective natural anti-browning inhibitors that could prevent PPO activity in the selected fruit and vegetable. Enzyme extracts from Annona muricata (Katu anoda) and Musa acuminata (Ash plantains), were prepared by homogenizing with analytical grade acetone, and pH of each enzyme extract was maintained at 7.0 using a phosphate buffer. The extracts of inhibitors were prepared using powdered ginger rhizomes and essential oil from the bark of Cinnamomum zeylanicum. Water extracts of ginger were prepared and the essential oil from Ceylon cinnamon bark was extracted using steam distillation method. Since the essential oil is not soluble in water, 0.1µl of cinnamon bark oil was mixed with 0.1µl of Triton X-100 emulsifier and 5.00 ml of water. The effect of each inhibitor on the PPO activity was investigated using catechol (0.1 mol dm-3) as the substrate and two samples of enzyme extracts prepared. The dosages of the prepared Cinnamon bark oil, and ginger (2 samples) which were used to measure the activity were 0.0035 g/ml, 0.091 g/ml and 0.087 g/ml respectively. The measurements of the inhibitory activity were obtained at a wavelength of 525 nm using the UV-visible spectrophotometer. The results evaluated thus revealed that % inhibition observed with cinnamon bark oil, and ginger for Annona muricata was 51.97%, and 60.90% respectively. The effects of cinnamon bark oil, and ginger extract on PPO activity of Musa acuminata were 49.51%, and 48.10%. The experimental findings thus revealed that Cinnamomum zeylanicum bark oil was a more effective inhibitor for PPO enzyme present in Musa acuminata and ginger was effective for PPO enzyme present in Annona muricata. Overall both the inhibitors were proven to be more effective towards the activities of PPO enzyme present in both samples. These inhibitors can thus be corroborated as effective, natural, non-toxic, anti-browning extracts, which when added to the above fruit and vegetable will increase the shelf life and also the acceptance of the product by the consumers.

Keywords: anti-browning agent, enzymatic browning, inhibitory activity, polyphenol oxidase

Procedia PDF Downloads 261
793 A Retrospective Study on the Age of Onset for Type 2 Diabetes Diagnosis

Authors: Mohamed A. Hammad, Dzul Azri Mohamed Noor, Syed Azhar Syed Sulaiman, Majed Ahmed Al-Mansoub, Muhammad Qamar

Abstract:

There is a progressive increase in the prevalence of early onset Type 2 diabetes mellitus. Early detection of Type 2 diabetes enhances the length and/or quality of life which might result from a reduction in the severity, frequency or prevent or delay of its long-term complications. The study aims to determine the onset age for the first diagnosis of Type 2 diabetes mellitus. A retrospective study conducted in the endocrine clinic at Hospital Pulau Pinang in Penang, Malaysia, January- December 2016. Records of 519 patients with Type 2 diabetes mellitus were screened to collect demographic data and determine the age of first-time diabetes mellitus diagnosis. Patients classified according to the age of diagnosis, gender, and ethnicity. The study included 519 patients with age (55.6±13.7) years, female 265 (51.1%) and male 254 (48.9%). The ethnicity distribution was Malay 191 (36.8%), Chinese 189 (36.4%) and Indian 139 (26.8%). The age of Type 2 diabetes diagnosis was (42±14.8) years. The female onset of diabetes mellitus was at age (41.5±13.7) years, while male (42.6±13.7) years. Distribution of diabetic onset by ethnicity was Malay at age (40.7±13.7) years, Chinese (43.2±13.7) years and Indian (42.3±13.7) years. Diabetic onset was classified by age as follow; ≤20 years’ cohort was 33 (6.4%) cases. Group >20- ≤40 years was 190 (36.6%) patients, and category >40- ≤60 years was 270 (52%) subjects. On the other hand, the group >60 years was 22 (4.2%) patients. The range of diagnosis was between 10 and 73 years old. Conclusion: Malay and female have an earlier onset of diabetes than Indian, Chinese and male. More than half of the patients had diabetes between 40 and 60 years old. Diabetes mellitus is becoming more common in younger age <40 years. The age at diagnosis of Type 2 diabetes mellitus has decreased with time.

Keywords: age of onset, diabetes diagnosis, diabetes mellitus, Malaysia, outpatients, type 2 diabetes, retrospective study

Procedia PDF Downloads 404
792 Offline High Voltage Diagnostic Test Findings on 15MVA Generator of Basochhu Hydropower Plant

Authors: Suprit Pradhan, Tshering Yangzom

Abstract:

Even with availability of the modern day online insulation diagnostic technologies like partial discharge monitoring, the measurements like Dissipation Factor (tanδ), DC High Voltage Insulation Currents, Polarization Index (PI) and Insulation Resistance Measurements are still widely used as a diagnostic tools to assess the condition of stator insulation in hydro power plants. To evaluate the condition of stator winding insulation in one of the generators that have been operated since 1999, diagnostic tests were performed on the stator bars of 15 MVA generators of Basochhu Hydropower Plant. This paper presents diagnostic study done on the data gathered from the measurements which were performed in 2015 and 2016 as part of regular maintenance as since its commissioning no proper aging data were maintained. Measurement results of Dissipation Factor, DC High Potential tests and Polarization Index are discussed with regard to their effectiveness in assessing the ageing condition of the stator insulation. After a brief review of the theoretical background, the strengths of each diagnostic method in detecting symptoms of insulation deterioration are identified. The interesting results observed from Basochhu Hydropower Plant is taken into consideration to conclude that Polarization Index and DC High Voltage Insulation current measurements are best suited for the detection of humidity and contamination problems and Dissipation Factor measurement is a robust indicator of long-term ageing caused by oxidative degradation.

Keywords: dissipation Factor (tanδ), polarization Index (PI), DC High Voltage Insulation Current, insulation resistance (IR), Tan Delta Tip-Up, dielectric absorption ratio

Procedia PDF Downloads 302
791 Monitoring and Evaluation of Web-Services Quality and Medium-Term Impact on E-Government Agencies' Efficiency

Authors: A. F. Huseynov, N. T. Mardanov, J. Y. Nakhchivanski

Abstract:

This practical research is aimed to improve the management quality and efficiency of public administration agencies providing e-services. The monitoring system developed will provide continuous review of the websites compliance with the selected indicators, their evaluation based on the selected indicators and ranking of services according to the quality criteria. The responsible departments in the government agencies were surveyed; the questionnaire includes issues of management and feedback, e-services provided, and the application of information systems. By analyzing the main affecting factors and barriers, the recommendations will be given that lead to the relevant decisions to strengthen the state agencies competencies for the management and the provision of their services. Component 1. E-services monitoring system. Three separate monitoring activities are proposed to be executed in parallel: Continuous tracing of e-government sites using built-in web-monitoring program; this program generates several quantitative values which are basically related to the technical characteristics and the performance of websites. The expert assessment of e-government sites in accordance with the two general criteria. Criterion 1. Technical quality of the site. Criterion 2. Usability/accessibility (load, see, use). Each high-level criterion is in turn subdivided into several sub-criteria, such as: the fonts and the color of the background (Is it readable?), W3C coding standards, availability of the Robots.txt and the site map, the search engine, the feedback/contact and the security mechanisms. The on-line survey of the users/citizens – a small group of questions embedded in the e-service websites. The questionnaires comprise of the information concerning navigation, users’ experience with the website (whether it was positive or negative), etc. Automated monitoring of web-sites by its own could not capture the whole evaluation process, and should therefore be seen as a complement to expert’s manual web evaluations. All of the separate results were integrated to provide the complete evaluation picture. Component 2. Assessment of the agencies/departments efficiency in providing e-government services. - the relevant indicators to evaluate the efficiency and the effectiveness of e-services were identified; - the survey was conducted in all the governmental organizations (ministries, committees and agencies) that provide electronic services for the citizens or the businesses; - the quantitative and qualitative measures are covering the following sections of activities: e-governance, e-services, the feedback from the users, the information systems at the agencies’ disposal. Main results: 1. The software program and the set of indicators for internet sites evaluation has been developed and the results of pilot monitoring have been presented. 2. The evaluation of the (internal) efficiency of the e-government agencies based on the survey results with the practical recommendations related to the human potential, the information systems used and e-services provided.

Keywords: e-government, web-sites monitoring, survey, internal efficiency

Procedia PDF Downloads 292
790 Valorization of Underutilized Fish Species Through a Multidisciplinary Approach

Authors: Tiziana Pepe, Gerardo Manfreda, Adriana Ianieri, Aniello Anastasio

Abstract:

The sustainable exploitation of marine biological resources is among the most important objectives of the EU's Common Fisheries Policy (CFP). Currently, Europe imports about 65% of its fish products, indicating that domestic production does not meet consumer demand. Despite the availability of numerous commercially significant fish species, European consumption is concentrated on a limited number of products (e.g., sea bass, sea bream, shrimp). Many native species, present in large quantities in the Mediterranean Sea, are little known to consumers and are therefore considered ‘fishing by-products’. All the data presented so far indicate a significant waste of local resources and the overexploitation of a few fish stocks. It is therefore necessary to develop strategies that guide the market towards sustainable conversion. The objective of this work was to valorize underutilized fish species of the Mediterranean Sea through a multidisciplinary approach. To this end, three fish species were sampled: Atlantic Horse Mackerel (Trachurus trachurus), Bogue (Boops boops), and Common Dolphinfish (Coryphaena hippurus). Nutritional properties (water %, fats, proteins, ashes, salts), physical/chemical properties (TVB-N, histamine, pH), and rheological properties (color, texture, viscosity) were analyzed. The analyses were conducted on both fillets and processing by-products. Additionally, mitochondrial DNA (mtDNA) was extracted from the muscle of each species. The mtDNA was then sequenced using the Illumina NGS technique. The analysis of nutritional properties classified the fillets of the sampled species as lean or semi-fat, as they had a fat content of less than 3%, while the by-products showed a higher lipid content (2.7-5%). The protein percentage for all fillets was 22-23%, while for processing by-products, the protein concentration was 18-19% for all species. Rheological analyses showed an increase in viscosity in saline solution in all species, indicating their potential suitability for industrial processing. High-quality and quantity complete mtDNA was extracted from all analyzed species. The complete mitochondrial genome sequences were successfully obtained and annotated. The results of this study suggest that all analyzed species are suitable for both human consumption and feed production. The sequencing of the complete mtDNA and its availability in international databases will be useful for accurate phylogenetic analysis and proper species identification, even in prepared and processed products. Underutilized fish species represent an important economic resource. Encouraging their consumption could limit the phenomenon of overfishing, protecting marine biodiversity. Furthermore, the valorization of these species will increase national fish production, supporting the local economy, cultural, and gastronomic tradition, and optimizing the exploitation of Mediterranean resources in accordance with the CFP.

Keywords: mtDNA, nutritional analysis, sustainable fisheries, underutilized fish species

Procedia PDF Downloads 13
789 DLtrace: Toward Understanding and Testing Deep Learning Information Flow in Deep Learning-Based Android Apps

Authors: Jie Zhang, Qianyu Guo, Tieyi Zhang, Zhiyong Feng, Xiaohong Li

Abstract:

With the widespread popularity of mobile devices and the development of artificial intelligence (AI), deep learning (DL) has been extensively applied in Android apps. Compared with traditional Android apps (traditional apps), deep learning based Android apps (DL-based apps) need to use more third-party application programming interfaces (APIs) to complete complex DL inference tasks. However, existing methods (e.g., FlowDroid) for detecting sensitive information leakage in Android apps cannot be directly used to detect DL-based apps as they are difficult to detect third-party APIs. To solve this problem, we design DLtrace; a new static information flow analysis tool that can effectively recognize third-party APIs. With our proposed trace and detection algorithms, DLtrace can also efficiently detect privacy leaks caused by sensitive APIs in DL-based apps. Moreover, using DLtrace, we summarize the non-sequential characteristics of DL inference tasks in DL-based apps and the specific functionalities provided by DL models for such apps. We propose two formal definitions to deal with the common polymorphism and anonymous inner-class problems in the Android static analyzer. We conducted an empirical assessment with DLtrace on 208 popular DL-based apps in the wild and found that 26.0% of the apps suffered from sensitive information leakage. Furthermore, DLtrace has a more robust performance than FlowDroid in detecting and identifying third-party APIs. The experimental results demonstrate that DLtrace expands FlowDroid in understanding DL-based apps and detecting security issues therein.

Keywords: mobile computing, deep learning apps, sensitive information, static analysis

Procedia PDF Downloads 156
788 Detecting Financial Bubbles Using Gap between Common Stocks and Preferred Stocks

Authors: Changju Lee, Seungmo Ku, Sondo Kim, Woojin Chang

Abstract:

How to detecting financial bubble? Addressing this simple question has been the focus of a vast amount of empirical research spanning almost half a century. However, financial bubble is hard to observe and varying over the time; there needs to be more research on this area. In this paper, we used abnormal difference between common stocks price and those preferred stocks price to explain financial bubble. First, we proposed the ‘W-index’ which indicates spread between common stocks and those preferred stocks in stock market. Second, to prove that this ‘W-index’ is valid for measuring financial bubble, we showed that there is an inverse relationship between this ‘W-index’ and S&P500 rate of return. Specifically, our hypothesis is that when ‘W-index’ is comparably higher than other periods, financial bubbles are added up in stock market and vice versa; according to our hypothesis, if investors made long term investments when ‘W-index’ is high, they would have negative rate of return; however, if investors made long term investments when ‘W-index’ is low, they would have positive rate of return. By comparing correlation values and adjusted R-squared values of between W-index and S&P500 return, VIX index and S&P500 return, and TED index and S&P500 return, we showed only W-index has significant relationship between S&P500 rate of return. In addition, we figured out how long investors should hold their investment position regard the effect of financial bubble. Using this W-index, investors could measure financial bubble in the market and invest with low risk.

Keywords: financial bubble detection, future return, forecasting, pairs trading, preferred stocks

Procedia PDF Downloads 361
787 De-Pigmentary Effect of Ayurvedic Treatment on Hyper-Pigmentation of Skin Due to Chloroquine: A Case Report

Authors: Sunil Kumar, Rajesh Sharma

Abstract:

Toxic epidermal necrolysis, pruritis, rashes, lichen planus like eruption, hyper pigmentation of skin are rare toxic effects of choloroquine used over a long time. Skin and mucus membrane hyper pigmentation is generally of a bluish black or grayish color and irreversible after discontinuation of the drug. According to Ayurveda, Dushivisha is the name given to any poisonous substance which is not fully endowed with the qualities of poison by nature (i.e. it acts as an impoverished or weak poison) and because of its mild potency, it remains in the body for many years causing various symptoms, one among them being discoloration of skin.The objective of this case report is to investigate the effect of Ayurvedic management of chloroquine induced hyper-pigmentation on the line of treatment of Dushivisha. Case Report: A 26-year-old female was suffering from hyper-pigmentation of the skin over the neck, forehead, temporo-mandibular joints, upper back and posterior aspect of both the arms since 8 years had history of taking Chloroquine came to Out Patient Department of National Institute of Ayurveda, Jaipur, India in Jan. 2015. The routine investigations (CBC, ESR, Eosinophil count) were within normal limits. Punch biopsy skin studied for histopathology under hematoxylin and eosin staining showed epidermis with hyper-pigmentation of the basal layer. In the papillary dermis as well as deep dermis there were scattered melanophages along with infiltration by mononuclear cells. There was no deposition of amyloid-like substances. These histopathological findings were suggestive of Chloroquine induced hyper-pigmentation. The case was treated on the line of treatment of Dushivisha and was given Vamana and Virechana (therapeutic emesis and purgation) every six months followed by Snehana karma (oleation therapy) with Panchatikta Ghrit and Swedana (sudation). Arogyavardhini Vati -1 g, Dushivishari Vati 500 mg, Mahamanjisthadi Quath 20 ml were given twelve hourly and Aragwadhadi Quath 25 ml at bed time orally. The patient started showing lightening of the pigments after six months and almost complete remission after 12 months of the treatment. Conclusion: This patient presented with the Dushivisha effect of Chloroquineandwas administered two relevant procedures from Panchakarma viz. Vamana and Virechana. Both Vamana and Virechanakarma here referred to Shodhana karma (purification procedures) eliminates accumulated toxins from the body. In this process, oleation dislodge the toxins from the tissues and sudation helps to bring them to the alimentary tract. The line of treatment did not target direct hypo pigmentary effects; rather aimed to eliminate the Dushivisha. This gave promising results in this condition.

Keywords: Ayurveda, chloroquine, Dushivisha, hyper-pigmentation

Procedia PDF Downloads 229
786 Geospatial Techniques and VHR Imagery Use for Identification and Classification of Slums in Gujrat City, Pakistan

Authors: Muhammad Ameer Nawaz Akram

Abstract:

The 21st century has been revealed that many individuals around the world are living in urban settlements than in rural zones. The evolution of numerous cities in emerging and newly developed countries is accompanied by the rise of slums. The precise definition of a slum varies countries to countries, but the universal harmony is that slums are dilapidated settlements facing severe poverty and have lacked access to sanitation, water, electricity, good living styles, and land tenure. The slum settlements always vary in unique patterns within and among the countries and cities. The core objective of this study is the spatial identification and classification of slums in Gujrat city Pakistan from very high-resolution GeoEye-1 (0.41m) satellite imagery. Slums were first identified using GPS for sample site identification and ground-truthing; through this process, 425 slums were identified. Then Object-Oriented Analysis (OOA) was applied to classify slums on digital image. Spatial analysis softwares, e.g., ArcGIS 10.3, Erdas Imagine 9.3, and Envi 5.1, were used for processing data and performing the analysis. Results show that OOA provides up to 90% accuracy for the identification of slums. Jalal Cheema and Allah Ho colonies are severely affected by slum settlements. The ratio of criminal activities is also higher here than in other areas. Slums are increasing with the passage of time in urban areas, and they will be like a hazardous problem in coming future. So now, the executive bodies need to make effective policies and move towards the amelioration process of the city.

Keywords: slums, GPS, satellite imagery, object oriented analysis, zonal change detection

Procedia PDF Downloads 124
785 Identification and Isolation of E. Coli O₁₅₇:H₇ From Water and Wastewater of Shahrood and Neka Cities by PCR Technique

Authors: Aliasghar Golmohammadian, Sona Rostampour Yasouri

Abstract:

One of the most important intestinal pathogenic strains is E. coli O₁₅₇:H₇. This pathogenic bacterium is transmitted to humans through water and food. E. coli O₁₅₇:H₇ is the main cause of Hemorrhagic colitis (HC), Hemolytic Uremic Syndrome (HUS), Thrombotic Thrombocytopenic Purpura (TTP) and in some cases death. Since E. coli O₁₅₇:H₇ can be transmitted through the consumption of different foods, including vegetables, agricultural products, and fresh dairy products, this study aims to identify and isolate E. coli O₁₅₇:H₇ from wastewater by PCR technique. One hundred twenty samples of water and wastewater were collected by Falcom Sterile from Shahrood and Neka cities. The samples were checked for colony formation after appropriate centrifugation and cultivation in the specific medium of Sorbitol MacConkey Agar (SMAC) and other diagnostic media of E. coli O₁₅₇:H₇. Also, the plates were observed macroscopically and microscopically. Then, the necessary phenotypic tests were performed on the colonies, and finally, after DNA extraction, the PCR technique was performed with specific primers related to rfbE and stx2 genes. The number of 5 samples (6%) out of all the samples examined were determined positive by PCR technique with observing the bands related to the mentioned genes on the agarose gel electrophoresis. PCR is a fast and accurate method to identify the bacteria E. coli O₁₅₇:H₇. Considering that E. coli bacteria is a resistant bacteria and survives in water and food for weeks and months, the PCR technique can provide the possibility of quick detection of contaminated water. Moreover, it helps people in the community control and prevent the transfer of bacteria to healthy and underground water and agricultural and even dairy products.

Keywords: E. coli O₁₅₇:H₇, PCR, water, wastewater

Procedia PDF Downloads 49
784 Virtual Process Hazard Analysis (Pha) Of a Nuclear Power Plant (Npp) Using Failure Mode and Effects Analysis (Fmea) Technique

Authors: Lormaine Anne A. Branzuela, Elysa V. Largo, Monet Concepcion M. Detras, Neil C. Concibido

Abstract:

The electricity demand is still increasing, and currently, the Philippine government is investigating the feasibility of operating the Bataan Nuclear Power Plant (BNPP) to address the country’s energy problem. However, the lack of process safety studies on BNPP focused on the effects of hazardous substances on the integrity of the structure, equipment, and other components, have made the plant operationalization questionable to the public. The three major nuclear power plant incidents – TMI-2, Chernobyl, and Fukushima – have made many people hesitant to include nuclear energy in the energy matrix. This study focused on the safety evaluation of possible operations of a nuclear power plant installed with a Pressurized Water Reactor (PWR), which is similar to BNPP. Failure Mode and Effects Analysis (FMEA) is one of the Process Hazard Analysis (PHA) techniques used for the identification of equipment failure modes and minimizing its consequences. Using the FMEA technique, this study was able to recognize 116 different failure modes in total. Upon computation and ranking of the risk priority number (RPN) and criticality rating (CR), it showed that failure of the reactor coolant pump due to earthquakes is the most critical failure mode. This hazard scenario could lead to a nuclear meltdown and radioactive release, as identified by the FMEA team. Safeguards and recommended risk reduction strategies to lower the RPN and CR were identified such that the effects are minimized, the likelihood of occurrence is reduced, and failure detection is improved.

Keywords: PHA, FMEA, nuclear power plant, bataan nuclear power plant

Procedia PDF Downloads 117
783 Complex Management of Arrhythmogenic Right Ventricular Dysplasia/Cardiomyopathy

Authors: Abdullah A. Al Qurashi, Hattan A. Hassani, Bader K. Alaslap

Abstract:

Arrhythmogenic Right Ventricular Dysplasia/Cardiomyopathy (ARVD/C) is an uncommon, inheritable cardiac disorder characterized by the progressive substitution of cardiac myocytes by fibro-fatty tissues. This pathologic substitution predisposes patients to ventricular arrhythmias and right ventricular failure. The underlying genetic defect predominantly involves genes encoding for desmosome proteins, particularly plakophilin-2 (PKP2). These aberrations lead to impaired cell adhesion, heightening the susceptibility to fibrofatty scarring under conditions of mechanical stress. Primarily, ARVD/C affects the right ventricle, but it can also compromise the left ventricle, potentially leading to biventricular heart failure. Clinical presentations can vary, spanning from asymptomatic individuals to those experiencing palpitations, syncopal episodes, and, in severe instances, sudden cardiac death. The establishment of a diagnostic criterion specifically tailored for ARVD/C significantly aids in its accurate diagnosis. Nevertheless, the task of early diagnosis is complicated by the disease's frequently asymptomatic initial stages, and the overall rarity of ARVD/C cases reported globally. In some cases, as exemplified by the adult female patient in this report, the disease may advance to terminal stages, rendering therapies like Ventricular Tachycardia (VT) ablation ineffective. This case underlines the necessity for increased awareness and understanding of ARVD/C to aid in its early detection and management. Through such efforts, we aim to decrease morbidity and mortality associated with this challenging cardiac disorder.

Keywords: arrhythmogenic right ventricular dysplasia, cardiac disease, interventional cardiology, cardiac electrophysiology

Procedia PDF Downloads 53
782 DeepNIC a Method to Transform Each Tabular Variable into an Independant Image Analyzable by Basic CNNs

Authors: Nguyen J. M., Lucas G., Ruan S., Digonnet H., Antonioli D.

Abstract:

Introduction: Deep Learning (DL) is a very powerful tool for analyzing image data. But for tabular data, it cannot compete with machine learning methods like XGBoost. The research question becomes: can tabular data be transformed into images that can be analyzed by simple CNNs (Convolutional Neuron Networks)? Will DL be the absolute tool for data classification? All current solutions consist in repositioning the variables in a 2x2 matrix using their correlation proximity. In doing so, it obtains an image whose pixels are the variables. We implement a technology, DeepNIC, that offers the possibility of obtaining an image for each variable, which can be analyzed by simple CNNs. Material and method: The 'ROP' (Regression OPtimized) model is a binary and atypical decision tree whose nodes are managed by a new artificial neuron, the Neurop. By positioning an artificial neuron in each node of the decision trees, it is possible to make an adjustment on a theoretically infinite number of variables at each node. From this new decision tree whose nodes are artificial neurons, we created the concept of a 'Random Forest of Perfect Trees' (RFPT), which disobeys Breiman's concepts by assembling very large numbers of small trees with no classification errors. From the results of the RFPT, we developed a family of 10 statistical information criteria, Nguyen Information Criterion (NICs), which evaluates in 3 dimensions the predictive quality of a variable: Performance, Complexity and Multiplicity of solution. A NIC is a probability that can be transformed into a grey level. The value of a NIC depends essentially on 2 super parameters used in Neurops. By varying these 2 super parameters, we obtain a 2x2 matrix of probabilities for each NIC. We can combine these 10 NICs with the functions AND, OR, and XOR. The total number of combinations is greater than 100,000. In total, we obtain for each variable an image of at least 1166x1167 pixels. The intensity of the pixels is proportional to the probability of the associated NIC. The color depends on the associated NIC. This image actually contains considerable information about the ability of the variable to make the prediction of Y, depending on the presence or absence of other variables. A basic CNNs model was trained for supervised classification. Results: The first results are impressive. Using the GSE22513 public data (Omic data set of markers of Taxane Sensitivity in Breast Cancer), DEEPNic outperformed other statistical methods, including XGBoost. We still need to generalize the comparison on several databases. Conclusion: The ability to transform any tabular variable into an image offers the possibility of merging image and tabular information in the same format. This opens up great perspectives in the analysis of metadata.

Keywords: tabular data, CNNs, NICs, DeepNICs, random forest of perfect trees, classification

Procedia PDF Downloads 108
781 Clinical Signs of Neonatal Calves in Experimental Colisepticemia

Authors: Samad Lotfollahzadeh

Abstract:

Escherichia coli (E.coli) is the most isolated bacteria from blood circulation of septicemic calves. Given the prevalence of septicemia in animals and its economic importance in veterinary practice, better understanding of changes in clinical signs following disease, may contribute to early detection of the disorder. The present study has been carried out to detect changes of clinical signs in induced sepsis in calves with E.coli. Colisepticemia has been induced in 10 twenty-day old healthy Holstein- Frisian calves with intravenous injection of 1.5 X 109 colony forming units (cfu) of O111: H8 strain of E.coli. Clinical signs including rectal temperature, heart rate, respiratory rate, shock, appetite, sucking reflex, feces consistency, general behavior, dehydration and standing ability were recorded in experimental calves during 24 hours after induction of colisepticemia. Blood culture was also carried out from calves four times during the experiment. ANOVA with repeated measure is used to see changes of calves’ clinical signs to experimental colisepticemia, and values of P≤ 0.05 was considered statistically significant. Mean values of rectal temperature and heart rate as well as median values of respiratory rate, appetite, suckling reflex, standing ability and feces consistency of experimental calves increased significantly during the study (P<0.05). In the present study, median value of shock score was not significantly increased in experimental calves (P> 0.05). The results of present study showed that total score of clinical signs in calves with experimental colisepticemia increased significantly, although the score of some clinical signs such as shock did not change significantly.

Keywords: calves, clinical signs scoring, E. coli O111:H8, experimental colisepticemia

Procedia PDF Downloads 368
780 A Robust and Efficient Segmentation Method Applied for Cardiac Left Ventricle with Abnormal Shapes

Authors: Peifei Zhu, Zisheng Li, Yasuki Kakishita, Mayumi Suzuki, Tomoaki Chono

Abstract:

Segmentation of left ventricle (LV) from cardiac ultrasound images provides a quantitative functional analysis of the heart to diagnose disease. Active Shape Model (ASM) is a widely used approach for LV segmentation but suffers from the drawback that initialization of the shape model is not sufficiently close to the target, especially when dealing with abnormal shapes in disease. In this work, a two-step framework is proposed to improve the accuracy and speed of the model-based segmentation. Firstly, a robust and efficient detector based on Hough forest is proposed to localize cardiac feature points, and such points are used to predict the initial fitting of the LV shape model. Secondly, to achieve more accurate and detailed segmentation, ASM is applied to further fit the LV shape model to the cardiac ultrasound image. The performance of the proposed method is evaluated on a dataset of 800 cardiac ultrasound images that are mostly of abnormal shapes. The proposed method is compared to several combinations of ASM and existing initialization methods. The experiment results demonstrate that the accuracy of feature point detection for initialization was improved by 40% compared to the existing methods. Moreover, the proposed method significantly reduces the number of necessary ASM fitting loops, thus speeding up the whole segmentation process. Therefore, the proposed method is able to achieve more accurate and efficient segmentation results and is applicable to unusual shapes of heart with cardiac diseases, such as left atrial enlargement.

Keywords: hough forest, active shape model, segmentation, cardiac left ventricle

Procedia PDF Downloads 330
779 Calculation of the Normalized Difference Vegetation Index and the Spectral Signature of Coffee Crops: Benefits of Image Filtering on Mixed Crops

Authors: Catalina Albornoz, Giacomo Barbieri

Abstract:

Crop monitoring has shown to reduce vulnerability to spreading plagues and pathologies in crops. Remote sensing with Unmanned Aerial Vehicles (UAVs) has made crop monitoring more precise, cost-efficient and accessible. Nowadays, remote monitoring involves calculating maps of vegetation indices by using different software that takes either Truecolor (RGB) or multispectral images as an input. These maps are then used to segment the crop into management zones. Finally, knowing the spectral signature of a crop (the reflected radiation as a function of wavelength) can be used as an input for decision-making and crop characterization. The calculation of vegetation indices using software such as Pix4D has high precision for monoculture plantations. However, this paper shows that using this software on mixed crops may lead to errors resulting in an incorrect segmentation of the field. Within this work, authors propose to filter all the elements different from the main crop before the calculation of vegetation indices and the spectral signature. A filter based on the Sobel method for border detection is used for filtering a coffee crop. Results show that segmentation into management zones changes with respect to the traditional situation in which a filter is not applied. In particular, it is shown how the values of the spectral signature change in up to 17% per spectral band. Future work will quantify the benefits of filtering through the comparison between in situ measurements and the calculated vegetation indices obtained through remote sensing.

Keywords: coffee, filtering, mixed crop, precision agriculture, remote sensing, spectral signature

Procedia PDF Downloads 379
778 Preparation of Frozen Bivalent Babesial (Babesia Bovis and Babesia Bigemina) Vaccine from Field Isolates and Evaluation of Its Efficacy in Calves

Authors: Muhammad Fiaz Qamar, Ahmad Faraz, Muhammad Arfan Zaman, Kazim Ali, Waleed Akram

Abstract:

Babesiosis is reflected as the most important disease of cattle that are transmitted by arthropods. In Pakistan, its prevalence is up to 29% in the cattle and buffalo population in different regions. Cattle show a long lasting and durable immunity by giving an infection of B.bovis, B. bigemina, or Babesiadivergens. this is used in cattle to immunize them in a few countries as anti-babesiosis vaccine. Development of frozen vaccine allows for complete testing after production of each batch, However, once thawed, its reduced its shelf life, frozen vaccines are more difficult to transport as well as expensive to produce as compared to chilled vaccine. The contamination of blood derived vaccine has the potential risk that makes pre-production and post-production quality control necessary. For the trail master seed production of whole blood frozen bivalent Babesia(Babesiabovis and Babesiabigemina), 100 blood samples of Babesial positive suspected cattle was taken and processed for separation microscopic detection and rectification by PCR. Vaccine passages were done to reduce the parasitaemiasis in live calves. After 8 passages, parasitemia of Babesia reduced from 80% to 15%. Infected donor calf’s blood was taken by jugular cannulation by using preservative free lithium heparin as an anticoagulant (5 International Units IU heparin/ml blood). In lab, parasite containing blood was mixed in equal volumes with 3 M glycerol in PBS supplemented with 5 mM glucose (final concentration of glycerol 1.5 M) at 37°C. The mixture was then equilibrized at 37°C for 30 minutes and were dispensed in required containers (e.g., 5 ml cryovials).

Keywords: distribution, babesia, primer sequences, PCV

Procedia PDF Downloads 92
777 Radio-Frequency Technologies for Sensing and Imaging

Authors: Cam Nguyen

Abstract:

Rapid, accurate, and safe sensing and imaging of physical quantities or structures finds many applications and is of significant interest to society. Sensing and imaging using radio-frequency (RF) techniques, particularly, has gone through significant development and subsequently established itself as a unique territory in the sensing world. RF sensing and imaging has played a critical role in providing us many sensing and imaging abilities beyond our human capabilities, benefiting both civilian and military applications - for example, from sensing abnormal conditions underneath some structures’ surfaces to detection and classification of concealed items, hidden activities, and buried objects. We present the developments of several sensing and imaging systems implementing RF technologies like ultra-wide band (UWB), synthetic-pulse, and interferometry. These systems are fabricated completely using RF integrated circuits. The UWB impulse system operates over multiple pulse durations from 450 to 1170 ps with 5.5-GHz RF bandwidth. It performs well through tests of various samples, demonstrating its usefulness for subsurface sensing. The synthetic-pulse system operating from 0.6 to 5.6 GHz can assess accurately subsurface structures. The synthetic-pulse system operating from 29.72-37.7 GHz demonstrates abilities for various surface and near-surface sensing such as profile mapping, liquid-level monitoring, and anti-personnel mine locating. The interferometric system operating at 35.6 GHz demonstrates its multi-functional capability for measurement of displacements and slow velocities. These RF sensors are attractive and useful for various surface and subsurface sensing applications. This paper was made possible by NPRP grant # 6-241-2-102 from the Qatar National Research Fund (a member of Qatar Foundation). The statements made herein are solely the responsibility of the authors.

Keywords: RF sensors, radars, surface sensing, subsurface sensing

Procedia PDF Downloads 304
776 Prediction of Damage to Cutting Tools in an Earth Pressure Balance Tunnel Boring Machine EPB TBM: A Case Study L3 Guadalajara Metro Line (Mexico)

Authors: Silvia Arrate, Waldo Salud, Eloy París

Abstract:

The wear of cutting tools is one of the most decisive elements when planning tunneling works, programming the maintenance stops and saving the optimum stock of spare parts during the evolution of the excavation. Being able to predict the behavior of cutting tools can give a very competitive advantage in terms of costs and excavation performance, optimized to the needs of the TBM itself. The incredible evolution of data science in recent years gives the option to implement it at the time of analyzing the key and most critical parameters related to machinery with the purpose of knowing how the cutting head is performing in front of the excavated ground. Taking this as a case study, Metro Line 3 of Guadalajara in Mexico will develop the feasibility of using Specific Energy versus data science applied over parameters of Torque, Penetration, and Contact Force, among others, to predict the behavior and status of cutting tools. The results obtained through both techniques are analyzed and verified in the function of the wear and the field situations observed in the excavation in order to determine its effectiveness regarding its predictive capacity. In conclusion, the possibilities and improvements offered by the application of digital tools and the programming of calculation algorithms for the analysis of wear of cutting head elements compared to purely empirical methods allow early detection of possible damage to cutting tools, which is reflected in optimization of excavation performance and a significant improvement in costs and deadlines.

Keywords: cutting tools, data science, prediction, TBM, wear

Procedia PDF Downloads 38
775 The Economic Burden of Breast Cancer on Women in Nigeria: Implication for Socio-Economic Development

Authors: Tolulope Allo, Mofoluwake P. Ajayi, Adenike E. Idowu, Emmanuel O. Amoo, Fadeke Esther Olu-Owolabi

Abstract:

Breast cancer which was more prevalent in Europe and America in the past is gradually being mirrored across the world today with greater economic burden on low and middle income countries (LMCs). Breast cancer is the most common cancer among women globally and current studies have shown that a woman dies with the diagnosis of breast cancer every thirteen minutes. The economic cost of breast cancer is overwhelming particularly for developing economies. While it causes billion of dollar in losses of national income, it pushes millions of people below poverty line. This study examined the economic burden of breast cancer on Nigerian women, its impacts on their standard of living and its effects on Nigeria’s socio economic development. The study adopts a qualitative research approach using the in-depth interview technique to elicit valuable information from respondents with cancer experience from the Southern part of Nigeria. Respondents constituted women in their reproductive age (15-49 years) that have experienced and survived cancer and also those that are currently receiving treatment. Excerpts from the interviews revealed that the cost of treatment is one of the major factors contributing to the late presentation of breast cancer incidences among women as many of them could not afford to pay for their own treatment. The study also revealed that many women prefer to explore other options such as herbal treatments and spiritual consultations which is less expensive and affordable. The study therefore concludes that breast cancer diagnosis and treatment should be subsidized by the government in order to facilitate easy access and affordability thereby promoting early detection and reducing the economic burden of treatment on women.

Keywords: breast cancer, development, economic burden, women

Procedia PDF Downloads 347
774 A Bayesian Parameter Identification Method for Thermorheological Complex Materials

Authors: Michael Anton Kraus, Miriam Schuster, Geralt Siebert, Jens Schneider

Abstract:

Polymers increasingly gained interest in construction materials over the last years in civil engineering applications. As polymeric materials typically show time- and temperature dependent material behavior, which is accounted for in the context of the theory of linear viscoelasticity. Within the context of this paper, the authors show, that some polymeric interlayers for laminated glass can not be considered as thermorheologically simple as they do not follow a simple TTSP, thus a methodology of identifying the thermorheologically complex constitutive bahavioir is needed. ‘Dynamical-Mechanical-Thermal-Analysis’ (DMTA) in tensile and shear mode as well as ‘Differential Scanning Caliometry’ (DSC) tests are carried out on the interlayer material ‘Ethylene-vinyl acetate’ (EVA). A navoel Bayesian framework for the Master Curving Process as well as the detection and parameter identification of the TTSPs along with their associated Prony-series is derived and applied to the EVA material data. To our best knowledge, this is the first time, an uncertainty quantification of the Prony-series in a Bayesian context is shown. Within this paper, we could successfully apply the derived Bayesian methodology to the EVA material data to gather meaningful Master Curves and TTSPs. Uncertainties occurring in this process can be well quantified. We found, that EVA needs two TTSPs with two associated Generalized Maxwell Models. As the methodology is kept general, the derived framework could be also applied to other thermorheologically complex polymers for parameter identification purposes.

Keywords: bayesian parameter identification, generalized Maxwell model, linear viscoelasticity, thermorheological complex

Procedia PDF Downloads 254
773 A New Multi-Target, Multi-Agent Search and Rescue Path Planning Approach

Authors: Jean Berger, Nassirou Lo, Martin Noel

Abstract:

Perfectly suited for natural or man-made emergency and disaster management situations such as flood, earthquakes, tornadoes, or tsunami, multi-target search path planning for a team of rescue agents is known to be computationally hard, and most techniques developed so far come short to successfully estimate optimality gap. A novel mixed-integer linear programming (MIP) formulation is proposed to optimally solve the multi-target multi-agent discrete search and rescue (SAR) path planning problem. Aimed at maximizing cumulative probability of successful target detection, it captures anticipated feedback information associated with possible observation outcomes resulting from projected path execution, while modeling agent discrete actions over all possible moving directions. Problem modeling further takes advantage of network representation to encompass decision variables, expedite compact constraint specification, and lead to substantial problem-solving speed-up. The proposed MIP approach uses CPLEX optimization machinery, efficiently computing near-optimal solutions for practical size problems, while giving a robust upper bound obtained from Lagrangean integrality constraint relaxation. Should eventually a target be positively detected during plan execution, a new problem instance would simply be reformulated from the current state, and then solved over the next decision cycle. A computational experiment shows the feasibility and the value of the proposed approach.

Keywords: search path planning, search and rescue, multi-agent, mixed-integer linear programming, optimization

Procedia PDF Downloads 362
772 Design of a Portable Shielding System for a Newly Installed NaI(Tl) Detector

Authors: Mayesha Tahsin, A.S. Mollah

Abstract:

Recently, a 1.5x1.5 inch NaI(Tl) detector based gamma-ray spectroscopy system has been installed in the laboratory of the Nuclear Science and Engineering Department of the Military Institute of Science and Technology for radioactivity detection purposes. The newly installed NaI(Tl) detector has a circular lead shield of 22 mm width. An important consideration of any gamma-ray spectroscopy is the minimization of natural background radiation not originating from the radioactive sample that is being measured. Natural background gamma-ray radiation comes from naturally occurring or man-made radionuclides in the environment or from cosmic sources. Moreover, the main problem with this system is that it is not suitable for measurements of radioactivity with a large sample container like Petridish or Marinelli beaker geometry. When any laboratory installs a new detector or/and new shield, it “must” first carry out quality and performance tests for the detector and shield. This paper describes a new portable shielding system with lead that can reduce the background radiation. Intensity of gamma radiation after passing the shielding will be calculated using shielding equation I=Ioe-µx where Io is initial intensity of the gamma source, I is intensity after passing through the shield, µ is linear attenuation coefficient of the shielding material, and x is the thickness of the shielding material. The height and width of the shielding will be selected in order to accommodate the large sample container. The detector will be surrounded by a 4π-geometry low activity lead shield. An additional 1.5 mm thick shield of tin and 1 mm thick shield of copper covering the inner part of the lead shielding will be added in order to remove the presence of characteristic X-rays from the lead shield.

Keywords: shield, NaI (Tl) detector, gamma radiation, intensity, linear attenuation coefficient

Procedia PDF Downloads 144
771 Code Embedding for Software Vulnerability Discovery Based on Semantic Information

Authors: Joseph Gear, Yue Xu, Ernest Foo, Praveen Gauravaran, Zahra Jadidi, Leonie Simpson

Abstract:

Deep learning methods have been seeing an increasing application to the long-standing security research goal of automatic vulnerability detection for source code. Attention, however, must still be paid to the task of producing vector representations for source code (code embeddings) as input for these deep learning models. Graphical representations of code, most predominantly Abstract Syntax Trees and Code Property Graphs, have received some use in this task of late; however, for very large graphs representing very large code snip- pets, learning becomes prohibitively computationally expensive. This expense may be reduced by intelligently pruning this input to only vulnerability-relevant information; however, little research in this area has been performed. Additionally, most existing work comprehends code based solely on the structure of the graph at the expense of the information contained by the node in the graph. This paper proposes Semantic-enhanced Code Embedding for Vulnerability Discovery (SCEVD), a deep learning model which uses semantic-based feature selection for its vulnerability classification model. It uses information from the nodes as well as the structure of the code graph in order to select features which are most indicative of the presence or absence of vulnerabilities. This model is implemented and experimentally tested using the SARD Juliet vulnerability test suite to determine its efficacy. It is able to improve on existing code graph feature selection methods, as demonstrated by its improved ability to discover vulnerabilities.

Keywords: code representation, deep learning, source code semantics, vulnerability discovery

Procedia PDF Downloads 147