Search results for: pixel entropy
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 565

Search results for: pixel entropy

115 Pressure-Controlled Dynamic Equations of the PFC Model: A Mathematical Formulation

Authors: Jatupon Em-Udom, Nirand Pisutha-Arnond

Abstract:

The phase-field-crystal, PFC, approach is a density-functional-type material model with an atomic resolution on a diffusive timescale. Spatially, the model incorporates periodic nature of crystal lattices and can naturally exhibit elasticity, plasticity and crystal defects such as grain boundaries and dislocations. Temporally, the model operates on a diffusive timescale which bypasses the need to resolve prohibitively small atomic-vibration time steps. The PFC model has been used to study many material phenomena such as grain growth, elastic and plastic deformations and solid-solid phase transformations. In this study, the pressure-controlled dynamic equation for the PFC model was developed to simulate a single-component system under externally applied pressure; these coupled equations are important for studies of deformable systems such as those under constant pressure. The formulation is based on the non-equilibrium thermodynamics and the thermodynamics of crystalline solids. To obtain the equations, the entropy variation around the equilibrium point was derived. Then the resulting driving forces and flux around the equilibrium were obtained and rewritten as conventional thermodynamic quantities. These dynamics equations are different from the recently-proposed equations; the equations in this study should provide more rigorous descriptions of the system dynamics under externally applied pressure.

Keywords: driving forces and flux, evolution equation, non equilibrium thermodynamics, Onsager’s reciprocal relation, phase field crystal model, thermodynamics of single-component solid

Procedia PDF Downloads 282
114 Cosmic Muon Tomography at the Wylfa Reactor Site Using an Anti-Neutrino Detector

Authors: Ronald Collins, Jonathon Coleman, Joel Dasari, George Holt, Carl Metelko, Matthew Murdoch, Alexander Morgan, Yan-Jie Schnellbach, Robert Mills, Gareth Edwards, Alexander Roberts

Abstract:

At the Wylfa Magnox Power Plant between 2014–2016, the VIDARR prototype anti-neutrino detector was deployed. It is comprised of extruded plastic scintillating bars measuring 4 cm × 1 cm × 152 cm and utilised wavelength shifting fibres (WLS) and multi-pixel photon counters (MPPCs) to detect and quantify radiation. During deployment, it took cosmic muon data in accidental coincidence with the anti-neutrino measurements with the power plant site buildings obscuring the muon sky. Cosmic muons have a significantly higher probability of being attenuated and/or absorbed by denser objects, and so one-sided cosmic muon tomography was utilised to image the reactor site buildings. In order to achieve clear building outlines, a control data set was taken at the University of Liverpool from 2016 – 2018, which had minimal occlusion of the cosmic muon flux by dense objects. By taking the ratio of these two data sets and using GEANT4 simulations, it is possible to perform a one-sided cosmic muon tomography analysis. This analysis can be used to discern specific buildings, building heights, and features at the Wylfa reactor site, including the reactor core/reactor core shielding using ∼ 3 hours worth of cosmic-ray detector live time. This result demonstrates the feasibility of using cosmic muon analysis to determine a segmented detector’s location with respect to surrounding buildings, assisted by aerial photography or satellite imagery.

Keywords: anti-neutrino, GEANT4, muon, tomography, occlusion

Procedia PDF Downloads 164
113 Research on the Efficiency and Driving Elements of Manufacturing Transformation and Upgrading in the Context of Digitization

Authors: Chen Zhang; Qiang Wang

Abstract:

With the rapid development of the new generation of digital technology, various industries have created more and more value by using digital technology, accelerating the digital transformation of various industries. The economic form of human society has evolved with the progress of technology, and in this context, the power conversion, transformation and upgrading of the manufacturing industry in terms of quality, efficiency and energy change has become a top priority. Based on the digitalization background, this paper analyzes the transformation and upgrading efficiency of the manufacturing industry and evaluates the impact of the driving factors, which have very important theoretical and practical significance. This paper utilizes qualitative research methods, entropy methods, data envelopment analysis methods and econometric models to explore the transformation and upgrading efficiency of manufacturing enterprises and driving factors. The study shows that the transformation and upgrading efficiency of the manufacturing industry shows a steady increase, and regions rich in natural resources and social resources provide certain resources for transformation and upgrading. The ability of scientific and technological innovation has been improved, but there is still much room for progress in the transformation of scientific and technological innovation achievements. Most manufacturing industries pay more attention to green manufacturing and sustainable development. In addition, based on the existing problems, this paper puts forward suggestions for improving infrastructure construction, developing the technological innovation capacity of enterprises, green production and sustainable development.

Keywords: digitization, manufacturing firms, transformation and upgrading, efficiency, driving factors

Procedia PDF Downloads 40
112 Comparing Accuracy of Semantic and Radiomics Features in Prognosis of Epidermal Growth Factor Receptor Mutation in Non-Small Cell Lung Cancer

Authors: Mahya Naghipoor

Abstract:

Purpose: Non-small cell lung cancer (NSCLC) is the most common lung cancer type. Epidermal growth factor receptor (EGFR) mutation is the main reason which causes NSCLC. Computed tomography (CT) is used for diagnosis and prognosis of lung cancers because of low price and little invasion. Semantic analyses of qualitative CT features are based on visual evaluation by radiologist. However, the naked eye ability may not assess all image features. On the other hand, radiomics provides the opportunity of quantitative analyses for CT images features. The aim of this review study was comparing accuracy of semantic and radiomics features in prognosis of EGFR mutation in NSCLC. Methods: For this purpose, the keywords including: non-small cell lung cancer, epidermal growth factor receptor mutation, semantic, radiomics, feature, receiver operating characteristics curve (ROC) and area under curve (AUC) were searched in PubMed and Google Scholar. Totally 29 papers were reviewed and the AUC of ROC analyses for semantic and radiomics features were compared. Results: The results showed that the reported AUC amounts for semantic features (ground glass opacity, shape, margins, lesion density and presence or absence of air bronchogram, emphysema and pleural effusion) were %41-%79. For radiomics features (kurtosis, skewness, entropy, texture, standard deviation (SD) and wavelet) the AUC values were found %50-%86. Conclusions: In conclusion, the accuracy of radiomics analysis is a little higher than semantic in prognosis of EGFR mutation in NSCLC.

Keywords: lung cancer, radiomics, computer tomography, mutation

Procedia PDF Downloads 131
111 A Research on the Coordinated Development of Chengdu-Chongqing Economic Circle under the Background of New Urbanization

Authors: Deng Tingting

Abstract:

The coordinated and integrated development of regions is an inevitable requirement for China to move towards high-quality, sustainable development. As one of the regions with the best economic foundation and the strongest economic strength in western China, it is a typical area with national importance and strong network connection characteristics in terms of the comprehensive effect of linking the inland hinterland and connecting the western and national urban networks. The integrated development of the Chengdu-Chongqing economic circle is of great strategic significance for the rapid and high-quality development of the western region. In the context of new urbanization, this paper takes 16 urban units within the economic circle as the research object, based on the 5-year panel data of population, regional economy, and spatial construction and development from 2016 to 2020, using the entropy method and Theil index to analyze the three target layers, and cause analysis. The research shows that there are temporal and spatial differences in the Chengdu-Chongqing economic circle, and there are significant differences between the core city and the surrounding cities. Therefore, by reforming and innovating the regional coordinated development mechanism, breaking administrative barriers, and strengthening the "polar nucleus" radiation function to release the driving force for economic development, especially in the gully areas of economic development belts, not only promote the coordinated development of internal regions but also promote the coordinated and sustainable development of the western region and take a high-quality development path.

Keywords: Chengdu-Chongqing economic circle, new urbanization, coordinated regional development, Theil Index

Procedia PDF Downloads 90
110 Comparati̇ve Study of Pi̇xel and Object-Based Image Classificati̇on Techni̇ques for Extracti̇on of Land Use/Land Cover Informati̇on

Authors: Mahesh Kumar Jat, Manisha Choudhary

Abstract:

Rapid population and economic growth resulted in changes in large-scale land use land cover (LULC) changes. Changes in the biophysical properties of the Earth's surface and its impact on climate are of primary concern nowadays. Different approaches, ranging from location-based relationships or modelling earth surface - atmospheric interaction through modelling techniques like surface energy balance (SEB) have been used in the recent past to examine the relationship between changes in Earth surface land cover and climatic characteristics like temperature and precipitation. A remote sensing-based model i.e., Surface Energy Balance Algorithm for Land (SEBAL), has been used to estimate the surface heat fluxes over Mahi Bajaj Sagar catchment (India) from 2001 to 2020. Landsat ETM and OLI satellite data are used to model the SEB of the area. Changes in observed precipitation and temperature, obtained from India Meteorological Department (IMD) have been correlated with changes in surface heat fluxes to understand the relative contributions of LULC change in changing these climatic variables. Results indicate a noticeable impact of LULC changes on climatic variables, which are aligned with respective changes in SEB components. Results suggest that precipitation increases at a rate of 20 mm/year. The maximum and minimum temperature decreases and increases at 0.007 ℃ /year and 0.02 ℃ /year, respectively. The average temperature increases at 0.009 ℃ /year. Changes in latent heat flux and sensible heat flux positively correlate with precipitation and temperature, respectively. Variation in surface heat fluxes influences the climate parameters and is an adequate reason for climate change. So, SEB modelling is helpful to understand the LULC change and its impact on climate.

Keywords: remote sensing, GIS, object based, classification

Procedia PDF Downloads 95
109 Heliport Remote Safeguard System Based on Real-Time Stereovision 3D Reconstruction Algorithm

Authors: Ł. Morawiński, C. Jasiński, M. Jurkiewicz, S. Bou Habib, M. Bondyra

Abstract:

With the development of optics, electronics, and computers, vision systems are increasingly used in various areas of life, science, and industry. Vision systems have a huge number of applications. They can be used in quality control, object detection, data reading, e.g., QR-code, etc. A large part of them is used for measurement purposes. Some of them make it possible to obtain a 3D reconstruction of the tested objects or measurement areas. 3D reconstruction algorithms are mostly based on creating depth maps from data that can be acquired from active or passive methods. Due to the specific appliance in airfield technology, only passive methods are applicable because of other existing systems working on the site, which can be blinded on most spectral levels. Furthermore, reconstruction is required to work long distances ranging from hundreds of meters to tens of kilometers with low loss of accuracy even with harsh conditions such as fog, rain, or snow. In response to those requirements, HRESS (Heliport REmote Safeguard System) was developed; which main part is a rotational head with a two-camera stereovision rig gathering images around the head in 360 degrees along with stereovision 3D reconstruction and point cloud combination. The sub-pixel analysis introduced in the HRESS system makes it possible to obtain an increased distance measurement resolution and accuracy of about 3% for distances over one kilometer. Ultimately, this leads to more accurate and reliable measurement data in the form of a point cloud. Moreover, the program algorithm introduces operations enabling the filtering of erroneously collected data in the point cloud. All activities from the programming, mechanical and optical side are aimed at obtaining the most accurate 3D reconstruction of the environment in the measurement area.

Keywords: airfield monitoring, artificial intelligence, stereovision, 3D reconstruction

Procedia PDF Downloads 96
108 Unveiling the Self-Assembly Behavior and Salt-Induced Morphological Transition of Double PEG-Tailed Unconventional Amphiphiles

Authors: Rita Ghosh, Joykrishna Dey

Abstract:

PEG-based amphiphiles are of tremendous importance for its widespread applications in pharmaceutics, household purposes, and drug delivery. Previously, a number of single PEG-tailed amphiphiles having significant applications have been reported from our group. Therefore, it was of immense interest to explore the properties and application potential of PEG-based double tailed amphiphiles. Herein, for the first time, two novel double PEG-tailed amphiphiles having different PEG chain lengths have been developed. The self-assembly behavior of the newly developed amphiphiles in aqueous buffer (pH 7.0) was thoroughly investigated at 25 oC by a number of techniques including, 1H-NMR, and steady-state and time-dependent fluorescence spectroscopy, dynamic light scattering, transmission electron microscopy, atomic force microscopy, and isothermal titration calorimetry. Despite having two polar PEG chains both molecules were found to have strong tendency to self-assemble in aqueous buffered solution above a very low concentration. Surprisingly, the amphiphiles were shown to form stable vesicles spontaneously at room temperature without any external stimuli. The results of calorimetric measurements showed that the vesicle formation is driven by the hydrophobic effect (positive entropy change) of the system, which is associated with the helix-to-random coil transition of the PEG chain. The spectroscopic data confirmed that the bilayer membrane of the vesicles is constituted by the PEG chains of the amphiphilic molecule. Interestingly, the vesicles were also found to exhibit structural transitions upon addition of salts in solution. These properties of the vesicles enable them as potential candidate for drug delivery.

Keywords: double-tailed amphiphiles, fluorescence, microscopy, PEG, vesicles

Procedia PDF Downloads 98
107 Recent Climate Variability and Crop Production in the Central Highlands of Ethiopia

Authors: Arragaw Alemayehu, Woldeamlak Bewket

Abstract:

The aim of this study was to understand the influence of current climate variability on crop production in the central highlands of Ethiopia. We used monthly rainfall and temperature data from 132 points each representing a pixel of 10×10 km. The data are reconstructions based on station records and meteorological satellite observations. Production data of the five major crops in the area were collected from the Central Statistical Agency for the period 2004-2013 and for the main cropping season, locally known as Meher. The production data are at the Enumeration Area (EA ) level and hence the best available dataset on crop production. The results show statistically significant decreasing trends in March–May (Belg) rainfall in the area. However, June – September (Kiremt) rainfall showed increasing trends in Efratana Gidim and Menz Gera Meder which the latter is statistically significant. Annual rainfall also showed positive trends in the area except Basona Werana where significant negative trends were observed. On the other hand, maximum and minimum temperatures showed warming trends in the study area. Correlation results have shown that crop production and area of cultivation have positive correlation with rainfall, and negative with temperature. When the trends in crop production are investigated, most crops showed negative trends and below average production was observed. Regression results have shown that rainfall was the most important determinant of crop production in the area. It is concluded that current climate variability has a significant influence on crop production in the area and any unfavorable change in the local climate in the future will have serious implications for household level food security. Efforts to adapt to the ongoing climate change should begin from tackling the current climate variability and take a climate risk management approach.

Keywords: central highlands, climate variability, crop production, Ethiopia, regression, trend

Procedia PDF Downloads 407
106 Improving Fault Tolerance and Load Balancing in Heterogeneous Grid Computing Using Fractal Transform

Authors: Saad M. Darwish, Adel A. El-Zoghabi, Moustafa F. Ashry

Abstract:

The popularity of the Internet and the availability of powerful computers and high-speed networks as low-cost commodity components are changing the way we use computers today. These technical opportunities have led to the possibility of using geographically distributed and multi-owner resources to solve large-scale problems in science, engineering, and commerce. Recent research on these topics has led to the emergence of a new paradigm known as Grid computing. To achieve the promising potentials of tremendous distributed resources, effective and efficient load balancing algorithms are fundamentally important. Unfortunately, load balancing algorithms in traditional parallel and distributed systems, which usually run on homogeneous and dedicated resources, cannot work well in the new circumstances. In this paper, the concept of a fast fractal transform in heterogeneous grid computing based on R-tree and the domain-range entropy is proposed to improve fault tolerance and load balancing algorithm by improve connectivity, communication delay, network bandwidth, resource availability, and resource unpredictability. A novel two-dimension figure of merit is suggested to describe the network effects on load balance and fault tolerance estimation. Fault tolerance is enhanced by adaptively decrease replication time and message cost while load balance is enhanced by adaptively decrease mean job response time. Experimental results show that the proposed method yields superior performance over other methods.

Keywords: Grid computing, load balancing, fault tolerance, R-tree, heterogeneous systems

Procedia PDF Downloads 459
105 Evaluation of IMERG Performance at Estimating the Rainfall Properties through Convective and Stratiform Rain Events in a Semi-Arid Region of Mexico

Authors: Eric Muñoz de la Torre, Julián González Trinidad, Efrén González Ramírez

Abstract:

Rain varies greatly in its duration, intensity, and spatial coverage, it is important to have sub-daily rainfall data for various applications, including risk prevention. However, the ground measurements are limited by the low and irregular density of rain gauges. An alternative to this problem are the Satellite Precipitation Products (SPPs) that use passive microwave and infrared sensors to estimate rainfall, as IMERG, however, these SPPs have to be validated before their application. The aim of this study is to evaluate the performance of the IMERG: Integrated Multi-satellitE Retrievals for Global Precipitation Measurament final run V06B SPP in a semi-arid region of Mexico, using 4 automatic rain gauges (pluviographs) sub-daily data of October 2019 and June to September 2021, using the Minimum inter-event Time (MIT) criterion to separate unique rain events with a dry period of 10 hrs. for the purpose of evaluating the rainfall properties (depth, duration and intensity). Point to pixel analysis, continuous, categorical, and volumetric statistical metrics were used. Results show that IMERG is capable to estimate the rainfall depth with a slight overestimation but is unable to identify the real duration and intensity of the rain events, showing large overestimations and underestimations, respectively. The study zone presented 80 to 85 % of convective rain events, the rest were stratiform rain events, classified by the depth magnitude variation of IMERG pixels and pluviographs. IMERG showed poorer performance at detecting the first ones but had a good performance at estimating stratiform rain events that are originated by Cold Fronts.

Keywords: IMERG, rainfall, rain gauge, remote sensing, statistical evaluation

Procedia PDF Downloads 39
104 Project Knowledge Harvesting: The Case of Improving Project Performance through Project Knowledge Sharing Framework

Authors: Eng Rima Al-Awadhi, Abdul Jaleel Tharayil

Abstract:

In a project-centric organization like KOC, managing the knowledge of the project is of critical importance to the success of the project and the organization. However, due to the very nature and complexity involved, each project engagement generates a lot of 'learnings' that need to be factored into while new projects are initiated and thus avoid repeating the same mistake. But, many a time these learnings are localized and remains as ‘tacit knowledge’ leading to scope re-work, schedule overrun, adjustment orders, concession requests and claims. While KOC follows an asset based organization structure, with a multi-cultural and multi-ethnic workforce and larger chunk of the work is carried out through complex, long term project engagement, diffusion of ‘learnings’ across assets while dealing with the natural entropy of the organization is of great significance. Considering the relatively higher number of mega projects, it's important that the issues raised during the project life cycle are centrally harvested, analyzed and the ‘learnings’ from these issues are shared, absorbed and are in-turn utilized to enhance and refine the existing process and practices, leading to improve the project performance. One of the many factors contributing to the successful completion of a project on time is the reduction in the number of variations or concessions triggered during the project life cycle. The project process integrated knowledge sharing framework discusses the knowledge harvesting methodology adopted, the challenges faced, learnings acquired and its impact on project performance. The framework facilitates the proactive identification of issues that may have an impact on the overall quality of the project and improve performance.

Keywords: knowledge harvesting, project integrated knowledge sharing, performance improvement, knowledge management, lessons learn

Procedia PDF Downloads 363
103 Empowering Transformers for Evidence-Based Medicine

Authors: Jinan Fiaidhi, Hashmath Shaik

Abstract:

Breaking the barrier for practicing evidence-based medicine relies on effective methods for rapidly identifying relevant evidence from the body of biomedical literature. An important challenge confronted by medical practitioners is the long time needed to browse, filter, summarize and compile information from different medical resources. Deep learning can help in solving this based on automatic question answering (Q&A) and transformers. However, Q&A and transformer technologies are not trained to answer clinical queries that can be used for evidence-based practice, nor can they respond to structured clinical questioning protocols like PICO (Patient/Problem, Intervention, Comparison and Outcome). This article describes the use of deep learning techniques for Q&A that are based on transformer models like BERT and GPT to answer PICO clinical questions that can be used for evidence-based practice extracted from sound medical research resources like PubMed. We are reporting acceptable clinical answers that are supported by findings from PubMed. Our transformer methods are reaching an acceptable state-of-the-art performance based on two staged bootstrapping processes involving filtering relevant articles followed by identifying articles that support the requested outcome expressed by the PICO question. Moreover, we are also reporting experimentations to empower our bootstrapping techniques with patch attention to the most important keywords in the clinical case and the PICO questions. Our bootstrapped patched with attention is showing relevancy of the evidence collected based on entropy metrics.

Keywords: automatic question answering, PICO questions, evidence-based medicine, generative models, LLM transformers

Procedia PDF Downloads 10
102 Analysis of Spatiotemporal Efficiency and Fairness of Railway Passenger Transport Network Based on Space Syntax: Taking Yangtze River Delta as an Example

Authors: Lin Dong, Fei Shi

Abstract:

Based on the railway network and the principles of space syntax, the study attempts to reconstruct the spatial relationship of the passenger network connections from space and time perspective. According to the travel time data of main stations in the Yangtze River Delta urban agglomeration obtained by the Internet, the topological drawing of railway network under different time sections is constructed. With the comprehensive index composed of connection and integration, the accessibility and network operation efficiency of the railway network in different time periods is calculated, while the fairness of the network is analyzed by the fairness indicators constructed with the integration and location entropy from the perspective of horizontal and vertical fairness respectively. From the analysis of the efficiency and fairness of the railway passenger transport network, the study finds: (1) There is a strong regularity in regional system accessibility change; (2) The problems of efficiency and fairness are different in different time periods; (3) The improvement of efficiency will lead to the decline of horizontal fairness to a certain extent, while from the perspective of vertical fairness, the supply-demand situation has changed smoothly with time; (4) The network connection efficiency of Shanghai, Jiangsu and Zhejiang regions is higher than that of the western regions such as Anqing and Chizhou; (5) The marginalization of Nantong, Yancheng, Yangzhou, Taizhou is obvious. The study explores the application of spatial syntactic theory in regional traffic analysis, in order to provide a reference for the development of urban agglomeration transportation network.

Keywords: spatial syntax, the Yangtze River Delta, railway passenger time, efficiency and fairness

Procedia PDF Downloads 112
101 Tool for Maxillary Sinus Quantification in Computed Tomography Exams

Authors: Guilherme Giacomini, Ana Luiza Menegatti Pavan, Allan Felipe Fattori Alves, Marcela de Oliveira, Fernando Antonio Bacchim Neto, José Ricardo de Arruda Miranda, Seizo Yamashita, Diana Rodrigues de Pina

Abstract:

The maxillary sinus (MS), part of the paranasal sinus complex, is one of the most enigmatic structures in modern humans. The literature has suggested that MSs function as olfaction accessories, to heat or humidify inspired air, for thermoregulation, to impart resonance to the voice and others. Thus, the real function of the MS is still uncertain. Furthermore, the MS anatomy is complex and varies from person to person. Many diseases may affect the development process of sinuses. The incidence of rhinosinusitis and other pathoses in the MS is comparatively high, so, volume analysis has clinical value. Providing volume values for MS could be helpful in evaluating the presence of any abnormality and could be used for treatment planning and evaluation of the outcome. The computed tomography (CT) has allowed a more exact assessment of this structure, which enables a quantitative analysis. However, this is not always possible in the clinical routine, and if possible, it involves much effort and/or time. Therefore, it is necessary to have a convenient, robust, and practical tool correlated with the MS volume, allowing clinical applicability. Nowadays, the available methods for MS segmentation are manual or semi-automatic. Additionally, manual methods present inter and intraindividual variability. Thus, the aim of this study was to develop an automatic tool to quantity the MS volume in CT scans of paranasal sinuses. This study was developed with ethical approval from the authors’ institutions and national review panels. The research involved 30 retrospective exams of University Hospital, Botucatu Medical School, São Paulo State University, Brazil. The tool for automatic MS quantification, developed in Matlab®, uses a hybrid method, combining different image processing techniques. For MS detection, the algorithm uses a Support Vector Machine (SVM), by features such as pixel value, spatial distribution, shape and others. The detected pixels are used as seed point for a region growing (RG) segmentation. Then, morphological operators are applied to reduce false-positive pixels, improving the segmentation accuracy. These steps are applied in all slices of CT exam, obtaining the MS volume. To evaluate the accuracy of the developed tool, the automatic method was compared with manual segmentation realized by an experienced radiologist. For comparison, we used Bland-Altman statistics, linear regression, and Jaccard similarity coefficient. From the statistical analyses for the comparison between both methods, the linear regression showed a strong association and low dispersion between variables. The Bland–Altman analyses showed no significant differences between the analyzed methods. The Jaccard similarity coefficient was > 0.90 in all exams. In conclusion, the developed tool to quantify MS volume proved to be robust, fast, and efficient, when compared with manual segmentation. Furthermore, it avoids the intra and inter-observer variations caused by manual and semi-automatic methods. As future work, the tool will be applied in clinical practice. Thus, it may be useful in the diagnosis and treatment determination of MS diseases. Providing volume values for MS could be helpful in evaluating the presence of any abnormality and could be used for treatment planning and evaluation of the outcome. The computed tomography (CT) has allowed a more exact assessment of this structure which enables a quantitative analysis. However, this is not always possible in the clinical routine, and if possible, it involves much effort and/or time. Therefore, it is necessary to have a convenient, robust and practical tool correlated with the MS volume, allowing clinical applicability. Nowadays, the available methods for MS segmentation are manual or semi-automatic. Additionally, manual methods present inter and intraindividual variability. Thus, the aim of this study was to develop an automatic tool to quantity the MS volume in CT scans of paranasal sinuses. This study was developed with ethical approval from the authors’ institutions and national review panels. The research involved 30 retrospective exams of University Hospital, Botucatu Medical School, São Paulo State University, Brazil. The tool for automatic MS quantification, developed in Matlab®, uses a hybrid method, combining different image processing techniques. For MS detection, the algorithm uses a Support Vector Machine (SVM), by features such as pixel value, spatial distribution, shape and others. The detected pixels are used as seed point for a region growing (RG) segmentation. Then, morphological operators are applied to reduce false-positive pixels, improving the segmentation accuracy. These steps are applied in all slices of CT exam, obtaining the MS volume. To evaluate the accuracy of the developed tool, the automatic method was compared with manual segmentation realized by an experienced radiologist. For comparison, we used Bland-Altman statistics, linear regression and Jaccard similarity coefficient. From the statistical analyses for the comparison between both methods, the linear regression showed a strong association and low dispersion between variables. The Bland–Altman analyses showed no significant differences between the analyzed methods. The Jaccard similarity coefficient was > 0.90 in all exams. In conclusion, the developed tool to automatically quantify MS volume proved to be robust, fast and efficient, when compared with manual segmentation. Furthermore, it avoids the intra and inter-observer variations caused by manual and semi-automatic methods. As future work, the tool will be applied in clinical practice. Thus, it may be useful in the diagnosis and treatment determination of MS diseases.

Keywords: maxillary sinus, support vector machine, region growing, volume quantification

Procedia PDF Downloads 483
100 Morphology and Permeability of Biomimetic Cellulose Triacetate-Impregnated Membranes: in situ Synchrotron Imaging and Experimental Studies

Authors: Amira Abdelrasoul

Abstract:

This study aimed to ascertain the controlled permeability of biomimetic cellulose triacetate (CTA) membranes by investigating the electrical oscillatory behavior across impregnated membranes (IM). The biomimetic CTA membranes were infused with a fatty acid to induce electrical oscillatory behavior and, hence, to ensure controlled permeability. In situ synchrotron radiation micro-computed tomography (SR-μCT) at the BioMedical Imaging and Therapy (BMIT) Beamline at the Canadian Light Source (CLS) was used to evaluate the main morphology of IMs compared to neat CTA membranes to ensure fatty acid impregnation inside the pores of the membrane matrices. A monochromatic beam at 20 keV was used for the visualization of the morphology of the membrane. The X-ray radiographs were recorded by means of a beam monitor AA-40 (500 μm LuAG scintillator, Hamamatsu, Japan) coupled with a high-resolution camera, providing a pixel size of 5.5 μm and a field of view (FOV) of 4.4 mm × 2.2 mm. Changes were evident in the phase transition temperatures of the impregnated CTA membrane at the melting temperature of the fatty acid. The pulsations of measured voltages were related to changes in the salt concentration of KCl in the vicinity of the electrode. Amplitudes and frequencies of voltage pulsations were dependent on the temperature and concentration of the KCl solution, which controlled the permeability of the biomimetic membranes. The presented smart biomimetic membrane successfully combined porous polymer support and impregnating liquid not only imitate the main barrier properties of the biological membranes but could be easily modified to achieve some new properties, such as facilitated and active transport, regulation by chemical, physical and pharmaceutical factors. These results open new frontiers for the facilitation and regulation of active transport and permeability through biomimetic smart membranes for a variety of biomedical and drug delivery applications.

Keywords: biomimetic, membrane, synchrotron, permeability, morphology

Procedia PDF Downloads 79
99 The Analysis of Spatial Development: Malekan City

Authors: Rahim Sarvar, Bahram Azadbakht, Samira Safaee

Abstract:

The leading goal of all planning is to attain sustainable development, regional balance, suitable distribution of activities, and maximum use of environmental capabilities in the process of development of regions. Intensive concentration of population and activities in one or some limited geographical locality is of main characteristics of most developing countries, especially Iran. Not considering the long-term programs and relying on temporary and superficial plans by people in charge of decision-making to attain their own objectives causes obstacles, resulting in unbalance development. The basic reason for these problems is to establish the development planning while economic aspects are merely considered and any attentions are not paid to social and regional feedbacks, what have been ending up to social and economic inequality, unbalanced distribution of development among the regions as well. In addition to study of special planning and structure of the county of Malekan, this research tries to achieve some other aims, i.e. recognition and introduction of approaches in order to utilize resources optimally, to distribute the population, activities, and facilities in optimum fashion, and to investigate and identify the spatial development potentials of the County. Based on documentary, descriptive, analytical, and field studies, this research employs maps to analyze the data, investigates the variables, and applies SPSS, Auto CAD, and Arc View software. The results show that the natural factors have a significant influence on spatial layout of settlements; distribution of facilities and functions are not equal among the rural districts of the county; and there is a spatial equivalence in the region area between population and number of settlements.

Keywords: development, entropy index, Malekan City, planning, regional equilibrium

Procedia PDF Downloads 414
98 Comparative Study of Sorption of Cr Ions and Dye Bezaktiv Yellow HE-4G with the Use of Adsorbents Natural Mixture of Olive Stone and Date Pits from Aqueous Solution

Authors: H. Aksas, H. Babaci, K. Louhab

Abstract:

In this paper, a comparative study of the adsorption of Chromium and dyes, onto mixture biosorbents, olive stones and date pits at different percentage was investigated in aqueous solution. The study of various parameters: Effect of contact time, pH, temperature and initial concentration shows that these materials possess a high affinity for the adsorption of chromium for the adsorption of dye bezaktiv yellow HE-4G. To deepen the comparative study of the adsorption of chromium and dye with the use of different blends of olive stones and date pits, the following models are studied: Langmuir, Freundlich isotherms and Dubinin- Radushkvich (D-R) were used as the adsorption equilibrium data model. Langmuir isotherm model was the most suitable for the adsorption of the dye bezaktiv HE-4G and the D-R model is most suitable for adsorption Chrome. The pseudo-first-order model, pseudo-second order and intraparticle diffusion were used to describe the adsorption kinetics. The apparent activation energy was found to be less than 8KJ/mol, which is characteristic of a controlled chemical reaction for the adsorption of two materials. t was noticed that adsorption of chromium and dye BEZAKTIV HE-YELLOW 4G follows the kinetics of the pseudo second order. The study of the effect of temperature was quantified by calculating various thermodynamic parameters such as Gibbs free energy, enthalpy and entropy changes. The resulting thermodynamic parameters indicate the endothermic nature of the adsorption of Cr (VI) ions and the dye Bezaktiv HE-4G. But these materials are very good adsorbents, as they represent a low cost. in addition, it has been noticed that the greater the quantity of olive stone in the mixture increases, the adsorption ability of the dye or chromium increases.

Keywords: chromium ions, anions dye, sorption, mixed adsorbents, olive stone, date pits

Procedia PDF Downloads 204
97 Physicochemical Properties and Thermal Inactivation of Polyphenol Oxidase of African Bush Mango (Irvingia Gabonensis) Fruit

Authors: Catherine Joke Adeseko

Abstract:

Enzymatic browning is an economically important disorder that degrades organoleptic properties and prevent the consumer from purchasing fresh fruit and vegetables. Prevention and control of enzymatic browning in fruit and its product is imperative. Therefore, this study sought to investigate the catalytic effect of polyphenol oxidase (PPO) in the adverse browning of African bush mango (Irvingia gabonensis) fruit peel and pulp. PPO was isolated and purified, and its physicochemical properties, such as the effect of pH with SDS, temperature, and thermodynamic studies, which invariably led to thermal inactivation of purified PPO at 80 °C, were evaluated. The pH and temperature optima of PPO were found at 7.0 and 50, respectively. There was a gradual increase in the activity of PPO as the pH increases. However, the enzyme exhibited a higher activity at neutral pH 7.0, while enzymatic inhibition was observed at acidic region, pH 2.0. The presence of SDS at pH 5.0 downward was found to inhibit the activity of PPO from the peel and pulp of I. gabonensis. The average value of enthalpy (ΔH), entropy (ΔS), and Gibbs free energy (ΔG) obtained at 20 min of incubation and temperature 30 – 80 °C were respectively 39.93 kJ.mol-1, 431.57 J.mol-1 .K-1 and -107.99 kJ.mol-1 for peel PPO, and 37.92 kJ.mol-1, -442.51J.mol-1.K-1, and -107.22 kJ.mol-1 for pulp PPO. Thermal inactivation of PPO from I. gabonensis exhibited a reduction in catalytic activity as the temperature and duration of heat inactivation increases using catechol, reflected by an increment in k value. The half-life of PPO (t1/2) decreases as the incubation temperature increases due to the instability of the enzyme at high temperatures and was higher in pulp than peel. Both D and Z values decrease with increase in temperature. The information from this study suggests processing parameters for controlling PPO in the potential industrial application of I. gabonensis fruit in order to prolong the shelf-life of this fruit for maximum utilization.

Keywords: enzymatic, browning, characterization, activity

Procedia PDF Downloads 59
96 Application of Federated Learning in the Health Care Sector for Malware Detection and Mitigation Using Software-Defined Networking Approach

Authors: A. Dinelka Panagoda, Bathiya Bandara, Chamod Wijetunga, Chathura Malinda, Lakmal Rupasinghe, Chethana Liyanapathirana

Abstract:

This research takes us forward with the concepts of Federated Learning and Software-Defined Networking (SDN) to introduce an efficient malware detection technique and provide a mitigation mechanism to give birth to a resilient and automated healthcare sector network system by also adding the feature of extended privacy preservation. Due to the daily transformation of new malware attacks on hospital Integrated Clinical Environment (ICEs), the healthcare industry is at an undefinable peak of never knowing its continuity direction. The state of blindness by the array of indispensable opportunities that new medical device inventions and their connected coordination offer daily, a factor that should be focused driven is not yet entirely understood by most healthcare operators and patients. This solution has the involvement of four clients in the form of hospital networks to build up the federated learning experimentation architectural structure with different geographical participation to reach the most reasonable accuracy rate with privacy preservation. While the logistic regression with cross-entropy conveys the detection, SDN comes in handy in the second half of the research to stack up the initial development phases of the system with malware mitigation based on policy implementation. The overall evaluation sums up with a system that proves the accuracy with the added privacy. It is no longer needed to continue with traditional centralized systems that offer almost everything but not privacy.

Keywords: software-defined network, federated learning, privacy, integrated clinical environment, decentralized learning, malware detection, malware mitigation

Procedia PDF Downloads 148
95 Optimizing Detection Methods for THz Bio-imaging Applications

Authors: C. Bolakis, I. S. Karanasiou, D. Grbovic, G. Karunasiri, N. Uzunoglu

Abstract:

A new approach for efficient detection of THz radiation in biomedical imaging applications is proposed. A double-layered absorber consisting of a 32 nm thick aluminum (Al) metallic layer, located on a glass medium (SiO2) of 1 mm thickness, was fabricated and used to design a fine-tuned absorber through a theoretical and finite element modeling process. The results indicate that the proposed low-cost, double-layered absorber can be tuned based on the metal layer sheet resistance and the thickness of various glass media taking advantage of the diversity of the absorption of the metal films in the desired THz domain (6 to 10 THz). It was found that the composite absorber could absorb up to 86% (a percentage exceeding the 50%, previously shown to be the highest achievable when using single thin metal layer) and reflect less than 1% of the incident THz power. This approach will enable monitoring of the transmission coefficient (THz transmission ‘’fingerprint’’) of the biosample with high accuracy, while also making the proposed double-layered absorber a good candidate for a microbolometer pixel’s active element. Based on the aforementioned promising results, a more sophisticated and effective double-layered absorber is under development. The glass medium has been substituted by diluted poly-si and the results were twofold: An absorption factor of 96% was reached and high TCR properties acquired. In addition, a generalization of these results and properties over the active frequency spectrum was achieved. Specifically, through the development of a theoretical equation having as input any arbitrary frequency in the IR spectrum (0.3 to 405.4 THz) and as output the appropriate thickness of the poly-si medium, the double-layered absorber retains the ability to absorb the 96% and reflects less than 1% of the incident power. As a result, through that post-optimization process and the spread spectrum frequency adjustment, the microbolometer detector efficiency could be further improved.

Keywords: bio-imaging, fine-tuned absorber, fingerprint, microbolometer

Procedia PDF Downloads 321
94 Synthesis and Physiochemical Properties of 3-Propanenitrile Imidazolium - Based Dual Functionalized Ionic Liquids Incorporating Dioctyl Sulfosuccinate Anion

Authors: Abobakr Khidir Ziyada, Cecilia Devi Wilfred

Abstract:

In the present work, a new series of 3-propanenitrile imidazolium-based Room Temperature Ionic Liquids (RTILs), incorporating dioctyl sulfosuccinate (DOSS) were prepared by reacting imidazole with acrylonitrile and then reacting the product with allyl chloride, 2-chloroethanol, and benzyl chloride. After the reaction had been completed, metathesis reaction was carried out using sodium dioctyl sulfosuccinate. The densities and viscosities of the present RTILs were measured at atmospheric pressure at T=293.15 to 353.15 K, the refractive index was measured at T=293.15 to 333.15 K, whereas, the start and decomposition temperatures were determined at heating rate 10°C. min^-1. The thermal expansion coefficient, densities at a range of temperatures and pressures, molecular volume, molar refraction, standard entropy and the lattice energy of these RTILs were also estimated. The present RTILs showed higher densities, similar refractive indices, and higher viscosities compared to the other 1-alkyl-3-propanenitrile imidazolium-based RTILs. The densities of the present synthesized RTILs are lower compared to the other nitrile-functionalized ILs. These present RTILs showed a weak temperature dependence on the thermal expansion coefficients, αp=5.0 × 10^−4 to 7.50 × 10−4 K^-1. Empirical correlations were proposed to represent the present data on the physical properties. The lattice energy for the present RTILs was similar to other nitrile–based imidazolium RTILs. The present RTILs showed very high molar refraction when compared similar RTILs incorporating other anions.

Keywords: dioctyl sulfosuccinate, nitrile ILs, 3-propanenitrile, anion, room temperature ionic liquids, RTIL

Procedia PDF Downloads 312
93 Derivation of Bathymetry Data Using Worldview-2 Multispectral Images in Shallow, Turbid and Saline Lake Acıgöl

Authors: Muhittin Karaman, Murat Budakoglu

Abstract:

In this study, derivation of lake bathymetry was evaluated using the high resolution Worldview-2 multispectral images in the very shallow hypersaline Lake Acıgöl which does not have a stable water table due to the wet-dry season changes and industrial usage. Every year, a great part of the lake water budget has been consumed for the industrial salt production in the evaporation ponds, which are generally located on the south and north shores of Lake Acıgöl. Therefore, determination of the water level changes from a perspective of remote sensing-based lake water by bathymetry studies has a great importance in the sustainability-control of the lake. While the water table interval is around 1 meter between dry and wet season, dissolved ion concentration, salinity and turbidity also show clear differences during these two distinct seasonal periods. At the same time, with the satellite data acquisition (June 9, 2013), a field study was conducted to collect the salinity values, Secchi disk depths and turbidity levels. Max depth, Secchi disk depth and salinity were determined as 1,7 m, 0,9 m and 43,11 ppt, respectively. Eight-band Worldview-2 image was corrected for atmospheric effects by ATCOR technique. For each sampling point in the image, mean reflectance values in 1*1, 3*3, 5*5, 7*7, 9*9, 11*11, 13*13, 15*15, 17*17, 19*19, 21*21, 51*51 pixel reflectance neighborhoods were calculated separately. A unique image has been derivated for each matrix resolution. Spectral values and depth relation were evaluated for these distinct resolution images. Correlation coefficients were determined for the 1x1 matrix: 0,98, 0,96, 0,95 and 0,90 for the 724 nm, 831 nm, 908 nm and 659 nm, respectively. While 15x5 matrix characteristics with 0,98, 0,97 and 0,97 correlation values for the 724 nm, 908 nm and 831 nm, respectively; 51x51 matrix shows 0,98, 0,97 and 0,96 correlation values for the 724 nm, 831 nm and 659 nm, respectively. Comparison of all matrix resolutions indicates that RedEdge band (724 nm) of the Worldview-2 satellite image has the best correlation with the saline shallow lake of Acıgöl in-situ depth.

Keywords: bathymetry, Worldview-2 satellite image, ATCOR technique, Lake Acıgöl, Denizli, Turkey

Procedia PDF Downloads 407
92 Mathematical Analysis of Variation in Inlet Shock Wave Angle on Specific Impulse of Scramjet Engine

Authors: Shrikant Ghadage

Abstract:

Study of shock waves generated in the Scramjet engine is typically restricted to pressure, temperature, density, entropy and Mach number variation across the shock wave. The present work discusses the impact of inlet shock wave angles on the specific impulse of the Scramjet engine. A mathematical analysis has done for the isentropic hypersonic flow of air flowing through a Scramjet with hydrogen fuel at an altitude of 30 km. Analysis has been done in order to get optimum shock wave angle to achieve maximum impulse. Since external drag has excluded from the analysis, the losses due to friction are not considered for the present analysis. When Mach number of the airflow at the entry of the nozzle reaches unity, then that flow is choked. This condition puts limitations on increasing the inlet shock wave angle. As inlet shock wave angle increases, speed of the flow entering into the nozzle decreases, which results in an increase in the specific impulse of the engine. When the speed of the flow at the entry of the nozzle reduces below sonic speed, then there is no further increase in the specific impulse of the engine. Here the Conclusion is the thrust and specific impulse of a scramjet engine, which increases gradually with an increase in inlet shock wave angle up to the condition when airflow speed reaches sonic velocity at the exit of the combustor. In addition to that, variation in drag force at the inlet of the scramjet and variation in hypersonic flow conditions at every stage of the scramjet also studied in order to understand variation on flow characteristics with respect to flow deflection angle. Essentially, it helps in designing inlet profile for the Scramjet engine to achieve optimum specific impulse.

Keywords: hypersonic flow, scramjet, shock waves, specific impulse, mathematical analysis

Procedia PDF Downloads 139
91 The Evolution of Spatio-Temporal Patterns of New-Type Urbanization in the Central Plains Economic Region in China

Authors: Sun fang, Zhang Wenxin

Abstract:

This paper establishes an evaluation index system for spatio-temporal patterns of urbanization, with the county as research unit. We use the Entropy Weight method, coefficient variance, the Theil index and ESDA-GIS to analyze spatial patterns and evolutionary characteristics of New-Type Urbanization in the Central Plains Economic Region (CPER) between 2000 and 2011. Results show that economic benefit, non-agricultural employment level and level of market development are the most important factors influencing the level of New-Type Urbanization in the CPER; overall regional differences in New-Type Urbanization have declined while spatial correlations have increased from 2000 to 2011. The overall spatial pattern has changed little, however; differences between the western and eastern areas of the CPER are clear, and the pattern of a strong west and weak east did not change significantly over the study period. Areas with high levels of New-Type Urbanization were mostly distributed along the Beijing-Guangzhou and LongHai Railways on both sides, a new influx of urbanization was tightly clustered around ZhengZhou in the Central Henan Urban Agglomeration, but this trend was found to be weakening slightly. The level of New-Type Urbanization in municipal districts was found to be much higher than it was in the county generally. Provincial borders experienced a lower rate of growth and a lower level of New-Type Urbanization than did any other areas, consistently forming clusters of cold spots and sub-cold spots. The analysis confirms that historical development, location, and diffusion effects of urban agglomeration are the main drivers of changes in New-Type Urbanization patterns in CPER.

Keywords: new-type urbanization, spatial pattern, central plains economic region, spatial evolution

Procedia PDF Downloads 269
90 Applications of Hyperspectral Remote Sensing: A Commercial Perspective

Authors: Tuba Zahra, Aakash Parekh

Abstract:

Hyperspectral remote sensing refers to imaging of objects or materials in narrow conspicuous spectral bands. Hyperspectral images (HSI) enable the extraction of spectral signatures for objects or materials observed. These images contain information about the reflectance of each pixel across the electromagnetic spectrum. It enables the acquisition of data simultaneously in hundreds of spectral bands with narrow bandwidths and can provide detailed contiguous spectral curves that traditional multispectral sensors cannot offer. The contiguous, narrow bandwidth of hyperspectral data facilitates the detailed surveying of Earth's surface features. This would otherwise not be possible with the relatively coarse bandwidths acquired by other types of imaging sensors. Hyperspectral imaging provides significantly higher spectral and spatial resolution. There are several use cases that represent the commercial applications of hyperspectral remote sensing. Each use case represents just one of the ways that hyperspectral satellite imagery can support operational efficiency in the respective vertical. There are some use cases that are specific to VNIR bands, while others are specific to SWIR bands. This paper discusses the different commercially viable use cases that are significant for HSI application areas, such as agriculture, mining, oil and gas, defense, environment, and climate, to name a few. Theoretically, there is n number of use cases for each of the application areas, but an attempt has been made to streamline the use cases depending upon economic feasibility and commercial viability and present a review of literature from this perspective. Some of the specific use cases with respect to agriculture are crop species (sub variety) detection, soil health mapping, pre-symptomatic crop disease detection, invasive species detection, crop condition optimization, yield estimation, and supply chain monitoring at scale. Similarly, each of the industry verticals has a specific commercially viable use case that is discussed in the paper in detail.

Keywords: agriculture, mining, oil and gas, defense, environment and climate, hyperspectral, VNIR, SWIR

Procedia PDF Downloads 47
89 Image-Based UAV Vertical Distance and Velocity Estimation Algorithm during the Vertical Landing Phase Using Low-Resolution Images

Authors: Seyed-Yaser Nabavi-Chashmi, Davood Asadi, Karim Ahmadi, Eren Demir

Abstract:

The landing phase of a UAV is very critical as there are many uncertainties in this phase, which can easily entail a hard landing or even a crash. In this paper, the estimation of relative distance and velocity to the ground, as one of the most important processes during the landing phase, is studied. Using accurate measurement sensors as an alternative approach can be very expensive for sensors like LIDAR, or with a limited operational range, for sensors like ultrasonic sensors. Additionally, absolute positioning systems like GPS or IMU cannot provide distance to the ground independently. The focus of this paper is to determine whether we can measure the relative distance and velocity of UAV and ground in the landing phase using just low-resolution images taken by a monocular camera. The Lucas-Konda feature detection technique is employed to extract the most suitable feature in a series of images taken during the UAV landing. Two different approaches based on Extended Kalman Filters (EKF) have been proposed, and their performance in estimation of the relative distance and velocity are compared. The first approach uses the kinematics of the UAV as the process and the calculated optical flow as the measurement; On the other hand, the second approach uses the feature’s projection on the camera plane (pixel position) as the measurement while employing both the kinematics of the UAV and the dynamics of variation of projected point as the process to estimate both relative distance and relative velocity. To verify the results, a sequence of low-quality images taken by a camera that is moving on a specifically developed testbed has been used to compare the performance of the proposed algorithm. The case studies show that the quality of images results in considerable noise, which reduces the performance of the first approach. On the other hand, using the projected feature position is much less sensitive to the noise and estimates the distance and velocity with relatively high accuracy. This approach also can be used to predict the future projected feature position, which can drastically decrease the computational workload, as an important criterion for real-time applications.

Keywords: altitude estimation, drone, image processing, trajectory planning

Procedia PDF Downloads 88
88 Conservation Planning of Paris Polyphylla Smith, an Important Medicinal Herb of the Indian Himalayan Region Using Predictive Distribution Modelling

Authors: Mohd Tariq, Shyamal K. Nandi, Indra D. Bhatt

Abstract:

Paris polyphylla Smith (Family- Liliaceae; English name-Love apple: Local name- Satuwa) is an important folk medicinal herb of the Indian subcontinent, being a source of number of bioactive compounds for drug formulation. The rhizomes are widely used as antihelmintic, antispasmodic, digestive stomachic, expectorant and vermifuge, antimicrobial, anti-inflammatory, heart and vascular malady, anti-fertility and sedative. Keeping in view of this, the species is being constantly removed from nature for trade and various pharmaceuticals purpose, as a result, the availability of the species in its natural habitat is decreasing. In this context, it would be pertinent to conserve this species and reintroduce them in its natural habitat. Predictive distribution modelling of this species was performed in Western Himalayan Region. One such recent method is Ecological Niche Modelling, also popularly known as Species distribution modelling, which uses computer algorithms to generate predictive maps of species distributions in a geographic space by correlating the point distributional data with a set of environmental raster data. In case of P. polyphylla, and to understand its potential distribution zones and setting up of artificial introductions, or selecting conservation sites, and conservation and management of their native habitat. Among the different districts of Uttarakhand (28°05ˈ-31°25ˈ N and 77°45ˈ-81°45ˈ E) Uttarkashi, Rudraprayag, Chamoli, Pauri Garhwal and some parts of Bageshwar, 'Maximum Entropy' (Maxent) has predicted wider potential distribution of P. polyphylla Smith. Distribution of P. polyphylla is mainly governed by Precipitation of Driest Quarter and Mean Diurnal Range i.e., 27.08% and 18.99% respectively which indicates that humidity (27%) and average temperature (19°C) might be suitable for better growth of Paris polyphylla.

Keywords: biodiversity conservation, Indian Himalayan region, Paris polyphylla, predictive distribution modelling

Procedia PDF Downloads 305
87 Exergy Analysis of a Vapor Absorption Refrigeration System Using Carbon Dioxide as Refrigerant

Authors: Samsher Gautam, Apoorva Roy, Bhuvan Aggarwal

Abstract:

Vapor absorption refrigeration systems can replace vapor compression systems in many applications as they can operate on a low-grade heat source and are environment-friendly. Widely used refrigerants such as CFCs and HFCs cause significant global warming. Natural refrigerants can be an alternative to them, among which carbon dioxide is promising for use in automotive air conditioning systems. Its inherent safety, ability to withstand high pressure and high heat transfer coefficient coupled with easy availability make it a likely choice for refrigerant. Various properties of the ionic liquid [bmim][PF₆], such as non-toxicity, stability over a wide temperature range and ability to dissolve gases like carbon dioxide, make it a suitable absorbent for a vapor absorption refrigeration system. In this paper, an absorption chiller consisting of a generator, condenser, evaporator and absorber was studied at an operating temperature of 70⁰C. A thermodynamic model was set up using the Peng-Robinson equations of state to predict the behavior of the refrigerant and absorbent pair at different points in the system. A MATLAB code was used to obtain the values of enthalpy and entropy at selected points in the system. The exergy destruction in each component and exergetic coefficient of performance (ECOP) of the system were calculated by performing an exergy analysis based on the second law of thermodynamics. Graphs were plotted between varying operating conditions and the ECOP obtained in each case. The effect of every component on the ECOP was examined. The exergetic coefficient of performance was found to be lesser than the coefficient of performance based on the first law of thermodynamics.

Keywords: [bmim][PF₆] as absorbent, carbon dioxide as refrigerant, exergy analysis, Peng-Robinson equations of state, vapor absorption refrigeration

Procedia PDF Downloads 257
86 Loss Function Optimization for CNN-Based Fingerprint Anti-Spoofing

Authors: Yehjune Heo

Abstract:

As biometric systems become widely deployed, the security of identification systems can be easily attacked by various spoof materials. This paper contributes to finding a reliable and practical anti-spoofing method using Convolutional Neural Networks (CNNs) based on the types of loss functions and optimizers. The types of CNNs used in this paper include AlexNet, VGGNet, and ResNet. By using various loss functions including Cross-Entropy, Center Loss, Cosine Proximity, and Hinge Loss, and various loss optimizers which include Adam, SGD, RMSProp, Adadelta, Adagrad, and Nadam, we obtained significant performance changes. We realize that choosing the correct loss function for each model is crucial since different loss functions lead to different errors on the same evaluation. By using a subset of the Livdet 2017 database, we validate our approach to compare the generalization power. It is important to note that we use a subset of LiveDet and the database is the same across all training and testing for each model. This way, we can compare the performance, in terms of generalization, for the unseen data across all different models. The best CNN (AlexNet) with the appropriate loss function and optimizers result in more than 3% of performance gain over the other CNN models with the default loss function and optimizer. In addition to the highest generalization performance, this paper also contains the models with high accuracy associated with parameters and mean average error rates to find the model that consumes the least memory and computation time for training and testing. Although AlexNet has less complexity over other CNN models, it is proven to be very efficient. For practical anti-spoofing systems, the deployed version should use a small amount of memory and should run very fast with high anti-spoofing performance. For our deployed version on smartphones, additional processing steps, such as quantization and pruning algorithms, have been applied in our final model.

Keywords: anti-spoofing, CNN, fingerprint recognition, loss function, optimizer

Procedia PDF Downloads 109