Search results for: realistic images
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 2877

Search results for: realistic images

1587 Probabilistic Damage Tolerance Methodology for Solid Fan Blades and Discs

Authors: Andrej Golowin, Viktor Denk, Axel Riepe

Abstract:

Solid fan blades and discs in aero engines are subjected to high combined low and high cycle fatigue loads especially around the contact areas between blade and disc. Therefore, special coatings (e.g. dry film lubricant) and surface treatments (e.g. shot peening or laser shock peening) are applied to increase the strength with respect to combined cyclic fatigue and fretting fatigue, but also to improve damage tolerance capability. The traditional deterministic damage tolerance assessment based on fracture mechanics analysis, which treats service damage as an initial crack, often gives overly conservative results especially in the presence of vibratory stresses. A probabilistic damage tolerance methodology using crack initiation data has been developed for fan discs exposed to relatively high vibratory stresses in cross- and tail-wind conditions at certain resonance speeds for limited time periods. This Monte-Carlo based method uses a damage databank from similar designs, measured vibration levels at typical aircraft operations and wind conditions and experimental crack initiation data derived from testing of artificially damaged specimens with representative surface treatment under combined fatigue conditions. The proposed methodology leads to a more realistic prediction of the minimum damage tolerance life for the most critical locations applicable to modern fan disc designs.

Keywords: combined fatigue, damage tolerance, engine, surface treatment

Procedia PDF Downloads 467
1586 Automatic Segmentation of Lung Pleura Based On Curvature Analysis

Authors: Sasidhar B., Bhaskar Rao N., Ramesh Babu D. R., Ravi Shankar M.

Abstract:

Segmentation of lung pleura is a preprocessing step in Computer-Aided Diagnosis (CAD) which helps in reducing false positives in detection of lung cancer. The existing methods fail in extraction of lung regions with the nodules at the pleura of the lungs. In this paper, a new method is proposed which segments lung regions with nodules at the pleura of the lungs based on curvature analysis and morphological operators. The proposed algorithm is tested on 06 patient’s dataset which consists of 60 images of Lung Image Database Consortium (LIDC) and the results are found to be satisfactory with 98.3% average overlap measure (AΩ).

Keywords: curvature analysis, image segmentation, morphological operators, thresholding

Procedia PDF Downloads 582
1585 A Non-Linear Eddy Viscosity Model for Turbulent Natural Convection in Geophysical Flows

Authors: J. P. Panda, K. Sasmal, H. V. Warrior

Abstract:

Eddy viscosity models in turbulence modeling can be mainly classified as linear and nonlinear models. Linear formulations are simple and require less computational resources but have the disadvantage that they cannot predict actual flow pattern in complex geophysical flows where streamline curvature and swirling motion are predominant. A constitutive equation of Reynolds stress anisotropy is adopted for the formulation of eddy viscosity including all the possible higher order terms quadratic in the mean velocity gradients, and a simplified model is developed for actual oceanic flows where only the vertical velocity gradients are important. The new model is incorporated into the one dimensional General Ocean Turbulence Model (GOTM). Two realistic oceanic test cases (OWS Papa and FLEX' 76) have been investigated. The new model predictions match well with the observational data and are better in comparison to the predictions of the two equation k-epsilon model. The proposed model can be easily incorporated in the three dimensional Princeton Ocean Model (POM) to simulate a wide range of oceanic processes. Practically, this model can be implemented in the coastal regions where trasverse shear induces higher vorticity, and for prediction of flow in estuaries and lakes, where depth is comparatively less. The model predictions of marine turbulence and other related data (e.g. Sea surface temperature, Surface heat flux and vertical temperature profile) can be utilized in short term ocean and climate forecasting and warning systems.

Keywords: Eddy viscosity, turbulence modeling, GOTM, CFD

Procedia PDF Downloads 183
1584 Biimodal Biometrics System Using Fusion of Iris and Fingerprint

Authors: Attallah Bilal, Hendel Fatiha

Abstract:

This paper proposes the bimodal biometrics system for identity verification iris and fingerprint, at matching score level architecture using weighted sum of score technique. The features are extracted from the pre processed images of iris and fingerprint. These features of a query image are compared with those of a database image to obtain matching scores. The individual scores generated after matching are passed to the fusion module. This module consists of three major steps i.e., normalization, generation of similarity score and fusion of weighted scores. The final score is then used to declare the person as genuine or an impostor. The system is tested on CASIA database and gives an overall accuracy of 91.04% with FAR of 2.58% and FRR of 8.34%.

Keywords: iris, fingerprint, sum rule, fusion

Procedia PDF Downloads 347
1583 Effect of Project Control Practices on the Performance of Building Construction Companies in Uganda: A Case Study of Kampala City

Authors: Tukundane Hillary

Abstract:

This research paper analytically evaluates the project control practice levels used by the building construction companies within Kampala, Uganda. The research also assesses the outcome of project control practices on the productivity of the companies. The research was performed to ascertain the current control practices among 160 respondents from various construction companies registered with the Uganda Registration Services Bureau. This research used amalgamation from multiple literature to obtain the variables. The research adopts 34 standard control practices from four vital project control duties: planning, monitoring, analyzing, and reporting. These project control tasks were organized using mean response ratings grounded on their relevance to the construction companies. Results showed that evaluating performance with the use of curves (4.32), timely access to information and encouragement (4.55), report representation using quantitative tools 4.75, and cost value comparison application during analysis (4.76) were rated least among the control practices. On the other hand, the top project control practices included formulation of the project schedule (8.88), Project feasibility validation (8.86), Budgeting for each activity (8.84), Key project route definition (8.81), Team awareness of the budget (8.77), Setting realistic targets for projects (8.50) and Consultation from subcontractors (8.74). From the results obtained by the sample respondents specified, it can be concluded that planning is the most vital project control task practiced in the building construction industry in Uganda. In addition, this research ascertained a substantial relationship between project control practices and the performance of building construction companies. Accordingly, this research recommends that project control practices be effectively observed by both contracting and consulting companies to enhance their overall performance and governance.

Keywords: cost value, project control, cost control, time control, project performance, control practices

Procedia PDF Downloads 49
1582 Sustainable Transboundary Water Management: Challenges and Good Practices of Cooperation in International River Basin Districts

Authors: Aleksandra Ibragimow, Moritz Albrecht, Eerika Albrecht

Abstract:

Close international cooperation between all countries within a river basin has become one of the key aspects of sustainable cross-border water management. This is due to the fact that water does not stop at administrative or political boundaries. Therefore, the preferred mode to protect and manage transnational water bodies is close cooperation between all countries and stakeholders within the natural hydrological unit of the river basin. However, past practices have demonstrated that combining interests of different countries and stakeholders with differing political systems and management approaches to environmental issues upstream as well as downstream can be challenging. The study focuses on particular problems and challenges of water management in international river basin districts by the example of the International Oder River Basin District. The Oder River is one of the largest cross-border rivers of the Baltic Sea basin passing through Poland, Germany, and the Czech Republic. Attention is directed towards the activities and the actions that were carried out during the Districts' first management cycle of transnational river basin management (2009-2015). The results show that actions of individual countries have been focused on the National Water Management Plans while a common appointment about identified supra-regional water management problems has not been solved, and conducted actions can be considered as preliminary and merely a basis for future management. This present state raises the question whether the achievement of main objectives of Water Framework Directive (2000/60/EC) can be a realistic task.

Keywords: International River Basin Districts, water management, water frameworkdirective, water management plans

Procedia PDF Downloads 299
1581 A Dynamic Cardiac Single Photon Emission Computer Tomography Using Conventional Gamma Camera to Estimate Coronary Flow Reserve

Authors: Maria Sciammarella, Uttam M. Shrestha, Youngho Seo, Grant T. Gullberg, Elias H. Botvinick

Abstract:

Background: Myocardial perfusion imaging (MPI) is typically performed with static imaging protocols and visually assessed for perfusion defects based on the relative intensity distribution. Dynamic cardiac SPECT, on the other hand, is a new imaging technique that is based on time varying information of radiotracer distribution, which permits quantification of myocardial blood flow (MBF). In this abstract, we report a progress and current status of dynamic cardiac SPECT using conventional gamma camera (Infinia Hawkeye 4, GE Healthcare) for estimation of myocardial blood flow and coronary flow reserve. Methods: A group of patients who had high risk of coronary artery disease was enrolled to evaluate our methodology. A low-dose/high-dose rest/pharmacologic-induced-stress protocol was implemented. A standard rest and a standard stress radionuclide dose of ⁹⁹ᵐTc-tetrofosmin (140 keV) was administered. The dynamic SPECT data for each patient were reconstructed using the standard 4-dimensional maximum likelihood expectation maximization (ML-EM) algorithm. Acquired data were used to estimate the myocardial blood flow (MBF). The correspondence between flow values in the main coronary vasculature with myocardial segments defined by the standardized myocardial segmentation and nomenclature were derived. The coronary flow reserve, CFR, was defined as the ratio of stress to rest MBF values. CFR values estimated with SPECT were also validated with dynamic PET. Results: The range of territorial MBF in LAD, RCA, and LCX was 0.44 ml/min/g to 3.81 ml/min/g. The MBF between estimated with PET and SPECT in the group of independent cohort of 7 patients showed statistically significant correlation, r = 0.71 (p < 0.001). But the corresponding CFR correlation was moderate r = 0.39 yet statistically significant (p = 0.037). The mean stress MBF value was significantly lower for angiographically abnormal than that for the normal (Normal Mean MBF = 2.49 ± 0.61, Abnormal Mean MBF = 1.43 ± 0. 0.62, P < .001). Conclusions: The visually assessed image findings in clinical SPECT are subjective, and may not reflect direct physiologic measures of coronary lesion. The MBF and CFR measured with dynamic SPECT are fully objective and available only with the data generated from the dynamic SPECT method. A quantitative approach such as measuring CFR using dynamic SPECT imaging is a better mode of diagnosing CAD than visual assessment of stress and rest images from static SPECT images Coronary Flow Reserve.

Keywords: dynamic SPECT, clinical SPECT/CT, selective coronary angiograph, ⁹⁹ᵐTc-Tetrofosmin

Procedia PDF Downloads 138
1580 Developing Value Chain of Synthetic Methane for Net-zero Carbon City Gas Supply in Japan

Authors: Ryota Kuzuki, Mitsuhiro Kohara, Noboru Kizuki, Satoshi Yoshida, Hidetaka Hirai, Yuta Nezasa

Abstract:

About fifty years have passed since Japan's gas supply industry became the first in the world to switch from coal and oil to LNG as a city gas feedstock. Since the Japanese government target of net-zero carbon emission in 2050 was announced in October 2020, it has now entered a new era of challenges to commit to the requirement for decarbonization. This paper describes the situation that synthetic methane, produced from renewable energy-derived hydrogen and recycled carbon, is a promising national policy of transition toward net-zero society. In November 2020, the Japan Gas Association announced the 'Carbon Neutral Challenge 2050' as a vision to contribute to the decarbonization of society by converting the city gas supply to carbon neutral. The key technologies is methanation. This paper shows that methanation is a realistic solution to contribute to the decarbonization of the whole country at a lower social cost, utilizing the supply chain that already exists, from LNG plants to burner chips. The challenges during the transition period (2030-2050), as CO2 captured from exhaust of thermal power plants and industrial factories are expected to be used, it is proposed that a system of guarantee of origin (GO) for H2 and CO2 should be established and harmonize international rules for calculating and allocating greenhouse gas emissions in the supply chain, a platform is also needed to manage tracking information on certified environmental values.

Keywords: synthetic methane, recycled carbon fuels, methanation, transition period, environmental value transfer platform

Procedia PDF Downloads 92
1579 Top-Down Influences to Multistable Perception: Evidence from Temporal Dynamics

Authors: Daria N. Podvigina, Tatiana V. Chernigovskaya

Abstract:

We have studied the temporal characteristics of bistable perception of the stimuli of two types: one involves alterations in a perceived depth and another one has an ambiguous content. We used the Necker lattice and lines of shadowed circles ambiguously perceived either as spheres or holes as stimuli of the first type. The Winson figure (the Eskimo/Indian picture) was a stimulus of the second type. We have analyzed how often the reversals occurred (reversal rate) and for how long each of the two interpretations, or percepts, was observed during one presentation (stability durations). For all three ambiguous images the reversal rate and the stability durations had similar values, which provide another evidence for a significant role of top-down processes in multistable perception.

Keywords: multistable perception, perceived depth, reversal rate, top-down processes

Procedia PDF Downloads 568
1578 An Assessment of Different Blade Tip Timing (BTT) Algorithms Using an Experimentally Validated Finite Element Model Simulator

Authors: Mohamed Mohamed, Philip Bonello, Peter Russhard

Abstract:

Blade Tip Timing (BTT) is a technology concerned with the estimation of both frequency and amplitude of rotating blades. A BTT system comprises two main parts: (a) the arrival time measurement system, and (b) the analysis algorithms. Simulators play an important role in the development of the analysis algorithms since they generate blade tip displacement data from the simulated blade vibration under controlled conditions. This enables an assessment of the performance of the different algorithms with respect to their ability to accurately reproduce the original simulated vibration. Such an assessment is usually not possible with real engine data since there is no practical alternative to BTT for blade vibration measurement. Most simulators used in the literature are based on a simple spring-mass-damper model to determine the vibration. In this work, a more realistic experimentally validated simulator based on the Finite Element (FE) model of a bladed disc (blisk) is first presented. It is then used to generate the necessary data for the assessment of different BTT algorithms. The FE modelling is validated using both a hammer test and two firewire cameras for the mode shapes. A number of autoregressive methods, fitting methods and state-of-the-art inverse methods (i.e. Russhard) are compared. All methods are compared with respect to both synchronous and asynchronous excitations with both single and simultaneous frequencies. The study assesses the applicability of each method for different conditions of vibration, amount of sampling data, and testing facilities, according to its performance and efficiency under these conditions.

Keywords: blade tip timing, blisk, finite element, vibration measurement

Procedia PDF Downloads 295
1577 Power-Sharing Politics: A Panacea to Conflict Resolution and Stability in Africa

Authors: Emmanuel Dangana Monday

Abstract:

Africa as a continent has been ravaged and bedeviled by series of political conflicts associated with politics and power-sharing maneuvering. As a result it has become the most unstable continent in the world in terms of power distribution and stable political culture. This paper examines the efficacy of conscious and deliberate power-sharing strategies to settle or resolve political conflicts in Africa in the arrangements of creation of states, revenue and resources allocation, and office distribution systems. The study is concerned with the spatial impact of conflicts generated in some renowned African countries in which power-sharing would have been a solution. Ethno-regional elite groups are identified as the major actors in the struggles for the distribution of territorial, economic and political powers in Africa. The struggle for power has become so intense that it has degenerated to conflicts and wars of inter and intra-political classes and parties respectively. Secondary data and deductive techniques were used in data collection and analysis. It is discovered that power-sharing has become an indispensable tool to curb the incessant political and power crisis in Africa. It is one of the finest tolerable modality of mediating elite’ competition, since it reflects the interests of both the dominant and the perceived marginalized groups. The study recommends that countries and regions of political, ethnic and religious differences in Africa should employed power-sharing strategy in order to avoid unnecessary political tension and resultant crisis. Interest groups should always come to the negotiation table to reach a realistic, durable and expected compromise to secure a peacefully resolute Africa.

Keywords: Africa, power-sharing, conflicts, politics and political stability

Procedia PDF Downloads 311
1576 Extraction of Urban Building Damage Using Spectral, Height and Corner Information

Authors: X. Wang

Abstract:

Timely and accurate information on urban building damage caused by earthquake is important basis for disaster assessment and emergency relief. Very high resolution (VHR) remotely sensed imagery containing abundant fine-scale information offers a large quantity of data for detecting and assessing urban building damage in the aftermath of earthquake disasters. However, the accuracy obtained using spectral features alone is comparatively low, since building damage, intact buildings and pavements are spectrally similar. Therefore, it is of great significance to detect urban building damage effectively using multi-source data. Considering that in general height or geometric structure of buildings change dramatically in the devastated areas, a novel multi-stage urban building damage detection method, using bi-temporal spectral, height and corner information, was proposed in this study. The pre-event height information was generated using stereo VHR images acquired from two different satellites, while the post-event height information was produced from airborne LiDAR data. The corner information was extracted from pre- and post-event panchromatic images. The proposed method can be summarized as follows. To reduce the classification errors caused by spectral similarity and errors in extracting height information, ground surface, shadows, and vegetation were first extracted using the post-event VHR image and height data and were masked out. Two different types of building damage were then extracted from the remaining areas: the height difference between pre- and post-event was used for detecting building damage showing significant height change; the difference in the density of corners between pre- and post-event was used for extracting building damage showing drastic change in geometric structure. The initial building damage result was generated by combining above two building damage results. Finally, a post-processing procedure was adopted to refine the obtained initial result. The proposed method was quantitatively evaluated and compared to two existing methods in Port au Prince, Haiti, which was heavily hit by an earthquake in January 2010, using pre-event GeoEye-1 image, pre-event WorldView-2 image, post-event QuickBird image and post-event LiDAR data. The results showed that the method proposed in this study significantly outperformed the two comparative methods in terms of urban building damage extraction accuracy. The proposed method provides a fast and reliable method to detect urban building collapse, which is also applicable to relevant applications.

Keywords: building damage, corner, earthquake, height, very high resolution (VHR)

Procedia PDF Downloads 201
1575 Preparing a Library of Abnormal Masses for Designing a Long-Lasting Anatomical Breast Phantom for Ultrasonography Training

Authors: Nasibullina A., Leonov D.

Abstract:

The ultrasonography method is actively used for the early diagnosis of various le-sions in the human body, including the mammary gland. The incidence of breast cancer has increased by more than 20%, and mortality by 14% since 2008. The correctness of the diagnosis often directly depends on the qualifications and expe-rience of a diagnostic medical sonographer. That is why special attention should be paid to the practical training of future specialists. Anatomical phantoms are ex-cellent teaching tools because they accurately imitate the characteristics of real hu-man tissues and organs. The purpose of this work is to create a breast phantom for practicing ultrasound diagnostic skills in grayscale and elastography imaging, as well as ultrasound-guided biopsy sampling. We used silicone-like compounds ranging from 3 to 17 on the Shore scale hardness units to simulate soft tissue and lesions. Impurities with experimentally selected concentrations were added to give the phantom the necessary attenuation and reflection parameters. We used 3D modeling programs and 3D printing with PLA plastic to create the casting mold. We developed a breast phantom with inclusions of varying shape, elasticity and echogenicity. After testing the created phantom in B-mode and elastography mode, we performed a survey asking 19 participants how realistic the sonograms of the phantom were. The results showed that the closest to real was the model of the cyst with 9.5 on the 0-10 similarity scale. Thus, the developed breast phantom can be used for ultrasonography, elastography, and ultrasound-guided biopsy training.

Keywords: breast ultrasound, mammary gland, mammography, training phantom, tissue-mimicking materials

Procedia PDF Downloads 68
1574 Medical Image Compression Based on Region of Interest: A Review

Authors: Sudeepti Dayal, Neelesh Gupta

Abstract:

In terms of transmission, bigger the size of any image, longer the time the channel takes for transmission. It is understood that the bandwidth of the channel is fixed. Therefore, if the size of an image is reduced, a larger number of data or images can be transmitted over the channel. Compression is the technique used to reduce the size of an image. In terms of storage, compression reduces the file size which it occupies on the disk. Any image is based on two parameters, region of interest and non-region of interest. There are several algorithms of compression that compress the data more economically. In this paper we have reviewed region of interest and non-region of interest based compression techniques and the algorithms which compress the image most efficiently.

Keywords: compression ratio, region of interest, DCT, DWT

Procedia PDF Downloads 359
1573 High Speed Image Rotation Algorithm

Authors: Hee-Choul Kwon, Hyungjin Cho, Heeyong Kwon

Abstract:

Image rotation is one of main pre-processing step in image processing or image pattern recognition. It is implemented with rotation matrix multiplication. However it requires lots of floating point arithmetic operations and trigonometric function calculations, so it takes long execution time. We propose a new high speed image rotation algorithm without two major time-consuming operations. We compare the proposed algorithm with the conventional rotation one with various size images. Experimental results show that the proposed algorithm is superior to the conventional rotation ones.

Keywords: high speed rotation operation, image processing, image rotation, pattern recognition, transformation matrix

Procedia PDF Downloads 486
1572 Pre-Processing of Ultrasonography Image Quality Improvement in Cases of Cervical Cancer Using Image Enhancement

Authors: Retno Supriyanti, Teguh Budiono, Yogi Ramadhani, Haris B. Widodo, Arwita Mulyawati

Abstract:

Cervical cancer is the leading cause of mortality in cancer-related diseases. In this diagnosis doctors usually perform several tests to determine the presence of cervical cancer in a patient. However, these checks require support equipment to get the results in more detail. One is by using ultrasonography. However, for the developing countries most of the existing ultrasonography has a low resolution. The goal of this research is to obtain abnormalities on low-resolution ultrasound images especially for cervical cancer case. In this paper, we emphasize our work to use Image Enhancement for pre-processing image quality improvement. The result shows that pre-processing stage is promising to support further analysis.

Keywords: cervical cancer, mortality, low-resolution, image enhancement.

Procedia PDF Downloads 615
1571 Anatomical Survey for Text Pattern Detection

Authors: S. Tehsin, S. Kausar

Abstract:

The ultimate aim of machine intelligence is to explore and materialize the human capabilities, one of which is the ability to detect various text objects within one or more images displayed on any canvas including prints, videos or electronic displays. Multimedia data has increased rapidly in past years. Textual information present in multimedia contains important information about the image/video content. However, it needs to technologically testify the commonly used human intelligence of detecting and differentiating the text within an image, for computers. Hence in this paper feature set based on anatomical study of human text detection system is proposed. Subsequent examination bears testimony to the fact that the features extracted proved instrumental to text detection.

Keywords: biologically inspired vision, content based retrieval, document analysis, text extraction

Procedia PDF Downloads 429
1570 Classification of Red, Green and Blue Values from Face Images Using k-NN Classifier to Predict the Skin or Non-Skin

Authors: Kemal Polat

Abstract:

In this study, it has been estimated whether there is skin by using RBG values obtained from the camera and k-nearest neighbor (k-NN) classifier. The dataset used in this study has an unbalanced distribution and a linearly non-separable structure. This problem can also be called a big data problem. The Skin dataset was taken from UCI machine learning repository. As the classifier, we have used the k-NN method to handle this big data problem. For k value of k-NN classifier, we have used as 1. To train and test the k-NN classifier, 50-50% training-testing partition has been used. As the performance metrics, TP rate, FP Rate, Precision, recall, f-measure and AUC values have been used to evaluate the performance of k-NN classifier. These obtained results are as follows: 0.999, 0.001, 0.999, 0.999, 0.999, and 1,00. As can be seen from the obtained results, this proposed method could be used to predict whether the image is skin or not.

Keywords: k-NN classifier, skin or non-skin classification, RGB values, classification

Procedia PDF Downloads 234
1569 AS-Geo: Arbitrary-Sized Image Geolocalization with Learnable Geometric Enhancement Resizer

Authors: Huayuan Lu, Chunfang Yang, Ma Zhu, Baojun Qi, Yaqiong Qiao, Jiangqian Xu

Abstract:

Image geolocalization has great application prospects in fields such as autonomous driving and virtual/augmented reality. In practical application scenarios, the size of the image to be located is not fixed; it is impractical to train different networks for all possible sizes. When its size does not match the size of the input of the descriptor extraction model, existing image geolocalization methods usually directly scale or crop the image in some common ways. This will result in the loss of some information important to the geolocalization task, thus affecting the performance of the image geolocalization method. For example, excessive down-sampling can lead to blurred building contour, and inappropriate cropping can lead to the loss of key semantic elements, resulting in incorrect geolocation results. To address this problem, this paper designs a learnable image resizer and proposes an arbitrary-sized image geolocation method. (1) The designed learnable image resizer employs the self-attention mechanism to enhance the geometric features of the resized image. Firstly, it applies bilinear interpolation to the input image and its feature maps to obtain the initial resized image and the resized feature maps. Then, SKNet (selective kernel net) is used to approximate the best receptive field, thus keeping the geometric shapes as the original image. And SENet (squeeze and extraction net) is used to automatically select the feature maps with strong contour information, enhancing the geometric features. Finally, the enhanced geometric features are fused with the initial resized image, to obtain the final resized images. (2) The proposed image geolocalization method embeds the above image resizer as a fronting layer of the descriptor extraction network. It not only enables the network to be compatible with arbitrary-sized input images but also enhances the geometric features that are crucial to the image geolocalization task. Moreover, the triplet attention mechanism is added after the first convolutional layer of the backbone network to optimize the utilization of geometric elements extracted by the first convolutional layer. Finally, the local features extracted by the backbone network are aggregated to form image descriptors for image geolocalization. The proposed method was evaluated on several mainstream datasets, such as Pittsburgh30K, Tokyo24/7, and Places365. The results show that the proposed method has excellent size compatibility and compares favorably to recently mainstream geolocalization methods.

Keywords: image geolocalization, self-attention mechanism, image resizer, geometric feature

Procedia PDF Downloads 197
1568 Immersive Environment as an Occupant-Centric Tool for Architecture Criticism and Architectural Education

Authors: Golnoush Rostami, Farzam Kharvari

Abstract:

In recent years, developments in the field of architectural education have resulted in a shift from conventional teaching methods to alternative state-of-the-art approaches in teaching methods and strategies. Criticism in architecture has been a key player both in the profession and education, but it has been mostly offered by renowned individuals. Hence, not only students or other professionals but also critics themselves may not have the option to experience buildings and rely on available 2D materials, such as images and plans, that may not result in a holistic understanding and evaluation of buildings. On the other hand, immersive environments provide students and professionals the opportunity to experience buildings virtually and reflect their evaluation by experiencing rather than judging based on 2D materials. Therefore, the aim of this study is to compare the effect of experiencing buildings in immersive environments and 2D drawings, including images and plans, on architecture criticism and architectural education. As a result, three buildings that have parametric brick facades were studied through 2D materials and in Unreal Engine v. 24 as an immersive environment among 22 architecture students that were selected using convenient sampling and were divided into two equal groups using simple random sampling. This study used mixed methods, including quantitative and qualitative methods; the quantitative section was carried out by a questionnaire, and deep interviews were used for the qualitative section. A questionnaire was developed for measuring three constructs, including privacy regulation based on Altman’s theory, the sufficiency of illuminance levels in the building, and the visual status of the view (visually appealing views based on obstructions that may have been caused by facades). Furthermore, participants had the opportunity to reflect their understanding and evaluation of the buildings in individual interviews. Accordingly, the collected data from the questionnaires were analyzed using independent t-test and descriptive analyses in IBM SPSS Statistics v. 26, and interviews were analyzed using the content analysis method. The results of the interviews showed that the participants who experienced the buildings in the immersive environment were able to have a thorough and more precise evaluation of the buildings in comparison to those who studied them through 2D materials. Moreover, the analyses of the respondents’ questionnaires revealed that there were statistically significant differences between measured constructs among the two groups. The outcome of this study suggests that integrating immersive environments into the profession and architectural education as an effective and efficient tool for architecture criticism is vital since these environments allow users to have a holistic evaluation of buildings for vigorous and sound criticism.

Keywords: immersive environments, architecture criticism, architectural education, occupant-centric evaluation, pre-occupancy evaluation

Procedia PDF Downloads 117
1567 Modeling of Strong Motion Generation Areas of the 2011 Tohoku, Japan Earthquake Using Modified Semi-Empirical Technique Incorporating Frequency Dependent Radiation Pattern Model

Authors: Sandeep, A. Joshi, Kamal, Piu Dhibar, Parveen Kumar

Abstract:

In the present work strong ground motion has been simulated using a modified semi-empirical technique (MSET), with frequency dependent radiation pattern model. Joshi et al. (2014) have modified the semi-empirical technique to incorporate the modeling of strong motion generation areas (SMGAs). A frequency dependent radiation pattern model is applied to simulate high frequency ground motion more precisely. Identified SMGAs (Kurahashi and Irikura 2012) of the 2011 Tohoku earthquake (Mw 9.0) were modeled using this modified technique. Records are simulated for both frequency dependent and constant radiation pattern function. Simulated records for both cases are compared with observed records in terms of peak ground acceleration and pseudo acceleration response spectra at different stations. Comparison of simulated and observed records in terms of root mean square error suggests that the method is capable of simulating record which matches in a wide frequency range for this earthquake and bears realistic appearance in terms of shape and strong motion parameters. The results confirm the efficacy and suitability of rupture model defined by five SMGAs for the developed modified technique.

Keywords: strong ground motion, semi-empirical, strong motion generation area, frequency dependent radiation pattern, 2011 Tohoku Earthquake

Procedia PDF Downloads 519
1566 On Phase Based Stereo Matching and Its Related Issues

Authors: András Rövid, Takeshi Hashimoto

Abstract:

The paper focuses on the problem of the point correspondence matching in stereo images. The proposed matching algorithm is based on the combination of simpler methods such as normalized sum of squared differences (NSSD) and a more complex phase correlation based approach, by considering the noise and other factors, as well. The speed of NSSD and the preciseness of the phase correlation together yield an efficient approach to find the best candidate point with sub-pixel accuracy in stereo image pairs. The task of the NSSD in this case is to approach the candidate pixel roughly. Afterwards the location of the candidate is refined by an enhanced phase correlation based method which in contrast to the NSSD has to run only once for each selected pixel.

Keywords: stereo matching, sub-pixel accuracy, phase correlation, SVD, NSSD

Procedia PDF Downloads 450
1565 Urban Search, Rescue and Rapid Field Assessment of Damaged and Collapsed Building Structures

Authors: Abid I. Abu-Tair, Gavin M. Wilde, John M. Kinuthia

Abstract:

Urban Search and Rescue (USAR) is a functional capability that has been developed to allow the United Kingdom Fire and Rescue Service to deal with ‘major incidents’ primarily involving structural collapse. The nature of the work undertaken by USAR means that staying out of a damaged or collapsed building structure is not usually an option for search and rescue personnel. As a result, there is always a risk that they could become victims. For this paper, a systematic and investigative review using desk research was undertaken to explore the role which structural engineering can play in assisting search and rescue personnel to conduct structural assessments when in the field. The focus is on how search and rescue personnel can assess damaged and collapsed building structures, not just in terms of the structural damage that may be countered, but also in relation to structural stability. Natural disasters, accidental emergencies, acts of terrorism and other extreme events can vary significantly in nature and ferocity, and can cause a wide variety of damage to building structures. It is not possible or, even realistic, to provide search and rescue personnel with definitive guidelines and procedures to assess damaged and collapsed building structures as there are too many variables to consider. However, understanding what implications damage may have upon the structural stability of a building structure will enable search and rescue personnel to judge better and quantify the risk from a life-safety standpoint. It is intended that this will allow search and rescue personnel to make informed decisions and ensure every effort is made to mitigate risk so that they do not become victims.

Keywords: damaged and collapsed building structures, life safety, quantifying risk, search and rescue personnel, structural assessments in the field

Procedia PDF Downloads 378
1564 Photovoltaic Modules Fault Diagnosis Using Low-Cost Integrated Sensors

Authors: Marjila Burhanzoi, Kenta Onohara, Tomoaki Ikegami

Abstract:

Faults in photovoltaic (PV) modules should be detected to the greatest extent as early as possible. For that conventional fault detection methods such as electrical characterization, visual inspection, infrared (IR) imaging, ultraviolet fluorescence and electroluminescence (EL) imaging are used, but they either fail to detect the location or category of fault, or they require expensive equipment and are not convenient for onsite application. Hence, these methods are not convenient to use for monitoring small-scale PV systems. Therefore, low cost and efficient inspection techniques with the ability of onsite application are indispensable for PV modules. In this study in order to establish efficient inspection technique, correlation between faults and magnetic flux density on the surface is of crystalline PV modules are investigated. Magnetic flux on the surface of normal and faulted PV modules is measured under the short circuit and illuminated conditions using two different sensor devices. One device is made of small integrated sensors namely 9-axis motion tracking sensor with a 3-axis electronic compass embedded, an IR temperature sensor, an optical laser position sensor and a microcontroller. This device measures the X, Y and Z components of the magnetic flux density (Bx, By and Bz) few mm above the surface of a PV module and outputs the data as line graphs in LabVIEW program. The second device is made of a laser optical sensor and two magnetic line sensor modules consisting 16 pieces of magnetic sensors. This device scans the magnetic field on the surface of PV module and outputs the data as a 3D surface plot of the magnetic flux intensity in a LabVIEW program. A PC equipped with LabVIEW software is used for data acquisition and analysis for both devices. To show the effectiveness of this method, measured results are compared to those of a normal reference module and their EL images. Through the experiments it was confirmed that the magnetic field in the faulted areas have different profiles which can be clearly identified in the measured plots. Measurement results showed a perfect correlation with the EL images and using position sensors it identified the exact location of faults. This method was applied on different modules and various faults were detected using it. The proposed method owns the ability of on-site measurement and real-time diagnosis. Since simple sensors are used to make the device, it is low cost and convenient to be sued by small-scale or residential PV system owners.

Keywords: fault diagnosis, fault location, integrated sensors, PV modules

Procedia PDF Downloads 209
1563 Quantum Engine Proposal using Two-level Atom Like Manipulation and Relativistic Motoring Control

Authors: Montree Bunruangses, Sonath Bhattacharyya, Somchat Sonasang, Preecha Yupapin

Abstract:

A two-level system is manipulated by a microstrip add-drop circuit configured as an atom like system for wave-particle behavior investigation when its traveling speed along the circuit perimeter is the speed of light. The entangled pair formed by the upper and lower sideband peaks is bound by the angular displacement, which is given by 0≤θ≤π/2. The control signals associated with 3-peak signal frequencies are applied by the external inputs via the microstrip add-drop multiplexer ports, where they are time functions without the space term involved. When a system satisfies the speed of light conditions, the mass term has been changed to energy based on the relativistic limit described by the Lorentz factor and Einstein equation. The different applied frequencies can be utilized to form the 3-phase torques that can be applied for quantum engines. The experiment will use the two-level system circuit and be conducted in the laboratory. The 3-phase torques will be recorded and investigated for quantum engine driving purpose. The obtained results will be compared to the simulation. The optimum amplification of torque can be obtained by the resonant successive filtering operation. Torque will be vanished when the system is balanced at the stopped position, where |Time|=0, which is required to be a system stability condition. It will be discussed for future applications. A larger device may be tested in the future for realistic use. A synchronous and asynchronous driven motor is also discussed for the warp drive use.

Keywords: quantum engine, relativistic motor, 3-phase torque, atomic engine

Procedia PDF Downloads 43
1562 Image Enhancement Algorithm of Photoacoustic Tomography Using Active Contour Filtering

Authors: Prasannakumar Palaniappan, Dong Ho Shin, Chul Gyu Song

Abstract:

The photoacoustic images are obtained from a custom developed linear array photoacoustic tomography system. The biological specimens are imitated by conducting phantom tests in order to retrieve a fully functional photoacoustic image. The acquired image undergoes the active region based contour filtering to remove the noise and accurately segment the object area for further processing. The universal back projection method is used as the image reconstruction algorithm. The active contour filtering is analyzed by evaluating the signal to noise ratio and comparing it with the other filtering methods.

Keywords: contour filtering, linear array, photoacoustic tomography, universal back projection

Procedia PDF Downloads 388
1561 Roof Material Detection Based on Object-Based Approach Using WorldView-2 Satellite Imagery

Authors: Ebrahim Taherzadeh, Helmi Z. M. Shafri, Kaveh Shahi

Abstract:

One of the most important tasks in urban area remote sensing is detection of impervious surface (IS), such as building roof and roads. However, detection of IS in heterogeneous areas still remains as one of the most challenging works. In this study, detection of concrete roof using an object-oriented approach was proposed. A new rule-based classification was developed to detect concrete roof tile. The proposed rule-based classification was applied to WorldView-2 image. Results showed that the proposed rule has good potential to predict concrete roof material from WorldView-2 images with 85% accuracy.

Keywords: object-based, roof material, concrete tile, WorldView-2

Procedia PDF Downloads 410
1560 Dose Saving and Image Quality Evaluation for Computed Tomography Head Scanning with Eye Protection

Authors: Yuan-Hao Lee, Chia-Wei Lee, Ming-Fang Lin, Tzu-Huei Wu, Chih-Hsiang Ko, Wing P. Chan

Abstract:

Computed tomography (CT) scan of the head is a good method for investigating cranial lesions. However, radiation-induced oxidative stress can be accumulated in the eyes and promote carcinogenesis and cataract. In this regard, we aimed to protect the eyes with barium sulfate shield(s) during CT scans and investigate the resultant image quality and radiation dose to the eye. Patients who underwent health examinations were selectively enrolled in this study in compliance with the protocol approved by the Ethics Committee of the Joint Institutional Review Board at Taipei Medical University. Participants’ brains were scanned with a water-based marker simultaneously by a multislice CT scanner (SOMATON Definition Flash) under a fixed tube current-time setting or automatic tube current modulation (TCM). The lens dose was measured by Gafchromic films, whose dose response curve was previously fitted using thermoluminescent dosimeters, with or without barium sulfate or bismuth-antimony shield laid above. For the assessment of image quality CT images at slice planes that exhibit the interested regions on the zygomatic, orbital and nasal bones of the head phantom as well as the water-based marker were used for calculating the signal-to-noise and contrast-to-noise ratios. The application of barium sulfate and bismuth-antimony shields decreased 24% and 47% of the lens dose on average, respectively. Under topogram-based TCM, the dose saving power of bismuth-antimony shield was mitigated whereas that of barium sulfate shield was enhanced. On the other hand, the signal-to-noise and contrast-to-noise ratios of DSCT images were decreased separately by barium sulfate and bismuth-antimony shield, resulting in an overall reduction of the CNR. In contrast, the integration of topogram-based TCM elevated signal difference between the ROIs on the zygomatic bones and eyeballs while preferentially decreasing the signal-to-noise ratios upon the use of barium sulfate shield. The results of this study indicate that the balance between eye exposure and image quality can be optimized by combining eye shields with topogram-based TCM on the multislice scanner. Eye shielding could change the photon attenuation characteristics of tissues that are close to the shield. The application of both shields on eye protection hence is not recommended for seeking intraorbital lesions.

Keywords: computed tomography, barium sulfate shield, dose saving, image quality

Procedia PDF Downloads 254
1559 A Study on Waste Management Policy in Minamata City Kumamoto Prefecture Japan

Authors: Qiannan Zhuo, Wanglin Yan

Abstract:

Minamata City and its citizens have been suffered from Minamata Disease, one of the worst environmental problems in Japan, since 1956. To mitigate the bad images brought by Minamata Disease, Minamata City has started a series of environmental friendly activities from 60 years ago. Garbage separation is the very beginning one. It has been already done for more than 20 years since Minamata citizens started to separate their garbage into more than 20 categories. In this research, the author evaluated the effectiveness of the waste management policy in Minamata city by analyzing the recycle rate and the landfill amount., and also pointed out the problems brought by it through the qualitative survey.

Keywords: Minamata City, households waste, garbage separation, recycle reduce reuse

Procedia PDF Downloads 224
1558 Characterization of 3D-MRP for Analyzing of Brain Balancing Index (BBI) Pattern

Authors: N. Fuad, M. N. Taib, R. Jailani, M. E. Marwan

Abstract:

This paper discusses on power spectral density (PSD) characteristics which are extracted from three-dimensional (3D) electroencephalogram (EEG) models. The EEG signal recording was conducted on 150 healthy subjects. Development of 3D EEG models involves pre-processing of raw EEG signals and construction of spectrogram images. Then, the values of maximum PSD were extracted as features from the model. These features are analysed using mean relative power (MRP) and different mean relative power (DMRP) technique to observe the pattern among different brain balancing indexes. The results showed that by implementing these techniques, the pattern of brain balancing indexes can be clearly observed. Some patterns are indicates between index 1 to index 5 for left frontal (LF) and right frontal (RF).

Keywords: power spectral density, 3D EEG model, brain balancing, mean relative power, different mean relative power

Procedia PDF Downloads 456