Search results for: cooling rate.
1876 Low Resolution Face Recognition Using Mixture of Experts
Authors: Fatemeh Behjati Ardakani, Fatemeh Khademian, Abbas Nowzari Dalini, Reza Ebrahimpour
Abstract:
Human activity is a major concern in a wide variety of applications, such as video surveillance, human computer interface and face image database management. Detecting and recognizing faces is a crucial step in these applications. Furthermore, major advancements and initiatives in security applications in the past years have propelled face recognition technology into the spotlight. The performance of existing face recognition systems declines significantly if the resolution of the face image falls below a certain level. This is especially critical in surveillance imagery where often, due to many reasons, only low-resolution video of faces is available. If these low-resolution images are passed to a face recognition system, the performance is usually unacceptable. Hence, resolution plays a key role in face recognition systems. In this paper we introduce a new low resolution face recognition system based on mixture of expert neural networks. In order to produce the low resolution input images we down-sampled the 48 × 48 ORL images to 12 × 12 ones using the nearest neighbor interpolation method and after that applying the bicubic interpolation method yields enhanced images which is given to the Principal Component Analysis feature extractor system. Comparison with some of the most related methods indicates that the proposed novel model yields excellent recognition rate in low resolution face recognition that is the recognition rate of 100% for the training set and 96.5% for the test set.Keywords: Low resolution face recognition, Multilayered neuralnetwork, Mixture of experts neural network, Principal componentanalysis, Bicubic interpolation, Nearest neighbor interpolation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17241875 Evaluating the Appropriateness of Passive Techniques Used in Achieving Thermal Comfort in Buildings: A Case of Lautech College of Health Sciences, Ogbomoso
Authors: Ilelabayo I. Adebisi, Yetunde R. Okeyinka, Abdulrasaq K. Ayinla
Abstract:
Architectural design is a complex process especially when the issue of user’s comfort, building sustainability and energy efficiency needs to be addressed. The current energy challenge and the seek for an environment where users will have a more physiological and psychological comfort in this part of the world have led various researchers to constantly explore the concept of passive design techniques. Passive techniques are design strategies used in regulating building indoor climates and improving users comfort without the use of energy driven devices. This paper describes and analyses the significance of passive techniques on indoor climates and their impact on thermal comfort of building users using LAUTECH College of health sciences Ogbomoso as a case study. The study aims at assessing the appropriateness of the passive strategies used in achieving comfort in their buildings with a view to evaluate their adequacy and effectiveness and suggesting how comfortable their building users are. This assessment was carried out through field survey and questionnaires and findings revealed that strategies such as Orientation, Spacing, Courtyards, window positioning and choice of landscape adopted are inadequate while only fins and roof overhangs are adequate. The finding also revealed that 72% of building occupants feel hot discomfort in their various spaces and hence have the urge to get fresh air from outside during work hours. The Mahoney table was used to provide appropriate architectural design recommendations to guide future designers in the study area.
Keywords: Energy challenge, passive cooling, techniques, thermal comfort, users comfort.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9011874 An Overview of Electronic Waste as Aggregate in Concrete
Authors: S. R. Shamili, C. Natarajan, J. Karthikeyan
Abstract:
Rapid growth of world population and widespread urbanization has remarkably increased the development of the construction industry which caused a huge demand for sand and gravels. Environmental problems occur when the rate of extraction of sand, gravels, and other materials exceeds the rate of generation of natural resources; therefore, an alternative source is essential to replace the materials used in concrete. Now-a-days, electronic products have become an integral part of daily life which provides more comfort, security, and ease of exchange of information. These electronic waste (E-Waste) materials have serious human health concerns and require extreme care in its disposal to avoid any adverse impacts. Disposal or dumping of these E-Wastes also causes major issues because it is highly complex to handle and often contains highly toxic chemicals such as lead, cadmium, mercury, beryllium, brominates flame retardants (BFRs), polyvinyl chloride (PVC), and phosphorus compounds. Hence, E-Waste can be incorporated in concrete to make a sustainable environment. This paper deals with the composition, preparation, properties, classification of E-Waste. All these processes avoid dumping to landfills whilst conserving natural aggregate resources, and providing a better environmental option. This paper also provides a detailed literature review on the behaviour of concrete with incorporation of E-Wastes. Many research shows the strong possibility of using E-Waste as a substitute of aggregates eventually it reduces the use of natural aggregates in concrete.
Keywords: Disposal, electronic waste, landfill, toxic chemicals.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 28401873 An Ising-based Model for the Spread of Infection
Authors: Christian P. Crisostomo, Chrysline Margus N. Piñol
Abstract:
A zero-field ferromagnetic Ising model is utilized to simulate the propagation of infection in a population that assumes a square lattice structure. The rate of infection increases with temperature. The disease spreads faster among individuals with low J values. Such effect, however, diminishes at higher temperatures.Keywords: Epidemiology, Ising model, lattice models
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20421872 Sub-Lethal Effects of Thiamethoxam and Pirimicarb on Life-Table Parameters of Diaeretiella rapae (Hymenoptera: Braconidae), Parasitoid of Lipaphis erysimi (Hemiptera: Aphididae)
Authors: Nastaran Rezaei, Mohammad Saeed Mossadegh, Farhan Kocheyli, Khalil Talebi Jahromi, Aurang Kavousi
Abstract:
Integrated Pest Management (IPM) aims to combine biological and chemical strategies and measures, hence highlighting the study of acute toxicity and sub-lethal effects of pesticides comprehensively. The present research focused on the side effects of thiamethoxam and pirimicarb sub-lethal concentrations on demographic parameters of Diaeretiella rapae (McIntosh Laboratory) (Hymenoptera: Braconidae). Adult parasitoids were exposed to LC25 of insecticides as well as distilled water as the control. The results showed that thiamethoxam adversely affected population parameters (r, λ, R0, T), adults' longevity, females' oviposition period and mean fecundity, and a similar trend was obtained for pirimicarb with the exception of generation time (T), the latter did not significantly change compared to the control. The intrinsic rate of increase (r) in the control and those treated with pirimicarb and thiamethoxam were 0.2801, 0.2064, 0.1525 days-1, respectively, and the sex ratio was biased toward females in all treatments. Furthermore, none of the insecticides influenced total pre-oviposition period (TPOP) and offspring emergence rate. In general, these results indicated that both insecticides potentially distort the demographic parameters of the parasitoid even at sub-lethal concentrations, and then they should not be considered for IPM program in the presence of D. rapae.
Keywords: Diaeretiella rapae, Lipaphis erysimi, life-table study, pirimicarb, thiamethoxam.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 7381871 Estimation of Attenuation and Phase Delay in Driving Voltage Waveform of a Digital-Noiseless, Ultra-High-Speed Image Sensor
Authors: V. T. S. Dao, T. G. Etoh, C. Vo Le, H. D. Nguyen, K. Takehara, T. Akino, K. Nishi
Abstract:
Since 2004, we have been developing an in-situ storage image sensor (ISIS) that captures more than 100 consecutive images at a frame rate of 10 Mfps with ultra-high sensitivity as well as the video camera for use with this ISIS. Currently, basic research is continuing in an attempt to increase the frame rate up to 100 Mfps and above. In order to suppress electro-magnetic noise at such high frequency, a digital-noiseless imaging transfer scheme has been developed utilizing solely sinusoidal driving voltages. This paper presents highly efficient-yet-accurate expressions to estimate attenuation as well as phase delay of driving voltages through RC networks of an ultra-high-speed image sensor. Elmore metric for a fundamental RC chain is employed as the first-order approximation. By application of dimensional analysis to SPICE data, we found a simple expression that significantly improves the accuracy of the approximation. Similarly, another simple closed-form model to estimate phase delay through fundamental RC networks is also obtained. Estimation error of both expressions is much less than previous works, only less 2% for most of the cases . The framework of this analysis can be extended to address similar issues of other VLSI structures.
Keywords: Dimensional Analysis, ISIS, Digital-noiseless, RC network, Attenuation, Phase Delay, Elmore model
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14541870 Pilot-Assisted Direct-Current Biased Optical Orthogonal Frequency Division Multiplexing Visible Light Communication System
Authors: Ayad A. Abdulkafi, Shahir F. Nawaf, Mohammed K. Hussein, Ibrahim K. Sileh, Fouad A. Abdulkafi
Abstract:
Visible light communication (VLC) is a new approach of optical wireless communication proposed to support the congested radio frequency (RF) spectrum. VLC systems are combined with orthogonal frequency division multiplexing (OFDM) to achieve high rate transmission and high spectral efficiency. In this paper, we investigate the Pilot-Assisted Channel Estimation for DC biased Optical OFDM (PACE-DCO-OFDM) systems to reduce the effects of the distortion on the transmitted signal. Least-square (LS) and linear minimum mean-squared error (LMMSE) estimators are implemented in MATLAB/Simulink to enhance the bit-error-rate (BER) of PACE-DCO-OFDM. Results show that DCO-OFDM system based on PACE scheme has achieved better BER performance compared to conventional system without pilot assisted channel estimation. Simulation results show that the proposed PACE-DCO-OFDM based on LMMSE algorithm can more accurately estimate the channel and achieves better BER performance when compared to the LS based PACE-DCO-OFDM and the traditional system without PACE. For the same signal to noise ratio (SNR) of 25 dB, the achieved BER is about 5×10-4 for LMMSE-PACE and 4.2×10-3 with LS-PACE while it is about 2×10-1 for system without PACE scheme.
Keywords: Channel estimation, OFDM, pilot-assist, VLC.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6691869 Image Transmission via Iterative Cellular-Turbo System
Authors: Ersin Gose, Kenan Buyukatak, Onur Osman, Osman N. Ucan
Abstract:
To compress, improve bit error performance and also enhance 2D images, a new scheme, called Iterative Cellular-Turbo System (IC-TS) is introduced. In IC-TS, the original image is partitioned into 2N quantization levels, where N is denoted as bit planes. Then each of the N-bit-plane is coded by Turbo encoder and transmitted over Additive White Gaussian Noise (AWGN) channel. At the receiver side, bit-planes are re-assembled taking into consideration of neighborhood relationship of pixels in 2-D images. Each of the noisy bit-plane values of the image is evaluated iteratively using IC-TS structure, which is composed of equalization block; Iterative Cellular Image Processing Algorithm (ICIPA) and Turbo decoder. In IC-TS, there is an iterative feedback link between ICIPA and Turbo decoder. ICIPA uses mean and standard deviation of estimated values of each pixel neighborhood. It has extra-ordinary satisfactory results of both Bit Error Rate (BER) and image enhancement performance for less than -1 dB Signal-to-Noise Ratio (SNR) values, compared to traditional turbo coding scheme and 2-D filtering, applied separately. Also, compression can be achieved by using IC-TS systems. In compression, less memory storage is used and data rate is increased up to N-1 times by simply choosing any number of bit slices, sacrificing resolution. Hence, it is concluded that IC-TS system will be a compromising approach in 2-D image transmission, recovery of noisy signals and image compression.
Keywords: Iterative Cellular Image Processing Algorithm (ICIPA), Turbo Coding, Iterative Cellular Turbo System (IC-TS), Image Compression.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18151868 An Active Solar Energy System to Supply Heating Demands of the Teaching Staff Dormitory of Islamic Azad University Ramhormoz Branch
Authors: M. Talebzadegan, S. Bina, I. Riazi
Abstract:
The purpose of this paper is to present an active solar energy system to supply heating demands of the teaching staff dormitory of the Islamic Azad University of Ramhormoz. The design takes into account the solar radiations and climate data of Ramhormoz town and is based on the daily warm water consumption for health demands of 450 residents of the dormitory, which is equal to 27000 lit of 50-C° water, and building heating requirements with an area of 3500 m² well-protected by heatproof materials. First, heating demands of the building were calculated, then a hybrid system made up of solar and fossil energies was developed and finally, the design was economically evaluated. Since there is only roof space for using 110 flat solar water heaters, the calculations were made to hybridize solar water heating system with heat pumping system in which solar energy contributes 67% of the heat generated. According to calculations, the net present value “N.P.V.” of revenue stream exceeds “N.P.V.” of cash paid off in this project over three years, which makes economically quite promising. The return of investment and payback period of the project is 4 years. Also, the internal rate of return (IRR) of the project was 25%, which exceeds bank rate of interest in Iran and emphasizes the desirability of the project.
Keywords: Solar energy, heat demand, renewable, pollution.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6881867 Image Analysis for Obturator Foramen Based on Marker-Controlled Watershed Segmentation and Zernike Moments
Authors: Seda Sahin, Emin Akata
Abstract:
Obturator Foramen is a specific structure in Pelvic bone images and recognition of it is a new concept in medical image processing. Moreover, segmentation of bone structures such as Obturator Foramen plays an essential role for clinical research in orthopedics. In this paper, we present a novel method to analyze the similarity between the substructures of the imaged region and a hand drawn template as a preprocessing step for computation of Pelvic bone rotation on hip radiographs. This method consists of integrated usage of Marker-controlled Watershed segmentation and Zernike moment feature descriptor and it is used to detect Obturator Foramen accurately. Marker-controlled Watershed segmentation is applied to separate Obturator Foramen from the background effectively. Then, Zernike moment feature descriptor is used to provide matching between binary template image and the segmented binary image for final extraction of Obturator Foramens. Finally, Pelvic bone rotation rate calculation for each hip radiograph is performed automatically to select and eliminate hip radiographs for further studies which depend on Pelvic bone angle measurements. The proposed method is tested on randomly selected 100 hip radiographs. The experimental results demonstrated that the proposed method is able to segment Obturator Foramen with 96% accuracy.Keywords: Medical image analysis, marker-controlled watershed segmentation, segmentation of bone structures on hip radiographs, pelvic bone rotation rate, zernike moment feature descriptor.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19931866 A Vehicular Visual Tracking System Incorporating Global Positioning System
Authors: Hsien-Chou Liao, Yu-Shiang Wang
Abstract:
Surveillance system is widely used in the traffic monitoring. The deployment of cameras is moving toward a ubiquitous camera (UbiCam) environment. In our previous study, a novel service, called GPS-VT, was firstly proposed by incorporating global positioning system (GPS) and visual tracking techniques for the UbiCam environment. The first prototype is called GODTA (GPS-based Moving Object Detection and Tracking Approach). For a moving person carried GPS-enabled mobile device, he can be tracking when he enters the field-of-view (FOV) of a camera according to his real-time GPS coordinate. In this paper, GPS-VT service is applied to the tracking of vehicles. The moving speed of a vehicle is much faster than a person. It means that the time passing through the FOV is much shorter than that of a person. Besides, the update interval of GPS coordinate is once per second, it is asynchronous with the frame rate of the real-time image. The above asynchronous is worsen by the network transmission delay. These factors are the main challenging to fulfill GPS-VT service on a vehicle.In order to overcome the influence of the above factors, a back-propagation neural network (BPNN) is used to predict the possible lane before the vehicle enters the FOV of a camera. Then, a template matching technique is used for the visual tracking of a target vehicle. The experimental result shows that the target vehicle can be located and tracking successfully. The success location rate of the implemented prototype is higher than that of the previous GODTA.Keywords: visual surveillance, visual tracking, globalpositioning system, intelligent transportation system
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19171865 Design and Performance Analysis of One Dimensional Zero Cross-Correlation Coding Technique for a Fixed Wavelength Hopping SAC-OCDMA
Authors: Satyasen Panda, Urmila Bhanja
Abstract:
This paper presents a SAC-OCDMA code with zero cross correlation property to minimize the Multiple Access Interface (MAI) as New Zero Cross Correlation code (NZCC), which is found to be more scalable compared to the other existing SAC-OCDMA codes. This NZCC code is constructed using address segment and data segment. In this work, the proposed NZCC code is implemented in an optical system using the Opti-System software for the spectral amplitude coded optical code-division multiple-access (SAC-OCDMA) scheme. The main contribution of the proposed NZCC code is the zero cross correlation, which reduces both the MAI and PIIN noises. The proposed NZCC code reveals properties of minimum cross-correlation, flexibility in selecting the code parameters and supports a large number of users, combined with high data rate and longer fiber length. Simulation results reveal that the optical code division multiple access system based on the proposed NZCC code accommodates maximum number of simultaneous users with higher data rate transmission, lower Bit Error Rates (BER) and longer travelling distance without any signal quality degradation, as compared to the former existing SAC-OCDMA codes.
Keywords: Cross Correlation, Optical Code Division Multiple Access, Spectral Amplitude Coding Optical Code Division Multiple Access, Multiple Access Interference, Phase Induced Intensity Noise, New Zero Cross Correlation code.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22431864 Analysis and Research of Two-Level Scheduling Profile for Open Real-Time System
Authors: Yongxian Jin, Jingzhou Huang
Abstract:
In an open real-time system environment, the coexistence of different kinds of real-time and non real-time applications makes the system scheduling mechanism face new requirements and challenges. One two-level scheduling scheme of the open real-time systems is introduced, and points out that hard and soft real-time applications are scheduled non-distinctively as the same type real-time applications, the Quality of Service (QoS) cannot be guaranteed. It has two flaws: The first, it can not differentiate scheduling priorities of hard and soft real-time applications, that is to say, it neglects characteristic differences between hard real-time applications and soft ones, so it does not suit a more complex real-time environment. The second, the worst case execution time of soft real-time applications cannot be predicted exactly, so it is not worth while to cost much spending in order to assure all soft real-time applications not to miss their deadlines, and doing that may cause resource wasting. In order to solve this problem, a novel two-level real-time scheduling mechanism (including scheduling profile and scheduling algorithm) which adds the process of dealing with soft real-time applications is proposed. Finally, we verify real-time scheduling mechanism from two aspects of theory and experiment. The results indicate that our scheduling mechanism can achieve the following objectives. (1) It can reflect the difference of priority when scheduling hard and soft real-time applications. (2) It can ensure schedulability of hard real-time applications, that is, their rate of missing deadline is 0. (3) The overall rate of missing deadline of soft real-time applications can be less than 1. (4) The deadline of a non-real-time application is not set, whereas the scheduling algorithm that server 0 S uses can avoid the “starvation" of jobs and increase QOS. By doing that, our scheduling mechanism is more compatible with different types of applications and it will be applied more widely.
Keywords: Hard real-time, two-level scheduling profile, open real-time system, non-distinctive schedule, soft real-time
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15681863 Effect of Needle Height on Discharge Coefficient and Cavitation Number
Authors: Azadeh Yazdi, Mohammadreza Nezamirad, Sepideh Amirahmadian, Nasim Sabetpour, Amirmasoud Hamedi
Abstract:
Cavitation inside diesel injector nozzle is investigated using Reynolds-Stress-Navier stokes equations. Schnerr-Sauer cavitation model is used for modeling cavitation inside diesel injector nozzle. The carrying fluid utilized in the current study is diesel fuel. The flow is verified at the beginning by comparing with the previous experimental data and it was found that K-Epsilon turbulent model could lead to a better accuracy comparing to K-Omega turbulent model. Moreover, mass flow rate obtained numerically is compared with the experimental value and discrepancy was found to be less than 5% - which shows the accuracy of the current results. Finally, a real-size four-hole nozzle is investigated and the flow inside it is visualized based on velocity profile, discharge coefficient and cavitation number. It was found that the mesh density could be reduced significantly by utilizing periodic boundary condition. Velocity contour at the mid nozzle showed that maximum value of velocity occurs at the end of the needle before entering the orifice area. Last but not least, at the same boundary conditions, when different needle heights were utilized, it was found that as needle height increases with an increase in cavitation number, discharge coefficient increases, while the mentioned increases is more tangible at smaller values of needle heights.
Keywords: cavitation, diesel fuel, CFD, real size nozzle, mass flow rate
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5241862 A Perceptually Optimized Wavelet Embedded Zero Tree Image Coder
Authors: A. Bajit, M. Nahid, A. Tamtaoui, E. H. Bouyakhf
Abstract:
In this paper, we propose a Perceptually Optimized Embedded ZeroTree Image Coder (POEZIC) that introduces a perceptual weighting to wavelet transform coefficients prior to control SPIHT encoding algorithm in order to reach a targeted bit rate with a perceptual quality improvement with respect to the coding quality obtained using the SPIHT algorithm only. The paper also, introduces a new objective quality metric based on a Psychovisual model that integrates the properties of the HVS that plays an important role in our POEZIC quality assessment. Our POEZIC coder is based on a vision model that incorporates various masking effects of human visual system HVS perception. Thus, our coder weights the wavelet coefficients based on that model and attempts to increase the perceptual quality for a given bit rate and observation distance. The perceptual weights for all wavelet subbands are computed based on 1) luminance masking and Contrast masking, 2) the contrast sensitivity function CSF to achieve the perceptual decomposition weighting, 3) the Wavelet Error Sensitivity WES used to reduce the perceptual quantization errors. The new perceptually optimized codec has the same complexity as the original SPIHT techniques. However, the experiments results show that our coder demonstrates very good performance in terms of quality measurement.
Keywords: DWT, linear-phase 9/7 filter, 9/7 Wavelets Error Sensitivity WES, CSF implementation approaches, JND Just Noticeable Difference, Luminance masking, Contrast masking, standard SPIHT, Objective Quality Measure, Probability Score PS.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20511861 Hybrid Equity Warrants Pricing Formulation under Stochastic Dynamics
Authors: Teh Raihana Nazirah Roslan, Siti Zulaiha Ibrahim, Sharmila Karim
Abstract:
A warrant is a financial contract that confers the right but not the obligation, to buy or sell a security at a certain price before expiration. The standard procedure to value equity warrants using call option pricing models such as the Black–Scholes model had been proven to contain many flaws, such as the assumption of constant interest rate and constant volatility. In fact, existing alternative models were found focusing more on demonstrating techniques for pricing, rather than empirical testing. Therefore, a mathematical model for pricing and analyzing equity warrants which comprises stochastic interest rate and stochastic volatility is essential to incorporate the dynamic relationships between the identified variables and illustrate the real market. Here, the aim is to develop dynamic pricing formulations for hybrid equity warrants by incorporating stochastic interest rates from the Cox-Ingersoll-Ross (CIR) model, along with stochastic volatility from the Heston model. The development of the model involves the derivations of stochastic differential equations that govern the model dynamics. The resulting equations which involve Cauchy problem and heat equations are then solved using partial differential equation approaches. The analytical pricing formulas obtained in this study comply with the form of analytical expressions embedded in the Black-Scholes model and other existing pricing models for equity warrants. This facilitates the practicality of this proposed formula for comparison purposes and further empirical study.
Keywords: Cox-Ingersoll-Ross model, equity warrants, Heston model, hybrid models, stochastic.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5861860 Computer Countenanced Diagnosis of Skin Nodule Detection and Histogram Augmentation: Extracting System for Skin Cancer
Authors: S. Zith Dey Babu, S. Kour, S. Verma, C. Verma, V. Pathania, A. Agrawal, V. Chaudhary, A. Manoj Puthur, R. Goyal, A. Pal, T. Danti Dey, A. Kumar, K. Wadhwa, O. Ved
Abstract:
Background: Skin cancer is now is the buzzing button in the field of medical science. The cyst's pandemic is drastically calibrating the body and well-being of the global village. Methods: The extracted image of the skin tumor cannot be used in one way for diagnosis. The stored image contains anarchies like the center. This approach will locate the forepart of an extracted appearance of skin. Partitioning image models has been presented to sort out the disturbance in the picture. Results: After completing partitioning, feature extraction has been formed by using genetic algorithm and finally, classification can be performed between the trained and test data to evaluate a large scale of an image that helps the doctors for the right prediction. To bring the improvisation of the existing system, we have set our objectives with an analysis. The efficiency of the natural selection process and the enriching histogram is essential in that respect. To reduce the false-positive rate or output, GA is performed with its accuracy. Conclusions: The objective of this task is to bring improvisation of effectiveness. GA is accomplishing its task with perfection to bring down the invalid-positive rate or outcome. The paper's mergeable portion conflicts with the composition of deep learning and medical image processing, which provides superior accuracy. Proportional types of handling create the reusability without any errors.
Keywords: Computer-aided system, detection, image segmentation, morphology.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5451859 Feasibility Study for a Castor oil Extraction Plant in South Africa
Authors: Mohamed Belaid, Edison Muzenda, Getrude Mitilene, Mansoor Mollagee
Abstract:
A feasibility study for the design and construction of a pilot plant for the extraction of castor oil in South Africa was conducted. The study emphasized the four critical aspects of project feasibility analysis, namely technical, financial, market and managerial aspects. The technical aspect involved research on existing oil extraction technologies, namely: mechanical pressing and solvent extraction, as well as assessment of the proposed production site for both short and long term viability of the project. The site is on the outskirts of Nkomazi village in the Mpumalanga province, where connections for water and electricity are currently underway, potential raw material supply proves to be reliable since the province is known for its commercial farming. The managerial aspect was evaluated based on the fact that the current producer of castor oil will be fully involved in the project while receiving training and technical assistance from Sasol Technology, the TSC and SEDA. Market and financial aspects were evaluated and the project was considered financially viable with a Net Present Value (NPV) of R2 731 687 and an Internal Rate of Return (IRR) of 18% at an annual interest rate of 10.5%. The payback time is 6years for analysis over the first 10 years with a net income of R1 971 000 in the first year. The project was thus found to be feasible with high chance of success while contributing to socio-economic development. It was recommended for lab tests to be conducted to establish process kinetics that would be used in the initial design of the plant.Keywords: Mechanical pressing, Net Present Value, Oilextraction, Project feasibility, Solvent extraction
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 60821858 Qanat (Subterranean Canal) Role in Traditional Cities and Settlements Formation of Hot-Arid Regions of Iran
Authors: Karim Shiraazi, Mahyar Asheghi Milani, Alireza Sadeghi, Eram Azami, Ahadollah Azami
Abstract:
A passive system "Qanat" is collection of some underground wells. A mother-well was dug in a place far from the city where they could reach to the water table maybe 100 meters underground, they dug other wells to direct water toward the city, with minimum possible gradient. Using the slope of the earth they could bring water close to the surface in the city. The source of water or the appearance of Qanat, land slope and the ownership lines are the important and effective factors in the formation of routes and the segment division of lands to the extent that making use of Qanat as the techniques of extracting underground waters creates a channel of routes with an organic order and hierarchy coinciding the slope of land and it also guides the Qanat waters in the tradition texture of salt desert and border provinces of it. Qanats are excavated in a specified distinction from each other. The quantity of water provided by Qanats depends on the kind of land, distance from mountain, geographical situation of them and the rate of water supply from the underground land. The rate of underground waters, possibility of Qanat excavation, number of Qanats and rate of their water supply from one hand and the quantity of cultivable fertile lands from the other hand are the important natural factors making the size of cities. In the same manner the cities with several Qanats have multi central textures. The location of cities is in direct relation with land quality, soil fertility and possibility of using underground water by excavating Qanats. Observing the allowable distance for Qanat watering is a determining factor for distance between villages and cities. Topography, land slope, soil quality, watering system, ownership, kind of cultivation, etc. are the effective factors in directing Qanats for excavation and guiding water toward the cultivable lands and it also causes the formation of different textures in land division of farming provinces. Several divisions such as orderly and wide, inorderly, thin and long, comb like, etc. are the introduction to organic order. And at the same time they are complete coincidence with environmental conditions in the typical development of ecological architecture and planning in the traditional cities and settlements order.Keywords: Qanat, Settlement Formation, Hot-Arid Region, Sustainable Development
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19201857 Agreement between Basal Metabolic Rate Measured by Bioelectrical Impedance Analysis and Estimated by Prediction Equations in Obese Groups
Authors: Orkide Donma, Mustafa M. Donma
Abstract:
Basal metabolic rate (BMR) is widely used and an accepted measure of energy expenditure. Its principal determinant is body mass. However, this parameter is also correlated with a variety of other factors. The objective of this study is to measure BMR and compare it with the values obtained from predictive equations in adults classified according to their body mass index (BMI) values. 276 adults were included into the scope of this study. Their age, height and weight values were recorded. Five groups were designed based on their BMI values. First group (n = 85) was composed of individuals with BMI values varying between 18.5 and 24.9 kg/m2. Those with BMI values varying from 25.0 to 29.9 kg/m2 constituted Group 2 (n = 90). Individuals with 30.0-34.9 kg/m2, 35.0-39.9 kg/m2, > 40.0 kg/m2 were included in Group 3 (n = 53), 4 (n = 28) and 5 (n = 20), respectively. The most commonly used equations to be compared with the measured BMR values were selected. For this purpose, the values were calculated by the use of four equations to predict BMR values, by name, introduced by Food and Agriculture Organization (FAO)/World Health Organization (WHO)/United Nations University (UNU), Harris and Benedict, Owen and Mifflin. Descriptive statistics, ANOVA, post-Hoc Tukey and Pearson’s correlation tests were performed by a statistical program designed for Windows (SPSS, version 16.0). p values smaller than 0.05 were accepted as statistically significant. Mean ± SD of groups 1, 2, 3, 4 and 5 for measured BMR in kcal were 1440.3 ± 210.0, 1618.8 ± 268.6, 1741.1 ± 345.2, 1853.1 ± 351.2 and 2028.0 ± 412.1, respectively. Upon evaluation of the comparison of means among groups, differences were highly significant between Group 1 and each of the remaining four groups. The values were increasing from Group 2 to Group 5. However, differences between Group 2 and Group 3, Group 3 and Group 4, Group 4 and Group 5 were not statistically significant. These insignificances were lost in predictive equations proposed by Harris and Benedict, FAO/WHO/UNU and Owen. For Mifflin, the insignificance was limited only to Group 4 and Group 5. Upon evaluation of the correlations of measured BMR and the estimated values computed from prediction equations, the lowest correlations between measured BMR and estimated BMR values were observed among the individuals within normal BMI range. The highest correlations were detected in individuals with BMI values varying between 30.0 and 34.9 kg/m2. Correlations between measured BMR values and BMR values calculated by FAO/WHO/UNU as well as Owen were the same and the highest. In all groups, the highest correlations were observed between BMR values calculated from Mifflin and Harris and Benedict equations using age as an additional parameter. In conclusion, the unique resemblance of the FAO/WHO/UNU and Owen equations were pointed out. However, mean values obtained from FAO/WHO/UNU were much closer to the measured BMR values. Besides, the highest correlations were found between BMR calculated from FAO/WHO/UNU and measured BMR. These findings suggested that FAO/WHO/UNU was the most reliable equation, which may be used in conditions when the measured BMR values are not available.Keywords: Adult, basal metabolic rate, FAO/WHO/UNU, obesity, prediction equations.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10111856 Effect of Reynolds Number and Concentration of Biopolymer (Gum Arabic) on Drag Reduction of Turbulent Flow in Circular Pipe
Authors: Kamaljit Singh Sokhal, Gangacharyulu Dasoraju, Vijaya Kumar Bulasara
Abstract:
Biopolymers are popular in many areas, like petrochemicals, food industry and agriculture due to their favorable properties like environment-friendly, availability, and cost. In this study, a biopolymer gum Arabic was used to find its effect on the pressure drop at various concentrations (100 ppm – 300 ppm) with various Reynolds numbers (10000 – 45000). A rheological study was also done by using the same concentrations to find the effect of the shear rate on the shear viscosity. Experiments were performed to find the effect of injection of gum Arabic directly near the boundary layer and to investigate its effect on the maximum possible drag reduction. Experiments were performed on a test section having i.d of 19.50 mm and length of 3045 mm. The polymer solution was injected from the top of the test section by using a peristaltic pump. The concentration of the polymer solution and the Reynolds number were used as parameters to get maximum possible drag reduction. Water was circulated through a centrifugal pump having a maximum 3000 rpm and the flow rate was measured by using rotameter. Results were validated by using Virk's maximum drag reduction asymptote. A maximum drag reduction of 62.15% was observed with the maximum concentration of gum Arabic, 300 ppm. The solution was circulated in the closed loop to find the effect of degradation of polymers with a number of cycles on the drag reduction percentage. It was observed that the injection of the polymer solution in the boundary layer was showing better results than premixed solutions.
Keywords: Drag reduction, shear viscosity, gum Arabic, injection point.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 7431855 Generative Adversarial Network Based Fingerprint Anti-Spoofing Limitations
Authors: Yehjune Heo
Abstract:
Fingerprint Anti-Spoofing approaches have been actively developed and applied in real-world applications. One of the main problems for Fingerprint Anti-Spoofing is not robust to unseen samples, especially in real-world scenarios. A possible solution will be to generate artificial, but realistic fingerprint samples and use them for training in order to achieve good generalization. This paper contains experimental and comparative results with currently popular GAN based methods and uses realistic synthesis of fingerprints in training in order to increase the performance. Among various GAN models, the most popular StyleGAN is used for the experiments. The CNN models were first trained with the dataset that did not contain generated fake images and the accuracy along with the mean average error rate were recorded. Then, the fake generated images (fake images of live fingerprints and fake images of spoof fingerprints) were each combined with the original images (real images of live fingerprints and real images of spoof fingerprints), and various CNN models were trained. The best performances for each CNN model, trained with the dataset of generated fake images and each time the accuracy and the mean average error rate, were recorded. We observe that current GAN based approaches need significant improvements for the Anti-Spoofing performance, although the overall quality of the synthesized fingerprints seems to be reasonable. We include the analysis of this performance degradation, especially with a small number of samples. In addition, we suggest several approaches towards improved generalization with a small number of samples, by focusing on what GAN based approaches should learn and should not learn.
Keywords: Anti-spoofing, CNN, fingerprint recognition, GAN.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5941854 The Solar Wall in the Italian Climates
Authors: F. Stazi, C. Di Perna, C. Filiaci, A. Stazi
Abstract:
Passive systems were born with the purpose of the greatest exploitation of solar energy in cold climates and high altitudes. They spread themselves until the 80-s all over the world without any attention to the specific climate and the summer behavior; this caused the deactivation of the systems due to a series of problems connected to the summer overheating, the complex management and the rising of the dust. Until today the European regulation limits only the winter consumptions without any attention to the summer behavior but, the recent European EN 15251 underlines the relevance of the indoor comfort, and the necessity of the analytic studies validation by monitoring case studies. In the porpose paper we demonstrate that the solar wall is an efficient system both from thermal comfort and energy saving point of view and it is the most suitable for our temperate climates because it can be used as a passive cooling sistem too. In particular the paper present an experimental and numerical analisys carried out on a case study with nine different solar passive systems in Ancona, Italy. We carried out a detailed study of the lodging provided by the solar wall by the monitoring and the evaluation of the indoor conditions. Analyzing the monitored data, on the base of recognized models of comfort (ISO, ASHRAE, Givoni-s BBCC), is emerged that the solar wall has an optimal behavior in the middle seasons. In winter phase this passive system gives more advantages in terms of energy consumptions than the other systems, because it gives greater heat gain and therefore smaller consumptions. In summer, when outside air temperature return in the mean seasonal value, the indoor comfort is optimal thanks to an efficient transversal ventilation activated from the same wall.Keywords: Building envelope, energy saving, passive solarwall, thermal comfort.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16541853 Catalytic Decomposition of Potassium Monopersulfate. The Kinetics
Authors: Olga Gimeno, Javier Rivas, Maria Carbajo, Teresa Borralho
Abstract:
Potassium monopersulfate has been decomposed in aqueous solution in the presence of Co(II). The process has been simulated by means of a mechanism based on elementary reactions. Rate constants have been taken from literature reports or, alternatively, assimilated to analogous reactions occurring in Fenton's chemistry. Several operating conditions have been successfully applied.
Keywords: Monopersulfate, Oxone®, Sulfate radicals, Water treatment
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19711852 Estimation of Bio-Kinetic Coefficients for Treatment of Brewery Wastewater
Authors: Abimbola M. Enitan, Josiah Adeyemo
Abstract:
Anaerobic modeling is a useful tool to describe and simulate the condition and behaviour of anaerobic treatment units for better effluent quality and biogas generation. The present investigation deals with the anaerobic treatment of brewery wastewater with varying organic loads. The chemical oxygen demand (COD) and total suspended solids (TSS) of the influent and effluent of the bioreactor were determined at various retention times to generate data for kinetic coefficients. The bio-kinetic coefficients in the modified Stover–Kincannon kinetic and methane generation models were determined to study the performance of anaerobic digestion process. At steady-state, the determination of the kinetic coefficient (K), the endogenous decay coefficient (Kd), the maximum growth rate of microorganisms (μmax), the growth yield coefficient (Y), ultimate methane yield (Bo), maximum utilization rate constant Umax and the saturation constant (KB) in the model were calculated to be 0.046 g/g COD, 0.083 (d¯¹), 0.117 (d-¹), 0.357 g/g, 0.516 (L CH4/gCODadded), 18.51 (g/L/day) and 13.64 (g/L/day) respectively. The outcome of this study will help in simulation of anaerobic model to predict usable methane and good effluent quality during the treatment of industrial wastewater. Thus, this will protect the environment, conserve natural resources, saves time and reduce cost incur by the industries for the discharge of untreated or partially treated wastewater. It will also contribute to a sustainable long-term clean development mechanism for the optimization of the methane produced from anaerobic degradation of waste in a close system.
Keywords: Brewery wastewater, methane generation model, environment, anaerobic modeling.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 42071851 Performance of BLDC Motor under Kalman Filter Sensorless Drive
Authors: Yuri Boiko, Ci Lin, Iluju Kiringa, Tet Yeap
Abstract:
The performance of a permanent magnet brushless direct current (BLDC) motor controlled by the Kalman filter based position-sensorless drive is studied in terms of its dependence from the system’s parameters variations. The effects of the system’s parameters changes on the dynamic behavior of state variables are verified. Simulated is the closed loop control scheme with Kalman filter in the feedback line. Distinguished are two separate data sampling modes in analyzing feedback output from the BLDC motor: (1) equal angular separation and (2) equal time intervals. In case (1), the data are collected via equal intervals of rotor’s angular position i, i.e. keeping = const. In case (2), the data collection time points ti are separated by equal sampling time intervals t = const. Demonstrated are the effects of the parameters changes on the sensorless control flow, in particular, reduction of the instability torque ripples, switching spikes, and torque load balancing. It is specifically shown that an efficient suppression of commutation induced instability torque ripples is an achievable selection of the sampling rate in the Kalman filter settings above a certain critical value. The computational cost of such suppression is shown to be higher for the motors with lower induction values of the windings.
Keywords: BLDC motor, Kalman filter, sensorless drive, state variables, instability torque ripples reduction, sampling rate.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 7311850 Quantifying the UK’s Future Thermal Electricity Generation Water Use: Regional Analysis
Authors: Daniel Murrant, Andrew Quinn, Lee Chapman
Abstract:
A growing population has led to increasing global water and energy demand. This demand, combined with the effects of climate change and an increasing need to maintain and protect the natural environment, represents a potentially severe threat to many national infrastructure systems. This has resulted in a considerable quantity of published material on the interdependencies that exist between the supply of water and the thermal generation of electricity, often known as the water-energy nexus. Focusing specifically on the UK, there is a growing concern that the future availability of water may at times constrain thermal electricity generation, and therefore hinder the UK in meeting its increasing demand for a secure, and affordable supply of low carbon electricity. To provide further information on the threat the water-energy nexus may pose to the UK’s energy system, this paper models the regional water demand of UK thermal electricity generation in 2030 and 2050. It uses the strategically important Energy Systems Modelling Environment model developed by the Energy Technologies Institute. Unlike previous research, this paper was able to use abstraction and consumption factors specific to UK power stations. It finds that by 2050 the South East, Yorkshire and Humber, the West Midlands and North West regions are those with the greatest freshwater demand and therefore most likely to suffer from a lack of resource. However, it finds that by 2050 it is the East, South West and East Midlands regions with the greatest total water (fresh, estuarine and seawater) demand and the most likely to be constrained by environmental standards.
Keywords: Water-energy nexus, water resources, abstraction, climate change, power station cooling.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15491849 Analysis of Climatic Strategies in Designing the Residential Buildings in Cold Dry Climate of Tabriz Metropolis to Reduce Air Pollution in Urban Environment
Authors: Shahryar Shaghaghi G., Paria Violette Shakiba , Gholamreza Irani
Abstract:
Nowadays, the earth is countered with serious problem of air pollution. This problem has been started from the industrial revolution and has been faster in recent years, so that leads the earth to ecological and environmental disaster. One of its results is the global warming problem and its related increase in global temperature. The most important factors in air pollution especially in urban environments are Automobiles and residential buildings that are the biggest consumers of the fossil energies, so that if the residential buildings as a big part of the consumers of such energies reduce their consumption rate, the air pollution will be decreased. Since Metropolises are the main centers of air pollution in the world, assessment and analysis of efficient strategies in decreasing air pollution in such cities, can lead to the desirable and suitable results and can solve the problem at least in critical level. Tabriz city is one of the most important metropolises in North west of Iran that about two million people are living there. for its situation in cold dry climate, has a high rate of fossil energies consumption that make air pollution in its urban environment. These two factors, being both metropolis and in cold dry climate, make this article try to analyze the strategies of climatic design in old districts of the city and use them in new districts of the future. These strategies can be used in this city and other similar cities and pave the way to reduce energy consumption and related air pollution to save whole world.Keywords: Air pollution, Urban Environment, Metropolis, Residential building, Fossil energies.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17861848 Exergy Based Performance Analysis of a Gas Turbine Unit at Various Ambient Conditions
Authors: Idris A. Elfeituri
Abstract:
This paper studies the effect of ambient conditions on the performance of a 285 MW gas turbine unit using the exergy concept. Based on the available exergy balance models developed, a computer program has been constructed to investigate the performance of the power plant under varying ambient temperature and relative humidity conditions. The variations of ambient temperature range from zero to 50 ºC and the relative humidity ranges from zero to 100%, while the unit load kept constant at 100% of the design load. The exergy destruction ratio and exergy efficiency are determined for each component and for the entire plant. The results show a moderate increase in the total exergy destruction ratio of the plant from 62.05% to 65.20%, while the overall exergy efficiency decrease from 38.2% to 34.8% as the ambient temperature increases from zero to 50 ºC at all relative humidity values. Furthermore, an increase of 1 ºC in ambient temperature leads to 0.063% increase in the total exergy destruction ratio and 0.07% decrease in the overall exergy efficiency. The relative humidity has a remarkable influence at higher ambient temperature values on the exergy destruction ratio of combustion chamber and on exergy loss ratio of the exhaust gas but almost no effect on the total exergy destruction ratio and overall exergy efficiency. At 50 ºC ambient temperature, the exergy destruction ratio of the combustion chamber increases from 30% to 52% while the exergy loss ratio of the exhaust gas decreases from 28% to 8% as the relative humidity increases from zero to 100%. In addition, exergy analysis reveals that the combustion chamber and exhaust gas are the main source of irreversibility in the gas turbine unit. It is also identified that the exergy efficiency and exergy destruction ratio are considerably dependent on the variations in the ambient air temperature and relative humidity. Therefore, the incorporation of the existing gas turbine plant with inlet air cooling and humidifier technologies should be considered seriously.
Keywords: Destruction, exergy, gas turbine, irreversibility, performance.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9071847 The DAQ Debugger for iFDAQ of the COMPASS Experiment
Authors: Y. Bai, M. Bodlak, V. Frolov, S. Huber, V. Jary, I. Konorov, D. Levit, J. Novy, D. Steffen, O. Subrt, M. Virius
Abstract:
In general, state-of-the-art Data Acquisition Systems (DAQ) in high energy physics experiments must satisfy high requirements in terms of reliability, efficiency and data rate capability. This paper presents the development and deployment of a debugging tool named DAQ Debugger for the intelligent, FPGA-based Data Acquisition System (iFDAQ) of the COMPASS experiment at CERN. Utilizing a hardware event builder, the iFDAQ is designed to be able to readout data at the average maximum rate of 1.5 GB/s of the experiment. In complex softwares, such as the iFDAQ, having thousands of lines of code, the debugging process is absolutely essential to reveal all software issues. Unfortunately, conventional debugging of the iFDAQ is not possible during the real data taking. The DAQ Debugger is a tool for identifying a problem, isolating the source of the problem, and then either correcting the problem or determining a way to work around it. It provides the layer for an easy integration to any process and has no impact on the process performance. Based on handling of system signals, the DAQ Debugger represents an alternative to conventional debuggers provided by most integrated development environments. Whenever problem occurs, it generates reports containing all necessary information important for a deeper investigation and analysis. The DAQ Debugger was fully incorporated to all processes in the iFDAQ during the run 2016. It helped to reveal remaining software issues and improved significantly the stability of the system in comparison with the previous run. In the paper, we present the DAQ Debugger from several insights and discuss it in a detailed way.Keywords: DAQ debugger, data acquisition system, FPGA, system signals, Qt framework.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 894