Search results for: large grain size
512 Green Synthesized Iron Oxide Nanoparticles: A Nano-Nutrient for the Growth and Enhancement of Flax (Linum usitatissimum L.) Plant
Authors: G. Karunakaran, M. Jagathambal, N. Van Minh, E. Kolesnikov, A. Gusev, O. V. Zakharova, E. V. Scripnikova, E. D. Vishnyakova, D. Kuznetsov
Abstract:
Iron oxide nanoparticles (Fe2O3NPs) are widely used in different applications due to its ecofriendly nature and biocompatibility. Hence, in this investigation, biosynthesized Fe2O3NPs influence on flax (Linum usitatissimum L.) plant was examined. The biosynthesized nanoparticles were found to be cubic phase which is confirmed by XRD analysis. FTIR analysis confirmed the presence of functional groups corresponding to the iron oxide nanoparticle. The elemental analysis also confirmed that the obtained nanoparticle is iron oxide nanoparticle. The scanning electron microscopy and the transmission electron microscopy confirm that the average particle size was around 56 nm. The effect of Fe2O3NPs on seed germination followed by biochemical analysis was carried out using standard methods. The results obtained after four days and 11 days of seed vigor studies showed that the seedling length (cm), average number of seedling with leaves, increase in root length (cm) was found to be enhanced on treatment with iron oxide nanoparticles when compared to control. A positive correlation was noticed with the dose of the nanoparticle and plant growth, which may be due to changes in metabolic activity. Hence, to evaluate the change in metabolic activity, peroxidase and catalase activities were estimated. It was clear from the observation that higher concentration of iron oxide nanoparticles (Fe2O3NPs 1000 mg/L) has enhanced peroxidase and catalase activities and in turn plant growth. Thus, this study clearly showed that biosynthesized iron oxide nanoparticles will be an effective nano-nutrient for agriculture applications.
Keywords: Catalase, fertilizer, iron oxide nanoparticles, Linum usitatissimum L., nano-nutrient, peroxidase.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1683511 Optimum Design of Steel Space Frames by Hybrid Teaching-Learning Based Optimization and Harmony Search Algorithms
Authors: Alper Akın, İbrahim Aydoğdu
Abstract:
This study presents a hybrid metaheuristic algorithm to obtain optimum designs for steel space buildings. The optimum design problem of three-dimensional steel frames is mathematically formulated according to provisions of LRFD-AISC (Load and Resistance factor design of American Institute of Steel Construction). Design constraints such as the strength requirements of structural members, the displacement limitations, the inter-story drift and the other structural constraints are derived from LRFD-AISC specification. In this study, a hybrid algorithm by using teachinglearning based optimization (TLBO) and harmony search (HS) algorithms is employed to solve the stated optimum design problem. These algorithms are two of the recent additions to metaheuristic techniques of numerical optimization and have been an efficient tool for solving discrete programming problems. Using these two algorithms in collaboration creates a more powerful tool and mitigates each other’s weaknesses. To demonstrate the powerful performance of presented hybrid algorithm, the optimum design of a large scale steel building is presented and the results are compared to the previously obtained results available in the literature.Keywords: Optimum structural design, hybrid techniques, teaching-learning based optimization, harmony search algorithm, minimum weight, steel space frame.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2450510 Iris Recognition Based On the Low Order Norms of Gradient Components
Authors: Iman A. Saad, Loay E. George
Abstract:
Iris pattern is an important biological feature of human body; it becomes very hot topic in both research and practical applications. In this paper, an algorithm is proposed for iris recognition and a simple, efficient and fast method is introduced to extract a set of discriminatory features using first order gradient operator applied on grayscale images. The gradient based features are robust, up to certain extents, against the variations may occur in contrast or brightness of iris image samples; the variations are mostly occur due lightening differences and camera changes. At first, the iris region is located, after that it is remapped to a rectangular area of size 360x60 pixels. Also, a new method is proposed for detecting eyelash and eyelid points; it depends on making image statistical analysis, to mark the eyelash and eyelid as a noise points. In order to cover the features localization (variation), the rectangular iris image is partitioned into N overlapped sub-images (blocks); then from each block a set of different average directional gradient densities values is calculated to be used as texture features vector. The applied gradient operators are taken along the horizontal, vertical and diagonal directions. The low order norms of gradient components were used to establish the feature vector. Euclidean distance based classifier was used as a matching metric for determining the degree of similarity between the features vector extracted from the tested iris image and template features vectors stored in the database. Experimental tests were performed using 2639 iris images from CASIA V4-Interival database, the attained recognition accuracy has reached up to 99.92%.
Keywords: Iris recognition, contrast stretching, gradient features, texture features, Euclidean metric.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1965509 Design of Low Power and High Speed Digital IIR Filter in 45nm with Optimized CSA for Digital Signal Processing Applications
Authors: G. Ramana Murthy, C. Senthilpari, P. Velrajkumar, Lim Tien Sze
Abstract:
In this paper, a design methodology to implement low-power and high-speed 2nd order recursive digital Infinite Impulse Response (IIR) filter has been proposed. Since IIR filters suffer from a large number of constant multiplications, the proposed method replaces the constant multiplications by using addition/subtraction and shift operations. The proposed new 6T adder cell is used as the Carry-Save Adder (CSA) to implement addition/subtraction operations in the design of recursive section IIR filter to reduce the propagation delay. Furthermore, high-level algorithms designed for the optimization of the number of CSA blocks are used to reduce the complexity of the IIR filter. The DSCH3 tool is used to generate the schematic of the proposed 6T CSA based shift-adds architecture design and it is analyzed by using Microwind CAD tool to synthesize low-complexity and high-speed IIR filters. The proposed design outperforms in terms of power, propagation delay, area and throughput when compared with MUX-12T, MCIT-7T based CSA adder filter design. It is observed from the experimental results that the proposed 6T based design method can find better IIR filter designs in terms of power and delay than those obtained by using efficient general multipliers.
Keywords: CSA Full Adder, Delay unit, IIR filter, Low-Power, PDP, Parametric Analysis, Propagation Delay, Throughput, VLSI.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3815508 Two Dimensionnal Model for Extraction Packed Column Simulation using Finite Element Method
Authors: N. Outili, A-H. Meniai
Abstract:
Modeling transfer phenomena in several chemical engineering operations leads to the resolution of partial differential equations systems. According to the complexity of the operations mechanisms, the equations present a nonlinear form and analytical solution became difficult, we have then to use numerical methods which are based on approximations in order to transform a differential system to an algebraic one.Finite element method is one of numerical methods which can be used to obtain an accurate solution in many complex cases of chemical engineering.The packed columns find a large application like contactor for liquid-liquid systems such solvent extraction. In the literature, the modeling of this type of equipment received less attention in comparison with the plate columns.A mathematical bidimensionnal model with radial and axial dispersion, simulating packed tower extraction behavior was developed and a partial differential equation was solved using the finite element method by adopting the Galerkine model. We developed a Mathcad program, which can be used for a similar equations and concentration profiles are obtained along the column. The influence of radial dispersion was prooved and it can-t be neglected, the results were compared with experimental concentration at the top of the column in the extraction system: acetone/toluene/water.Keywords: finite element method, Galerkine method, liquidliquid extraction modelling, packed column simulation, two dimensional model
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1690507 A Hybrid Neural Network and Traditional Approach for Forecasting Lumpy Demand
Authors: A. Nasiri Pour, B. Rostami Tabar, A.Rahimzadeh
Abstract:
Accurate demand forecasting is one of the most key issues in inventory management of spare parts. The problem of modeling future consumption becomes especially difficult for lumpy patterns, which characterized by intervals in which there is no demand and, periods with actual demand occurrences with large variation in demand levels. However, many of the forecasting methods may perform poorly when demand for an item is lumpy. In this study based on the characteristic of lumpy demand patterns of spare parts a hybrid forecasting approach has been developed, which use a multi-layered perceptron neural network and a traditional recursive method for forecasting future demands. In the described approach the multi-layered perceptron are adapted to forecast occurrences of non-zero demands, and then a conventional recursive method is used to estimate the quantity of non-zero demands. In order to evaluate the performance of the proposed approach, their forecasts were compared to those obtained by using Syntetos & Boylan approximation, recently employed multi-layered perceptron neural network, generalized regression neural network and elman recurrent neural network in this area. The models were applied to forecast future demand of spare parts of Arak Petrochemical Company in Iran, using 30 types of real data sets. The results indicate that the forecasts obtained by using our proposed mode are superior to those obtained by using other methods.Keywords: Lumpy Demand, Neural Network, Forecasting, Hybrid Approach.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2680506 Solar and Wind Energy Potential Study of Lower Sindh, Pakistan for Power Generation
Authors: M. Akhlaque Ahmed, Sidra A. Shaikh, Maliha A. Siddiqui
Abstract:
Global and diffuse solar radiation on horizontal surface of Lower Sindh, namely Karachi, Hyderabad, Nawabshah were carried out using sunshine hour data of the area to assess the feasibility of solar energy utilization for power generation in Sindh province. The results obtained show a large variation in the direct and diffuse component of solar radiation in summer and winter months in Lower Sindh (50% direct and 50% diffuse for Karachi and Hyderabad). In Nawabshah area, the contribution of diffuse solar radiation is low during the monsoon months, July and August. The KT value of Nawabshah indicates a clear sky throughout almost the entire year. The percentage of diffuse radiation does not exceed more than 20%. In Nawabshah, the appearance of cloud is rare even during the monsoon months. The estimated values indicate that Nawabshah has high solar potential, whereas Karachi and Hyderabad have low solar potential. During the monsoon months the Lower part of Sindh can utilize the hybrid system with wind power. Near Karachi and Hyderabad, the wind speed ranges between 6.2 m/sec to 6.9 m/sec. A wind corridor exists near Karachi, Hyderabad, Gharo, Keti Bander and Shah Bander. The short fall of solar can be compensated by wind because in the monsoon months of July and August, wind speeds are higher in the Lower region of Sindh.Keywords: Hybrid power system, power generation, solar and wind energy potential, Lower Sindh.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1801505 Computational Design of Inhibitory Agents of BMP-Noggin Interaction to Promote Osteogenesis
Authors: Shaila Ahmed, Raghu Prasad Rao Metpally, Sreedhara Sangadala, Boojala Vijay B Reddy
Abstract:
Bone growth factors, such as Bone Morphogenic Protein-2 (BMP-2) have been approved by the FDA to replace grafting for some surgical interventions, but the high dose requirement limits its use in patients. Noggin, an extracellular protein, blocks the effect of BMP-2 by binding to BMP. Preventing the BMP-2/noggin interaction will help increase the free concentration of BMP-2 and therefore should enhance its efficacy to induce bone formation. The work presented here involves computational design of novel small molecule inhibitory agents of BMP-2/noggin interaction, based on our current understanding of BMP-2, and its known putative ligands (receptors and antagonists). A successful acquisition of such an inhibitory agent of BMP-2/noggin interaction would allow clinicians to reduce the dose required of BMP-2 protein in clinical applications to promote osteogenesis. The available crystal structures of the BMPs, its receptors, and the binding partner noggin were analyzed to identify the critical residues involved in their interaction. In presenting this study, LUDI de novo design method was utilized to perform virtual screening of a large number of compounds from a commercially available library against the binding sites of noggin to identify the lead chemical compounds that could potentially block BMP-noggin interaction with a high specificity.Keywords: Transforming growth factor-beta, Bone morphogenic proteins, Noggin, LUDI de novo design method, CAP small molecules.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1920504 Assessing the Potential of a Waste Material for Cement Replacement and the Effect of Its Fineness in Soft Soil Stabilisation
Authors: Hassnen M. Jafer, W. Atherton, F. Ruddock, E. Loffil
Abstract:
This paper represents the results of experimental work to investigate the suitability of a waste material (WM) for soft soil stabilisation. In addition, the effect of particle size distribution (PSD) of the waste material on its performance as a soil stabiliser was investigated. The WM used in this study is produced from the incineration processes in domestic energy power plant and it is available in two different grades of fineness (coarse waste material (CWM) and fine waste material (FWM)). An intermediate plasticity silty clayey soil with medium organic matter content has been used in this study. The suitability of the CWM and FWM to improve the physical and engineering properties of the selected soil was evaluated dependant on the results obtained from the consistency limits, compaction characteristics (optimum moisture content (OMC) and maximum dry density (MDD)); along with the unconfined compressive strength test (UCS). Different percentages of CWM were added to the soft soil (3, 6, 9, 12 and 15%) to produce various admixtures. Then the UCS test was carried out on specimens under different curing periods (zero, 7, 14, and 28 days) to find the optimum percentage of CWM. The optimum and other two percentages (either side of the optimum content) were used for FWM to evaluate the effect of the fineness of the WM on UCS of the stabilised soil. Results indicated that both types of the WM used in this study improved the physical properties of the soft soil where the index of plasticity (IP) was decreased significantly. IP was decreased from 21 to 13.64 and 13.10 with 12% of CWM and 15% of FWM respectively. The results of the unconfined compressive strength test indicated that 12% of CWM was the optimum and this percentage developed the UCS value from 202kPa to 500kPa for 28 days cured samples, which is equal, approximately 2.5 times the UCS value for untreated soil. Moreover, this percentage provided 1.4 times the value of UCS for stabilized soil-CWA by using FWM which recorded just under 700kPa after 28 days curing.
Keywords: Soft soil stabilisation, waste materials, fineness, and unconfined compressive strength.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2643503 BeamGA Median: A Hybrid Heuristic Search Approach
Authors: Ghada Badr, Manar Hosny, Nuha Bintayyash, Eman Albilali, Souad Larabi Marie-Sainte
Abstract:
The median problem is significantly applied to derive the most reasonable rearrangement phylogenetic tree for many species. More specifically, the problem is concerned with finding a permutation that minimizes the sum of distances between itself and a set of three signed permutations. Genomes with equal number of genes but different order can be represented as permutations. In this paper, an algorithm, namely BeamGA median, is proposed that combines a heuristic search approach (local beam) as an initialization step to generate a number of solutions, and then a Genetic Algorithm (GA) is applied in order to refine the solutions, aiming to achieve a better median with the smallest possible reversal distance from the three original permutations. In this approach, any genome rearrangement distance can be applied. In this paper, we use the reversal distance. To the best of our knowledge, the proposed approach was not applied before for solving the median problem. Our approach considers true biological evolution scenario by applying the concept of common intervals during the GA optimization process. This allows us to imitate a true biological behavior and enhance genetic approach time convergence. We were able to handle permutations with a large number of genes, within an acceptable time performance and with same or better accuracy as compared to existing algorithms.Keywords: Median problem, phylogenetic tree, permutation, genetic algorithm, beam search, genome rearrangement distance.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 979502 Standard Deviation of Mean and Variance of Rows and Columns of Images for CBIR
Authors: H. B. Kekre, Kavita Patil
Abstract:
This paper describes a novel and effective approach to content-based image retrieval (CBIR) that represents each image in the database by a vector of feature values called “Standard deviation of mean vectors of color distribution of rows and columns of images for CBIR". In many areas of commerce, government, academia, and hospitals, large collections of digital images are being created. This paper describes the approach that uses contents as feature vector for retrieval of similar images. There are several classes of features that are used to specify queries: colour, texture, shape, spatial layout. Colour features are often easily obtained directly from the pixel intensities. In this paper feature extraction is done for the texture descriptor that is 'variance' and 'Variance of Variances'. First standard deviation of each row and column mean is calculated for R, G, and B planes. These six values are obtained for one image which acts as a feature vector. Secondly we calculate variance of the row and column of R, G and B planes of an image. Then six standard deviations of these variance sequences are calculated to form a feature vector of dimension six. We applied our approach to a database of 300 BMP images. We have determined the capability of automatic indexing by analyzing image content: color and texture as features and by applying a similarity measure Euclidean distance.
Keywords: Standard deviation Image retrieval, color distribution, Variance, Variance of Variance, Euclidean distance.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3746501 Closed Greenhouse Production Systems for Smart Plant Production in Urban Areas
Authors: U. Schmidt, D. Dannehl, I. Schuch, J. Suhl, T. Rocksch, R. Salazar-Moreno, E. Fitz-Rodrigues, A. Rojano Aquilar, I. Lopez Cruz, G. Navas Gomez, R. A. Abraham, L. C. Irineo, N. G. Gilberto
Abstract:
The integration of agricultural production systems into urban areas is a challenge for the coming decades. Because of increasing greenhouse gas emission and rising resource consumption as well as costs in animal husbandry, the dietary habits of people in the 21st century have to focus on herbal foods. Intensive plant cultivation systems in large cities and megacities require a smart coupling of information, material and energy flow with the urban infrastructure in terms of Horticulture 4.0. In recent years, many puzzle pieces have been developed for these closed processes at the Humboldt University. To compile these for an urban plant production, it has to be optimized and networked with urban infrastructure systems. In the field of heat energy production, it was shown that with closed greenhouse technology and patented heat exchange and storage technology energy can be provided for heating and domestic hot water supply in the city. Closed water circuits can be drastically reducing the water requirements of plant production in urban areas. Ion sensitive sensors and new disinfection methods can help keep circulating nutrient solutions in the system for a longer time in urban plant production greenhouses.
Keywords: Semi closed, greenhouses, urban farming, solar heat collector, closed water cycles, aquaponics.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 771500 Numerical Modelling of Dust Propagation in the Atmosphere of Tbilisi City in Case of Western Background Light Air
Authors: N. Gigauri, V. Kukhalashvili, A. Surmava, L. Intskirveli, L. Gverdtsiteli
Abstract:
Tbilisi, a large city of the South Caucasus, is a junction point connecting Asia and Europe, Russia and republics of the Asia Minor. Over the last years, its atmosphere has been experienced an increasing anthropogenic load. Numerical modeling method is used for study of Tbilisi atmospheric air pollution. By means of 3D non-linear non-steady numerical model a peculiarity of city atmosphere pollution is investigated during background western light air. Dust concentration spatial and time changes are determined. There are identified the zones of high, average and less pollution, dust accumulation areas, transfer directions etc. By numerical modeling, there is shown that the process of air pollution by the dust proceeds in four stages, and they depend on the intensity of motor traffic, the micro-relief of the city, and the location of city mains. In the interval of time 06:00-09:00 the intensive growth, 09:00-15:00 a constancy or weak decrease, 18:00-21:00 an increase, and from 21:00 to 06:00 a reduction of the dust concentrations take place. The highly polluted areas are located in the vicinity of the city center and at some peripherical territories of the city, where the maximum dust concentration at 9PM is equal to 2 maximum allowable concentrations. The similar investigations conducted in case of various meteorological situations will enable us to compile the map of background urban pollution and to elaborate practical measures for ambient air protection.
Keywords: Numerical modelling, source of pollution, dust propagation, western light air.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 489499 Urban Greenery in the Greatest Polish Cities: Analysis of Spatial Concentration
Authors: Elżbieta Antczak
Abstract:
Cities offer important opportunities for economic development and for expanding access to basic services, including health care and education, for large numbers of people. Moreover, green areas (as an integral part of sustainable urban development) present a major opportunity for improving urban environments, quality of lives and livelihoods. This paper examines, using spatial concentration and spatial taxonomic measures, regional diversification of greenery in the cities of Poland. The analysis includes location quotients, Lorenz curve, Locational Gini Index, and the synthetic index of greenery and spatial statistics tools: (1) To verify the occurrence of strong concentration or dispersion of the phenomenon in time and space depending on the variable category, and, (2) To study if the level of greenery depends on the spatial autocorrelation. The data includes the greatest Polish cities, categories of the urban greenery (parks, lawns, street greenery, and green areas on housing estates, cemeteries, and forests) and the time span 2004-2015. According to the obtained estimations, most of cites in Poland are already taking measures to become greener. However, in the country there are still many barriers to well-balanced urban greenery development (e.g. uncontrolled urban sprawl, poor management as well as lack of spatial urban planning systems).
Keywords: Greenery, urban areas, regional spatial diversification and concentration, spatial taxonomic measure.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1253498 Identifying Quality Islamic Content in Community Question Answering Sites
Authors: Rabia Bibi, Muhammad Shahzad Faisal, Khalid Iqbal, Atif Inayat
Abstract:
Internet is growing rapidly and new community-based content is added by people every second. With this fast-growing community-based content, if a user requires answers of particular questions, then reviews are required from experts or community. However, it is difficult to get quality answers. The Muslim community all over the world is seeking help to get their questions and issues discussed to get answers. Online web portals of religious schools and community-based question answering sites are two big platforms to solve the issues of users. In the case of religious schools, there are experts and qualified religious scholars (mufti) who can give the expert opinion. However, the quality of community-based content cannot be guaranteed as it may not be an answer that satisfies the question of a user. Users on CQA sites may include spammers or individual criticizing the questioner instead of providing useful answers. In this paper, we research strategies to naturally distinguish the right content. As an experiment, we concentrate on Yahoo! Answers, and Quora, popular online QA sites, where questions are asked, answered, edited, and organized by a large community of users. We present the classification of data to categorize both relevant and irrelevant answers. Specifically, we demonstrate that the proposed framework can isolate quality answers from the rest with an exactness near that of people.
Keywords: Community-based question and answering, evaluation and prediction of quality answer, answer classification, Islamic content, answer ranking.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 82497 Faster Pedestrian Recognition Using Deformable Part Models
Authors: Alessandro Preziosi, Antonio Prioletti, Luca Castangia
Abstract:
Deformable part models achieve high precision in pedestrian recognition, but all publicly available implementations are too slow for real-time applications. We implemented a deformable part model algorithm fast enough for real-time use by exploiting information about the camera position and orientation. This implementation is both faster and more precise than alternative DPM implementations. These results are obtained by computing convolutions in the frequency domain and using lookup tables to speed up feature computation. This approach is almost an order of magnitude faster than the reference DPM implementation, with no loss in precision. Knowing the position of the camera with respect to horizon it is also possible prune many hypotheses based on their size and location. The range of acceptable sizes and positions is set by looking at the statistical distribution of bounding boxes in labelled images. With this approach it is not needed to compute the entire feature pyramid: for example higher resolution features are only needed near the horizon. This results in an increase in mean average precision of 5% and an increase in speed by a factor of two. Furthermore, to reduce misdetections involving small pedestrians near the horizon, input images are supersampled near the horizon. Supersampling the image at 1.5 times the original scale, results in an increase in precision of about 4%. The implementation was tested against the public KITTI dataset, obtaining an 8% improvement in mean average precision over the best performing DPM-based method. By allowing for a small loss in precision computational time can be easily brought down to our target of 100ms per image, reaching a solution that is faster and still more precise than all publicly available DPM implementations.Keywords: Autonomous vehicles, deformable part model, dpm, pedestrian recognition.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1397496 Finite Volume Method for Flow Prediction Using Unstructured Meshes
Authors: Juhee Lee, Yongjun Lee
Abstract:
In designing a low-energy-consuming buildings, the heat transfer through a large glass or wall becomes critical. Multiple layers of the window glasses and walls are employed for the high insulation. The gravity driven air flow between window glasses or wall layers is a natural heat convection phenomenon being a key of the heat transfer. For the first step of the natural heat transfer analysis, in this study the development and application of a finite volume method for the numerical computation of viscous incompressible flows is presented. It will become a part of the natural convection analysis with high-order scheme, multi-grid method, and dual-time step in the future. A finite volume method based on a fully-implicit second-order is used to discretize and solve the fluid flow on unstructured grids composed of arbitrary-shaped cells. The integrations of the governing equation are discretised in the finite volume manner using a collocated arrangement of variables. The convergence of the SIMPLE segregated algorithm for the solution of the coupled nonlinear algebraic equations is accelerated by using a sparse matrix solver such as BiCGSTAB. The method used in the present study is verified by applying it to some flows for which either the numerical solution is known or the solution can be obtained using another numerical technique available in the other researches. The accuracy of the method is assessed through the grid refinement.
Keywords: Finite volume method, fluid flow, laminar flow, unstructured grid.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1847495 Dynamic Features Selection for Heart Disease Classification
Authors: Walid MOUDANI
Abstract:
The healthcare environment is generally perceived as being information rich yet knowledge poor. However, there is a lack of effective analysis tools to discover hidden relationships and trends in data. In fact, valuable knowledge can be discovered from application of data mining techniques in healthcare system. In this study, a proficient methodology for the extraction of significant patterns from the Coronary Heart Disease warehouses for heart attack prediction, which unfortunately continues to be a leading cause of mortality in the whole world, has been presented. For this purpose, we propose to enumerate dynamically the optimal subsets of the reduced features of high interest by using rough sets technique associated to dynamic programming. Therefore, we propose to validate the classification using Random Forest (RF) decision tree to identify the risky heart disease cases. This work is based on a large amount of data collected from several clinical institutions based on the medical profile of patient. Moreover, the experts- knowledge in this field has been taken into consideration in order to define the disease, its risk factors, and to establish significant knowledge relationships among the medical factors. A computer-aided system is developed for this purpose based on a population of 525 adults. The performance of the proposed model is analyzed and evaluated based on set of benchmark techniques applied in this classification problem.Keywords: Multi-Classifier Decisions Tree, Features Reduction, Dynamic Programming, Rough Sets.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2533494 Oracle JDE Enterprise One ERP Implementation: A Case Study
Authors: Abhimanyu Pati, Krishna Kumar Veluri
Abstract:
The paper intends to bring out a real life experience encountered during actual implementation of a large scale Tier-1 Enterprise Resource Planning (ERP) system in a multi-location, discrete manufacturing organization in India, involved in manufacturing of auto components and aggregates. The business complexities, prior to the implementation of ERP, include multi-product with hierarchical product structures, geographically distributed multiple plant locations with disparate business practices, lack of inter-plant broadband connectivity, existence of disparate legacy applications for different business functions, and non-standardized codifications of products, machines, employees, and accounts apart from others. On the other hand, the manufacturing environment consisted of processes like Assemble-to-Order (ATO), Make-to-Stock (MTS), and Engineer-to-Order (ETO) with a mix of discrete and process operations. The paper has highlighted various business plan areas and concerns, prior to the implementation, with specific focus on strategic issues and objectives. Subsequently, it has dealt with the complete process of ERP implementation, starting from strategic planning, project planning, resource mobilization, and finally, the program execution. The step-by-step process provides a very good learning opportunity about the implementation methodology. At the end, various organizational challenges and lessons emerged, which will act as guidelines and checklist for organizations to successfully align and implement ERP and achieve their business objectives.
Keywords: ERP, ATO, MTS, ETO, discrete manufacturing, strategic planning.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1806493 Effect of Cow bone and Groundnut Shell Reinforced in Epoxy Resin on the Mechanical Properties and Microstructure of the Composites
Authors: O. I. Rufai, G. I. Lawal, B. O. Bolasodun, S. I. Durowaye, J. O. Etoh
Abstract:
It is an established fact that polymers have several physical limitations such as low stiffness and low resistance to impact on loading. Hence, polymers do not usually have requisite mechanical strength for application in various fields. The reinforcement by high strength fibers provides the polymer substantially enhanced mechanical properties and makes them more suitable for a large number of diverse applications. This research evaluates the effects of particulate Cow bone and Groundnut shell additions on the mechanical properties and microstructure of cow bone and groundnut shell reinforced epoxy composite in order to assess the possibility of using it as a material for engineering applications. Cow bone and groundnut shell particles reinforced with epoxy (CBRPC and GSRPC) was prepared by varying the cow bone and groundnut shell particles from 0-25 wt% with 5 wt% intervals. A Hybrid of the Cow bone and Groundnut shell (HGSCB) reinforce with epoxy was also prepared. The mechanical properties of the developed composites were investigated. Optical microscopy was used to examine the microstructure of the composites. The results revealed that mechanical properties did not increase uniformly with additions in filler but exhibited maximum properties at specific percentages of filler additions. From the Microscopic evaluation, it was discovered that homogeneity decreases with increase in % filler, this could be due to poor interfacial bonding.Keywords: Groundnut shell reinforced polymer composite (GSRPC), Cow bone reinforced polymer composite (CBRPC), Hybrid of ground nutshell and cowbone (HGSCB).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3062492 Productivity Effect of Urea Deep Placement Technology: An Empirical Analysis from Irrigation Rice Farmers in the Northern Region of Ghana
Authors: Shaibu Baanni Azumah, Ignatius Tindjina, Stella Obanyi, Tara N. Wood
Abstract:
This study examined the effect of Urea Deep Placement (UDP) technology on the output of irrigated rice farmers in the northern region of Ghana. Multi-stage sampling technique was used to select 142 rice farmers from the Golinga and Bontanga irrigation schemes, around Tamale. A treatment effect model was estimated at two stages; firstly, to determine the factors that influenced farmers’ decision to adopt the UDP technology and secondly, to determine the effect of the adoption of the UDP technology on the output of rice farmers. The significant variables that influenced rice farmers’ adoption of the UPD technology were sex of the farmer, land ownership, off-farm activity, extension service, farmer group participation and training. The results also revealed that farm size and the adoption of UDP technology significantly influenced the output of rice farmers in the northern region of Ghana. In addition to the potential of the technology to improve yields, it also presents an employment opportunity for women and youth, who are engaged in the deep placement of Urea Super Granules (USG), as well as in the transplantation of rice. It is recommended that the government of Ghana work closely with the IFDC to embed the UDP technology in the national agricultural programmes and policies. The study also recommends an effective collaboration between the government, through the Ministry of Food and Agriculture (MoFA) and the International Fertilizer Development Center (IFDC) to train agricultural extension agents on UDP technology in the rice producing areas of the country.
Keywords: Northern Ghana, output, irrigation rice farmers, treatment effect model, urea deep placement.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1129491 Fluorescence Quenching as an Efficient Tool for Sensing Application: Study on the Fluorescence Quenching of Naphthalimide Dye by Graphene Oxide
Authors: Sanaz Seraj, Shohre Rouhani
Abstract:
Recently, graphene has gained much attention because of its unique optical, mechanical, electrical, and thermal properties. Graphene has been used as a key material in the technological applications in various areas such as sensors, drug delivery, super capacitors, transparent conductor, and solar cell. It has a superior quenching efficiency for various fluorophores. Based on these unique properties, the optical sensors with graphene materials as the energy acceptors have demonstrated great success in recent years. During quenching, the emission of a fluorophore is perturbed by a quencher which can be a substrate or biomolecule, and due to this phenomenon, fluorophore-quencher has been used for selective detection of target molecules. Among fluorescence dyes, 1,8-naphthalimide is well known for its typical intramolecular charge transfer (ICT) and photo-induced charge transfer (PET) fluorophore, strong absorption and emission in the visible region, high photo stability, and large Stokes shift. Derivatives of 1,8-naphthalimides have found applications in some areas, especially fluorescence sensors. Herein, the fluorescence quenching of graphene oxide has been carried out on a naphthalimide dye as a fluorescent probe model. The quenching ability of graphene oxide on naphthalimide dye was studied by UV-VIS and fluorescence spectroscopy. This study showed that graphene is an efficient quencher for fluorescent dyes. Therefore, it can be used as a suitable candidate sensing platform. To the best of our knowledge, studies on the quenching and absorption of naphthalimide dyes by graphene oxide are rare.
Keywords: Fluorescence, graphene oxide, naphthalimide dye, quenching.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 757490 Identifying Autism Spectrum Disorder Using Optimization-Based Clustering
Authors: Sharifah Mousli, Sona Taheri, Jiayuan He
Abstract:
Autism spectrum disorder (ASD) is a complex developmental condition involving persistent difficulties with social communication, restricted interests, and repetitive behavior. The challenges associated with ASD can interfere with an affected individual’s ability to function in social, academic, and employment settings. Although there is no effective medication known to treat ASD, to our best knowledge, early intervention can significantly improve an affected individual’s overall development. Hence, an accurate diagnosis of ASD at an early phase is essential. The use of machine learning approaches improves and speeds up the diagnosis of ASD. In this paper, we focus on the application of unsupervised clustering methods in ASD, as a large volume of ASD data generated through hospitals, therapy centers, and mobile applications has no pre-existing labels. We conduct a comparative analysis using seven clustering approaches, such as K-means, agglomerative hierarchical, model-based, fuzzy-C-means, affinity propagation, self organizing maps, linear vector quantisation – as well as the recently developed optimization-based clustering (COMSEP-Clust) approach. We evaluate the performances of the clustering methods extensively on real-world ASD datasets encompassing different age groups: toddlers, children, adolescents, and adults. Our experimental results suggest that the COMSEP-Clust approach outperforms the other seven methods in recognizing ASD with well-separated clusters.
Keywords: Autism spectrum disorder, clustering, optimization, unsupervised machine learning.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 422489 A Review of the Antecedents and Consequences of Employee Engagementc
Authors: Ibrahim Hamidu Magem
Abstract:
Employee engagement has continued to gain popularity among practitioners, consultants and academicians recent years. This is due to the fact that the engaged employees are central to organizational success in today’s highly competitive and rapidly changing business environment. Employee engagement depicts a situation whereby employee’s harnessed themselves to their work roles. The importance of employee engagement to organizations cannot be overemphasized in today’s rapidly changing business environment. Organizations both large and small are constantly striving to improve their performance, retain employees, reduce absenteeism, and create loyal customers among others. To be able to achieve these organizations need a team of highly engaged employees. In line with this, the study attempts to provide a valuable framework for understanding the antecedents and consequences of employee engagement in organizations. The paper categorizes the antecedents of employee engagement into individual and organizational factors which it is assumed that the existence of such factors could result into engaged employees that will be of benefit to organizations. Therefore, it is recommended that organizations should revisit and redesign its employee engagement system to enable them attain their organizational goals and objectives. In addition, organizations should note that engagement is personal but organizational engagement programmes should be about everyone in the organization. The findings from this paper adds to existing studies about employee engagement and also provide awareness to academics and practitioners about the importance of employee engagement to improve organizations efficiency and effectiveness, as well as to impact to overall firm performance.
Keywords: Antecedent, employee engagement, job involvement, organization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1573488 Thermal Performance of an Air Heating Storing System
Authors: Mohammed A. Elhaj, Jamal S. Yassin
Abstract:
Owing to the lack of synchronization between the solar energy availability and the heat demands in a specific application, the energy storing sub-system is necessary to maintain the continuity of thermal process. The present work is dealing with an active solar heating storing system in which an air solar collector is connected to storing unit where this energy is distributed and provided to the heated space in a controlled manner. The solar collector is a box type absorber where the air flows between a number of vanes attached between the collector absorber and the bottom plate. This design can improve the efficiency due to increasing the heat transfer area exposed to the flowing air, as well as the heat conduction through the metal vanes from the top absorbing surface. The storing unit is a packed bed type where the air is coming from the air collector and circulated through the bed in order to add/remove the energy through the charging / discharging processes, respectively. The major advantage of the packed bed storage is its high degree of thermal stratification. Numerical solution of the packed bed energy storage is considered through dividing the bed into a number of equal segments for the bed particles and solved the energy equation for each segment depending on the neighbor ones. The studied design and performance parameters in the developed simulation model including, particle size, void fraction, etc. The final results showed that the collector efficiency was fluctuated between 55%-61% in winter season (January) under the climatic conditions of Misurata in Libya. Maximum temperature of 52ºC is attained at the top of the bed while the lower one is 25ºC at the end of the charging process of hot air into the bed. This distribution can satisfy the required load for the most house heating in Libya.
Keywords: Solar energy, thermal process, performance, collector, packed bed, numerical analysis, simulation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1998487 Modeling Spatial Distributions of Point and Nonpoint Source Pollution Loadings in the Great Lakes Watersheds
Authors: Chansheng He, Carlo DeMarchi
Abstract:
A physically based, spatially-distributed water quality model is being developed to simulate spatial and temporal distributions of material transport in the Great Lakes Watersheds of the U.S. Multiple databases of meteorology, land use, topography, hydrography, soils, agricultural statistics, and water quality were used to estimate nonpoint source loading potential in the study watersheds. Animal manure production was computed from tabulations of animals by zip code area for the census years of 1987, 1992, 1997, and 2002. Relative chemical loadings for agricultural land use were calculated from fertilizer and pesticide estimates by crop for the same periods. Comparison of these estimates to the monitored total phosphorous load indicates that both point and nonpoint sources are major contributors to the total nutrient loads in the study watersheds, with nonpoint sources being the largest contributor, particularly in the rural watersheds. These estimates are used as the input to the distributed water quality model for simulating pollutant transport through surface and subsurface processes to Great Lakes waters. Visualization and GIS interfaces are developed to visualize the spatial and temporal distribution of the pollutant transport in support of water management programs.
Keywords: Distributed Large Basin Runoff Model, Great LakesWatersheds, nonpoint source pollution, and point sources.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1534486 Pragati Node Popularity (PNP) Approach to Identify Congestion Hot Spots in MPLS
Authors: E. Ramaraj, A. Padmapriya
Abstract:
In large Internet backbones, Service Providers typically have to explicitly manage the traffic flows in order to optimize the use of network resources. This process is often referred to as Traffic Engineering (TE). Common objectives of traffic engineering include balance traffic distribution across the network and avoiding congestion hot spots. Raj P H and SVK Raja designed the Bayesian network approach to identify congestion hors pots in MPLS. In this approach for every node in the network the Conditional Probability Distribution (CPD) is specified. Based on the CPD the congestion hot spots are identified. Then the traffic can be distributed so that no link in the network is either over utilized or under utilized. Although the Bayesian network approach has been implemented in operational networks, it has a number of well known scaling issues. This paper proposes a new approach, which we call the Pragati (means Progress) Node Popularity (PNP) approach to identify the congestion hot spots with the network topology alone. In the new Pragati Node Popularity approach, IP routing runs natively over the physical topology rather than depending on the CPD of each node as in Bayesian network. We first illustrate our approach with a simple network, then present a formal analysis of the Pragati Node Popularity approach. Our PNP approach shows that for any given network of Bayesian approach, it exactly identifies the same result with minimum efforts. We further extend the result to a more generic one: for any network topology and even though the network is loopy. A theoretical insight of our result is that the optimal routing is always shortest path routing with respect to some considerations of hot spots in the networks.Keywords: Conditional Probability Distribution, Congestion hotspots, Operational Networks, Traffic Engineering.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1988485 Prediction of Road Accidents in Qatar by 2022
Authors: M. Abou-Amouna, A. Radwan, L. Al-kuwari, A. Hammuda, K. Al-Khalifa
Abstract:
There is growing concern over increasing incidences of road accidents and consequent loss of human life in Qatar. In light to the future planned event in Qatar, World Cup 2022; Qatar should put into consideration the future deaths caused by road accidents, and past trends should be considered to give a reasonable picture of what may happen in the future. Qatar roads should be arranged and paved in a way that accommodate high capacity of the population in that time, since then there will be a huge number of visitors from the world. Qatar should also consider the risk issues of road accidents raised in that period, and plan to maintain high level to safety strategies. According to the increase in the number of road accidents in Qatar from 1995 until 2012, an analysis of elements affecting and causing road accidents will be effectively studied. This paper aims to identify and criticize the factors that have high effect on causing road accidents in the state of Qatar, and predict the total number of road accidents in Qatar 2022. Alternative methods are discussed and the most applicable ones according to the previous researches are selected for further studies. The methods that satisfy the existing case in Qatar were the multiple linear regression model (MLR) and artificial neutral network (ANN). Those methods are analyzed and their findings are compared. We conclude that by using MLR the number of accidents in 2022 will become 355,226 accidents, and by using ANN 216,264 accidents. We conclude that MLR gave better results than ANN because the artificial neutral network doesn’t fit data with large range varieties.
Keywords: Road Safety, Prediction, Accident, Model, Qatar.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2632484 Study on Landscape Pattern Evolution of Ecological-Living-Industrial Land in Plateau Mountainous Area: A Case Study of Yuxi City, Yunnan Province
Authors: Ying Pan, Li Wu, Jing Zhou, Lan Li
Abstract:
The coordination and development of ecological-living-industrial land uses are the premise foundations for the formulation and implementation of the current land space planning, and more attention should be paid to plateau mountainous areas. This research is based on spatial analysis technology and landscape pattern index method taking Yuxi city, a typical mountainous plateau as the research area. By using relevant software such as ArcGIS10.5, Fragstats 4.2 and the four remote sensing images of Yuxi city in 1980, 1995, 2005 and 2015, the temporal-spatial evolution and differentiation pattern of ecological-living-industrial land applications have been discussed. The research results show that: (1) From the perspective of land use type change, ecological land of Yuxi city has been the main source of land from 1980 to 2015, which totally occupies more than 78%. During this period, the spatial structure of the ecological-living-industrial land changed significantly, namely, the living land. Its land area increased significantly from 0.83% of the total area in 1980 to 1.25% in 2015, the change range of ecological land and industrial land is relatively small. (2) In terms of land use landscape pattern transfer matrix, from 1980 to 2015, the industrial land and ecological land in Yuxi city have been gradually transferred to living land. (3) In the aspect of landscape pattern changes, various landscape pattern indexes of Yuxi city indicate that the fragmentation degree of landscape pattern of the ecological-living-industrial land in this region is increasing. The degree of agglomeration goes down, and the landscape types have changed from being relatively simple to relatively rich. The landscape is more diverse, but the patch size is uneven, meanwhile, the integrity of the ecological space is destroyed.
Keywords: Ecological-living-industrial land, spatio-temporal evolution, landscape pattern, plateau mountainous area.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 682483 A Rule-based Approach for Anomaly Detection in Subscriber Usage Pattern
Authors: Rupesh K. Gopal, Saroj K. Meher
Abstract:
In this report we present a rule-based approach to detect anomalous telephone calls. The method described here uses subscriber usage CDR (call detail record) data sampled over two observation periods: study period and test period. The study period contains call records of customers- non-anomalous behaviour. Customers are first grouped according to their similar usage behaviour (like, average number of local calls per week, etc). For customers in each group, we develop a probabilistic model to describe their usage. Next, we use maximum likelihood estimation (MLE) to estimate the parameters of the calling behaviour. Then we determine thresholds by calculating acceptable change within a group. MLE is used on the data in the test period to estimate the parameters of the calling behaviour. These parameters are compared against thresholds. Any deviation beyond the threshold is used to raise an alarm. This method has the advantage of identifying local anomalies as compared to techniques which identify global anomalies. The method is tested for 90 days of study data and 10 days of test data of telecom customers. For medium to large deviations in the data in test window, the method is able to identify 90% of anomalous usage with less than 1% false alarm rate.Keywords: Subscription fraud, fraud detection, anomalydetection, maximum likelihood estimation, rule based systems.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2813