Search results for: optimization algorithms
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 4682

Search results for: optimization algorithms

992 Supply Chain Network Design for Perishable Products in Developing Countries

Authors: Abhishek Jain, Kavish Kejriwal, V. Balaji Rao, Abhigna Chavda

Abstract:

Increasing environmental and social concerns are forcing companies to take a fresh view of the impact of supply chain operations on environment and society when designing a supply chain. A challenging task in today’s food industry is the distribution of high-quality food items throughout the food supply chain. Improper storage and unwanted transportation are the major hurdles in food supply chain and can be tackled by making dynamic storage facility location decisions with the distribution network. Since food supply chain in India is one of the biggest supply chains in the world, the companies should also consider environmental impact caused by the supply chain. This project proposes a multi-objective optimization model by integrating sustainability in decision-making, on distribution in a food supply chain network (SCN). A Multi-Objective Mixed-Integer Linear Programming (MOMILP) model between overall cost and environmental impact caused by the SCN is formulated for the problem. The goal of MOMILP is to determine the pareto solutions for overall cost and environmental impact caused by the supply chain. This is solved by using GAMS with CPLEX as third party solver. The outcomes of the project are pareto solutions for overall cost and environmental impact, facilities to be operated and the amount to be transferred to each warehouse during the time horizon.

Keywords: multi-objective mixed linear programming, food supply chain network, GAMS, multi-product, multi-period, environment

Procedia PDF Downloads 296
991 Heliport Remote Safeguard System Based on Real-Time Stereovision 3D Reconstruction Algorithm

Authors: Ł. Morawiński, C. Jasiński, M. Jurkiewicz, S. Bou Habib, M. Bondyra

Abstract:

With the development of optics, electronics, and computers, vision systems are increasingly used in various areas of life, science, and industry. Vision systems have a huge number of applications. They can be used in quality control, object detection, data reading, e.g., QR-code, etc. A large part of them is used for measurement purposes. Some of them make it possible to obtain a 3D reconstruction of the tested objects or measurement areas. 3D reconstruction algorithms are mostly based on creating depth maps from data that can be acquired from active or passive methods. Due to the specific appliance in airfield technology, only passive methods are applicable because of other existing systems working on the site, which can be blinded on most spectral levels. Furthermore, reconstruction is required to work long distances ranging from hundreds of meters to tens of kilometers with low loss of accuracy even with harsh conditions such as fog, rain, or snow. In response to those requirements, HRESS (Heliport REmote Safeguard System) was developed; which main part is a rotational head with a two-camera stereovision rig gathering images around the head in 360 degrees along with stereovision 3D reconstruction and point cloud combination. The sub-pixel analysis introduced in the HRESS system makes it possible to obtain an increased distance measurement resolution and accuracy of about 3% for distances over one kilometer. Ultimately, this leads to more accurate and reliable measurement data in the form of a point cloud. Moreover, the program algorithm introduces operations enabling the filtering of erroneously collected data in the point cloud. All activities from the programming, mechanical and optical side are aimed at obtaining the most accurate 3D reconstruction of the environment in the measurement area.

Keywords: airfield monitoring, artificial intelligence, stereovision, 3D reconstruction

Procedia PDF Downloads 96
990 A Novel Heuristic for Analysis of Large Datasets by Selecting Wrapper-Based Features

Authors: Bushra Zafar, Usman Qamar

Abstract:

Large data sample size and dimensions render the effectiveness of conventional data mining methodologies. A data mining technique are important tools for collection of knowledgeable information from variety of databases and provides supervised learning in the form of classification to design models to describe vital data classes while structure of the classifier is based on class attribute. Classification efficiency and accuracy are often influenced to great extent by noisy and undesirable features in real application data sets. The inherent natures of data set greatly masks its quality analysis and leave us with quite few practical approaches to use. To our knowledge first time, we present a new approach for investigation of structure and quality of datasets by providing a targeted analysis of localization of noisy and irrelevant features of data sets. Machine learning is based primarily on feature selection as pre-processing step which offers us to select few features from number of features as a subset by reducing the space according to certain evaluation criterion. The primary objective of this study is to trim down the scope of the given data sample by searching a small set of important features which may results into good classification performance. For this purpose, a heuristic for wrapper-based feature selection using genetic algorithm and for discriminative feature selection an external classifier are used. Selection of feature based on its number of occurrence in the chosen chromosomes. Sample dataset has been used to demonstrate proposed idea effectively. A proposed method has improved average accuracy of different datasets is about 95%. Experimental results illustrate that proposed algorithm increases the accuracy of prediction of different diseases.

Keywords: data mining, generic algorithm, KNN algorithms, wrapper based feature selection

Procedia PDF Downloads 297
989 The Review of Permanent Downhole Monitoring System

Authors: Jing Hu, Dong Yang

Abstract:

With the increasingly difficult development and operating environment of exploration, there are many new challenges and difficulties in developing and exploiting oil and gas resources. These include the ability to dynamically monitor wells and provide data and assurance for the completion and production of high-cost and complex wells. A key technology in providing these assurances and maximizing oilfield profitability is real-time permanent reservoir monitoring. The emergence of optical fiber sensing systems has gradually begun to replace traditional electronic systems. Traditional temperature sensors can only achieve single-point temperature monitoring, but fiber optic sensing systems based on the Bragg grating principle have a high level of reliability, accuracy, stability, and resolution, enabling cost-effective monitoring, which can be done in real-time, anytime, and without well intervention. Continuous data acquisition is performed along the entire wellbore. The integrated package with the downhole pressure gauge, packer, and surface system can also realize real-time dynamic monitoring of the pressure in some sections of the downhole, avoiding oil well intervention and eliminating the production delay and operational risks of conventional surveys. Real-time information obtained through permanent optical fibers can also provide critical reservoir monitoring data for production and recovery optimization.

Keywords: PDHM, optical fiber, coiled tubing, photoelectric composite cable, digital-oilfield

Procedia PDF Downloads 57
988 Visualization Tool for EEG Signal Segmentation

Authors: Sweeti, Anoop Kant Godiyal, Neha Singh, Sneh Anand, B. K. Panigrahi, Jayasree Santhosh

Abstract:

This work is about developing a tool for visualization and segmentation of Electroencephalograph (EEG) signals based on frequency domain features. Change in the frequency domain characteristics are correlated with change in mental state of the subject under study. Proposed algorithm provides a way to represent the change in the mental states using the different frequency band powers in form of segmented EEG signal. Many segmentation algorithms have been suggested in literature having application in brain computer interface, epilepsy and cognition studies that have been used for data classification. But the proposed method focusses mainly on the better presentation of signal and that’s why it could be a good utilization tool for clinician. Algorithm performs the basic filtering using band pass and notch filters in the range of 0.1-45 Hz. Advanced filtering is then performed by principal component analysis and wavelet transform based de-noising method. Frequency domain features are used for segmentation; considering the fact that the spectrum power of different frequency bands describes the mental state of the subject. Two sliding windows are further used for segmentation; one provides the time scale and other assigns the segmentation rule. The segmented data is displayed second by second successively with different color codes. Segment’s length can be selected as per need of the objective. Proposed algorithm has been tested on the EEG data set obtained from University of California in San Diego’s online data repository. Proposed tool gives a better visualization of the signal in form of segmented epochs of desired length representing the power spectrum variation in data. The algorithm is designed in such a way that it takes the data points with respect to the sampling frequency for each time frame and so it can be improved to use in real time visualization with desired epoch length.

Keywords: de-noising, multi-channel data, PCA, power spectra, segmentation

Procedia PDF Downloads 370
987 Sustainable Project Management Necessarily Implemented in the Chinese Wine Market Due to Climate Variation

Authors: Ruixin Zhang, Joel Carboni, Songchenchen Gong

Abstract:

Since the Sustainable Development Goals (SDGs) officially became the 17 development goals set by the United Nations in 2015, it has become an inevitable trend in project management development globally. Since Sustainability and glob-alization are the main focus and trends in the 21st century, project management contains system-based optimization, and or-ganizational humanities, environmental protection, and economic development. As a populous country globally, with the advanced development of economy and technology, China becomes one of the biggest markets in the wine industry. However, the develop-ment of society also brings specific environmental issues. Climate changes have already brought severe impacts on the Chinese wine market, including consumer behavior, wine production activities, and organizational humanities. Therefore, the implementation of sustainable project management in Chinese wine market is essential. Surveys based analysis is the primary method to interpret how the climate variation effect the Chinese wine market and the importance of sustainable project management implementation for green market growth in China. This paper proposes the CWW Conceptual model that can be used in the wine industry, the new 7 Drivers Model, and SPM Framework to interpret the main drivers that impact project management implementation in the wine industry and to offer the directions to wine companies in China which would help them to achieve the green growth.

Keywords: project management, sustainability, green growth, climate changes, Chinese wine market

Procedia PDF Downloads 103
986 A Statistical Approach to Predict and Classify the Commercial Hatchability of Chickens Using Extrinsic Parameters of Breeders and Eggs

Authors: M. S. Wickramarachchi, L. S. Nawarathna, C. M. B. Dematawewa

Abstract:

Hatchery performance is critical for the profitability of poultry breeder operations. Some extrinsic parameters of eggs and breeders cause to increase or decrease the hatchability. This study aims to identify the affecting extrinsic parameters on the commercial hatchability of local chicken's eggs and determine the most efficient classification model with a hatchability rate greater than 90%. In this study, seven extrinsic parameters were considered: egg weight, moisture loss, breeders age, number of fertilised eggs, shell width, shell length, and shell thickness. Multiple linear regression was performed to determine the most influencing variable on hatchability. First, the correlation between each parameter and hatchability were checked. Then a multiple regression model was developed, and the accuracy of the fitted model was evaluated. Linear Discriminant Analysis (LDA), Classification and Regression Trees (CART), k-Nearest Neighbors (kNN), Support Vector Machines (SVM) with a linear kernel, and Random Forest (RF) algorithms were applied to classify the hatchability. This grouping process was conducted using binary classification techniques. Hatchability was negatively correlated with egg weight, breeders' age, shell width, shell length, and positive correlations were identified with moisture loss, number of fertilised eggs, and shell thickness. Multiple linear regression models were more accurate than single linear models regarding the highest coefficient of determination (R²) with 94% and minimum AIC and BIC values. According to the classification results, RF, CART, and kNN had performed the highest accuracy values 0.99, 0.975, and 0.972, respectively, for the commercial hatchery process. Therefore, the RF is the most appropriate machine learning algorithm for classifying the breeder outcomes, which are economically profitable or not, in a commercial hatchery.

Keywords: classification models, egg weight, fertilised eggs, multiple linear regression

Procedia PDF Downloads 63
985 A Machine Learning Approach for Detecting and Locating Hardware Trojans

Authors: Kaiwen Zheng, Wanting Zhou, Nan Tang, Lei Li, Yuanhang He

Abstract:

The integrated circuit industry has become a cornerstone of the information society, finding widespread application in areas such as industry, communication, medicine, and aerospace. However, with the increasing complexity of integrated circuits, Hardware Trojans (HTs) implanted by attackers have become a significant threat to their security. In this paper, we proposed a hardware trojan detection method for large-scale circuits. As HTs introduce physical characteristic changes such as structure, area, and power consumption as additional redundant circuits, we proposed a machine-learning-based hardware trojan detection method based on the physical characteristics of gate-level netlists. This method transforms the hardware trojan detection problem into a machine-learning binary classification problem based on physical characteristics, greatly improving detection speed. To address the problem of imbalanced data, where the number of pure circuit samples is far less than that of HTs circuit samples, we used the SMOTETomek algorithm to expand the dataset and further improve the performance of the classifier. We used three machine learning algorithms, K-Nearest Neighbors, Random Forest, and Support Vector Machine, to train and validate benchmark circuits on Trust-Hub, and all achieved good results. In our case studies based on AES encryption circuits provided by trust-hub, the test results showed the effectiveness of the proposed method. To further validate the method’s effectiveness for detecting variant HTs, we designed variant HTs using open-source HTs. The proposed method can guarantee robust detection accuracy in the millisecond level detection time for IC, and FPGA design flows and has good detection performance for library variant HTs.

Keywords: hardware trojans, physical properties, machine learning, hardware security

Procedia PDF Downloads 113
984 Optimization of the Drinking Water Treatment Process

Authors: M. Farhaoui, M. Derraz

Abstract:

Problem statement: In the water treatment processes, the coagulation and flocculation processes produce sludge according to the level of the water turbidity. The aluminum sulfate is the most common coagulant used in water treatment plants of Morocco as well as many countries. It is difficult to manage the sludge produced by the treatment plant. However, it can be used in the process to improve the quality of the treated water and reduce the aluminum sulfate dose. Approach: In this study, the effectiveness of sludge was evaluated at different turbidity levels (low, medium, and high turbidity) and coagulant dosage to find optimal operational conditions. The influence of settling time was also studied. A set of jar test experiments was conducted to find the sludge and aluminum sulfate dosages in order to improve the produced water quality for different turbidity levels. Results: Results demonstrated that using sludge produced by the treatment plant can improve the quality of the produced water and reduce the aluminum sulfate using. The aluminum sulfate dosage can be reduced from 40 to 50% according to the turbidity level (10, 20 and 40 NTU). Conclusions/Recommendations: Results show that sludge can be used in order to reduce the aluminum sulfate dosage and improve the quality of treated water. The highest turbidity removal efficiency is observed within 6 mg/l of aluminum sulfate and 35 mg/l of sludge in low turbidity, 20 mg/l of aluminum sulfate and 50 mg/l of sludge in medium turbidity and 20 mg/l of aluminum sulfate and 60 mg/l of sludge in high turbidity. The turbidity removal efficiency is 97.56%, 98.96% and 99.47% respectively for low, medium and high turbidity levels.

Keywords: coagulation process, coagulant dose, sludge, turbidity removal

Procedia PDF Downloads 301
983 Fabrication of Pure and Doped MAPbI3 Thin Films by One Step Chemical Vapor Deposition Method for Energy Harvesting Applications

Authors: S. V. N. Pammi, Soon-Gil Yoon

Abstract:

In the present study, we report a facile chemical vapor deposition (CVD) method for Perovskite MAPbI3 thin films by doping with Br and Cl. We performed a systematic optimization of CVD parameters such as deposition temperature, working pressure and annealing time and temperature to obtain high-quality films of CH3NH3PbI3, CH3NH3PbI3-xBrx and CH3NH3PbI3-xClx perovskite. Scanning electron microscopy and X-ray Diffraction pattern showed that the perovskite films have a large grain size when compared to traditional spin coated thin films. To the best of our knowledge, there are very few reports on highly quality perovskite thin films by various doping such as Br and Cl using one step CVD and there is scope for significant improvement in device efficiency. In addition, their band-gap can be conveniently and widely tuned via doping process. This deposition process produces perovskite thin films with large grain size, long diffusion length and high surface coverage. The enhancement of the output power, CH3NH3PbI3 (MAPbI3) dye films when compared to spin coated films and enhancement in output power by doping in doped films was demonstrated in detail. The facile one-step method for deposition of perovskite thin films shows a potential candidate for photovoltaic and energy harvesting applications.

Keywords: perovskite thin films, chemical vapor deposition, energy harvesting, photovoltaics

Procedia PDF Downloads 281
982 Fusion Neutron Generator Dosimetry and Applications for Medical, Security, and Industry

Authors: Kaouther Bergaui, Nafaa Reguigui, Charles Gary

Abstract:

Characterization and the applications of deuterium-deuterium (DD) neutron generator developed by Adelphie technology and acquired by the National Centre of Nuclear Science and Technology (NCNST) were presented in this work. We study the performance of the neutron generator in terms of neutron yield, production efficiency, and the ionic current as a function of the acceleration voltage at various RF powers. We provide the design and optimization of the PGNAA chamber and thus give insight into the capabilities of the planned PGNAA facility. Additional non-destructive techniques were studied employing the DD neutron generator, such as PGNAA and neutron radiography: The PGNAA is used for determining the concentration of 10B in Si and SiO2 matrices by using a germanium detector HPGe and the results obtained are compared with PGNAA system using a Sodium Iodide detector (NaI (Tl)); Neutron radiography facility was tested and simulated, using a camera device CCD and simulated by the Monte Carlo code; and the explosive detection system (EDS) also simulated using the Monte Carlo code. The study allows us to show that the new models of DD neutron generators are feasible and that superior-quality neutron beams could be produced and used for various applications. The feasibility of Boron neutron capture therapy (BNCT) for cancer treatment using a neutron generator was assessed by optimizing Beam Shaping Assembly (BSA) on a phantom using Monte-Carlo (MCNP6) simulations.

Keywords: neutron generator deuterium-deuterium, Monte Carlo method, radiation, neutron flux, neutron activation analysis, born, neutron radiography, explosive detection, BNCT

Procedia PDF Downloads 164
981 Feature Evaluation Based on Random Subspace and Multiple-K Ensemble

Authors: Jaehong Yu, Seoung Bum Kim

Abstract:

Clustering analysis can facilitate the extraction of intrinsic patterns in a dataset and reveal its natural groupings without requiring class information. For effective clustering analysis in high dimensional datasets, unsupervised dimensionality reduction is an important task. Unsupervised dimensionality reduction can generally be achieved by feature extraction or feature selection. In many situations, feature selection methods are more appropriate than feature extraction methods because of their clear interpretation with respect to the original features. The unsupervised feature selection can be categorized as feature subset selection and feature ranking method, and we focused on unsupervised feature ranking methods which evaluate the features based on their importance scores. Recently, several unsupervised feature ranking methods were developed based on ensemble approaches to achieve their higher accuracy and stability. However, most of the ensemble-based feature ranking methods require the true number of clusters. Furthermore, these algorithms evaluate the feature importance depending on the ensemble clustering solution, and they produce undesirable evaluation results if the clustering solutions are inaccurate. To address these limitations, we proposed an ensemble-based feature ranking method with random subspace and multiple-k ensemble (FRRM). The proposed FRRM algorithm evaluates the importance of each feature with the random subspace ensemble, and all evaluation results are combined with the ensemble importance scores. Moreover, FRRM does not require the determination of the true number of clusters in advance through the use of the multiple-k ensemble idea. Experiments on various benchmark datasets were conducted to examine the properties of the proposed FRRM algorithm and to compare its performance with that of existing feature ranking methods. The experimental results demonstrated that the proposed FRRM outperformed the competitors.

Keywords: clustering analysis, multiple-k ensemble, random subspace-based feature evaluation, unsupervised feature ranking

Procedia PDF Downloads 309
980 The Influence of Different Flux Patterns on Magnetic Losses in Electric Machine Cores

Authors: Natheer Alatawneh

Abstract:

The finite element analysis of magnetic fields in electromagnetic devices shows that the machine cores experience different flux patterns including alternating and rotating fields. The rotating fields are generated in different configurations range between circular and elliptical with different ratios between the major and minor axis of the flux locus. Experimental measurements on electrical steel exposed to different flux patterns disclose different magnetic losses in the samples under test. Consequently, electric machines require special attention during the cores loss calculation process to consider the flux patterns. In this study, a circular rotational single sheet tester is employed to measure the core losses in electric steel sample of M36G29. The sample was exposed to alternating field, circular field, and elliptical fields with axis ratios of 0.2, 0.4, 0.6 and 0.8. The measured data was implemented on 6-4 switched reluctance motor at three different frequencies of interest to the industry as 60 Hz, 400 Hz, and 1 kHz. The results disclose a high margin of error that may occur during the loss calculations if the flux patterns issue is neglected. The error in different parts of the machine associated with considering the flux patterns can be around 50%, 10%, and 2% at 60Hz, 400Hz, and 1 kHz, respectively. The future work will focus on the optimization of machine geometrical shape which has a primary effect on the flux pattern in order to minimize the magnetic losses in machine cores.

Keywords: alternating core losses, electric machines, finite element analysis, rotational core losses

Procedia PDF Downloads 231
979 From Creativity to Innovation: Tracking Rejected Ideas

Authors: Lisete Barlach, Guilherme Ary Plonski

Abstract:

Innovative ideas are not always synonymous with business opportunities. Any idea can be creative and not recognized as a potential project in which money and time will be invested, among other resources. Even in firms that promote and enhance innovation, there are two 'check-points', the first corresponding to the acknowledgment of the idea as creative and the second, its consideration as a business opportunity. Both the recognition of new business opportunities or new ideas involve cognitive and psychological frameworks which provide individuals with a basis for noticing connections between seemingly independent events or trends as if they were 'connecting the dots'. It also involves prototypes-representing the most typical member of a certain category–functioning as 'templates' for this recognition. There is a general assumption that these kinds of evaluation processes develop through experience, explaining why expertise plays a central role in this process: the more experienced a professional, the easier for him (her) to identify new opportunities in business. But, paradoxically, an increase in expertise can lead to the inflexibility of thought due to automation of procedures. And, besides this, other cognitive biases can also be present, because new ideas or business opportunities generally depend on heuristics, rather than on established algorithms. The paper presents a literature review about the Einstellung effect by tracking famous cases of rejected ideas, extracted from historical records. It also presents the results of empirical research, with data upon rejected ideas gathered from two different environments: projects rejected during first semester of 2017 at a large incubator center in Sao Paulo and ideas proposed by employees that were rejected by a well-known business company, at its Brazilian headquarter. There is an implicit assumption that Einstellung effect tends to be more and more present in contemporaneity, due to time pressure upon decision-making and idea generation process. The analysis discusses desirability, viability, and feasibility as elements that affect decision-making.

Keywords: cognitive biases, Einstellung effect, recognition of business opportunities, rejected ideas

Procedia PDF Downloads 170
978 Analytical Slope Stability Analysis Based on the Statistical Characterization of Soil Shear Strength

Authors: Bernardo C. P. Albuquerque, Darym J. F. Campos

Abstract:

Increasing our ability to solve complex engineering problems is directly related to the processing capacity of computers. By means of such equipments, one is able to fast and accurately run numerical algorithms. Besides the increasing interest in numerical simulations, probabilistic approaches are also of great importance. This way, statistical tools have shown their relevance to the modelling of practical engineering problems. In general, statistical approaches to such problems consider that the random variables involved follow a normal distribution. This assumption tends to provide incorrect results when skew data is present since normal distributions are symmetric about their means. Thus, in order to visualize and quantify this aspect, 9 statistical distributions (symmetric and skew) have been considered to model a hypothetical slope stability problem. The data modeled is the friction angle of a superficial soil in Brasilia, Brazil. Despite the apparent universality, the normal distribution did not qualify as the best fit. In the present effort, data obtained in consolidated-drained triaxial tests and saturated direct shear tests have been modeled and used to analytically derive the probability density function (PDF) of the safety factor of a hypothetical slope based on Mohr-Coulomb rupture criterion. Therefore, based on this analysis, it is possible to explicitly derive the failure probability considering the friction angle as a random variable. Furthermore, it is possible to compare the stability analysis when the friction angle is modelled as a Dagum distribution (distribution that presented the best fit to the histogram) and as a Normal distribution. This comparison leads to relevant differences when analyzed in light of the risk management.

Keywords: statistical slope stability analysis, skew distributions, probability of failure, functions of random variables

Procedia PDF Downloads 311
977 Conservation Planning of Paris Polyphylla Smith, an Important Medicinal Herb of the Indian Himalayan Region Using Predictive Distribution Modelling

Authors: Mohd Tariq, Shyamal K. Nandi, Indra D. Bhatt

Abstract:

Paris polyphylla Smith (Family- Liliaceae; English name-Love apple: Local name- Satuwa) is an important folk medicinal herb of the Indian subcontinent, being a source of number of bioactive compounds for drug formulation. The rhizomes are widely used as antihelmintic, antispasmodic, digestive stomachic, expectorant and vermifuge, antimicrobial, anti-inflammatory, heart and vascular malady, anti-fertility and sedative. Keeping in view of this, the species is being constantly removed from nature for trade and various pharmaceuticals purpose, as a result, the availability of the species in its natural habitat is decreasing. In this context, it would be pertinent to conserve this species and reintroduce them in its natural habitat. Predictive distribution modelling of this species was performed in Western Himalayan Region. One such recent method is Ecological Niche Modelling, also popularly known as Species distribution modelling, which uses computer algorithms to generate predictive maps of species distributions in a geographic space by correlating the point distributional data with a set of environmental raster data. In case of P. polyphylla, and to understand its potential distribution zones and setting up of artificial introductions, or selecting conservation sites, and conservation and management of their native habitat. Among the different districts of Uttarakhand (28°05ˈ-31°25ˈ N and 77°45ˈ-81°45ˈ E) Uttarkashi, Rudraprayag, Chamoli, Pauri Garhwal and some parts of Bageshwar, 'Maximum Entropy' (Maxent) has predicted wider potential distribution of P. polyphylla Smith. Distribution of P. polyphylla is mainly governed by Precipitation of Driest Quarter and Mean Diurnal Range i.e., 27.08% and 18.99% respectively which indicates that humidity (27%) and average temperature (19°C) might be suitable for better growth of Paris polyphylla.

Keywords: biodiversity conservation, Indian Himalayan region, Paris polyphylla, predictive distribution modelling

Procedia PDF Downloads 305
976 Preparation of Corn Flour Based Extruded Product and Evaluate Its Physical Characteristics

Authors: C. S. Saini

Abstract:

The composite flour blend consisting of corn, pearl millet, black gram and wheat bran in the ratio of 80:5:10:5 was taken to prepare the extruded product and their effect on physical properties of extrudate was studied. The extrusion process was conducted in laboratory by using twin screw extruder. The physical characteristics evaluated include lateral expansion, bulk density, water absorption index, water solubility index, rehydration ratio and moisture retention. The Central Composite Rotatable Design (CCRD) was used to decide the level of processing variables i.e. feed moisture content (%), screw speed (rpm), and barrel temperature (oC) for the experiment. The data obtained after extrusion process were analyzed by using response surface methodology. A second order polynomial model for the dependent variables was established to fit the experimental data. The numerical optimization studies resulted in 127°C of barrel temperature, 246 rpm of screw speed, and 14.5% of feed moisture as optimum variables to produce acceptable extruded product. The responses predicted by the software for the optimum process condition resulted in lateral expansion 126 %, bulk density 0.28 g/cm3, water absorption index 4.10 g/g, water solubility index 39.90 %, rehydration ratio 544 % and moisture retention 11.90 % with 75 % desirability.

Keywords: black gram, corn flour, extrusion, physical characteristics

Procedia PDF Downloads 453
975 Case Analysis of Bamboo Based Social Enterprises in India-Improving Profitability and Sustainability

Authors: Priyal Motwani

Abstract:

The current market for bamboo products in India is about Rs. 21000 crores and is highly unorganised and fragmented. In this study, we have closely analysed the structure and functions of a major bamboo craft based organisation in Kerela, India and elaborated about its value chain, product mix, pricing strategy and supply chain, collaborations and competitive landscape. We have identified six major bottlenecks that are prevalent in such organisations, based on the Indian context, in relation to their product mix, asset management, and supply chain- corresponding waste management and retail network. The study has identified that the target customers for the bamboo based products and alternative revenue streams (eco-tourism, microenterprises, training), by carrying out secondary and primary research (5000 sample space), that can boost the existing revenue by 150%. We have then recommended an optimum product mix-covering premium, medium and low valued processing, for medium sized bamboo based organisations, in accordance with their capacity to maximize their revenue potential. After studying such organisations and their counter parts, the study has established an optimum retail network, considering B2B, B2C physical and online retail, to maximize their sales to their target groups. On the basis of the results obtained from the analysis of the future and present trends, our study gives recommendations to improve the revenue potential of bamboo based organisation in India and promote sustainability.

Keywords: bamboo, bottlenecks, optimization, product mix, retail network, value chain

Procedia PDF Downloads 192
974 Effect of Concentration Level and Moisture Content on the Detection and Quantification of Nickel in Clay Agricultural Soil in Lebanon

Authors: Layan Moussa, Darine Salam, Samir Mustapha

Abstract:

Heavy metal contamination in agricultural soils in Lebanon poses serious environmental and health problems. Intensive efforts are employed to improve existing quantification methods of heavy metals in contaminated environments since conventional detection techniques have shown to be time-consuming, tedious, and costly. The implication of hyperspectral remote sensing in this field is possible and promising. However, factors impacting the efficiency of hyperspectral imaging in detecting and quantifying heavy metals in agricultural soils were not thoroughly studied. This study proposes to assess the use of hyperspectral imaging for the detection of Ni in agricultural clay soil collected from the Bekaa Valley, a major agricultural area in Lebanon, under different contamination levels and soil moisture content. Soil samples were contaminated with Ni, with concentrations ranging from 150 mg/kg to 4000 mg/kg. On the other hand, soil with background contamination was subjected to increased moisture levels varying from 5 to 75%. Hyperspectral imaging was used to detect and quantify Ni contamination in the soil at different contamination levels and moisture content. IBM SPSS statistical software was used to develop models that predict the concentration of Ni and moisture content in agricultural soil. The models were constructed using linear regression algorithms. The spectral curves obtained reflected an inverse correlation between both Ni concentration and moisture content with respect to reflectance. On the other hand, the models developed resulted in high values of predicted R2 of 0.763 for Ni concentration and 0.854 for moisture content. Those predictions stated that Ni presence was well expressed near 2200 nm and that of moisture was at 1900 nm. The results from this study would allow us to define the potential of using the hyperspectral imaging (HSI) technique as a reliable and cost-effective alternative for heavy metal pollution detection in contaminated soils and soil moisture prediction.

Keywords: heavy metals, hyperspectral imaging, moisture content, soil contamination

Procedia PDF Downloads 72
973 Quantitative Evaluation of Supported Catalysts Key Properties from Electron Tomography Studies: Assessing Accuracy Using Material-Realistic 3D-Models

Authors: Ainouna Bouziane

Abstract:

The ability of Electron Tomography to recover the 3D structure of catalysts, with spatial resolution in the subnanometer scale, has been widely explored and reviewed in the last decades. A variety of experimental techniques, based either on Transmission Electron Microscopy (TEM) or Scanning Transmission Electron Microscopy (STEM) have been used to reveal different features of nanostructured catalysts in 3D, but High Angle Annular Dark Field imaging in STEM mode (HAADF-STEM) stands out as the most frequently used, given its chemical sensitivity and avoidance of imaging artifacts related to diffraction phenomena when dealing with crystalline materials. In this regard, our group has developed a methodology that combines image denoising by undecimated wavelet transforms (UWT) with automated, advanced segmentation procedures and parameter selection methods using CS-TVM (Compressed Sensing-total variation minimization) algorithms to reveal more reliable quantitative information out of the 3D characterization studies. However, evaluating the accuracy of the magnitudes estimated from the segmented volumes is also an important issue that has not been properly addressed yet, because a perfectly known reference is needed. The problem particularly complicates in the case of multicomponent material systems. To tackle this key question, we have developed a methodology that incorporates volume reconstruction/segmentation methods. In particular, we have established an approach to evaluate, in quantitative terms, the accuracy of TVM reconstructions, which considers the influence of relevant experimental parameters like the range of tilt angles, image noise level or object orientation. The approach is based on the analysis of material-realistic, 3D phantoms, which include the most relevant features of the system under analysis.

Keywords: electron tomography, supported catalysts, nanometrology, error assessment

Procedia PDF Downloads 57
972 A Novel Epitope Prediction for Vaccine Designing against Ebola Viral Envelope Proteins

Authors: Manju Kanu, Subrata Sinha, Surabhi Johari

Abstract:

Viral proteins of Ebola viruses belong to one of the best studied viruses; however no effective prevention against EBOV has been developed. Epitope-based vaccines provide a new strategy for prophylactic and therapeutic application of pathogen-specific immunity. A critical requirement of this strategy is the identification and selection of T-cell epitopes that act as vaccine targets. This study describes current methodologies for the selection process, with Ebola virus as a model system. Hence great challenge in the field of ebola virus research is to design universal vaccine. A combination of publicly available bioinformatics algorithms and computational tools are used to screen and select antigen sequences as potential T-cell epitopes of supertypes Human Leukocyte Antigen (HLA) alleles. MUSCLE and MOTIF tools were used to find out most conserved peptide sequences of viral proteins. Immunoinformatics tools were used for prediction of immunogenic peptides of viral proteins in zaire strains of Ebola virus. Putative epitopes for viral proteins (VP) were predicted from conserved peptide sequences of VP. Three tools NetCTL 1.2, BIMAS and Syfpeithi were used to predict the Class I putative epitopes while three tools, ProPred, IEDB-SMM-align and NetMHCII 2.2 were used to predict the Class II putative epitopes. B cell epitopes were predicted by BCPREDS 1.0. Immunogenic peptides were identified and selected manually by putative epitopes predicted from online tools individually for both MHC classes. Finally sequences of predicted peptides for both MHC classes were looked for common region which was selected as common immunogenic peptide. The immunogenic peptides were found for viral proteins of Ebola virus: epitopes FLESGAVKY, SSLAKHGEY. These predicted peptides could be promising candidates to be used as target for vaccine design.

Keywords: epitope, b cell, immunogenicity, ebola

Procedia PDF Downloads 286
971 Durian Marker Kit for Durian (Durio zibethinus Murr.) Identity

Authors: Emma K. Sales

Abstract:

Durian is the flagship fruit of Mindanao and there is an abundance of several cultivars with many confusing identities/ names. The project was conducted to develop procedure for reliable and rapid detection and sorting of durian planting materials. Moreover, it is also aimed to establish specific genetic or DNA markers for routine testing and authentication of durian cultivars in question. The project developed molecular procedures for routine testing. SSR primers were also screened and identified for their utility in discriminating durian cultivars collected. Results of the study showed the following accomplishments; 1. Twenty (29) SSR primers were selected and identified based on their ability to discriminate durian cultivars, 2. Optimized and established standard procedure for identification and authentication of Durian cultivars 3. Genetic profile of durian is now available at Biotech Unit. Our results demonstrate the relevance of using molecular techniques in evaluating and identifying durian clones. The most polymorphic primers tested in this study could be useful tools for detecting variation even at the early stage of the plant especially for commercial purposes. The process developed combines the efficiency of the microsatellites development process with the optimization of non-radioactive detection process resulting in a user-friendly protocol that can be performed in two (2) weeks and easily incorporated into laboratories about to start microsatellite development projects. This can be of great importance to extend microsatellite analyses to other crop species where minimal genetic information is currently available. With this, the University can now be a service laboratory for routine testing and authentication of durian clones.

Keywords: DNA, SSR analysis, genotype, genetic diversity, cultivars

Procedia PDF Downloads 429
970 Cleaner Production Options for Fishery Wastes Around Lake Tana-Ethiopia

Authors: Abate Getnet Demisash, Beshatu Taye Hatew, Ababo Geleta Gudisa

Abstract:

As consumption trends of fish are rising in Ethiopia, assessment of the environmental performance of Fisheries becomes vital. Hence, Cleaner Production Assessment was conducted on Lake Tana No.1 Fish Supply Association. This paper focuses on determining the characteristics, quantity and setting up cleaner production option for the site with experimental investigation. The survey analysis showed that illegal waste dumping in Lake Tana is common practice in the area and some of the main reasons raised were they have no option than doing this for discharging fish wastes. Quantifying a fish waste by examination of records at the point of generation resulted in generation rate of 72,822.61 kg per year which is a significant amount of waste and needs management system. The result of the proximate analysis showed high free fat content of about 12.33% and this was a good candidate for the production of biodiesel that has been set as an option for fish waste utilization. Among the different waste management options, waste reduction by product optimization which involves biodiesel production was chosen as a potential method. Laboratory scale experiments were performed to produce renewable energy source from the wastes. The resulting biodiesel was characterized and found to have a density of 0.756kg/L, viscosity 0.24p and 153°C flash points which shows the product has values in compliance with American Society for Testing and Materials (ASTM) standards.

Keywords: biodiesel, cleaner production, renewable energy, clean energy, waste to energy

Procedia PDF Downloads 111
969 Technology Valuation of Unconventional Gas R&D Project Using Real Option Approach

Authors: Young Yoon, Jinsoo Kim

Abstract:

The adoption of information and communication technologies (ICT) in all industry is growing under industry 4.0. Many oil companies also are increasingly adopting ICT to improve the efficiency of existing operations, take more accurate and quicker decision making and reduce entire cost by optimization. It is true that ICT is playing an important role in the process of unconventional oil and gas development and companies must take advantage of ICT to gain competitive advantage. In this study, real option approach has been applied to Unconventional gas R&D project to evaluate ICT of them. Many unconventional gas reserves such as shale gas and coal-bed methane(CBM) has developed due to technological improvement and high energy price. There are many uncertainties in unconventional development on the three stage(Exploration, Development, Production). The traditional quantitative benefits-cost method, such as net present value(NPV) is not sufficient for capturing ICT value. We attempted to evaluate the ICT valuation by applying the compound option model; the model is applied to real CBM project case, showing how it consider uncertainties. Variables are treated as uncertain and a Monte Carlo simulation is performed to consider variables effect. Acknowledgement—This work was supported by the Energy Efficiency & Resources Core Technology Program of the Korea Institute of Energy Technology Evaluation and Planning (KETEP) granted financial resource from the Ministry of Trade, Industry & Energy, Republic of Korea (No. 20152510101880) and by the National Research Foundation of Korea Grant funded by the Korean Government (NRF-205S1A3A2046684).

Keywords: information and communication technologies, R&D, real option, unconventional gas

Procedia PDF Downloads 207
968 Physical Characterization of a Watershed for Correlation with Parameters of Thomas Hydrological Model and Its Application in Iber Hidrodinamic Model

Authors: Carlos Caro, Ernest Blade, Nestor Rojas

Abstract:

This study determined the relationship between basic geo-technical parameters and parameters of the hydro logical model Thomas for water balance of rural watersheds, as a methodological calibration application, applicable in distributed models as IBER model, which represents a distributed system simulation models for unsteady flow numerical free surface. There was an exploration in 25 points (on 15 sub) basin of Rio Piedras (Boy.) obtaining soil samples, to which geo-technical characterization was performed by laboratory tests. Thomas model has a physical characterization of the input area by only four parameters (a, b, c, d). Achieve measurable relationship between geo technical parameters and 4 values of hydro logical parameters helps to determine subsurface, underground and surface flow more agile manner. It is intended in this way to reach some solutions regarding limits initial model parameters on the basis of Thomas geo-technical characterization. In hydro geological models of rural watersheds, calibration is an important process in the characterization of the study area. This step can require a significant computational cost and time, especially if the initial values or parameters before calibration are outside of the geo-technical reality. A better approach in these initial values means optimization of these process through a geo-technical materials area, where is obtained an important approach to the study as in the starting range of variation for the calibration parameters.

Keywords: distributed hydrology, hydrological and geotechnical characterization, Iber model

Procedia PDF Downloads 488
967 Removal of Heavy Metals by KOH Activated Diplotaxis harra Biomass: Experimental Design Optimization

Authors: H. Tounsadi, A. Khalidi, M. Abdennouri, N. Barka

Abstract:

The objective of this study was to produce high quality activated carbons from Diplotaxis harra biomass by potassium hydroxide activation and their application for heavy metals removal. To reduce the number of experiments, full factorial experimental design at two levels were carried out to occur optimal preparation conditions and better conditions for the removal of cadmium and cobalt ions from aqueous solutions. The influence of different variables during the activation process, such as carbonization temperature, activation temperature, activation time and impregnation ratio (g KOH/g carbon) have been investigated, and the best production conditions were determined. The experimental results showed that removal of cadmium and cobalt ions onto activated carbons was more sensitive to methylene blue index instead of iodine number. Although, the removal of cadmium and cobalt ions is more influenced by activation temperature with a negative effect followed by the impregnation ratio with a positive impact. Based on the statistical data, the best conditions for the removal of cadmium and cobalt by prepared activated carbons have been established. The maximum iodine number and methylene blue index obtained under these conditions and the greater sorption capacities for cadmium and cobalt were investigated. These sorption capacities were greater than those of a commercial activated carbon used in water treatment.

Keywords: activated carbon, cadmium, cobalt, Diplotaxis harra, experimental design, potassium hydroxide

Procedia PDF Downloads 175
966 The Design of Fire in Tube Boiler

Authors: Yoftahe Nigussie

Abstract:

This report presents a final year project pertaining to the design of Fire tube boiler for the purpose of producing saturated steam. The objective of the project is to produce saturated steam for different purpose with a capacity of 2000kg/h at 12bar design pressure by performing a design of a higher performance fire tube boiler that considered the requirements of cost minimization and parameters improvement. This is mostly done in selection of appropriate material for component parts, construction materials and production methods in different steps of analysis. In the analysis process, most of the design parameters are obtained by iterating with related formulas like selection of diameter of tubes with overall heat transfer coefficient optimization, and the other selections are also as like considered. The number of passes is two because of the size and area of the tubes and shell. As the analysis express by using heavy oil fuel no6 with a higher heating value of 44000kJ/kg and lower heating value of 41300kJ/kg and the amount of fuel consumed 140.37kg/hr. and produce 1610kw of heat with efficiency of 85.25%. The flow of the fluid is a cross flow because of its own advantage and the arrangement of the tube in-side the shell is welded with the tube sheet, and the tube sheet is attached with the shell and the end by using a gasket and weld. The design of the shell, using European Standard code section, is as like pressure vessel by considering the weight, including content and the supplementary accessories such as lifting lugs, openings, ends, man hole and supports with detail and assembly drawing.

Keywords: steam generation, external treatment, internal treatment, steam velocity

Procedia PDF Downloads 48
965 Estimation of Bio-Kinetic Coefficients for Treatment of Brewery Wastewater

Authors: Abimbola M. Enitan, J. Adeyemo

Abstract:

Anaerobic modeling is a useful tool to describe and simulate the condition and behaviour of anaerobic treatment units for better effluent quality and biogas generation. The present investigation deals with the anaerobic treatment of brewery wastewater with varying organic loads. The chemical oxygen demand (COD) and total suspended solids (TSS) of the influent and effluent of the bioreactor were determined at various retention times to generate data for kinetic coefficients. The bio-kinetic coefficients in the modified Stover–Kincannon kinetic and methane generation models were determined to study the performance of anaerobic digestion process. At steady-state, the determination of the kinetic coefficient (K), the endogenous decay coefficient (Kd), the maximum growth rate of microorganisms (µmax), the growth yield coefficient (Y), ultimate methane yield (Bo), maximum utilization rate constant Umax and the saturation constant (KB) in the model were calculated to be 0.046 g/g COD, 0.083 (dˉ¹), 0.117 (d-¹), 0.357 g/g, 0.516 (L CH4/gCODadded), 18.51 (g/L/day) and 13.64 (g/L/day) respectively. The outcome of this study will help in simulation of anaerobic model to predict usable methane and good effluent quality during the treatment of industrial wastewater. Thus, this will protect the environment, conserve natural resources, saves time and reduce cost incur by the industries for the discharge of untreated or partially treated wastewater. It will also contribute to a sustainable long-term clean development mechanism for the optimization of the methane produced from anaerobic degradation of waste in a close system.

Keywords: brewery wastewater, methane generation model, environment, anaerobic modeling

Procedia PDF Downloads 230
964 An Assessment of Floodplain Vegetation Response to Groundwater Changes Using the Soil & Water Assessment Tool Hydrological Model, Geographic Information System, and Machine Learning in the Southeast Australian River Basin

Authors: Newton Muhury, Armando A. Apan, Tek N. Marasani, Gebiaw T. Ayele

Abstract:

The changing climate has degraded freshwater availability in Australia that influencing vegetation growth to a great extent. This study assessed the vegetation responses to groundwater using Terra’s moderate resolution imaging spectroradiometer (MODIS), Normalised Difference Vegetation Index (NDVI), and soil water content (SWC). A hydrological model, SWAT, has been set up in a southeast Australian river catchment for groundwater analysis. The model was calibrated and validated against monthly streamflow from 2001 to 2006 and 2007 to 2010, respectively. The SWAT simulated soil water content for 43 sub-basins and monthly MODIS NDVI data for three different types of vegetation (forest, shrub, and grass) were applied in the machine learning tool, Waikato Environment for Knowledge Analysis (WEKA), using two supervised machine learning algorithms, i.e., support vector machine (SVM) and random forest (RF). The assessment shows that different types of vegetation response and soil water content vary in the dry and wet seasons. The WEKA model generated high positive relationships (r = 0.76, 0.73, and 0.81) between NDVI values of all vegetation in the sub-basins against soil water content (SWC), the groundwater flow (GW), and the combination of these two variables, respectively, during the dry season. However, these responses were reduced by 36.8% (r = 0.48) and 13.6% (r = 0.63) against GW and SWC, respectively, in the wet season. Although the rainfall pattern is highly variable in the study area, the summer rainfall is very effective for the growth of the grass vegetation type. This study has enriched our knowledge of vegetation responses to groundwater in each season, which will facilitate better floodplain vegetation management.

Keywords: ArcSWAT, machine learning, floodplain vegetation, MODIS NDVI, groundwater

Procedia PDF Downloads 68
963 Optimization of Friction Stir Welding Parameters for Joining Aluminium Alloys using Response Surface Methodology and Artificial Neural Network

Authors: A. M. Khourshid, A. M. El-Kassas, I. Sabry

Abstract:

The objective of this work was to investigate the mechanical properties in order to demonstrate the feasibility of friction stir welding for joining Al 6061 aluminium alloys. Welding was performed on pipe with different thickness (2, 3 and 4 mm), five rotational speeds (485, 710, 910, 1120 and 1400 rpm) and a traverse speed of 4mm/min. This work focuses on two methods which are artificial neural networks using software and Response Surface Methodology (RSM) to predict the tensile strength, the percentage of elongation and hardness of friction stir welded 6061 aluminium alloy. An Artificial Neural Network (ANN) model was developed for the analysis of the friction stir welding parameters of 6061 pipe. Tensile strength, the percentage of elongation and hardness of weld joints were predicted by taking the parameters tool rotation speed, material thickness and axial force as a function. A comparison was made between measured and predicted data. Response Surface Methodology (RSM) was also developed and the values obtained for the response tensile strength, the percentage of elongation and hardness are compared with measured values. The effect of FSW process parameters on mechanical properties of 6061 aluminium alloy has been analysed in detail.

Keywords: friction stir welding, aluminium alloy, response surface methodology, artificial neural network

Procedia PDF Downloads 262