Search results for: Risk Management Process
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 8130

Search results for: Risk Management Process

1500 Adaptive Control Strategy of Robot Polishing Force Based on Position Impedance

Authors: Wang Zhan-Xi, Zhang Yi-Ming, Chen Hang, Wang Gang

Abstract:

Manual polishing has problems such as high labor intensity, low production efficiency and difficulty in guaranteeing the consistency of polishing quality. The use of robot polishing instead of manual polishing can effectively avoid these problems. Polishing force directly affects the quality of polishing, so accurate tracking and control of polishing force is one of the most important conditions for improving the accuracy of robot polishing. The traditional force control strategy is difficult to adapt to the strong coupling of force control and position control during the robot polishing process. Therefore, based on the analysis of force-based impedance control and position-based impedance control, this paper proposed a type of adaptive controller. Based on force feedback control of active compliance control, the controller can adaptively estimate the stiffness and position of the external environment and eliminate the steady-state force error produced by traditional impedance control. The simulation results of the model show that the adaptive controller has good adaptability to changing environmental positions and environmental stiffness, and can accurately track and control polishing force.

Keywords: robot polishing, force feedback, impedance control, adaptive control

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 599
1499 Reversible Watermarking on Stereo Image Sequences

Authors: John N. Ellinas

Abstract:

In this paper, a new reversible watermarking method is presented that reduces the size of a stereoscopic image sequence while keeping its content visible. The proposed technique embeds the residuals of the right frames to the corresponding frames of the left sequence, halving the total capacity. The residual frames may result in after a disparity compensated procedure between the two video streams or by a joint motion and disparity compensation. The residuals are usually lossy compressed before embedding because of the limited embedding capacity of the left frames. The watermarked frames are visible at a high quality and at any instant the stereoscopic video may be recovered by an inverse process. In fact, the left frames may be exactly recovered whereas the right ones are slightly distorted as the residuals are not embedded intact. The employed embedding method reorders the left frame into an array of consecutive pixel pairs and embeds a number of bits according to their intensity difference. In this way, it hides a number of bits in intensity smooth areas and most of the data in textured areas where resulting distortions are less visible. The experimental evaluation demonstrates that the proposed scheme is quite effective.

Keywords: Stereoscopic video, Reversible watermarking, Disparity compensation, Joint compensation

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1419
1498 Shape Restoration of the Left Ventricle

Authors: May-Ling Tan, Yi Su, Chi-Wan Lim, Liang Zhong, Ru-San Tan

Abstract:

This paper describes an automatic algorithm to restore the shape of three-dimensional (3D) left ventricle (LV) models created from magnetic resonance imaging (MRI) data using a geometry-driven optimization approach. Our basic premise is to restore the LV shape such that the LV epicardial surface is smooth after the restoration. A geometrical measure known as the Minimum Principle Curvature (κ2) is used to assess the smoothness of the LV. This measure is used to construct the objective function of a two-step optimization process. The objective of the optimization is to achieve a smooth epicardial shape by iterative in-plane translation of the MRI slices. Quantitatively, this yields a minimum sum in terms of the magnitude of κ 2, when κ2 is negative. A limited memory quasi-Newton algorithm, L-BFGS-B, is used to solve the optimization problem. We tested our algorithm on an in vitro theoretical LV model and 10 in vivo patient-specific models which contain significant motion artifacts. The results show that our method is able to automatically restore the shape of LV models back to smoothness without altering the general shape of the model. The magnitudes of in-plane translations are also consistent with existing registration techniques and experimental findings.

Keywords: Magnetic Resonance Imaging, Left Ventricle, ShapeRestoration, Principle Curvature, Optimization

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1626
1497 Modeling and Parametric Study for CO2/CH4 Separation Using Membrane Processes

Authors: Faizan Ahmad, Lau Kok Keong, Azmi Mohd. Shariff

Abstract:

The upgrading of low quality crude natural gas (NG) is attracting interest due to high demand of pipeline-grade gas in recent years. Membrane processes are commercially proven technology for the removal of impurities like carbon dioxide from NG. In this work, cross flow mathematical model has been suggested to be incorporated with ASPEN HYSYS as a user defined unit operation in order to design the membrane system for CO2/CH4 separation. The effect of operating conditions (such as feed composition and pressure) and membrane selectivity on the design parameters (methane recovery and total membrane area required for the separation) has been studied for different design configurations. These configurations include single stage (with and without recycle) and double stage membrane systems (with and without permeate or retentate recycle). It is shown that methane recovery can be improved by recycling permeate or retentate stream as well as by using double stage membrane systems. The ASPEN HYSYS user defined unit operation proposed in the study has potential to be applied for complex membrane system design and optimization.

Keywords: CO2/CH4 Separation, Membrane Process, Membrane modeling, Natural Gas Processing

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3844
1496 A Multi-Objective Methodology for Selecting Lean Initiatives in Modular Construction Companies

Authors: Saba Shams Bidhendi, Steven Goh, Andrew Wandel

Abstract:

The implementation of lean manufacturing initiatives has produced significant impacts in improving operational performance and reducing manufacturing wastes in the production process. However, selecting an appropriate set of lean strategies is critical to avoid misapplication of the lean manufacturing techniques and consequential increase in non-value-adding activities. To the author’s best knowledge, there is currently no methodology to select lean strategies that considers their impacts on manufacturing wastes and performance metrics simultaneously. In this research, a multi-objective methodology is proposed that suggests an appropriate set of lean initiatives based on their impacts on performance metrics and manufacturing wastes and within manufacturers’ resource limitation. The proposed methodology in this research suggests the best set of lean initiatives for implementation that have highest impacts on identified critical performance metrics and manufacturing wastes. Therefore, manufacturers can assure that implementing suggested lean tools improves their production performance and reduces manufacturing wastes at the same time. A case study was conducted to show the effectiveness and validate the proposed model and methodologies.

Keywords: Lean manufacturing, Lean strategies, manufacturing wastes, manufacturing performance metrics, decision making, optimisation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 775
1495 Atmospheric Fluid Bed Gasification of Different Biomass Fuels

Authors: Martin Lisý, Marek Baláš, Michal Špiláček, Zdeněk Skála

Abstract:

This paper shortly describes various types of biomass and a growing number of facilities utilizing the biomass in the Czech Republic. The considerable part of this paper deals with energy parameters of the most frequently used types of biomass and results of their gasification testing. Sixteen most used "Czech" woody plants and grasses were selected; raw, element and biochemical analyses were performed and basic calorimetric values, ash composition, and ash characteristic temperatures were identified. Later, each biofuel was tested in a fluidized bed gasifier. The essential part of this paper provides results of the gasification of selected biomass types. Operating conditions are described in detail with a focus on individual fuels properties. Gas composition and impurities content are also identified. In terms of operating conditions and gas quality, the essential difference occurred mainly between woody plants and grasses. The woody plants were evaluated as more suitable fuels for fluidized bed gasifiers. Testing results significantly help with a decision-making process regarding suitability of energy plants for growing and with a selection of optimal biomass-treatment technology.

Keywords: Biomass Growing, Biomass Types, Gasification.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1917
1494 Definition and Core Components of the Role-Partner Allocation Problem in Collaborative Networks

Authors: J. Andrade-Garda, A. Anguera, J. Ares-Casal, M. Hidalgo-Lorenzo, J.-A. Lara, D. Lizcano, S. Suárez-Garaboa

Abstract:

In the current constantly changing economic context, collaborative networks allow partners to undertake projects that would not be possible if attempted by them individually. These projects usually involve the performance of a group of tasks (named roles) that have to be distributed among the partners. Thus, an allocation/matching problem arises that will be referred to as Role-Partner Allocation problem. In real life this situation is addressed by negotiation between partners in order to reach ad hoc agreements. Besides taking a long time and being hard work, both historical evidence and economic analysis show that such approach is not recommended. Instead, the allocation process should be automated by means of a centralized matching scheme. However, as a preliminary step to start the search for such a matching mechanism (or even the development of a new one), the problem and its core components must be specified. To this end, this paper establishes (i) the definition of the problem and its constraints, (ii) the key features of the involved elements (i.e., roles and partners); and (iii) how to create preference lists both for roles and partners. Only this way it will be possible to conduct subsequent methodological research on the solution method.     

Keywords: Collaborative network, matching, partner, preference list, role.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 810
1493 Supply Chain Competitiveness: A Review of Select Enablers

Authors: Ajay Verma, Nitin Seth

Abstract:

In present days market environment, supply chains has to be competitive enough to handle pressures like varying customer’s expectations, low cost high quality products to be delivered at the minimum time and the most important is throat cutting competition at world wide scale. Recently, supply chain competitiveness has been, therefore, accepted as one of the most important philosophies in the supply chain literature. Various researchers and practitioners have tried to identify and implement enablers in supply chains which can bring competitiveness in the supply chains i.e. supply chain competitiveness. The purpose of this paper is to suggest select enablers for supply chain competitiveness in the Indian manufacturing sector using an integrated approach of literature review and exploratory interviews with eminent professionals from the supply chain area in various industries, academia and research. The aim of the paper is to highlight the important enablers in the area of supply chain competitiveness and suggest recommendations to the industry and managers of manufacturing supply chains with a view to encourage more efforts in this wider domain area.

Keywords: Supply Chain Competitiveness, Supply Chain Collaboration, Coordination, Quality Management, Cost.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2291
1492 Optimization of Soybean Oil by Modified Supercritical Carbon Dioxide

Authors: N. R. Putra, A. H. Abdul Aziz, A. S. Zaini, Z. Idham, F. Idrus, M. Z. Bin Zullyadini, M. A. Che Yunus

Abstract:

The content of omega-3 in soybean oil is important in the development of infants and is an alternative for the omega-3 in fish oils. The investigation of extraction of soybean oil is needed to obtain the bioactive compound in the extract. Supercritical carbon dioxide extraction is modern and green technology to extract herbs and plants to obtain high quality extract due to high diffusivity and solubility of the solvent. The aim of this study was to obtain the optimum condition of soybean oil extraction by modified supercritical carbon dioxide. The soybean oil was extracted by using modified supercritical carbon dioxide (SC-CO2) under the temperatures of 40, 60, 80 °C, pressures of 150, 250, 350 Bar, and constant flow-rate of 10 g/min as the parameters of extraction processes. An experimental design was performed in order to optimize three important parameters of SC-CO2 extraction which are pressure (X1), temperature (X2) to achieve optimum yields of soybean oil. Box Behnken Design was applied for experimental design. From the optimization process, the optimum condition of extraction of soybean oil was obtained at pressure 338 Bar and temperature 80 °C with oil yield of 2.713 g. Effect of pressure is significant on the extraction of soybean oil by modified supercritical carbon dioxide. Increasing of pressure will increase the oil yield of soybean oil.

Keywords: Soybean oil, SC-CO2 extraction, yield, optimization.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 924
1491 Effect of Curing Temperature on Mechanical Properties of Jute Fiber Reinforced Polylactic Acid Based Green Composite

Authors: Sehijpal Singh Khangura, Jai Inder Preet Singh, Vikas Dhawan

Abstract:

Global warming, growing awareness of the environment, waste management issues, dwindling fossil resources, and rising oil prices resulted to increase the research in the materials that are friendly to our health and environment. Due to these reasons, green products are increasingly being promoted for sustainable development. In this work, fully biodegradable green composites have been developed using jute fibers as reinforcement and poly lactic acid as matrix material by film stacking technique. The effect of curing temperature during development of composites ranging from 160 °C, 170 °C, 180 °C and 190 °C was investigated for various mechanical properties. Results obtained from various tests indicate that impact strength decreases with an increase in curing temperature, but tensile and flexural strength increases till 180 °C, thereafter both the properties decrease. This study gives an optimum curing temperature for the development of jute/PLA composites.

Keywords: Natural fibers, polymer matrix composites, jute, compression molding, biodegradation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1047
1490 Medical Image Fusion Based On Redundant Wavelet Transform and Morphological Processing

Authors: P. S. Gomathi, B. Kalaavathi

Abstract:

The process in which the complementary information from multiple images is integrated to provide composite image that contains more information than the original input images is called image fusion. Medical image fusion provides useful information from multimodality medical images that provides additional information to the doctor for diagnosis of diseases in a better way. This paper represents the wavelet based medical image fusion algorithm on different multimodality medical images. In order to fuse the medical images, images are decomposed using Redundant Wavelet Transform (RWT). The high frequency coefficients are convolved with morphological operator followed by the maximum-selection (MS) rule. The low frequency coefficients are processed by MS rule. The reconstructed image is obtained by inverse RWT. The quantitative measures which includes Mean, Standard Deviation, Average Gradient, Spatial frequency, Edge based Similarity Measures are considered for evaluating the fused images. The performance of this proposed method is compared with Pixel averaging, PCA, and DWT fusion methods. When compared with conventional methods, the proposed framework provides better performance for analysis of multimodality medical images.

Keywords: Discrete Wavelet Transform (DWT), Image Fusion, Morphological Processing, Redundant Wavelet Transform (RWT).

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2146
1489 Cadmium Filter Cake of a Hydrometallurgical Zinc Smelter as a New Source for the Biological Synthesis of CdS Quantum Dots

Authors: Mehran Bakhshi, Mohammad Raouf Hosseini, Mohammadhosein Rahimi

Abstract:

The cadmium sulfide nanoparticles were synthesized from the nickel-cadmium cake of a hydrometallurgical zinc producing plant and sodium sulfide as Cd2+ and S-2 sources, respectively. Also, the synthesis process was performed by using the secretions of Bacillus licheniformis as bio-surfactant. Initially, in order to obtain a cadmium rich solution, two following steps were carried out: 1) Alkaline leaching for the removal of zinc oxide from the cake, and 2) acidic leaching to dissolve cadmium from the remained solid residue. Afterward, the obtained CdSO4 solution was used for the nanoparticle biosynthesis. Nanoparticles were characterized by the energy dispersive spectroscopy (EDS) and X-ray diffraction (XRD) to confirm the formation of CdS crystals with cubic structure. Also, transmission electron microscopy (TEM) was applied to determine the particle sizes which were in 2-10 nm range. Moreover, the presence of the protein containing bio-surfactants was approved by using infrared analysis (FTIR). In addition, the absorbance below 400 nm confirms quantum particles’ size. Finally, it was shown that valuable CdS quantum dots could be obtained from the industrial waste products via environment-friendly biological approaches.

Keywords: Biosynthesis, cadmium cake, cadmium sulfide, nanoparticle, zinc smelter.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1521
1488 Geometric Contrast of a 3D Model Obtained by Means of Digital Photogrametry with a Quasimetric Camera on UAV Classical Methods

Authors: Julio Manuel de Luis Ruiz, Javier Sedano Cibrián, Rubén Pérez Álvarez, Raúl Pereda García, Cristina Diego Soroa

Abstract:

Nowadays, the use of drones has been extended to practically any human activity. One of the main applications is focused on the surveying field. In this regard, software programs that process the images captured by the sensor from the drone in an almost automatic way have been developed and commercialized, but they only allow contrasting the results through control points. This work proposes the contrast of a 3D model obtained from a flight developed by a drone and a non-metric camera (due to its low cost), with a second model that is obtained by means of the historically-endorsed classical methods. In addition to this, the contrast is developed over a certain territory with a significant unevenness, so as to test the model generated with photogrammetry, and considering that photogrammetry with drones finds more difficulties in terms of accuracy in this kind of situations. Distances, heights, surfaces and volumes are measured on the basis of the 3D models generated, and the results are contrasted. The differences are about 0.2% for the measurement of distances and heights, 0.3% for surfaces and 0.6% when measuring volumes. Although they are not important, they do not meet the order of magnitude that is presented by salespeople.

Keywords: Accuracy, classical topographic, 3D model, photogrammetry, UAV.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 512
1487 Trap Assisted Tunneling Model for Gate Current in Nano Scale MOSFET with High-K Gate Dielectrics

Authors: Ashwani K. Rana, Narottam Chand, Vinod Kapoor

Abstract:

This paper presents a new compact analytical model of the gate leakage current in high-k based nano scale MOSFET by assuming a two-step inelastic trap-assisted tunneling (ITAT) process as the conduction mechanism. This model is based on an inelastic trap-assisted tunneling (ITAT) mechanism combined with a semiempirical gate leakage current formulation in the BSIM 4 model. The gate tunneling currents have been calculated as a function of gate voltage for different gate dielectrics structures such as HfO2, Al2O3 and Si3N4 with EOT (equivalent oxide thickness) of 1.0 nm. The proposed model is compared and contrasted with santaurus simulation results to verify the accuracy of the model and excellent agreement is found between the analytical and simulated data. It is observed that proposed analytical model is suitable for different highk gate dielectrics simply by adjusting two fitting parameters. It was also shown that gate leakages reduced with the introduction of high-k gate dielectric in place of SiO2.

Keywords: Analytical model, High-k gate dielectrics, inelastic trap assisted tunneling, metal–oxide–semiconductor (MOS) devices.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3291
1486 An Image Encryption Method with Magnitude and Phase Manipulation using Carrier Images

Authors: S. R. M. Prasanna, Y. V. Subba Rao, A. Mitra

Abstract:

We describe an effective method for image encryption which employs magnitude and phase manipulation using carrier images. Although it involves traditional methods like magnitude and phase encryptions, the novelty of this work lies in deploying the concept of carrier images for encryption purpose. To this end, a carrier image is randomly chosen from a set of stored images. One dimensional (1-D) discrete Fourier transform (DFT) is then carried out on the original image to be encrypted along with the carrier image. Row wise spectral addition and scaling is performed between the magnitude spectra of the original and carrier images by randomly selecting the rows. Similarly, row wise phase addition and scaling is performed between the original and carrier images phase spectra by randomly selecting the rows. The encrypted image obtained by these two operations is further subjected to one more level of magnitude and phase manipulation using another randomly chosen carrier image by 1-D DFT along the columns. The resulting encrypted image is found to be fully distorted, resulting in increasing the robustness of the proposed work. Further, applying the reverse process at the receiver, the decrypted image is found to be distortionless.

Keywords: Encryption, Carrier images, Magnitude manipulation, Phase manipulation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1588
1485 Using Structural Equation Modeling in Causal Relationship Design for Balanced-Scorecards' Strategic Map

Authors: A. Saghaei, R. Ghasemi

Abstract:

Through 1980s, management accounting researchers described the increasing irrelevance of traditional control and performance measurement systems. The Balanced Scorecard (BSC) is a critical business tool for a lot of organizations. It is a performance measurement system which translates mission and strategy into objectives. Strategy map approach is a development variant of BSC in which some necessary causal relations must be established. To recognize these relations, experts usually use experience. It is also possible to utilize regression for the same purpose. Structural Equation Modeling (SEM), which is one of the most powerful methods of multivariate data analysis, obtains more appropriate results than traditional methods such as regression. In the present paper, we propose SEM for the first time to identify the relations between objectives in the strategy map, and a test to measure the importance of relations. In SEM, factor analysis and test of hypotheses are done in the same analysis. SEM is known to be better than other techniques at supporting analysis and reporting. Our approach provides a framework which permits the experts to design the strategy map by applying a comprehensive and scientific method together with their experience. Therefore this scheme is a more reliable method in comparison with the previously established methods.

Keywords: BSC, SEM, Strategy map.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2694
1484 Problems of Innovation Development of Wireless Data Transfer Branch in the Cellular Market of Kazakhstan

Authors: Yessengeldy Kuanyshpayev

Abstract:

Now in some countries of the world the cellular market is on the point of saturation, in others - positive dynamics of development kept on. The reasons for it are also different, but there are united by their general susceptibility to innovation changes, if they are really innovative. If to take as an example the cellular market of Kazakhstan it is defined by the low percent of smart phones at consumers, the low population density, undercapacity of the 3G channel, and absence of universal access to the LTE technology that limits dynamical growth of this branch. These moments are aggravated by failures of starting commercial projects by private companies which prevent to be implemented and widely adopted to a new product among consumers. The object of the research is possible integration of wireless and program technologies at which introduction the idea can regenerate in an innovation. The analysis of existing projects in the market and the possible union of the technologies through a prism of theoretical bases of innovative activity shows that efficiency of the company by development and introduction of innovations is possible only thanks to strict observance of all terms and conditions of the innovative process which main term is profit. Despite that fact that on a global scale the innovativeness issue of companies is very popular, there are no researches about possibility of innovative breaks in the field of wireless access to the Internet in the cellular market of Kazakhstan.

Keywords: Cellular market, commercialization, innovation, the effectiveness of company.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2065
1483 A Taxonomy of Behavior for a Medical Coordinator by Utlizing Leadership Styles

Authors: Aryana Collins Jackson, Elisabetta Bevacqua, Pierre De Loor, Ronan Querrec

Abstract:

This paper presents a taxonomy of non-technical skills, communicative intentions, and behavior for an individual acting as a medical coordinator. In medical emergency situations, a leader among the group is imperative to both patient health and team emotional and mental health. Situational Leadership is used to make clear and easy-to-follow guidelines for behavior depending on circumstantial factors. Low-level leadership behaviors belonging to two different styles, directive and supporting, are identified from literature and are included in the proposed taxonomy. The high-level information in the taxonomy consists of the necessary non-technical skills belonging to a medical coordinator: situation awareness, decision making, task management, and teamwork. Finally, communicative intentions, dimensions, and functions are included. Thus this work brings high-level and low-level information - medical non-technical skills, communication capabilities, and leadership behavior - into a single versatile taxonomy of behavior.

Keywords: Medical, leadership styles, taxonomy, human behavior.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 613
1482 A Real Time Ultra-Wideband Location System for Smart Healthcare

Authors: Mingyang Sun, Guozheng Yan, Dasheng Liu, Lei Yang

Abstract:

Driven by the demand of intelligent monitoring in rehabilitation centers or hospitals, a high accuracy real-time location system based on UWB (ultra-wideband) technology was proposed. The system measures precise location of a specific person, traces his movement and visualizes his trajectory on the screen for doctors or administrators. Therefore, doctors could view the position of the patient at any time and find them immediately and exactly when something emergent happens. In our design process, different algorithms were discussed, and their errors were analyzed. In addition, we discussed about a , simple but effective way of correcting the antenna delay error, which turned out to be effective. By choosing the best algorithm and correcting errors with corresponding methods, the system attained a good accuracy. Experiments indicated that the ranging error of the system is lower than 7 cm, the locating error is lower than 20 cm, and the refresh rate exceeds 5 times per second. In future works, by embedding the system in wearable IoT (Internet of Things) devices, it could provide not only physical parameters, but also the activity status of the patient, which would help doctors a lot in performing healthcare.

Keywords: Intelligent monitoring, IoT devices, real-time location, smart healthcare, ultra-wideband technology.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 878
1481 Land Suitability Analysis for Maize Production in Egbeda Local Government Area of Oyo State Using GIS Techniques

Authors: Abegunde Linda, Adedeji Oluwatola, Tope-Ajayi Opeyemi

Abstract:

Maize constitutes a major agrarian production for use by the vast population but despite its economic importance; it has not been produced to meet the economic needs of the country. Achieving optimum yield in maize can meaningfully be supported by land suitability analysis in order to guarantee self-sufficiency for future production optimization. This study examines land suitability for maize production through the analysis of the physicochemical variations in soil properties and other land attributes over space using a Geographic Information System (GIS) framework. Physicochemical parameters of importance selected include slope, landuse, physical and chemical properties of the soil, and climatic variables. Landsat imagery was used to categorize the landuse, Shuttle Radar Topographic Mapping (SRTM) generated the slope and soil samples were analyzed for its physical and chemical components. Suitability was categorized into highly, moderately and marginally suitable based on Food and Agricultural Organisation (FAO) classification, using the Analytical Hierarchy Process (AHP) technique of GIS. This result can be used by small scale farmers for efficient decision making in the allocation of land for maize production.

Keywords: AHP, GIS, MCE, Suitability, Zea mays.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3877
1480 Adaptation Actions in Companies as Theoretical and Practical Aspects: A Case Study of a Food Ingredients and Additives Producer

Authors: Maja Sajdak

Abstract:

The aim of this article is to identify the measures companies undertake in order to adapt to the environment as well as discussing their diversity and effectiveness. The research methods used in the study include an in-depth analysis of the literature and a case study, which helps to illustrate the issue in question. Referring to the concept of agility, which is firmly embedded in the theory of strategic management and has been developed with the aim of adapting to the environment and its changes, the paper first examines different types of adaptation measures for companies. Then the issue under discussion is illustrated with the example of the company Hortimex. This company is an eminent representative of the world’s leading manufacturers of food additives and ingredients. The company was established in 1988 and is a family business, which in practice means that it conducts business in a responsible manner, observing the law and respecting the interests of society and the environment. The company’s mission is to develop a market in Poland for the products and solutions offered by their partners and to share their knowledge of additives in food production and consumption.

Keywords: Adaptation measures, agile company, flexibility, unanticipated changes.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1479
1479 Ultrasound Assisted Cooling Crystallization of Lactose Monohydrate

Authors: Sanjaykumar R. Patel, Parth R. Kayastha

Abstract:

α-lactose monohydrate is widely used in the pharmaceutical industries as an inactive substance that acts as a vehicle or a medium for a drug or other active substance. It is a byproduct of dairy industries, and the recovery of lactose from whey not only boosts the improvement of the economics of whey utilization but also causes a reduction in pollution as lactose recovery can reduce the BOD of whey by more than 80%. In the present study, levels of process parameters were kept as initial lactose concentration (30-50% w/w), sonication amplitude (20-40%), sonication time (2-6 hours), and crystallization temperature (10-20 oC) for the recovery of lactose in ultrasound assisted cooling crystallization. In comparison with cooling crystallization, the use of ultrasound enhanced the lactose recovery by 39.17% (w/w). The parameters were optimized for the lactose recovery using Taguchi Method. The optimum conditions found were initial lactose concentration at level 3 (50% w/w), amplitude of sonication at level 2 (40%), the sonication time at level 3 (6 hours), and crystallization temperature at level 1 (10 °C). The maximum recovery was found to be 85.85% at the optimum conditions. Sonication time and the initial lactose concentration were found to be significant parameters for the lactose recovery.

Keywords: Crystallization, Taguchi method, ultrasound, lactose.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 983
1478 Defining Programming Problems as Learning Objects

Authors: José Paulo Leal, Ricardo Queirós

Abstract:

Standards for learning objects focus primarily on content presentation. They were already extended to support automatic evaluation but it is limited to exercises with a predefined set of answers. The existing standards lack the metadata required by specialized evaluators to handle types of exercises with an indefinite set of solutions. To address this issue existing learning object standards were extended to the particular requirements of a specialized domain. A definition of programming problems as learning objects, compatible both with Learning Management Systems and with systems performing automatic evaluation of programs, is presented in this paper. The proposed definition includes metadata that cannot be conveniently represented using existing standards, such as: the type of automatic evaluation; the requirements of the evaluation engine; and the roles of different assets - tests cases, program solutions, etc. The EduJudge project and its main services are also presented as a case study on the use of the proposed definition of programming problems as learning objects.

Keywords: Content Packaging, eLearning Services, Interoperability, Learning Objects.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1545
1477 Prediction Modeling of Compression Properties of a Knitted Sportswear Fabric Using Response Surface Method

Authors: Jawairia Umar, Tanveer Hussain, Zulfiqar Ali, Muhammad Maqsood

Abstract:

Different knitted structures and knitted parameters play a vital role in the stretch and recovery management of compression sportswear in addition to the materials use to generate this stretch and recovery behavior of the fabric. The present work was planned to predict the different performance indicators of a compression sportswear fabric with some ground parameters i.e. base yarn stitch length (polyester as base yarn and spandex as plating yarn involve to make a compression fabric) and linear density of the spandex which is a key material of any sportswear fabric. The prediction models were generated by response surface method for performance indicators such as stretch & recovery percentage, compression generated by the garment on body, total elongation on application of high power force and load generated on certain percentage extension in fabric. Certain physical properties of the fabric were also modeled using these two parameters.

Keywords: Compression, sportswear, stretch and recovery, statistical model, kikuhime.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2038
1476 A Monte Carlo Method to Data Stream Analysis

Authors: Kittisak Kerdprasop, Nittaya Kerdprasop, Pairote Sattayatham

Abstract:

Data stream analysis is the process of computing various summaries and derived values from large amounts of data which are continuously generated at a rapid rate. The nature of a stream does not allow a revisit on each data element. Furthermore, data processing must be fast to produce timely analysis results. These requirements impose constraints on the design of the algorithms to balance correctness against timely responses. Several techniques have been proposed over the past few years to address these challenges. These techniques can be categorized as either dataoriented or task-oriented. The data-oriented approach analyzes a subset of data or a smaller transformed representation, whereas taskoriented scheme solves the problem directly via approximation techniques. We propose a hybrid approach to tackle the data stream analysis problem. The data stream has been both statistically transformed to a smaller size and computationally approximated its characteristics. We adopt a Monte Carlo method in the approximation step. The data reduction has been performed horizontally and vertically through our EMR sampling method. The proposed method is analyzed by a series of experiments. We apply our algorithm on clustering and classification tasks to evaluate the utility of our approach.

Keywords: Data Stream, Monte Carlo, Sampling, DensityEstimation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1409
1475 Land Use Change Detection Using Remote Sensing and GIS

Authors: Naser Ahmadi Sani, Karim Solaimani, Lida Razaghnia, Jalal Zandi

Abstract:

In recent decades, rapid and incorrect changes in land-use have been associated with consequences such as natural resources degradation and environmental pollution. Detecting changes in land-use is one of the tools for natural resource management and assessment of changes in ecosystems. The target of this research is studying the land-use changes in Haraz basin with an area of 677000 hectares in a 15 years period (1996 to 2011) using LANDSAT data. Therefore, the quality of the images was first evaluated. Various enhancement methods for creating synthetic bonds were used in the analysis. Separate training sites were selected for each image. Then the images of each period were classified in 9 classes using supervised classification method and the maximum likelihood algorithm. Finally, the changes were extracted in GIS environment. The results showed that these changes are an alarm for the HARAZ basin status in future. The reason is that 27% of the area has been changed, which is related to changing the range lands to bare land and dry farming and also changing the dense forest to sparse forest, horticulture, farming land and residential area.

Keywords: HARAZ Basin, Change Detection, Land-use, Satellite Data.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2309
1474 A Robust Optimization Model for the Single-Depot Capacitated Location-Routing Problem

Authors: Abdolsalam Ghaderi

Abstract:

In this paper, the single-depot capacitated location-routing problem under uncertainty is presented. The problem aims to find the optimal location of a single depot and the routing of vehicles to serve the customers when the parameters may change under different circumstances. This problem has many applications, especially in the area of supply chain management and distribution systems. To get closer to real-world situations, travel time of vehicles, the fixed cost of vehicles usage and customers’ demand are considered as a source of uncertainty. A combined approach including robust optimization and stochastic programming was presented to deal with the uncertainty in the problem at hand. For this purpose, a mixed integer programming model is developed and a heuristic algorithm based on Variable Neighborhood Search(VNS) is presented to solve the model. Finally, the computational results are presented and future research directions are discussed.

Keywords: Location-routing problem, robust optimization, Stochastic Programming, variable neighborhood search.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 744
1473 SAP: A Smart Amusement Park System for Tourist Services

Authors: Pei-Chun Lee, Sheng-Shih Wang, Pei-Hsuan Ku

Abstract:

Many existing amusement parks have been operated with assistance of a variety of information and communications technologies to design friendly and efficient service systems for tourists. However, these systems leave various levels of decisions to tourists to make by themselves. This incurs pressure on tourists and thereby bringing negative experience in their tour. This paper proposes a smart amusement park system to offer each tourist the GPS-based customized plan without tourists making decisions by themselves. The proposed system consists of the mobile app subsystem, the central subsystem, and the detecting/counting subsystem. The mobile app subsystem interacts with the central subsystem. The central subsystem performs the necessary computing and database management of the proposed system. The detecting/counting subsystem aims to detect and compute the number of visitors to an attraction. Experimental results show that the proposed system can not only work well, but also provide an innovative business operating model for owners of amusement parks.

Keywords: Amusement park, location-based service (LBS), mobile app, tourist service.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3190
1472 Fingerprint Identification using Discretization Technique

Authors: W. Y. Leng, S. M. Shamsuddin

Abstract:

Fingerprint based identification system; one of a well known biometric system in the area of pattern recognition and has always been under study through its important role in forensic science that could help government criminal justice community. In this paper, we proposed an identification framework of individuals by means of fingerprint. Different from the most conventional fingerprint identification frameworks the extracted Geometrical element features (GEFs) will go through a Discretization process. The intention of Discretization in this study is to attain individual unique features that could reflect the individual varianceness in order to discriminate one person from another. Previously, Discretization has been shown a particularly efficient identification on English handwriting with accuracy of 99.9% and on discrimination of twins- handwriting with accuracy of 98%. Due to its high discriminative power, this method is adopted into this framework as an independent based method to seek for the accuracy of fingerprint identification. Finally the experimental result shows that the accuracy rate of identification of the proposed system using Discretization is 100% for FVC2000, 93% for FVC2002 and 89.7% for FVC2004 which is much better than the conventional or the existing fingerprint identification system (72% for FVC2000, 26% for FVC2002 and 32.8% for FVC2004). The result indicates that Discretization approach manages to boost up the classification effectively, and therefore prove to be suitable for other biometric features besides handwriting and fingerprint.

Keywords: Discretization, fingerprint identification, geometrical features, pattern recognition

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2346
1471 Auto Tuning PID Controller based on Improved Genetic Algorithm for Reverse Osmosis Plant

Authors: Jin-Sung Kim, Jin-Hwan Kim, Ji-Mo Park, Sung-Man Park, Won-Yong Choe, Hoon Heo

Abstract:

An optimal control of Reverse Osmosis (RO) plant is studied in this paper utilizing the auto tuning concept in conjunction with PID controller. A control scheme composing an auto tuning stochastic technique based on an improved Genetic Algorithm (GA) is proposed. For better evaluation of the process in GA, objective function defined newly in sense of root mean square error has been used. Also in order to achieve better performance of GA, more pureness and longer period of random number generation in operation are sought. The main improvement is made by replacing the uniform distribution random number generator in conventional GA technique to newly designed hybrid random generator composed of Cauchy distribution and linear congruential generator, which provides independent and different random numbers at each individual steps in Genetic operation. The performance of newly proposed GA tuned controller is compared with those of conventional ones via simulation.

Keywords: Genetic Algorithm, Auto tuning, Hybrid random number generator, Reverse Osmosis, PID controller

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3117