Search results for: experience quality
367 Implementation of a Paraconsistent-Fuzzy Digital PID Controller in a Level Control Process
Authors: H. M. Côrtes, J. I. Da Silva Filho, M. F. Blos, B. S. Zanon
Abstract:
In a modern society the factor corresponding to the increase in the level of quality in industrial production demand new techniques of control and machinery automation. In this context, this work presents the implementation of a Paraconsistent-Fuzzy Digital PID controller. The controller is based on the treatment of inconsistencies both in the Paraconsistent Logic and in the Fuzzy Logic. Paraconsistent analysis is performed on the signals applied to the system inputs using concepts from the Paraconsistent Annotated Logic with annotation of two values (PAL2v). The signals resulting from the paraconsistent analysis are two values defined as Dc - Degree of Certainty and Dct - Degree of Contradiction, which receive a treatment according to the Fuzzy Logic theory, and the resulting output of the logic actions is a single value called the crisp value, which is used to control dynamic system. Through an example, it was demonstrated the application of the proposed model. Initially, the Paraconsistent-Fuzzy Digital PID controller was built and tested in an isolated MATLAB environment and then compared to the equivalent Digital PID function of this software for standard step excitation. After this step, a level control plant was modeled to execute the controller function on a physical model, making the tests closer to the actual. For this, the control parameters (proportional, integral and derivative) were determined for the configuration of the conventional Digital PID controller and of the Paraconsistent-Fuzzy Digital PID, and the control meshes in MATLAB were assembled with the respective transfer function of the plant. Finally, the results of the comparison of the level control process between the Paraconsistent-Fuzzy Digital PID controller and the conventional Digital PID controller were presented.
Keywords: Fuzzy logic, paraconsistent annotated logic, level control, digital PID.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1237366 Effect of Strain and Storage Period on Some Qualitative and Quantitative Traits of Table Eggs
Authors: Hani N. Hermiz, Sukar H. Ali
Abstract:
This study include the effect of strain and storage period and their interaction on some quantitative and qualitative traits and percentages of the egg components in the eggs collected at the start of production (at age 24 weeks). Eggs were divided into three storage periods (1, 7 and 14) days under refrigerator temperature (5- 7)0C. Fifty seven eggs obtained randomly from each strain including Isa Brown and Lohman White. General Linear Model within SAS programme was used to analyze the collected data and correlations between the studied traits were calculated for each strain.Average egg weight (EW), Haugh Unit (HU), yolk index (YI), yolk % (HP), albumin % (AP) and yolk to albumin ratio (YAR) was 56.629 gm, 87.968 %, 0.493, 22.13%, 67.74% and 32.76 respectively. Egg produced from ISA Brown surpassed those produced by Lohman White significantly (P<0.01) in EW (59.337 vs. 53.921 g) and AP (68.46 vs. 67.02 %), while Lohman White surpassed ISA Brown significantly (P<0.01) in HU (91.998 against 83.939 %), YI (0.498 against 0.487), YP (22.83 against 21.44%) and YAR (34.12 against 31.40). Storage period did not have any significant effect on EW and YI. Increasing the storage period caused a significant (P<0.01) decrease in HU. A non-significant increasing in YP and significant decreasing in AP % due to increasing storage period caused a significant increasing in YAR. The interaction between strain and storage period affect EW, HU and YI significantly (P <0.01), while its effect on YP, AP and YAR was not significant. Highest and significant (P<0.01) correlation was recorded between YP with YAR (0.99) in both strains, while the lowest values were between AP with YAR and being -0.97 and -0.95 in ISA Brown and Lohman White, respectively. The conclusion: increasing storage period caused a few decreasing in egg weight and this enabling the consumer to store eggs without any damage. Because of using the albumin in many food industries, so it is very important to focus on its weight. The correlations between some of the studied traits were significant, which means that selection for any trait will improve other traits.Keywords: Quality, Quantity, Storage period, Strain, Table egg
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1659365 Energy Efficient Transmission of Image over DWT-OFDM System
Authors: Lakshmi Pujitha Dachuri, Nalini Uppala
Abstract:
In many applications retransmissions of lost packets are not permitted. OFDM is a multi-carrier modulation scheme having excellent performance which allows overlapping in frequency domain. With OFDM there is a simple way of dealing with multipath relatively simple DSP algorithms.
In this paper, an image frame is compressed using DWT, and the compressed data is arranged in data vectors, each with equal number of coefficients. These vectors are quantized and binary coded to get the bit steams, which are then packetized and intelligently mapped to the OFDM system. Based on one-bit channel state information at the transmitter, the descriptions in order of descending priority are assigned to the currently good channels such that poorer sub-channels can only affect the lesser important data vectors. We consider only one-bit channel state information available at the transmitter, informing only about the sub-channels to be good or bad. For a good sub-channel, instantaneous received power should be greater than a threshold Pth. Otherwise, the sub-channel is in fading state and considered bad for that batch of coefficients. In order to reduce the system power consumption, the mapped descriptions onto the bad sub channels are dropped at the transmitter. The binary channel state information gives an opportunity to map the bit streams intelligently and to save a reasonable amount of power. By using MAT LAB simulation we can analysis the performance of our proposed scheme, in terms of system energy saving without compromising the received quality in terms of peak signal-noise ratio.
Keywords: Binary channel state, Channel state feedback, DWT-OFDM system, Energy saving, Fading broadcast channel.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2812364 Thermal Method for Testing Small Chemisorbents Samples on the Base of Potassium Superoxide
Authors: Pavel V. Balabanov, Daria A. Liubimova, Aleksandr P. Savenkov
Abstract:
The increase of technogenic and natural accidents, accompanied by air pollution, for example, by combustion products, leads to the necessity of respiratory protection. This work is devoted to the development of a calorimetric method and a device which allows investigating quickly the kinetics of carbon dioxide sorption by chemisorbents on the base of potassium superoxide in order to assess the protective properties of respiratory protective closed circuit apparatus. The features of the traditional approach for determining the sorption properties in a thin layer of chemisorbent are described, as well as methods and devices, which can be used for the sorption kinetics study. The authors developed an approach (as opposed to the traditional approach) based on the power measurement of internal heat sources in the chemisorbent layer. The emergence of the heat sources is a result of exothermic reaction of carbon dioxide sorption. This approach eliminates the necessity of chemical analysis of samples and can significantly reduce the time and material expenses during chemisorbents testing. Error of determining the volume fraction of adsorbed carbon dioxide by the developed method does not exceed 12%. Taking into account the efficiency of the method, we consider that it is a good alternative to traditional methods of chemical analysis under the assessment of the protection sorbents quality.
Keywords: Carbon dioxide chemisorption, exothermic reaction, internal heat sources, respiratory protective apparatus.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1696363 Simulation Based VLSI Implementation of Fast Efficient Lossless Image Compression System Using Adjusted Binary Code & Golumb Rice Code
Authors: N. Muthukumaran, R. Ravi
Abstract:
The Simulation based VLSI Implementation of FELICS (Fast Efficient Lossless Image Compression System) Algorithm is proposed to provide the lossless image compression and is implemented in simulation oriented VLSI (Very Large Scale Integrated). To analysis the performance of Lossless image compression and to reduce the image without losing image quality and then implemented in VLSI based FELICS algorithm. In FELICS algorithm, which consists of simplified adjusted binary code for Image compression and these compression image is converted in pixel and then implemented in VLSI domain. This parameter is used to achieve high processing speed and minimize the area and power. The simplified adjusted binary code reduces the number of arithmetic operation and achieved high processing speed. The color difference preprocessing is also proposed to improve coding efficiency with simple arithmetic operation. Although VLSI based FELICS Algorithm provides effective solution for hardware architecture design for regular pipelining data flow parallelism with four stages. With two level parallelisms, consecutive pixels can be classified into even and odd samples and the individual hardware engine is dedicated for each one. This method can be further enhanced by multilevel parallelisms.
Keywords: Image compression, Pixel, Compression Ratio, Adjusted Binary code, Golumb Rice code, High Definition display, VLSI Implementation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2073362 A Reference Framework Integrating Lean and Green Principles within Supply Chain Management
Authors: M. Bortolini, E. Ferrari, F. G. Galizia, C. Mora
Abstract:
In the last decades, an increasing set of companies adopted lean philosophy to improve their productivity and efficiency promoting the so-called continuous improvement concept, reducing waste of time and cutting off no-value added activities. In parallel, increasing attention rises toward green practice and management through the spread of the green supply chain pattern, to minimise landfilled waste, drained wastewater and pollutant emissions. Starting from a review on contributions deepening lean and green principles applied to supply chain management, the most relevant drivers to measure the performance of industrial processes are pointed out. Specific attention is paid on the role of cost because it is of key importance and it crosses both lean and green principles. This analysis leads to figure out an original reference framework for integrating lean and green principles in designing and managing supply chains. The proposed framework supports the application, to the whole value chain or to parts of it, e.g. distribution network, assembly system, job-shop, storage system etc., of the lean-green integrated perspective. Evidences show that the combination of the lean and green practices lead to great results, higher than the sum of the performances from their separate application. Lean thinking has beneficial effects on green practices and, at the same time, methods allowing environmental savings generate positive effects on time reduction and process quality increase.
Keywords: Environmental sustainability, green supply chain, integrated framework, lean thinking, supply chain management.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2784361 A Large Ion Collider Experiment (ALICE) Diffractive Detector Control System for RUN-II at the Large Hadron Collider
Authors: J. C. Cabanillas-Noris, M. I. Martínez-Hernández, I. León-Monzón
Abstract:
The selection of diffractive events in the ALICE experiment during the first data taking period (RUN-I) of the Large Hadron Collider (LHC) was limited by the range over which rapidity gaps occur. It would be possible to achieve better measurements by expanding the range in which the production of particles can be detected. For this purpose, the ALICE Diffractive (AD0) detector has been installed and commissioned for the second phase (RUN-II). Any new detector should be able to take the data synchronously with all other detectors and be operated through the ALICE central systems. One of the key elements that must be developed for the AD0 detector is the Detector Control System (DCS). The DCS must be designed to operate safely and correctly this detector. Furthermore, the DCS must also provide optimum operating conditions for the acquisition and storage of physics data and ensure these are of the highest quality. The operation of AD0 implies the configuration of about 200 parameters, from electronics settings and power supply levels to the archiving of operating conditions data and the generation of safety alerts. It also includes the automation of procedures to get the AD0 detector ready for taking data in the appropriate conditions for the different run types in ALICE. The performance of AD0 detector depends on a certain number of parameters such as the nominal voltages for each photomultiplier tube (PMT), their threshold levels to accept or reject the incoming pulses, the definition of triggers, etc. All these parameters define the efficiency of AD0 and they have to be monitored and controlled through AD0 DCS. Finally, AD0 DCS provides the operator with multiple interfaces to execute these tasks. They are realized as operating panels and scripts running in the background. These features are implemented on a SCADA software platform as a distributed control system which integrates to the global control system of the ALICE experiment.Keywords: AD0, ALICE, DCS, LHC.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1398360 Chronic Patients- Prescription Refill Intentions
Authors: Ching - Yi Lee, I-Hsiung Tseng, Feng-Chuan Pan
Abstract:
Environment today is featured with aging population, increasing prevalence of chronic disease and complex of medical treatment. Safe use of pharmaceutics relied very much on the efforts made by both the health- related organizations and as well as the government agencies. As far as the specialization concern in providing health services to the patients, the government actively issued and implemented the divisions of medical treatment and pharmaceutical to improve the quality of care and to reduce medication errors and ensure public health. Pharmaceutical sub-sector policy has been implemented for 13 years. This study attempts to explore the factors that affect the patients- behavior intention of refilling a prescription from a NHIB pharmacy. Samples were those patients refilling their prescriptions with the case NHIB pharmacies. A self-administered questionnaire was used to collect respondents- information while the patients or family members visit the pharmacy for the refilling. 1,200 questionnaires were dispatched in 37 pharmacies that randomly selected from Pingtung City, Dongkang, Chaozhou, Hengchun areas. 732 responses were gained with 604 valid samples for further analyses. Results of data analyses indicated that respondents- attitude, subjective norm, perceived behavior control and behavior intentions toward refilling behavior varied from some demographic variables to another. This research also suggested adding actual behavior, either by a self-report or observed, into the research.Keywords: Separation of dispensing and prescribing, prescriptions refill slip, NHIB contracted pharmacy, drug safety, theory of planned behavior.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1360359 Exploring Students’ Self-Evaluation on Their Learning Outcomes through an Integrated Cumulative Grade Point Average Reporting Mechanism
Authors: Suriyani Ariffin, Nor Aziah Alias, Khairil Iskandar Othman, Haslinda Yusoff
Abstract:
An Integrated Cumulative Grade Point Average (iCGPA) is a mechanism and strategy to ensure the curriculum of an academic programme is constructively aligned to the expected learning outcomes and student performance based on the attainment of those learning outcomes that is reported objectively in a spider web. Much effort and time has been spent to develop a viable mechanism and trains academics to utilize the platform for reporting. The question is: How well do learners conceive the idea of their achievement via iCGPA and whether quality learner attributes have been nurtured through the iCGPA mechanism? This paper presents the architecture of an integrated CGPA mechanism purported to address a holistic evaluation from the evaluation of courses learning outcomes to aligned programme learning outcomes attainment. The paper then discusses the students’ understanding of the mechanism and evaluation of their achievement from the generated spider web. A set of questionnaires were distributed to a group of students with iCGPA reporting and frequency analysis was used to compare the perspectives of students on their performance. In addition, the questionnaire also explored how they conceive the idea of an integrated, holistic reporting and how it generates their motivation to improve. The iCGPA group was found to be receptive to what they have achieved throughout their study period. They agreed that the achievement level generated from their spider web allows them to develop intervention and enhance the programme learning outcomes before they graduate.
Keywords: Learning outcomes attainment, iCGPA, programme learning outcomes, spider web, iCGPA reporting skills.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 777358 Participation in IAEA Proficiency Test to Analyse Cobalt, Strontium and Caesium in Seawater Using Direct Counting and Radiochemical Techniques
Authors: S. Visetpotjanakit, C. Khrautongkieo
Abstract:
Radiation monitoring in the environment and foodstuffs is one of the main responsibilities of Office of Atoms for Peace (OAP) as the nuclear regulatory body of Thailand. The main goal of the OAP is to assure the safety of the Thai people and environment from any radiological incidents. Various radioanalytical methods have been developed to monitor radiation and radionuclides in the environmental and foodstuff samples. To validate our analytical performance, several proficiency test exercises from the International Atomic Energy Agency (IAEA) have been performed. Here, the results of a proficiency test exercise referred to as the Proficiency Test for Tritium, Cobalt, Strontium and Caesium Isotopes in Seawater 2017 (IAEA-RML-2017-01) are presented. All radionuclides excepting ³H were analysed using various radioanalytical methods, i.e. direct gamma-ray counting for determining ⁶⁰Co, ¹³⁴Cs and ¹³⁷Cs and developed radiochemical techniques for analysing ¹³⁴Cs, ¹³⁷Cs using AMP pre-concentration technique and 90Sr using di-(2-ethylhexyl) phosphoric acid (HDEHP) liquid extraction technique. The analysis results were submitted to IAEA. All results passed IAEA criteria, i.e. accuracy, precision and trueness and obtained ‘Accepted’ statuses. These confirm the data quality from the OAP environmental radiation laboratory to monitor radiation in the environment.
Keywords: International atomic energy agency, proficiency test, radiation monitoring, seawater.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 824357 Effect of Architecture and Operating Conditions of Vehicle on Bulb Lifetime in Automotive
Authors: Hatice Özbek, Caner Çil, Ahmet Rodoplu
Abstract:
Automotive lighting is the leading function in the configuration of vehicle architecture. Especially headlights and taillights from external lighting functions are among the structures that determine the stylistic character of the vehicle. At the same time, the fact that lighting functions are related to many other functions brings along difficulties in design. Customers expect maximum quality from the vehicle. In these circumstances, it is necessary to make designs that aim to keep the performance of bulbs with limited working lives at the highest level. With this study, the factors that influence the working lives of filament lamps were examined and bulb explosions that can occur sooner than anticipated in the future were prevented while the vehicle was still in the design phase by determining the relations with electrical, dynamical and static variables. Especially the filaments of the bulbs used in the front lighting of the vehicle are deformed in a shorter time due to the high voltage requirement. In addition to this, rear lighting lamps vibrate as a result of the tailgate opening and closing and cause the filaments to be exposed to high stress. With this study, the findings that cause bulb explosions were evaluated. Among the most important findings: 1. The structure of the cables to the lighting functions of the vehicle and the effect of the voltage values are drawn; 2. The effect of the vibration to bulb throughout the life of the vehicle; 3 The effect of the loads carried to bulb while the vehicle doors are opened and closed. At the end of the study, the maximum performance was established in the bulb lifetimes with the optimum changes made in the vehicle architecture based on the findings obtained.
Keywords: Vehicle architecture, automotive lighting functions, filament lamps, bulb lifetime.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 782356 A New Model to Perform Preliminary Evaluations of Complex Systems for the Production of Energy for Buildings: Case Study
Authors: Roberto de Lieto Vollaro, Emanuele de Lieto Vollaro, Gianluca Coltrinari
Abstract:
The building sector is responsible, in many industrialized countries, for about 40% of the total energy requirements, so it seems necessary to devote some efforts in this area in order to achieve a significant reduction of energy consumption and of greenhouse gases emissions. The paper presents a study aiming at providing a design methodology able to identify the best configuration of the system building/plant, from a technical, economic and environmentally point of view. Normally, the classical approach involves a building's energy loads analysis under steady state conditions, and subsequent selection of measures aimed at improving the energy performance, based on previous experience made by architects and engineers in the design team. Instead, the proposed approach uses a sequence of two wellknown scientifically validated calculation methods (TRNSYS and RETScreen), that allow quite a detailed feasibility analysis. To assess the validity of the calculation model, an existing, historical building in Central Italy, that will be the object of restoration and preservative redevelopment, was selected as a casestudy. The building is made of a basement and three floors, with a total floor area of about 3,000 square meters. The first step has been the determination of the heating and cooling energy loads of the building in a dynamic regime by means, which allows simulating the real energy needs of the building in function of its use. Traditional methodologies, based as they are on steady-state conditions, cannot faithfully reproduce the effects of varying climatic conditions and of inertial properties of the structure. With this model is possible to obtain quite accurate and reliable results that allow identifying effective combinations building-HVAC system. The second step has consisted of using output data obtained as input to the calculation model, which enables to compare different system configurations from the energy, environmental and financial point of view, with an analysis of investment, and operation and maintenance costs, so allowing determining the economic benefit of possible interventions. The classical methodology often leads to the choice of conventional plant systems, while our calculation model provides a financial-economic assessment for innovative energy systems and low environmental impact. Computational analysis can help in the design phase, particularly in the case of complex structures with centralized plant systems, by comparing the data returned by the calculation model for different design options.
Keywords: Energy, Buildings, Systems, Evaluation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2029355 Production of Biocomposites Using Chars Obtained by Co-Pyrolysis of Olive Pomace with Plastic Wastes
Authors: Esra Yel, Tabriz Aslanov, Merve Sogancioglu, Suheyla Kocaman, Gulnare Ahmetli
Abstract:
The disposal of waste plastics has become a major worldwide environmental problem. Pyrolysis of waste plastics is one of the routes to waste minimization and recycling that has been gaining interest. In pyrolysis, the pyrolysed material is separated into gas, liquid (both are fuel) and solid (char) products. All fractions have utilities and economical value depending upon their characteristics. The first objective of this study is to determine the co-pyrolysis product fractions of waste HDPE- (high density polyethylene) and LDPE (low density polyethylene)-olive pomace (OP) and to determine the qualities of the solid product char. Chars obtained at 700 °C pyrolysis were used in biocomposite preparation as additive. As the second objective, the effects of char on biocomposite quality were investigated. Pyrolysis runs were performed at temperature 700 °C with heating rates of 5 °C/min. Biocomposites were prepared by mixing of chars with bisphenol-F type epoxy resin in various wt%. Biocomposite properties were determined by measuring electrical conductivity, surface hardness, Young’s modulus and tensile strength of the composites. The best electrical conductivity results were obtained with HDPE-OP char. For HDPE-OP char and LDPE-OP char, compared to neat epoxy, the tensile strength values of the composites increased by 102% and 78%, respectively, at 10% char dose. The hardness measurements showed similar results to the tensile tests, since there is a correlation between the hardness and the tensile strength.Keywords: Pyrolysis, olive pomace, char, biocomposite, PE plastics.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1906354 New Simultaneous High Performance Liquid Chromatographic Method for Determination of NSAIDs and Opioid Analgesics in Advanced Drug Delivery Systems and Human Plasma
Authors: Asad Ullah Madni, Mahmood Ahmad, Naveed Akhtar, Muhammad Usman
Abstract:
A new and cost effective RP-HPLC method was developed and validated for simultaneous analysis of non steroidal anti inflammatory dugs Diclofenac sodium (DFS), Flurbiprofen (FLP) and an opioid analgesic Tramadol (TMD) in advanced drug delivery systems (Liposome and Microcapsules), marketed brands and human plasma. Isocratic system was employed for the flow of mobile phase consisting of 10 mM sodium dihydrogen phosphate buffer and acetonitrile in molar ratio of 67: 33 with adjusted pH of 3.2. The stationary phase was hypersil ODS column (C18, 250×4.6 mm i.d., 5 μm) with controlled temperature of 30 C°. DFS in liposomes, microcapsules and marketed drug products was determined in range of 99.76-99.84%. FLP and TMD in microcapsules and brands formulation were 99.78 - 99.94 % and 99.80 - 99.82 %, respectively. Single step liquid-liquid extraction procedure using combination of acetonitrile and trichloroacetic acid (TCA) as protein precipitating agent was employed. The detection limits (at S/N ratio 3) of quality control solutions and plasma samples were 10, 20, and 20 ng/ml for DFS, FLP and TMD, respectively. The Assay was acceptable in linear dynamic range. All other validation parameters were found in limits of FDA and ICH method validation guidelines. The proposed method is sensitive, accurate and precise and could be applicable for routine analysis in pharmaceutical industry as well as in human plasma samples for bioequivalence and pharmacokinetics studies.Keywords: Diclofenac Sodium, Flurbiprofen, Tramadol, HPLCUV detection, Validation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1859353 Remittances and the Changing Roles of Women in Laos
Authors: N. Southiseng, J. Walsh
Abstract:
Prior to 1975, women in Laos suffered from having reduced levels of power over decision-making in their families and in their communities. This has had a negative impact on their ability to develop their own identities. Their roles were identified as being responsible for household activities and making preparations for their marriage. Many women lost opportunities to get educated and access the outdoor work that might have empowered them to improve their situations. So far, no accurate figures of either emigrants or return migrants have been compiled but it appears that most of them were women, and it was women who most and more frequently remitted money home. However, very few recent studies have addressed the relationship between remittances and the roles of women in Laos. This study, therefore, aims at redressing to some extent the deficiencies in knowledge. Qualitative techniques were used to gather data, including individual in-depth interviews and direct observation in combination with the content analysis method. Forty women in Vientiane Municipality and Savannakhet province were individually interviewed. It was found that the monetary remittance was typically used for family security and well-being; on fungible activities; on economic and business activities; and on community development, especially concerning hospitality and providing daily household necessities. Remittances played important roles in improving many respondents- livelihoods and positively changed their identities in families and communities. Women became empowered as they were able to start commercial businesses, rather than taking care of (just) housework, children and elders. Interviews indicated that 92.5% of the respondents their quality of lives improved, 90% felt happier in their families and 82.5% felt conflicts in their families were reduced.Keywords: Laos, Monetary Remittances, Social Remittance, Women's Empowerment.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2142352 Potential of Sunflower (Helianthus annuus L.) for Phytoremediation of Soils Contaminated with Heavy Metals
Authors: Violina R. Angelova, Mariana N. Perifanova-Nemska, Galina P. Uzunova, Krasimir I. Ivanov, Huu Q. Lee
Abstract:
A field study was conducted to evaluate the efficacy of the sunflower (Helianthus annuus L.) for phytoremediation of contaminated soils. The experiment was performed on an agricultural field contaminated by the Non-Ferrous-Metal Works near Plovdiv, Bulgaria. Field experiments with a randomized, complete block design with five treatments (control, compost amendments added at 20 and 40 t/daa, and vemicompost amendments added at 20 and 40 t/daa) were carried out. The accumulation of heavy metals in the sunflower plant and the quality of the sunflower oil (heavy metals and fatty acid composition) were determined. The tested organic amendments significantly influenced the uptake of Pb, Zn and Cd by the sunflower plant. The incorporation of 40 t/decare of compost and 20 t/decare of vermicompost to the soil led to an increase in the ability of the sunflower to take up and accumulate Cd, Pb and Zn. Sunflower can be subjected to the accumulators of Pb, Zn and Cd and can be successfully used for phytoremediation of contaminated soils with heavy metals. The 40 t/daa compost treatment led to a decrease in heavy metal content in sunflower oil to below the regulated limits. Oil content and fatty acids composition were affected by compost and vermicompost amendment treatments. Adding compost and vermicompost increased the oil content in the seeds. Adding organic amendments increased the content of stearic, palmitoleic and oleic acids, and reduced the content of palmitic and gadoleic acids in sunflower oil. The possibility of further industrial processing of seeds to oil and use of the obtained oil will make sunflowers economically interesting crops for farmers of phytoremediation technology.Keywords: Heavy metals, organic amendments, phytoremediation, sunflower.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3936351 Evaluation of Sensitometric Properties of Radiographic Films at Different Processing Solutions
Authors: Mojiri M, Ghazi Khanloo Sani K, Moghim Beigi A
Abstract:
The aim of this study was to compare the sensitometric properties of commonly used radiographic films processed with chemical solutions in different workload hospitals. The effect of different processing conditions on induced densities on radiologic films was investigated. Two accessible double emulsions Fuji and Kodak films were exposed with 11-step wedge and processed with Champion and CPAC processing solutions. The mentioned films provided in both workloads centers, high and low. Our findings displays that the speed and contrast of Kodak filmscreen in both work load (high and low) is higher than Fuji filmscreen for both processing solutions. However there was significant differences in films contrast for both workloads when CPAC solution had been used (p=0.000 and 0.028). The results showed base plus fog density for Kodak film was lower than Fuji. Generally Champion processing solution caused more speed and contrast for investigated films in different conditions and there was significant differences in 95% confidence level between two used processing solutions (p=0.01). Low base plus fog density for Kodak films provide more visibility and accuracy and higher contrast results in using lower exposure factors to obtain better quality in resulting radiographs. In this study we found an economic advantages since Champion solution and Kodak film are used while it makes lower patient dose. Thus, in a radiologic facility any change in film processor/processing cycle or chemistry should be carefully investigated before radiological procedures of patients are acquired.Keywords: Sensitometry, densitometry, Radiographic film, processing solution
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1628350 Identification of Critical Success Factors in Non-Formal Service Sector Using Delphi Technique
Authors: Amol A. Talankar, Prakash Verma, Nitin Seth
Abstract:
The purpose of this study is to identify the critical success factors (CSFs) for the effective implementation of Six Sigma in non-formal service Sectors.
Based on the survey of literature, the critical success factors (CSFs) for Six Sigma have been identified and are assessed for their importance in Non-formal service sector using Delphi Technique. These selected CSFs were put forth to the panel of expert to cluster them and prepare cognitive map to establish their relationship.
All the critical success factors examined and obtained from the review of literature have been assessed for their importance with respect to their contribution to Six Sigma effectiveness in non formal service sector.
The study is limited to the non-formal service sectors involved in the organization of religious festival only. However, the similar exercise can be conducted for broader sample of other non-formal service sectors like temple/ashram management, religious tours management etc.
The research suggests an approach to identify CSFs of Six Sigma for Non-formal service sector. All the CSFs of the formal service sector will not be applicable to Non-formal services, hence opinion of experts was sought to add or delete the CSFs. In the first round of Delphi, the panel of experts has suggested, two new CSFs-“competitive benchmarking (F19) and resident’s involvement (F28)”, which were added for assessment in the next round of Delphi. One of the CSFs-“fulltime six sigma personnel (F15)” has been omitted in proposed clusters of CSFs for non-formal organization, as it is practically impossible to deploy full time trained Six Sigma recruits.
Keywords: Critical success factors (CSFs), Quality assurance, non-formal service sectors, Six Sigma.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2453349 Generative Adversarial Network Based Fingerprint Anti-Spoofing Limitations
Authors: Yehjune Heo
Abstract:
Fingerprint Anti-Spoofing approaches have been actively developed and applied in real-world applications. One of the main problems for Fingerprint Anti-Spoofing is not robust to unseen samples, especially in real-world scenarios. A possible solution will be to generate artificial, but realistic fingerprint samples and use them for training in order to achieve good generalization. This paper contains experimental and comparative results with currently popular GAN based methods and uses realistic synthesis of fingerprints in training in order to increase the performance. Among various GAN models, the most popular StyleGAN is used for the experiments. The CNN models were first trained with the dataset that did not contain generated fake images and the accuracy along with the mean average error rate were recorded. Then, the fake generated images (fake images of live fingerprints and fake images of spoof fingerprints) were each combined with the original images (real images of live fingerprints and real images of spoof fingerprints), and various CNN models were trained. The best performances for each CNN model, trained with the dataset of generated fake images and each time the accuracy and the mean average error rate, were recorded. We observe that current GAN based approaches need significant improvements for the Anti-Spoofing performance, although the overall quality of the synthesized fingerprints seems to be reasonable. We include the analysis of this performance degradation, especially with a small number of samples. In addition, we suggest several approaches towards improved generalization with a small number of samples, by focusing on what GAN based approaches should learn and should not learn.
Keywords: Anti-spoofing, CNN, fingerprint recognition, GAN.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 594348 Biometrical Comparison of Artemia urmiana Günther, 1899 (Crustacea: Anostraca) Cysts between Rainy and Drought Years (1994-2003/4) from Urmia Lake, Iran
Authors: A. Asem, N. Rastegar-Pouyani, P. De Los Rios, R. Manaffar, F. Mohebbi
Abstract:
Nowadays, biometrical characterizations of Artemia cysts are used as one of the most important factors in the study of Artemia populations and intraspecific particularity; meanwhile these characters can be used as economical indices. For example, typically high hatching efficiency is possible due to the small diameter of cysts (high number per gram); therefore small diameter of cysts show someway high quality of cysts. This study was performed during a ten year period, including two different ecological conditions: rainy and drought. It is important from two different aspects because it covers alteration of A. urmiana during ten years also its variation in the best and worst environmental situations in which salinity increased from 173.8 ppt in 1994 to 280.8 ppt in 2003/4. In this study the biometrical raw data of Artemia urmiana cysts at seven stations from the Urmia Lake in 1994 and their seven identical locations at 26 studied stations in 2003/4 were reanalyzed again and compared together. Biometrical comparison of untreated and decapsulated cysts in each of the seven similar stations showed a highly significant variation between 1994 and 2003/4. Based on this study, in whole stations the untreated and decapsulated cysts from 1994 were larger than cysts of 2003/4 without any exception. But there was no logical relationship between salinity and chorion thickness in the Urmia Lake. With regard to PCA analyses the stations of two different studied years certainly have been separated with factor 1 from each other. In conclusion, the interaction between genetic and environmental factors can determine and explain variation in the range of cysts diameter in Artemia.Keywords: Artemia urmiana, Biometry, Cyst, Urmia Lake
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3685347 Multi Response Optimization in Drilling Al6063/SiC/15% Metal Matrix Composite
Authors: Hari Singh, Abhishek Kamboj, Sudhir Kumar
Abstract:
This investigation proposes a grey-based Taguchi method to solve the multi-response problems. The grey-based Taguchi method is based on the Taguchi’s design of experimental method, and adopts grey relational analysis (GRA) to transfer multi-response problems into single-response problems. In this investigation, an attempt has been made to optimize the drilling process parameters considering weighted output response characteristics using grey relational analysis. The output response characteristics considered are surface roughness, burr height and hole diameter error under the experimental conditions of cutting speed, feed rate, step angle, and cutting environment. The drilling experiments were conducted using L27 orthogonal array. A combination of orthogonal array, design of experiments and grey relational analysis was used to ascertain best possible drilling process parameters that give minimum surface roughness, burr height and hole diameter error. The results reveal that combination of Taguchi design of experiment and grey relational analysis improves surface quality of drilled hole.
Keywords: Metal matrix composite, Drilling, Optimization, step drill, Surface roughness, burr height, hole diameter error.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3255346 Development of a Robust Supply Chain for Dynamic Operating Environment
Authors: Shilan Li, Ivan Arokiam, Peter Jarvis, Wendy Garner, Gazelleh Moradi, Stuart Wakefield
Abstract:
Development of a Robust Supply Chain for Dynamic Operating Environment as we move further into the twenty first century, organisations are under increasing pressure to deliver a high product variation at a reasonable cost without compromise in quality. In a number of cases this will take the form of a customised or high variety low volume manufacturing system that requires prudent management of resources, among a number of functions, to achieve competitive advantage. Purchasing and Supply Chain management is one of such function and due to the substantial interaction with external elements needs to be strategically managed. This requires a number of primary and supporting tools that will enable the appropriate decisions to be made rapidly. This capability is especially vital in a dynamic environment as it provides a pivotal role in increasing the profit margin of the product. The management of this function can be challenging by itself and even more for Small and Medium Enterprises (SMEs) due to the limited resources and expertise available at their disposal. This paper discusses the development of tools and concepts towards effectively managing the purchasing and supply chain function. The developed tools and concepts will provide a cost effective way of managing this function within SMEs. The paper further shows the use of these tools within Contechs, a manufacturer of luxury boat interiors, and the associated benefits achieved as a result of this implementation. Finally a generic framework towards use in such environments is presented.Keywords: Lean, Supply Chain, High variety Low volume, Small and Medium Enterprises.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1461345 The Design of a Vehicle Traffic Flow Prediction Model for a Gauteng Freeway Based on an Ensemble of Multi-Layer Perceptron
Authors: Tebogo Emma Makaba, Barnabas Ndlovu Gatsheni
Abstract:
The cities of Johannesburg and Pretoria both located in the Gauteng province are separated by a distance of 58 km. The traffic queues on the Ben Schoeman freeway which connects these two cities can stretch for almost 1.5 km. Vehicle traffic congestion impacts negatively on the business and the commuter’s quality of life. The goal of this paper is to identify variables that influence the flow of traffic and to design a vehicle traffic prediction model, which will predict the traffic flow pattern in advance. The model will unable motorist to be able to make appropriate travel decisions ahead of time. The data used was collected by Mikro’s Traffic Monitoring (MTM). Multi-Layer perceptron (MLP) was used individually to construct the model and the MLP was also combined with Bagging ensemble method to training the data. The cross—validation method was used for evaluating the models. The results obtained from the techniques were compared using predictive and prediction costs. The cost was computed using combination of the loss matrix and the confusion matrix. The predicted models designed shows that the status of the traffic flow on the freeway can be predicted using the following parameters travel time, average speed, traffic volume and day of month. The implications of this work is that commuters will be able to spend less time travelling on the route and spend time with their families. The logistics industry will save more than twice what they are currently spending.Keywords: Bagging ensemble methods, confusion matrix, multi-layer perceptron, vehicle traffic flow.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1777344 Effects of Alternative Opportunities and Compensation on Turnover Intention of Singapore PMET
Authors: Han Guan Chew, Keith Yong Ngee Ng, Shan-Wei Fan
Abstract:
In Singapore, talent retention is one of the most persistent and real issue companies have to grapple with due to the tight labour market. Being resource-scarce, Singapore depends solely on its talented pool of high quality human resource to sustain its competitive advantage in the global economy. But the complex and multifaceted nature of turnover phenomenon makes the prescription of effective talent retention strategies in such a competitive labour market very challenging, especially when it comes to monetary incentives, companies struggle to answer the question of “How much is enough?” By examining the interactive effects of perceived alternative employment opportunities, annual salary and satisfaction with compensation on the turnover intention of 102 Singapore Professionals, Managers, Executives and Technicians (PMET) through correlation analyses and multiple regressions, important insights into the psyche of the Singapore talent pool can be drawn. It is found that annual salary influence turnover intention indirectly through mediation and moderation effects on PMET’s satisfaction on compensation. PMET are also found to be heavily swayed by better external opportunities. This implies that talent retention strategies should not adopt a purely monetary based blanket approach but rather a comprehensive and holistic one that considers the dynamics of prevailing market conditions.
Keywords: Employee Turnover, High Performers, Knowledge Workers, Perceived Alternative Employment Opportunities Salary, Satisfaction on Compensation, Singapore PMET, Talent Retention.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3541343 Study on Mitigation Measures of Gumti Hydro Power Plant Using Analytic Hierarchy Process and Concordance Analysis Techniques
Authors: K. Majumdar, S. Datta
Abstract:
Electricity is recognized as fundamental to industrialization and improving the quality of life of the people. Harnessing the immense untapped hydropower potential in Tripura region opens avenues for growth and provides an opportunity to improve the well-being of the people of the region, while making substantial contribution to the national economy. Gumti hydro power plant generates power to mitigate the crisis of power in Tripura, India. The first unit of hydro power plant (5MW) was commissioned in June 1976 & another two units of 5 MW was commissioned simultaneously. But out of 15MW capacity at present only 8MW- 9MW power is produced from Gumti hydro power plant during rainy season. But during lean season the production reduces to 0.5MW due to shortage of water. Now, it is essential to implement some mitigation measures so that the further atrocities can be prevented and originality will be possible to restore. The decision making ability of the Analytic Hierarchy Process (AHP) and Concordance Analysis Techniques (CAT) are utilized to identify the better decision or solution to the present problem. Some related attributes are identified by the method of surveying within the experts and the available reports and literatures. Similar criteria are removed and ultimately seven relevant ones are identified. All the attributes are compared with each other and rated accordingly to their importance over the other with the help of Pair wise Comparison Matrix. In the present investigation different mitigation measures are identified and compared to find the best suitable alternative which can solve the present uncertainties involving the existence of the Gumti Hydro Power Plant.
Keywords: Concordance Analysis Techniques, Analytic Hierarchy Process, Hydro Power.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1990342 Co-payment Strategies for Chronic Medications: A Qualitative and Comparative Analysis at European Level
Authors: Pedro M. Abreu, Bruno R. Mendes
Abstract:
The management of pharmacotherapy and the process of dispensing medicines is becoming critical in clinical pharmacy due to the increase of incidence and prevalence of chronic diseases, the complexity and customization of therapeutic regimens, the introduction of innovative and more expensive medicines, the unbalanced relation between expenditure and revenue as well as due to the lack of rationalization associated with medication use. For these reasons, co-payments emerged in Europe in the 70s and have been applied over the past few years in healthcare. Co-payments lead to a rationing and rationalization of user’s access under healthcare services and products, and simultaneously, to a qualification and improvement of the services and products for the end-user. This analysis, under hospital practices particularly and co-payment strategies in general, was carried out on all the European regions and identified four reference countries, that apply repeatedly this tool and with different approaches. The structure, content and adaptation of European co-payments were analyzed through 7 qualitative attributes and 19 performance indicators, and the results expressed in a scorecard, allowing to conclude that the German models (total score of 68,2% and 63,6% in both elected co-payments) can collect more compliance and effectiveness, the English models (total score of 50%) can be more accessible, and the French models (total score of 50%) can be more adequate to the socio-economic and legal framework. Other European models did not show the same quality and/or performance, so were not taken as a standard in the future design of co-payments strategies. In this sense, we can see in the co-payments a strategy not only to moderate the consumption of healthcare products and services, but especially to improve them, as well as a strategy to increment the value that the end-user assigns to these services and products, such as medicines.
Keywords: Clinical pharmacy, co-payments, healthcare, medicines.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1359341 Monitorization of Junction Temperature Using a Thermal-Test-Device
Authors: B. Arzhanov, A. Correia, P. Delgado, J. Meireles
Abstract:
Due to the higher power loss levels in electronic components, the thermal design of PCBs (Printed Circuit Boards) of an assembled device becomes one of the most important quality factors in electronics. Nonetheless, some of leading causes of the microelectronic component failures are due to higher temperatures, the leakages or thermal-mechanical stress, which is a concern, is the reliability of microelectronic packages. This article presents an experimental approach to measure the junction temperature of exposed pad packages. The implemented solution is in a prototype phase, using a temperature-sensitive parameter (TSP) to measure temperature directly on the die, validating the numeric results provided by the Mechanical APDL (Ansys Parametric Design Language) under same conditions. The physical device-under-test is composed by a Thermal Test Chip (TTC-1002) and assembly in a QFN cavity, soldered to a test-board according to JEDEC Standards. Monitoring the voltage drop across a forward-biased diode, is an indirectly method but accurate to obtain the junction temperature of QFN component with an applied power range between 0,3W to 1.5W. The temperature distributions on the PCB test-board and QFN cavity surface were monitored by an infra-red thermal camera (Goby-384) controlled and images processed by the Xeneth software. The article provides a set-up to monitorize in real-time the junction temperature of ICs, namely devices with the exposed pad package (i.e. QFN). Presenting the PCB layout parameters that the designer should use to improve thermal performance, and evaluate the impact of voids in solder interface in the device junction temperature.
Keywords: Quad Flat No-Lead packages, exposed pads, junction temperature, thermal management, measurements.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1712340 Prediction of Product Size Distribution of a Vertical Stirred Mill Based on Breakage Kinetics
Authors: C. R. Danielle, S. Erik, T. Patrick, M. Hugh
Abstract:
In the last decade there has been an increase in demand for fine grinding due to the depletion of coarse-grained orebodies and an increase of processing fine disseminated minerals and complex orebodies. These ores have provided new challenges in concentrator design because fine and ultra-fine grinding is required to achieve acceptable recovery rates. Therefore, the correct design of a grinding circuit is important for minimizing unit costs and increasing product quality. The use of ball mills for grinding in fine size ranges is inefficient and, therefore, vertical stirred grinding mills are becoming increasingly popular in the mineral processing industry due to its already known high energy efficiency. This work presents a hypothesis of a methodology to predict the product size distribution of a vertical stirred mill using a Bond ball mill. The Population Balance Model (PBM) was used to empirically analyze the performance of a vertical mill and a Bond ball mill. The breakage parameters obtained for both grinding mills are compared to determine the possibility of predicting the product size distribution of a vertical mill based on the results obtained from the Bond ball mill. The biggest advantage of this methodology is that most of the minerals processing laboratories already have a Bond ball mill to perform the tests suggested in this study. Preliminary results show the possibility of predicting the performance of a laboratory vertical stirred mill using a Bond ball mill.
Keywords: Bond ball mill, population balance model, product size distribution, vertical stirred mill.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1149339 Automatic Classification of Lung Diseases from CT Images
Authors: Abobaker Mohammed Qasem Farhan, Shangming Yang, Mohammed Al-Nehari
Abstract:
Pneumonia is a kind of lung disease that creates congestion in the chest. Such pneumonic conditions lead to loss of life due to the severity of high congestion. Pneumonic lung disease is caused by viral pneumonia, bacterial pneumonia, or COVID-19 induced pneumonia. The early prediction and classification of such lung diseases help reduce the mortality rate. We propose the automatic Computer-Aided Diagnosis (CAD) system in this paper using the deep learning approach. The proposed CAD system takes input from raw computerized tomography (CT) scans of the patient's chest and automatically predicts disease classification. We designed the Hybrid Deep Learning Algorithm (HDLA) to improve accuracy and reduce processing requirements. The raw CT scans are pre-processed first to enhance their quality for further analysis. We then applied a hybrid model that consists of automatic feature extraction and classification. We propose the robust 2D Convolutional Neural Network (CNN) model to extract the automatic features from the pre-processed CT image. This CNN model assures feature learning with extremely effective 1D feature extraction for each input CT image. The outcome of the 2D CNN model is then normalized using the Min-Max technique. The second step of the proposed hybrid model is related to training and classification using different classifiers. The simulation outcomes using the publicly available dataset prove the robustness and efficiency of the proposed model compared to state-of-art algorithms.
Keywords: CT scans, COVID-19, deep learning, image processing, pneumonia, lung disease.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 610338 Application of Griddization Management to Construction Hazard Management
Authors: Lingzhi Li, Jiankun Zhang, Tiantian Gu
Abstract:
Hazard management that can prevent fatal accidents and property losses is a fundamental process during the buildings’ construction stage. However, due to lack of safety supervision resources and operational pressures, the conduction of hazard management is poor and ineffective in China. In order to improve the quality of construction safety management, it is critical to explore the use of information technologies to ensure that the process of hazard management is efficient and effective. After exploring the existing problems of construction hazard management in China, this paper develops the griddization management model for construction hazard management. First, following the knowledge grid infrastructure, the griddization computing infrastructure for construction hazards management is designed which includes five layers: resource entity layer, information management layer, task management layer, knowledge transformation layer and application layer. This infrastructure will be as the technical support for realizing grid management. Second, this study divides the construction hazards into grids through city level, district level and construction site level according to grid principles. Last, a griddization management process including hazard identification, assessment and control is developed. Meanwhile, all stakeholders of construction safety management, such as owners, contractors, supervision organizations and government departments, should take the corresponding responsibilities in this process. Finally, a case study based on actual construction hazard identification, assessment and control is used to validate the effectiveness and efficiency of the proposed griddization management model. The advantage of this designed model is to realize information sharing and cooperative management between various safety management departments.
Keywords: Construction hazard, grid management, griddization computing, process.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1579