Search results for: First Order Ordinary Differential Equations
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 6540

Search results for: First Order Ordinary Differential Equations

600 Optimal Duty-Cycle Modulation Scheme for Analog-To-Digital Conversion Systems

Authors: G. Sonfack, J. Mbihi, B. Lonla Moffo

Abstract:

This paper presents an optimal duty-cycle modulation (ODCM) scheme for analog-to-digital conversion (ADC) systems. The overall ODCM-Based ADC problem is decoupled into optimal DCM and digital filtering sub-problems, while taking into account constraints of mutual design parameters between the two. Using a set of three lemmas and four morphological theorems, the ODCM sub-problem is modelled as a nonlinear cost function with nonlinear constraints. Then, a weighted least pth norm of the error between ideal and predicted frequency responses is used as a cost function for the digital filtering sub-problem. In addition, MATLAB fmincon and MATLAB iirlnorm tools are used as optimal DCM and least pth norm solvers respectively. Furthermore, the virtual simulation scheme of an overall prototyping ODCM-based ADC system is implemented and well tested with the help of Simulink tool according to relevant set of design data, i.e., 3 KHz of modulating bandwidth, 172 KHz of maximum modulation frequency and 25 MHZ of sampling frequency. Finally, the results obtained and presented show that the ODCM-based ADC achieves under 3 KHz of modulating bandwidth: 57 dBc of SINAD (signal-to-noise and distorsion), 58 dB of SFDR (Surpious free dynamic range) -80 dBc of THD (total harmonic distorsion), and 10 bits of minimum resolution. These performance levels appear to be a great challenge within the class of oversampling ADC topologies, with 2nd order IIR (infinite impulse response) decimation filter.

Keywords: Digital IIR filter, morphological lemmas and theorems, optimal DCM-based DAC, virtual simulation, weighted least pth norm.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 900
599 Students' Perception of Virtual Learning Environment (VLE) Skills in Setting up the Simulator Welding Technology

Authors: Mohd Afif Md Nasir, Faizal Amin NurYunus, Jamaluddin Hashim, Abd Samad Hassan Basari, A. Halim Sahelan

Abstract:

The aim of this study is to identify the suitability of Virtual Learning Environment (VLE) in welding simulator application towards Computer-Based Training (CBT) in developing skills upon new students at the Advanced Technology Training Center (ADTEC) Batu Pahat, Johor, Malaysia and GIATMARA, Batu Pahat, Johor, Malaysia. The significance of the study is to create a computer-based skills development approach in welding technology among new students in ADTEC and GIATMARA as well as to cultivate the elements of general skills among them. This study is also important in elevating the number of individual knowledge workers (K-workers) working in manufacturing industry in order to achieve a national vision which is to be an industrial nation in the year of 2020. The design of the study is a survey type of research which using questionnaires as the instruments and some 136 students from ADTEC and GIATMARA were interviewed. Descriptive analysis is used to identify the frequency and mean values. The findings of the study show that the welding technology has developed skills in the students because of the application of VLE simulated at a high level and the respondents agreed that the skills could be embedded through the application of the VLE simulator. In summary, the VLE simulator is suitable in welding skills development training in terms of exposing new students with the relevant characteristics of welding skills and at the same time spurring the students’ interest towards learning more about the skills.

Keywords: Computer-Based Training (CBT), knowledge workers (K-workers), virtual learning environment, welding simulator, welding technology.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2070
598 A Comparative Analysis Approach Based on Fuzzy AHP, TOPSIS and PROMETHEE for the Selection Problem of GSCM Solutions

Authors: Omar Boutkhoum, Mohamed Hanine, Abdessadek Bendarag

Abstract:

Sustainable economic growth is nowadays driving firms to extend toward the adoption of many green supply chain management (GSCM) solutions. However, the evaluation and selection of these solutions is a matter of concern that needs very serious decisions, involving complexity owing to the presence of various associated factors. To resolve this problem, a comparative analysis approach based on multi-criteria decision-making methods is proposed for adequate evaluation of sustainable supply chain management solutions. In the present paper, we propose an integrated decision-making model based on FAHP (Fuzzy Analytic Hierarchy Process), TOPSIS (Technique for Order of Preference by Similarity to Ideal Solution) and PROMETHEE (Preference Ranking Organisation METHod for Enrichment Evaluations) to contribute to a better understanding and development of new sustainable strategies for industrial organizations. Due to the varied importance of the selected criteria, FAHP is used to identify the evaluation criteria and assign the importance weights for each criterion, while TOPSIS and PROMETHEE methods employ these weighted criteria as inputs to evaluate and rank the alternatives. The main objective is to provide a comparative analysis based on TOPSIS and PROMETHEE processes to help make sound and reasoned decisions related to the selection problem of GSCM solution.

Keywords: GSCM solutions, multi-criteria analysis, FAHP, TOPSIS, PROMETHEE, decision support system.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 905
597 A Study of the Planning and Designing of the Built Environment under the Green Transit-Oriented Development

Authors: Wann-Ming Wey

Abstract:

In recent years, the problems of global climate change and natural disasters have induced the concerns and attentions of environmental sustainability issues for the public. Aside from the environmental planning efforts done for human environment, Transit-Oriented Development (TOD) has been widely used as one of the future solutions for the sustainable city development. In order to be more consistent with the urban sustainable development, the development of the built environment planning based on the concept of Green TOD which combines both TOD and Green Urbanism is adapted here. The connotation of the urban development under the green TOD including the design toward environment protect, the maximum enhancement resources and the efficiency of energy use, use technology to construct green buildings and protected areas, natural ecosystems and communities linked, etc. Green TOD is not only to provide the solution to urban traffic problems, but to direct more sustainable and greener consideration for future urban development planning and design. In this study, we use both the TOD and Green Urbanism concepts to proceed to the study of the built environment planning and design. Fuzzy Delphi Technique (FDT) is utilized to screen suitable criteria of the green TOD. Furthermore, Fuzzy Analytic Network Process (FANP) and Quality Function Deployment (QFD) were then developed to evaluate the criteria and prioritize the alternatives. The study results can be regarded as the future guidelines of the built environment planning and designing under green TOD development in Taiwan.

Keywords: Green transit-oriented development, built environment, fuzzy Delphi technique, quality function deployment, fuzzy analytic network process.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1493
596 Managing Iterations in Product Design and Development

Authors: K. Aravindhan, Trishit Bandyopadhyay, Mahesh Mehendale, Supriya Kumar De

Abstract:

The inherent iterative nature of product design and development poses significant challenge to reduce the product design and development time (PD). In order to shorten the time to market, organizations have adopted concurrent development where multiple specialized tasks and design activities are carried out in parallel. Iterative nature of work coupled with the overlap of activities can result in unpredictable time to completion and significant rework. Many of the products have missed the time to market window due to unanticipated or rather unplanned iteration and rework. The iterative and often overlapped processes introduce greater amounts of ambiguity in design and development, where the traditional methods and tools of project management provide less value. In this context, identifying critical metrics to understand the iteration probability is an open research area where significant contribution can be made given that iteration has been the key driver of cost and schedule risk in PD projects. Two important questions that the proposed study attempts to address are: Can we predict and identify the number of iterations in a product development flow? Can we provide managerial insights for a better control over iteration? The proposal introduces the concept of decision points and using this concept intends to develop metrics that can provide managerial insights into iteration predictability. By characterizing the product development flow as a network of decision points, the proposed research intends to delve further into iteration probability and attempts to provide more clarity.

Keywords: Decision Points, Iteration, Product Design, Rework.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2168
595 ZnS and Graphene Quantum Dots Nanocomposite as Potential Electron Acceptor for Photovoltaics

Authors: S. M. Giripunje, Shikha Jindal

Abstract:

Zinc sulphide (ZnS) quantum dots (QDs) were synthesized successfully via simple sonochemical method. X-ray diffraction (XRD), scanning electron microscopy (SEM) and high resolution transmission electron microscopy (HRTEM) analysis revealed the average size of QDs of the order of 3.7 nm. The band gap of the QDs was tuned to 5.2 eV by optimizing the synthesis parameters. UV-Vis absorption spectra of ZnS QD confirm the quantum confinement effect. Fourier transform infrared (FTIR) analysis confirmed the formation of single phase ZnS QDs. To fabricate the diode, blend of ZnS QDs and P3HT was prepared and the heterojunction of PEDOT:PSS and the blend was formed by spin coating on indium tin oxide (ITO) coated glass substrate. The diode behaviour of the heterojunction was analysed, wherein the ideality factor was found to be 2.53 with turn on voltage 0.75 V and the barrier height was found to be 1.429 eV. ZnS-Graphene QDs nanocomposite was characterised for the surface morphological study. It was found that the synthesized ZnS QDs appear as quasi spherical particles on the graphene sheets. The average particle size of ZnS-graphene nanocomposite QDs was found to be 8.4 nm. From voltage-current characteristics of ZnS-graphene nanocomposites, it is observed that the conductivity of the composite increases by 104 times the conductivity of ZnS QDs. Thus the addition of graphene QDs in ZnS QDs enhances the mobility of the charge carriers in the composite material. Thus, the graphene QDs, with high specific area for a large interface, high mobility and tunable band gap, show a great potential as an electron-acceptors in photovoltaic devices.

Keywords: Graphene, mobility, nanocomposites, photovoltaics, quantum dots, zinc sulphide.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1381
594 Engineering of E-Learning Content Creation: Case Study for African Countries

Authors: María-Dolores Afonso-Suárez, Nayra Pumar-Carreras, Juan Ruiz-Alzola

Abstract:

This research addresses the use of an e-Learning creation methodology for learning objects. Throughout the process, indicators are being gathered, to determine if it responds to the main objectives of an engineering discipline. These parameters will also indicate if it is necessary to review the creation cycle and readjust any phase. Within the project developed for this study, apart from the use of structured methods, there has been a central objective: the establishment of a learning atmosphere. A place where all the professionals involved are able to collaborate, plan, solve problems and determine guides to follow in order to develop creative and innovative solutions. It has been outlined as a blended learning program with an assessment plan that proposes face to face lessons, coaching, collaboration, multimedia and web based learning objects as well as support resources. The project has been drawn as a long term task, the pilot teaching actions designed provide the preliminary results object of study. This methodology is been used in the creation of learning content for the African countries of Senegal, Mauritania and Cape Verde. It has been developed within the framework of the MACbioIDi, an Interreg European project for the International cooperation and development. The educational area of this project is focused in the training and advice of professionals of the medicine as well as engineers in the use of applications of medical imaging technology, specifically the 3DSlicer application and the Open Anatomy Browser.

Keywords: Teaching contents engineering, e-learning, blended learning, international cooperation, 3DSlicer, open anatomy browser.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1013
593 Hospital Waste Management Practices: A Case Study in Iran

Authors: M. Farzadkia, S. Jorfi

Abstract:

Hospital waste is a category of waste consisting of infectious and non-infectious waste, which pose environmental and health risks. Therefore, special planning and management is required, due to the potential hazards of them. The lack of valid and comprehensive information regarding the generation and management of hospital waste in Iran is one of the most important problems in this field. This research aimed to evaluate hospital waste management efficiency in Karaj city, Iran. The four greatest hospitals in Karaj city had been selected in this cross-sectional study. Site observations and interviews with employees were implemented. The data was gathered based on the hospital waste management questionnaire which was designed by World Health Organization for developing countries. Collected Data had been analyzed using SPSS software. The average of solid waste which was generated per bed was 2.78 kg, which included 90% of domestic waste and 10% of infectious waste. Based on the quantitative analysis of general and infectious waste in these hospitals, the highest contributors of general waste were consisting of food waste (37.39%), while textile (28.06%) were the highest contributors of the infectious waste. According to the information contained in the questionnaires, the main defects of waste management in these hospitals were; inadequate staff in waste management sector, poorly disinfection of solid waste containers and temporary storage locations, and a lack of proper infectious waste treatment. According to the results of this research, waste management in these hospitals were far from optimum conditions. In order to improve the existing conditions, mentioned problems must be solved quickly, and planning for continuous monitoring in the waste management field in these hospitals should be established.

Keywords: Waste management, hospital wastes, solid wastes, Iran.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2118
592 An Assessment of Water Pollution of the Beshar River Aquatic Ecosystems

Authors: Amir Eghbal Khajeh Rahimi, Fardin Boustani, Omid Tabiee, Masoud Hashemi

Abstract:

The Beshar River is one of the most important aquatic ecosystems in the upstream of the Karun watershed in south of Iran which is affected by point and non point pollutant sources . This study was done in order to evaluate the effects of pollutants activities on the water quality of the Beshar river and its aquatic ecosystems. This river is approximately 190 km in length and situated at the geographical positions of 51° 20´ to 51° 48´ E and 30° 18´ to 30° 52´ N it is one of the most important aquatic ecosystems of Kohkiloye and Boyerahmad province in south-west Iran. In this research project, five study stations were selected to examine water pollution in the Beshar River systems. Human activity is now one of the most important factors affecting on hydrology and water quality of the Beshar river. Humans use large amounts of resources to sustain various standards of living, although measures of sustainability are highly variable depending on how sustainability is defined. The Beshar river ecosystems are particularly sensitive and vulnerable to human activities. Therefore, to determine the impact of human activities on the Beshar River, the most important water quality parameters such as pH, dissolve oxygen (DO), Biological Oxygen Demand (BOD5), Total Dissolve Solids (TDS), Nitrates (NO3-N) and Phosphates (PO4) were estimated at the five stations. As the results show, the most important pollution index parameters such as BOD5, NO3 and PO4 increase and DO and pH decrease according to human activities (P<0.05). However, due to pollutant degradation and dilution, pollution index parameters improve downstream sampling stations.

Keywords: Human activities, Water pollution, Beshar River, Iran.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1890
591 A Study on Algorithm Fusion for Recognition and Tracking of Moving Robot

Authors: Jungho Choi, Youngwan Cho

Abstract:

This paper presents an algorithm for the recognition and tracking of moving objects, 1/10 scale model car is used to verify performance of the algorithm. Presented algorithm for the recognition and tracking of moving objects in the paper is as follows. SURF algorithm is merged with Lucas-Kanade algorithm. SURF algorithm has strong performance on contrast, size, rotation changes and it recognizes objects but it is slow due to many computational complexities. Processing speed of Lucas-Kanade algorithm is fast but the recognition of objects is impossible. Its optical flow compares the previous and current frames so that can track the movement of a pixel. The fusion algorithm is created in order to solve problems which occurred using the Kalman Filter to estimate the position and the accumulated error compensation algorithm was implemented. Kalman filter is used to create presented algorithm to complement problems that is occurred when fusion two algorithms. Kalman filter is used to estimate next location, compensate for the accumulated error. The resolution of the camera (Vision Sensor) is fixed to be 640x480. To verify the performance of the fusion algorithm, test is compared to SURF algorithm under three situations, driving straight, curve, and recognizing cars behind the obstacles. Situation similar to the actual is possible using a model vehicle. Proposed fusion algorithm showed superior performance and accuracy than the existing object recognition and tracking algorithms. We will improve the performance of the algorithm, so that you can experiment with the images of the actual road environment.

Keywords: SURF, Optical Flow Lucas-Kanade, Kalman Filter, object recognition, object tracking.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2265
590 An Advanced Stereo Vision Based Obstacle Detection with a Robust Shadow Removal Technique

Authors: Saeid Fazli, Hajar Mohammadi D., Payman Moallem

Abstract:

This paper presents a robust method to detect obstacles in stereo images using shadow removal technique and color information. Stereo vision based obstacle detection is an algorithm that aims to detect and compute obstacle depth using stereo matching and disparity map. The proposed advanced method is divided into three phases, the first phase is detecting obstacles and removing shadows, the second one is matching and the last phase is depth computing. We propose a robust method for detecting obstacles in stereo images using a shadow removal technique based on color information in HIS space, at the first phase. In this paper we use Normalized Cross Correlation (NCC) function matching with a 5 × 5 window and prepare an empty matching table τ and start growing disparity components by drawing a seed s from S which is computed using canny edge detector, and adding it to τ. In this way we achieve higher performance than the previous works [2,17]. A fast stereo matching algorithm is proposed that visits only a small fraction of disparity space in order to find a semi-dense disparity map. It works by growing from a small set of correspondence seeds. The obstacle identified in phase one which appears in the disparity map of phase two enters to the third phase of depth computing. Finally, experimental results are presented to show the effectiveness of the proposed method.

Keywords: obstacle detection, stereo vision, shadowremoval, color, stereo matching

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2047
589 Growth and Stomatal Responses of Bread Wheat Genotypes in Tolerance to Salt Stress

Authors: Afrasyab Rahnama, Kazem Poustini, Reza Tavakkol-Afshari, Afshin Tavakoli

Abstract:

Plant growth is affected by the osmotic stress as well as toxicity of salt in leaves. In order to study of salt stress effects on stomatal conductance and growth rate and relationship between them as wells osmotic and Na+-specific effects on these traits, four bread wheat genotypes differing in salt tolerance were selected. Salinity was applied when the leaf 4 was fully expanded. Sodium (Na+) concentrations in flag leaf blade at 3 salinity levels (0, 100 and 200 mM NaCl) were measured. Salt-tolerant genotypes showed higher stomatal conductance and growth rate compared to salt-sensitive ones. After 10 and 20 days exposure to salt, stomatal conductance and relative growth rate were reduced, but the reduction was greater in sensitive genotypes. Growth rate was reduced severely in the first period (1-10 days) of salt commencements and it was due to osmotic effect of salt not Na+ toxicity. In the second period (11-20 days) after salt treatment growth reduced only when salt accumulated to toxic concentrations in the leaves. A positive relationship between stomatal conductance and relative growth rate showed that stomatal conductance can be a reliable indicator of growth rate, and finally can be considered as a sensitive indicator of the osmotic stress. It seems 20 days after salinity, the major effect of salt, especially at low to moderate salinity levels on growth properties was due to the osmotic effect of salt, not to Na+-specific effects within the plant.

Keywords: Osmotic stress, relative growth rate, stomatal conductance, wheat.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2255
588 Tool Wear of Titanium/Tungsten/Silicon/Aluminum-based-coated end Mill Cutters in Millin Hardened Steel

Authors: Tadahiro Wada, Koji Iwamoto

Abstract:

In turning hardened steel, polycrystalline cubic boron nitride (cBN) compacts are widely used, due to their higher hardness and higher thermal conductivity. However, in milling hardened steel, fracture of cBN cutting tools readily occurs because they have poor fracture toughness. Therefore, coated cemented carbide tools, which have good fracture toughness and wear resistance, are generally widely used. In this study, hardened steel (ASTM D2, JIS SKD11, 60HRC) was milled with three physical vapor deposition (PVD)-coated cemented carbide end mill cutters in order to determine effective tool materials for cutting hardened steel at high cutting speeds. The coating films used were (Ti,W)N/(Ti,W,Si)N and (Ti,W)N/(Ti,W,Si,Al)N coating films. (Ti,W,Si,Al)N is a new type of coating film. The inner layer of the (Ti,W)N/(Ti,W,Si)N and (Ti,W)N/(Ti,W,Si,Al)N coating system is (Ti,W)N coating film, and the outer layer is (Ti,W,Si)N and (Ti,W,Si,Al)N coating films, respectively. Furthermore, commercial (Ti,Al)N-based coating film was also used. The following results were obtained: (1) In milling hardened steel at a cutting speed of 3.33 m/s, the tool wear width of the (Ti,W)N/(Ti,W,Si,Al)N-coated tool was smaller than that of the (Ti,W)N/(Ti,W,Si)N-coated tool. And, compared with the commercial (Ti,Al)N, the tool wear width of the (Ti,W)N/(Ti,W,Si,Al)N-coated tool was smaller than that of the (Ti,Al)N-coated tool. (2) The tool wear of the (Ti,W)N/(Ti,W,Si,Al)N-coated tool increased with an increase in cutting speed. (3) The (Ti,W)N/(Ti,W,Si,Al)N-coated cemented carbide was an effective tool material for high-speed cutting below a cutting speed of 3.33 m/s.

Keywords: cutting, physical vapor deposition (PVD) coating system, hardened steel, tool wear

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2027
587 Forming Limit Analysis of DP600-800 Steels

Authors: M. C. Cardoso, L. P. Moreira

Abstract:

In this work, the plastic behaviour of cold-rolled zinc coated dual-phase steel sheets DP600 and DP800 grades is firstly investigated with the help of uniaxial, hydraulic bulge and Forming Limit Curve (FLC) tests. The uniaxial tensile tests were performed in three angular orientations with respect to the rolling direction to evaluate the strain-hardening and plastic anisotropy. True stressstrain curves at large strains were determined from hydraulic bulge testing and fitted to a work-hardening equation. The limit strains are defined at both localized necking and fracture conditions according to Nakajima’s hemispherical punch procedure. Also, an elasto-plastic localization model is proposed in order to predict strain and stress based forming limit curves. The investigated dual-phase sheets showed a good formability in the biaxial stretching and drawing FLC regions. For both DP600 and DP800 sheets, the corresponding numerical predictions overestimated and underestimated the experimental limit strains in the biaxial stretching and drawing FLC regions, respectively. This can be attributed to the restricted failure necking condition adopted in the numerical model, which is not suitable to describe the tensile and shear fracture mechanisms in advanced high strength steels under equibiaxial and biaxial stretching conditions.

Keywords: Advanced high strength steels, forming limit curve, numerical modeling, sheet metal forming.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3431
586 Near-Infrared Hyperspectral Imaging Spectroscopy to Detect Microplastics and Pieces of Plastic in Almond Flour

Authors: H. Apaza, L. Chévez, H. Loro

Abstract:

Plastic and microplastic pollution in human food chain is a big problem for human health that requires more elaborated techniques that can identify their presences in different kinds of food. Hyperspectral imaging technique is an optical technique than can detect the presence of different elements in an image and can be used to detect plastics and microplastics in a scene. To do this statistical techniques are required that need to be evaluated and compared in order to find the more efficient ones. In this work, two problems related to the presence of plastics are addressed, the first is to detect and identify pieces of plastic immersed in almond seeds, and the second problem is to detect and quantify microplastic in almond flour. To do this we make use of the analysis hyperspectral images taken in the range of 900 to 1700 nm using 4 unmixing techniques of hyperspectral imaging which are: least squares unmixing (LSU), non-negatively constrained least squares unmixing (NCLSU), fully constrained least squares unmixing (FCLSU), and scaled constrained least squares unmixing (SCLSU). NCLSU, FCLSU, SCLSU techniques manage to find the region where the plastic is found and also manage to quantify the amount of microplastic contained in the almond flour. The SCLSU technique estimated a 13.03% abundance of microplastics and 86.97% of almond flour compared to 16.66% of microplastics and 83.33% abundance of almond flour prepared for the experiment. Results show the feasibility of applying near-infrared hyperspectral image analysis for the detection of plastic contaminants in food.

Keywords: Food, plastic, microplastic, NIR hyperspectral imaging, unmixing.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1145
585 Digital Automatic Gain Control Integrated on WLAN Platform

Authors: Emilija Miletic, Milos Krstic, Maxim Piz, Michael Methfessel

Abstract:

In this work we present a solution for DAGC (Digital Automatic Gain Control) in WLAN receivers compatible to IEEE 802.11a/g standard. Those standards define communication in 5/2.4 GHz band using Orthogonal Frequency Division Multiplexing OFDM modulation scheme. WLAN Transceiver that we have used enables gain control over Low Noise Amplifier (LNA) and a Variable Gain Amplifier (VGA). The control over those signals is performed in our digital baseband processor using dedicated hardware block DAGC. DAGC in this process is used to automatically control the VGA and LNA in order to achieve better signal-to-noise ratio, decrease FER (Frame Error Rate) and hold the average power of the baseband signal close to the desired set point. DAGC function in baseband processor is done in few steps: measuring power levels of baseband samples of an RF signal,accumulating the differences between the measured power level and actual gain setting, adjusting a gain factor of the accumulation, and applying the adjusted gain factor the baseband values. Based on the measurement results of RSSI signal dependence to input power we have concluded that this digital AGC can be implemented applying the simple linearization of the RSSI. This solution is very simple but also effective and reduces complexity and power consumption of the DAGC. This DAGC is implemented and tested both in FPGA and in ASIC as a part of our WLAN baseband processor. Finally, we have integrated this circuit in a compact WLAN PCMCIA board based on MAC and baseband ASIC chips designed from us.

Keywords: WLAN, AGC, RSSI, baseband processor

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3927
584 Drivers of Customer Satisfaction in an Industrial Company from Marketing Aspect

Authors: M. Arefi, A.M. Amini, K. Fallahi

Abstract:

One of the basic concepts in marketing is the concept of meeting customers- needs. Since customer satisfaction is essential for lasting survival and development of a business, screening and observing customer satisfaction and recognizing its underlying factors must be one of the key activities of every business. The purpose of this study is to recognize the drivers that effect customer satisfaction in a business-to-business situation in order to improve marketing activities. We conducted a survey in which 93 business customers of a manufacturer of Diesel Generator in Iran participated and they talked about their ideas and satisfaction of supplier-s services related to its products. We developed the measures for drivers of satisfaction first by as investigative research (by means of feedback from executives and customers of sponsoring firm). Then based on these measures, we created a mail survey, and asked the respondents to explain their opinion about the sponsoring firm which was a supplier of diesel generator and similar products. Furthermore, the survey required the participants to mention their functional areas and their company features. In Conclusion we found that there are three drivers for customer satisfaction, which are reliability, information about product, and commercial features. Buyers/users from different functional areas attribute different degree of importance to the last two drivers. For instance, people from buying and management areas believe that commercial features are more important than information about products. But people in engineering, maintenance and production areas believe that having information about products is more important than commercial aspects. Marketing experts should consider the attribute of customers regarding information about the product and commercial features to improve market share.

Keywords: B2B, Customer satisfaction, Commercial, Industry.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2641
583 The use of Hormone Auxin in the Different Period Growth on Yield Components of Plant Vetch

Authors: Almas Tayebi, Tayeb Saki Nejad, Alireza Shoukofar

Abstract:

The trial in the city, located 170 kilometers from the Iranian city of Ahvaz was Omidiyeh. The main factor in this project includes 4 levels in control (without hormones), use of hormones in the seed, vegetative and flowering stage respectively. And sub-plots included 3 varieties of vetch in three levels, with local names, was the jewel in the study of light and Auxin in the vegetative and reproductive different times in different varieties of vetch was investigated. This test has been taken in the plots in a randomized complete block with four replications. In order to study the effects of the hormone Auxin in the growth stages (seed, vegetative and flowering) to control (no hormone Auxin) on three local varieties of vetch, the essence of light and plant height, number of pods per plant, seed number The pods, seeds per plant, grain weight, grain yield, plant dry weight and protein content were measured. Among the vetch varieties for plant height, number of pods per plant, a seed per plant, grain weight, grain yield, and plant dry weight and protein levels of 1 percent of plant and seed number per pod per plant at 5% level of There was no significant difference. Interactions for grain yield per plant, grain yield and protein levels of 1 percent and the number of seeds per pod and seed weight are significant differences in levels 5 and plant height and plant dry weight of the interaction were INFLUENCE There was no significant difference in them.

Keywords: Auxin hormones, various periods of growth, production components, vetch

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2149
582 Performance Analysis of Digital Signal Processors Using SMV Benchmark

Authors: Erh-Wen Hu, Cyril S. Ku, Andrew T. Russo, Bogong Su, Jian Wang

Abstract:

Unlike general-purpose processors, digital signal processors (DSP processors) are strongly application-dependent. To meet the needs for diverse applications, a wide variety of DSP processors based on different architectures ranging from the traditional to VLIW have been introduced to the market over the years. The functionality, performance, and cost of these processors vary over a wide range. In order to select a processor that meets the design criteria for an application, processor performance is usually the major concern for digital signal processing (DSP) application developers. Performance data are also essential for the designers of DSP processors to improve their design. Consequently, several DSP performance benchmarks have been proposed over the past decade or so. However, none of these benchmarks seem to have included recent new DSP applications. In this paper, we use a new benchmark that we recently developed to compare the performance of popular DSP processors from Texas Instruments and StarCore. The new benchmark is based on the Selectable Mode Vocoder (SMV), a speech-coding program from the recent third generation (3G) wireless voice applications. All benchmark kernels are compiled by the compilers of the respective DSP processors and run on their simulators. Weighted arithmetic mean of clock cycles and arithmetic mean of code size are used to compare the performance of five DSP processors. In addition, we studied how the performance of a processor is affected by code structure, features of processor architecture and optimization of compiler. The extensive experimental data gathered, analyzed, and presented in this paper should be helpful for DSP processor and compiler designers to meet their specific design goals.

Keywords: digital signal processors, DSP benchmark, instruction level parallelism, modified cyclomatic complexity, performance analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1579
581 The Evaluation and the Comparison of the Effect of Without Engine Power and Power Mechanical Systems on Rice Weed

Authors: F. E. Cherati, T. Naij , A. Amraei, Y. Hosseinpor

Abstract:

In order to study the influence of different methods of controlling weeds such as mechanical weeding and mechanical weeder efficiency analysis in mechanical cultivation conditions, in farming year of 2011 an experiment was done in a farm in coupling and development of technology center in Haraz,Iran. The treatments consisted of (I) control treatment: where no weeding was done, (II) use of mechanical weeding without engine and (III) power mechanical weeding. Results showed that experimental treatments had significantly different effects (p=0.05) on yield traits and number of filled grains per panicle, while treatments had the significant effects on grain weight and dry weight of weeds in the first, second and third weeding methods at 1% of confidence level. Treatment (II) had its most significant effect on number of filled grains per panicle and yield performance standpoint, which was 3705.97 kg ha-1 in its highest peak. Treatment (III) was ranked as second influential with 3559.8 kg ha-1. In addition, under (I) treatments, 2364.73 kg ha-1 of yield produced. The minimum dry weights of weeds in all weeding methods were related to the treatment (II), (III) and (I), respectively. The correlation coefficient analysis showed that total yield had a significant positive correlation with the panicle grain yield per plant (r= 0.55*) and the number of grains per panicle-1 (r= 0.57*) and the number of filled grains (r= 0.63*). Total rice yield also had negative correlation of r= -0. 64* with weed dry weight at second weed sampling time (17 DAT). The weed dry weight at third and fourth sampling times (24 and 40 DAT) had negative correlations of -0.65** and r=-0.61* with rice yield, respectively.

Keywords: Dry weight, without engine mechanical weeder, power mechanical weeder, yield rice.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1427
580 Image Magnification Using Adaptive Interpolationby Pixel Level Data-Dependent Geometrical Shapes

Authors: Muhammad Sajjad, Naveed Khattak, Noman Jafri

Abstract:

World has entered in 21st century. The technology of computer graphics and digital cameras is prevalent. High resolution display and printer are available. Therefore high resolution images are needed in order to produce high quality display images and high quality prints. However, since high resolution images are not usually provided, there is a need to magnify the original images. One common difficulty in the previous magnification techniques is that of preserving details, i.e. edges and at the same time smoothing the data for not introducing the spurious artefacts. A definitive solution to this is still an open issue. In this paper an image magnification using adaptive interpolation by pixel level data-dependent geometrical shapes is proposed that tries to take into account information about the edges (sharp luminance variations) and smoothness of the image. It calculate threshold, classify interpolation region in the form of geometrical shapes and then assign suitable values inside interpolation region to the undefined pixels while preserving the sharp luminance variations and smoothness at the same time. The results of proposed technique has been compared qualitatively and quantitatively with five other techniques. In which the qualitative results show that the proposed method beats completely the Nearest Neighbouring (NN), bilinear(BL) and bicubic(BC) interpolation. The quantitative results are competitive and consistent with NN, BL, BC and others.

Keywords: Adaptive, digital image processing, imagemagnification, interpolation, geometrical shapes, qualitative &quantitative analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1773
579 Economic Assessment of Green House for Cultivation of Float Based Seedling Production in India

Authors: Srinath Ramakkrushnan, Aswathaman Vijayan

Abstract:

In conventional seedling production, the seedlings are being grown in the open field under natural conditions. Here they are susceptible to sudden changes in climate were their quality and yield is affected. Quality seedlings are essential for good growth and performance of crops in main field; they serve as a foundation for the economic returns to the farmer. Producing quality seedling demands usage of hybrid seeds as they have the ability to result in better yield, greater uniformity, improved color, disease resistance, and so forth. Hybrid seed production poses major operational challenge and its seed use efficiency plays an important role. Thus in order to overcome the difficulties currently present in conventional seedling production and to efficiently use hybrid seeds, ITC Limited Agri Business Divisions - Sustainability Cell as conceptualized a novel method of seedling production unit for farmers in West Godavari District of Andhra Pradesh. The “Green House based Float Seedling" methodology aims at a protected cultivation technique wherein the micro climate surrounding the plant/seedling body is controlled partially or fully as per the requirement of the species. This paper reports on the techno economic evaluation of green house for cultivation of float based seedling production with experimental results that was attained from the pilot implementation in West Godavari District, Rajahmundry region of India.

Keywords: Economic Assessment, Float Seedling, Green House, ITC Limited, Payback period.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4176
578 Hybrid Recovery of Copper and Silver from PV Ribbon and Ag Finger of EOL Solar Panels

Authors: T. Patcharawit, C. Kansomket, N. Wongnaree, W. Kritsrikan, T. Yingnakorn, S. Khumkoa

Abstract:

Recovery of pure copper and silver from end-of-life photovoltaic (PV) panels was investigated in this paper using an effective hybrid pyro-hydrometallurgical process. In the first step of waste treatment, solar panel waste was first dismantled to obtain a PV sheet to be cut and calcined at 500 °C, to separate out PV ribbon from glass cullet, ash, and volatile while the silicon wafer containing silver finger was collected for recovery. In the second step of metal recovery, copper recovery from PV ribbon was via 1-3 M HCl leaching with SnCl₂ and H₂O₂ additions in order to remove the tin-lead coating on the ribbon. The leached copper band was cleaned and subsequently melted as an anode for the next step of electrorefining. Stainless steel was set as the cathode with CuSO₄ as an electrolyte, and at a potential of 0.2 V, high purity copper of 99.93% was obtained at 96.11% recovery after 24 hours. For silver recovery, the silicon wafer containing silver finger was leached using HNO₃ at 1-4 M in an ultrasonic bath. In the next step of precipitation, silver chloride was then obtained and subsequently reduced by sucrose and NaOH to give silver powder prior to oxy-acetylene melting to finally obtain pure silver metal. The integrated recycling process is considered to be economical, providing effective recovery of high purity metals such as copper and silver while other materials such as aluminum, copper wire, glass cullet can also be recovered to be reused commercially. Compounds such as PbCl₂ and SnO₂ obtained can also be recovered to enter the market.

Keywords: Electrorefining, leaching, calcination, PV ribbon, silver finger, solar panel.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 427
577 Performance Evaluation of Parallel Surface Modeling and Generation on Actual and Virtual Multicore Systems

Authors: Nyeng P. Gyang

Abstract:

Even though past, current and future trends suggest that multicore and cloud computing systems are increasingly prevalent/ubiquitous, this class of parallel systems is nonetheless underutilized, in general, and barely used for research on employing parallel Delaunay triangulation for parallel surface modeling and generation, in particular. The performances, of actual/physical and virtual/cloud multicore systems/machines, at executing various algorithms, which implement various parallelization strategies of the incremental insertion technique of the Delaunay triangulation algorithm, were evaluated. T-tests were run on the data collected, in order to determine whether various performance metrics differences (including execution time, speedup and efficiency) were statistically significant. Results show that the actual machine is approximately twice faster than the virtual machine at executing the same programs for the various parallelization strategies. Results, which furnish the scalability behaviors of the various parallelization strategies, also show that some of the differences between the performances of these systems, during different runs of the algorithms on the systems, were statistically significant. A few pseudo superlinear speedup results, which were computed from the raw data collected, are not true superlinear speedup values. These pseudo superlinear speedup values, which arise as a result of one way of computing speedups, disappear and give way to asymmetric speedups, which are the accurate kind of speedups that occur in the experiments performed.

Keywords: Cloud computing systems, multicore systems, parallel delaunay triangulation, parallel surface modeling and generation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 857
576 Application of Griddization Management to Construction Hazard Management

Authors: Lingzhi Li, Jiankun Zhang, Tiantian Gu

Abstract:

Hazard management that can prevent fatal accidents and property losses is a fundamental process during the buildings’ construction stage. However, due to lack of safety supervision resources and operational pressures, the conduction of hazard management is poor and ineffective in China. In order to improve the quality of construction safety management, it is critical to explore the use of information technologies to ensure that the process of hazard management is efficient and effective. After exploring the existing problems of construction hazard management in China, this paper develops the griddization management model for construction hazard management. First, following the knowledge grid infrastructure, the griddization computing infrastructure for construction hazards management is designed which includes five layers: resource entity layer, information management layer, task management layer, knowledge transformation layer and application layer. This infrastructure will be as the technical support for realizing grid management. Second, this study divides the construction hazards into grids through city level, district level and construction site level according to grid principles. Last, a griddization management process including hazard identification, assessment and control is developed. Meanwhile, all stakeholders of construction safety management, such as owners, contractors, supervision organizations and government departments, should take the corresponding responsibilities in this process. Finally, a case study based on actual construction hazard identification, assessment and control is used to validate the effectiveness and efficiency of the proposed griddization management model. The advantage of this designed model is to realize information sharing and cooperative management between various safety management departments.

Keywords: Construction hazard, grid management, griddization computing, process.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1549
575 ICT for Smart Appliances: Current Technology and Identification of Future ICT Trend

Authors: Abubakar Uba Ibrahim, Ibrahim Haruna Shanono

Abstract:

Smart metering and demand response are gaining ground in industrial and residential applications. Smart Appliances have been given concern towards achieving Smart home. The success of Smart grid development relies on the successful implementation of Information and Communication Technology (ICT) in power sector. Smart Appliances have been the technology under development and many new contributions to its realization have been reported in the last few years. The role of ICT here is to capture data in real time, thereby allowing bi-directional flow of information/data between producing and utilization point; that lead a way for the attainment of Smart appliances where home appliances can communicate between themselves and provide a self-control (switch on and off) using the signal (information) obtained from the grid. This paper depicts the background on ICT for smart appliances paying a particular attention to the current technology and identifying the future ICT trends for load monitoring through which smart appliances can be achieved to facilitate an efficient smart home system which promote demand response program. This paper grouped and reviewed the recent contributions, in order to establish the current state of the art and trends of the technology, so that the reader can be provided with a comprehensive and insightful review of where ICT for smart appliances stands and is heading to. The paper also presents a brief overview of communication types, and then narrowed the discussion to the load monitoring (Non-intrusive Appliances Load Monitoring ‘NALM’). Finally, some future trends and challenges in the further development of the ICT framework are discussed to motivate future contributions that address open problems and explore new possibilities.

Keywords: Communication technology between appliances, demand response, load monitoring, smart appliances and smart grid.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2519
574 Image Ranking to Assist Object Labeling for Training Detection Models

Authors: Tonislav Ivanov, Oleksii Nedashkivskyi, Denis Babeshko, Vadim Pinskiy, Matthew Putman

Abstract:

Training a machine learning model for object detection that generalizes well is known to benefit from a training dataset with diverse examples. However, training datasets usually contain many repeats of common examples of a class and lack rarely seen examples. This is due to the process commonly used during human annotation where a person would proceed sequentially through a list of images labeling a sufficiently high total number of examples. Instead, the method presented involves an active process where, after the initial labeling of several images is completed, the next subset of images for labeling is selected by an algorithm. This process of algorithmic image selection and manual labeling continues in an iterative fashion. The algorithm used for the image selection is a deep learning algorithm, based on the U-shaped architecture, which quantifies the presence of unseen data in each image in order to find images that contain the most novel examples. Moreover, the location of the unseen data in each image is highlighted, aiding the labeler in spotting these examples. Experiments performed using semiconductor wafer data show that labeling a subset of the data, curated by this algorithm, resulted in a model with a better performance than a model produced from sequentially labeling the same amount of data. Also, similar performance is achieved compared to a model trained on exhaustive labeling of the whole dataset. Overall, the proposed approach results in a dataset that has a diverse set of examples per class as well as more balanced classes, which proves beneficial when training a deep learning model.

Keywords: Computer vision, deep learning, object detection, semiconductor.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 777
573 The Study of the Correlation of Proactive Coping and Retirement Planning: An Example of Senior Civil Servants in Taiwan

Authors: Ya-Hui Lee, Chien-Hung Hsieh, Ching-Yi Lu

Abstract:

Demographic aging is the major problem that Taiwanese society is facing, and retirement life adaptation is the most concerning issue. In recent years, studies have suggested that in order to have successful aging and retirement planning, a view for the future is necessary. In Taiwan, civil servants receive better pensions and retirement benefits than do other industries. Therefore, their retirement preparation is considerably more significant than other senior groups in Taiwan. The purpose of this study is to understand the correlation of proactive coping and retirement planning of senior civil servants in Taiwan. The method is conducted by questionnaire surveys, with 342 valid questionnaires collected. The results of this study are: 1. The background variables of the interviewees, including age, perceived economic statuses, and retirement statuses, are all significantly related to their proactive coping and retirement planning. 2. Regarding age, the interviewees with ages 55 and above have better proactive coping and retirement planning than those with ages 45 and below. 3. In the aspect of perceived economic statuses, the participants who feel “very good” economic statuses have better proactive coping ability and retirement readiness than those who feel “bad” and “very bad”. 4. Retirees have better proactive coping and retirement planning than those who are still working. 5. Monthly income is significant in retirement planning only. The participants’ retirement planning would be better if they have higher incomes. Furthermore, the participants’ retirement planning would be better if their revenue were €1453~€1937, than if their revenue were below €968. 6. There are positive correlations between proactive coping and retirement planning. 7. Proactive coping can predict retirement planning. The result of this study will be provided as references to the Taiwan government for educational retirement planning policies.

Keywords: Civil servants, proactive coping, retirement planning.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1740
572 Unsteady Flow of an Incompressible Viscous Electrically Conducting Fluid in Tube of Elliptical Cross Section under the Influence of Magnetic Field

Authors: Sanjay Baburao Kulkarni

Abstract:

Exact solution of an unsteady flow of elastico-viscous electrically conducting fluid through a porous media in a tube of elliptical cross section under the influence of constant pressure gradient and magnetic field has been obtained in this paper. Initially, the flow is generated by a constant pressure gradient. After attaining the steady state, the pressure gradient is suddenly withdrawn and the resulting fluid motion in a tube of elliptical cross section by taking into account of the transverse magnetic field and porosity factor of the bounding surface is investigated. The problem is solved in twostages the first stage is a steady motion in tube under the influence of a constant pressure gradient, the second stage concern with an unsteady motion. The problem is solved employing separation of variables technique. The results are expressed in terms of a nondimensional porosity parameter (K), magnetic parameter (m) and elastico-viscosity parameter (β), which depends on the Non- Newtonian coefficient. The flow parameters are found to be identical with that of Newtonian case as elastic-viscosity parameter and magnetic parameter tends to zero and porosity tends to infinity. It is seen that the effect of elastico-viscosity parameter, magnetic parameter and the porosity parameter of the bounding surface has significant effect on the velocity parameter.

Keywords: Elastico-viscous fluid, Elliptic cross-section, Porous media, Second order fluids.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1908
571 Forest Risk and Vulnerability Assessment: A Case Study from East Bokaro Coal Mining Area in India

Authors: Sujata Upgupta, Prasoon Kumar Singh

Abstract:

The expansion of large scale coal mining into forest areas is a potential hazard for the local biodiversity and wildlife. The objective of this study is to provide a picture of the threat that coal mining poses to the forests of the East Bokaro landscape. The vulnerable forest areas at risk have been assessed and the priority areas for conservation have been presented. The forested areas at risk in the current scenario have been assessed and compared with the past conditions using classification and buffer based overlay approach. Forest vulnerability has been assessed using an analytical framework based on systematic indicators and composite vulnerability index values. The results indicate that more than 4 km2 of forests have been lost from 1973 to 2016. Large patches of forests have been diverted for coal mining projects. Forests in the northern part of the coal field within 1-3 km radius around the coal mines are at immediate risk. The original contiguous forests have been converted into fragmented and degraded forest patches. Most of the collieries are located within or very close to the forests thus threatening the biodiversity and hydrology of the surrounding regions. Based on the vulnerability values estimated, it was concluded that more than 90% of the forested grids in East Bokaro are highly vulnerable to mining. The forests in the sub-districts of Bermo and Chandrapura have been identified as the most vulnerable to coal mining activities. This case study would add to the capacity of the forest managers and mine managers to address the risk and vulnerability of forests at a small landscape level in order to achieve sustainable development.

Keywords: Coal mining, forest, indicators, vulnerability.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1137