Search results for: Exponentially Weighted Moving Average
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 2110

Search results for: Exponentially Weighted Moving Average

1750 Numerical Study of Mixed Convection Coupled to Radiation in a Square Cavity with a Lid-Driven

Authors: Mohamed Amine Belmiloud, Nord Eddine Sad Chemloul

Abstract:

In this study, we investigated numerically heat transfer by mixed convection coupled to radiation in a square cavity; the upper horizontal wall is movable. The purpose of this study is to see the influence of the emissivity ε and the varying of the Richardson number Ri on the variation of average Nusselt number Nu. The vertical walls of the cavity are differentially heated, the left wall is maintained at a uniform temperature higher than the right wall, and the two horizontal walls are adiabatic. The finite volume method is used for solving the dimensionless Governing Equations. Emissivity values used in this study are ranged between 0 and 1, the Richardson number in the range 0.1 to 10. The Rayleigh number is fixed to Ra=104 and the Prandtl number is maintained constant Pr=0.71. Streamlines, isothermal lines and the average Nusselt number are presented according to the surface emissivity. The results of this study show that the Richardson number Ri and emissivity ε affect the average Nusselt number.

Keywords: Numerical study, mixed convection, square cavity, wall emissivity, lid-driven.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2209
1749 Characteristics of Speed Dispersion in Urban Expressway

Authors: Fujian Wang, Shubin Ruan, Meiwei Dai

Abstract:

Speed dispersion has tight relation to traffic safety. In this paper, several kinds of indicating parameters (the standard speed deviation, the coefficient of variation, the deviation of V85 and V15, the mean speed deviations, and the difference between adjacent car speeds) are applied to investigate the characteristics of speed dispersion, where V85 and V15 are 85th and 15th percentile speed, respectively. Their relationships are into full investigations and the results show that: there exists a positive relation (linear) between mean speed and the deviation of V85 and V15; while a negative relation (quadratic) between traffic flow and standard speed deviation. The mean speed deviation grows exponentially with mean speed while the absolute speed deviation between adjacent cars grows linearly with the headway. The results provide some basic information for traffic management.

Keywords: Headway, indicating parameters, speed dispersion, urban expressway.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1367
1748 Stochastic Scheduling to Minimize Expected Lateness in Multiple Identical Machines

Authors: Ghulam Zakria, Zailin Guan , Yasser Riaz Awan, Wan Lizhi

Abstract:

There are many real world problems in which parameters like the arrival time of new jobs, failure of resources, and completion time of jobs change continuously. This paper tackles the problem of scheduling jobs with random due dates on multiple identical machines in a stochastic environment. First to assign jobs to different machine centers LPT scheduling methods have been used, after that the particular sequence of jobs to be processed on the machine have been found using simple stochastic techniques. The performance parameter under consideration has been the maximum lateness concerning the stochastic due dates which are independent and exponentially distributed. At the end a relevant problem has been solved using the techniques in the paper..

Keywords: Quantity Production Flow Shop, LPT Scheduling, Stochastic Scheduling, Maximum Lateness, Random Due Dates

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1364
1747 Multichannel Scheme under Max-Min Fairness Environment for Cognitive Radio Networks

Authors: Hans R. Márquez, Cesar Hernández, Ingrid Páez

Abstract:

This paper develops a multiple channel assignment model, which allows to take advantage of spectrum opportunities in cognitive radio networks in the most efficient way. The developed scheme allows making several assignments of available and frequency adjacent channel, which require a bigger bandwidth, under an equality environment. The hybrid assignment model it is made by two algorithms, one that makes the ranking and selects available frequency channels and the other one in charge of establishing the Max-Min Fairness for not restrict the spectrum opportunities for all the other secondary users, who also claim to make transmissions. Measurements made were done for average bandwidth, average delay, as well as fairness computation for several channel assignments. Reached results were evaluated with experimental spectrum occupational data from captured GSM frequency band. The developed model shows evidence of improvement in spectrum opportunity use and a wider average transmission bandwidth for each secondary user, maintaining equality criteria in channel assignment.

Keywords: Bandwidth, fairness, multichannel, secondary users.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1733
1746 Border Limited Adaptive Subdivision Based On Triangle Meshes

Authors: Pichayut Peerasathien, Hiroshi Nagahashi

Abstract:

Subdivision is a method to create a smooth surface from a coarse mesh by subdividing the entire mesh. The conventional ways to compute and render surfaces are inconvenient both in terms of memory and computational time as the number of meshes will increase exponentially. An adaptive subdivision is the way to reduce the computational time and memory by subdividing only certain selected areas. In this paper, a new adaptive subdivision method for triangle meshes is introduced. This method defines a new adaptive subdivision rules by considering the properties of each triangle's neighbors and is embedded in a traditional Loop's subdivision. It prevents some undesirable side effects that appear in the conventional adaptive ways. Models that were subdivided by our method are compared with other adaptive subdivision methods

Keywords: Subdivision, loop subdivision, handle cracks, smooth surface.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1621
1745 Objects Extraction by Cooperating Optical Flow, Edge Detection and Region Growing Procedures

Authors: C. Lodato, S. Lopes

Abstract:

The image segmentation method described in this paper has been developed as a pre-processing stage to be used in methodologies and tools for video/image indexing and retrieval by content. This method solves the problem of whole objects extraction from background and it produces images of single complete objects from videos or photos. The extracted images are used for calculating the object visual features necessary for both indexing and retrieval processes. The segmentation algorithm is based on the cooperation among an optical flow evaluation method, edge detection and region growing procedures. The optical flow estimator belongs to the class of differential methods. It permits to detect motions ranging from a fraction of a pixel to a few pixels per frame, achieving good results in presence of noise without the need of a filtering pre-processing stage and includes a specialised model for moving object detection. The first task of the presented method exploits the cues from motion analysis for moving areas detection. Objects and background are then refined using respectively edge detection and seeded region growing procedures. All the tasks are iteratively performed until objects and background are completely resolved. The method has been applied to a variety of indoor and outdoor scenes where objects of different type and shape are represented on variously textured background.

Keywords: Image Segmentation, Motion Detection, Object Extraction, Optical Flow

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1730
1744 Estimation of Bayesian Sample Size for Binomial Proportions Using Areas P-tolerance with Lowest Posterior Loss

Authors: H. Bevrani, N. Najafi

Abstract:

This paper uses p-tolerance with the lowest posterior loss, quadratic loss function, average length criteria, average coverage criteria, and worst outcome criterion for computing of sample size to estimate proportion in Binomial probability function with Beta prior distribution. The proposed methodology is examined, and its effectiveness is shown.

Keywords: Bayesian inference, Beta-binomial Distribution, LPLcriteria, quadratic loss function.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1721
1743 Evaluating the Understanding of the University Students (Basic Sciences and Engineering) about the Numerical Representation of the Average Rate of Change

Authors: Saeid Haghjoo, Ebrahim Reyhani, Fahimeh Kolahdouz

Abstract:

The present study aimed to evaluate the understanding of the students in Tehran universities (Iran) about the numerical representation of the average rate of change based on the Structure of Observed Learning Outcomes (SOLO). In the present descriptive-survey research, the statistical population included undergraduate students (basic sciences and engineering) in the universities of Tehran. The samples were 604 students selected by random multi-stage clustering. The measurement tool was a task whose face and content validity was confirmed by math and mathematics education professors. Using Cronbach's Alpha criterion, the reliability coefficient of the task was obtained 0.95, which verified its reliability. The collected data were analyzed by descriptive statistics and inferential statistics (chi-squared and independent t-tests) under SPSS-24 software. According to the SOLO model in the prestructural, unistructural, and multistructural levels, basic science students had a higher percentage of understanding than that of engineering students, although the outcome was inverse at the relational level. However, there was no significant difference in the average understanding of both groups. The results indicated that students failed to have a proper understanding of the numerical representation of the average rate of change, in addition to missconceptions when using physics formulas in solving the problem. In addition, multiple solutions were derived along with their dominant methods during the qualitative analysis. The current research proposed to focus on the context problems with approximate calculations and numerical representation, using software and connection common relations between math and physics in the teaching process of teachers and professors.

Keywords: Average rate of change, context problems, derivative, numerical representation, SOLO taxonomy.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 710
1742 Comparison of Composite Programming and Compromise Programming for Aircraft Selection Problem Using Multiple Criteria Decision Making Analysis Method

Authors: C. Ardil

Abstract:

In this paper, the comparison of composite programming and compromise programming for the aircraft selection problem is discussed using the multiple criteria decision analysis method. The decision making process requires the prior definition and fulfillment of certain factors, especially when it comes to complex areas such as aircraft selection problems. The proposed technique gives more efficient results by extending the composite programming and compromise programming, which are widely used in modeling multiple criteria decisions. The proposed model is applied to a practical decision problem for evaluating and selecting aircraft problems.A selection of aircraft was made based on the proposed approach developed in the field of multiple criteria decision making. The model presented is solved by using the following methods: composite programming, and compromise programming. The importance values of the weight coefficients of the criteria are calculated using the mean weight method. The evaluation and ranking of aircraft are carried out using the composite programming and compromise programming methods. In order to determine the stability of the model and the ability to apply the developed composite programming and compromise programming approach, the paper analyzes its sensitivity, which involves changing the value of the coefficient λ and q in the first part. The second part of the sensitivity analysis relates to the application of different multiple criteria decision making methods, composite programming and compromise programming. In addition, in the third part of the sensitivity analysis, the Spearman correlation coefficient of the ranks obtained was calculated which confirms the applicability of all the proposed approaches.

Keywords: composite programming, compromise programming, additive weighted model, multiplicative weighted model, multiple criteria decision making analysis, MCDMA, aircraft selection

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 625
1741 Some Physical Properties of Musk Lime (Citrus Microcarpa)

Authors: M.H.R.O. Abdullah, P.E. Ch'ng, N.A. Yunus

Abstract:

Some physical properties of musk lime (Citrus microcarpa) were determined in this study. The average moisture content (wet basis) of the fruit was found to be 85.10 (±0.72) %. The mean of length, width and thickness of the fruit was 26.36 (±0.97), 26.40 (±1.04) and 25.26 (±0.94) mm respectively. The average value for geometric mean diameter, sphericity, aspect ratio, mass, surface area, volume, true density, bulk density and porosity was 26.00 (±0.82) mm, 98.67 (±2.04) %, 100.23 (±3.28) %, 10.007 (±0.878) g, 2125.07 (±133.93) mm2, 8800.00 (±731.82) mm3, 1002.87 (±39.16) kgm-3, 501.70 (±22.58) kgm-3 and 49.89 (±3.15) % respectively. The coefficient of static friction on four types of structural surface was found to be varying from 0.238 (±0.025) for glass to 0.247 (±0.024) for steel surface.

Keywords: Musk lime, Citrus microcarpa, physical properties.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3257
1740 Linear Prediction System in Measuring Glucose Level in Blood

Authors: Intan Maisarah Abd Rahim, Herlina Abdul Rahim, Rashidah Ghazali

Abstract:

Diabetes is a medical condition that can lead to various diseases such as stroke, heart disease, blindness and obesity. In clinical practice, the concern of the diabetic patients towards the blood glucose examination is rather alarming as some of the individual describing it as something painful with pinprick and pinch. As for some patient with high level of glucose level, pricking the fingers multiple times a day with the conventional glucose meter for close monitoring can be tiresome, time consuming and painful. With these concerns, several non-invasive techniques were used by researchers in measuring the glucose level in blood, including ultrasonic sensor implementation, multisensory systems, absorbance of transmittance, bio-impedance, voltage intensity, and thermography. This paper is discussing the application of the near-infrared (NIR) spectroscopy as a non-invasive method in measuring the glucose level and the implementation of the linear system identification model in predicting the output data for the NIR measurement. In this study, the wavelengths considered are at the 1450 nm and 1950 nm. Both of these wavelengths showed the most reliable information on the glucose presence in blood. Then, the linear Autoregressive Moving Average Exogenous model (ARMAX) model with both un-regularized and regularized methods was implemented in predicting the output result for the NIR measurement in order to investigate the practicality of the linear system in this study. However, the result showed only 50.11% accuracy obtained from the system which is far from the satisfying results that should be obtained.

Keywords: Diabetes, glucose level, linear, near-infrared (NIR), non-invasive, prediction system.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 836
1739 High Resolution Images: Segmenting, Extracting Information and GIS Integration

Authors: Erick López-Ornelas

Abstract:

As the world changes more rapidly, the demand for update information for resource management, environment monitoring, planning are increasing exponentially. Integration of Remote Sensing with GIS technology will significantly promote the ability for addressing these concerns. This paper presents an alternative way of update GIS applications using image processing and high resolution images. We show a method of high-resolution image segmentation using graphs and morphological operations, where a preprocessing step (watershed operation) is required. A morphological process is then applied using the opening and closing operations. After this segmentation we can extract significant cartographic elements such as urban areas, streets or green areas. The result of this segmentation and this extraction is then used to update GIS applications. Some examples are shown using aerial photography.

Keywords: GIS, Remote Sensing, image segmentation, feature extraction.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1611
1738 Simulating a Single-Server Queue using the Q – Simulator

Authors: Irene K. Amponsah, Bennony K. Gordor, Francis Dogbey

Abstract:

This paper introduces a technique for simulating a single-server exponential queuing system. The technique called the Q-Simulator is a computer program which can simulate the effect of traffic intensity on all system average quantities given the arrival and/or service rates. The Q-Simulator has three phases namely: the formula based method, the uncontrolled simulation, and the controlled simulation. The Q-Simulator generates graphs (crystal solutions) for all results of the simulation or calculation and can be used to estimate desirable average quantities such as waiting times, queue lengths, etc.

Keywords: Automation system-Simulator, Simulation, Singleserver exponential system

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2267
1737 Detection of Action Potentials in the Presence of Noise Using Phase-Space Techniques

Authors: Christopher Paterson, Richard Curry, Alan Purvis, Simon Johnson

Abstract:

Emerging Bio-engineering fields such as Brain Computer Interfaces, neuroprothesis devices and modeling and simulation of neural networks have led to increased research activity in algorithms for the detection, isolation and classification of Action Potentials (AP) from noisy data trains. Current techniques in the field of 'unsupervised no-prior knowledge' biosignal processing include energy operators, wavelet detection and adaptive thresholding. These tend to bias towards larger AP waveforms, AP may be missed due to deviations in spike shape and frequency and correlated noise spectrums can cause false detection. Also, such algorithms tend to suffer from large computational expense. A new signal detection technique based upon the ideas of phasespace diagrams and trajectories is proposed based upon the use of a delayed copy of the AP to highlight discontinuities relative to background noise. This idea has been used to create algorithms that are computationally inexpensive and address the above problems. Distinct AP have been picked out and manually classified from real physiological data recorded from a cockroach. To facilitate testing of the new technique, an Auto Regressive Moving Average (ARMA) noise model has been constructed bases upon background noise of the recordings. Along with the AP classification means this model enables generation of realistic neuronal data sets at arbitrary signal to noise ratio (SNR).

Keywords: Action potential detection, Low SNR, Phase spacediagrams/trajectories, Unsupervised/no-prior knowledge.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1613
1736 Optimal Temperature and Duration for Dabbing Customers with the Massage Compressed Packs Reported from Customers’ Perception

Authors: Wichan Lertlop, Boonyarat Chaleephay

Abstract:

The objective of this research was to study the appropriate thermal level and time for dabbing customers with the massage compressed pack reported from their perception. The investigation was conducted by comparing different angles of tilted heads done by the customers together with their perception before and after the dabbing. The variables included different temperature of the compressed packs and different dabbing duration. Samples in this study included volunteers who got massage therapy and dabbing with hot compressed packs by traditional Thai medical students. The experiment was conducted during January to June 2013. The research tool consisted of angle meters, stop watches, thermometers, and massage compressed packs. The customers were interviewed for their perceptions before and after the dabbing. The results showed that:

  1. There was a difference of the average angles of tilted heads before and after the dabbing.
  2. There was no difference of the average angles at different temperatures but constant duration.
  3. There was no difference of the average angles at different durations.
  4. The customers reported relaxation no matter what the various temperatures and various dabbing durations were. However, they reported too hot at the temperature 70oC and over.

Keywords: Massage, Therapy, Therapeutic Systems and Technologies.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1654
1735 Route Training in Mobile Robotics through System Identification

Authors: Roberto Iglesias, Theocharis Kyriacou, Ulrich Nehmzow, Steve Billings

Abstract:

Fundamental sensor-motor couplings form the backbone of most mobile robot control tasks, and often need to be implemented fast, efficiently and nevertheless reliably. Machine learning techniques are therefore often used to obtain the desired sensor-motor competences. In this paper we present an alternative to established machine learning methods such as artificial neural networks, that is very fast, easy to implement, and has the distinct advantage that it generates transparent, analysable sensor-motor couplings: system identification through nonlinear polynomial mapping. This work, which is part of the RobotMODIC project at the universities of Essex and Sheffield, aims to develop a theoretical understanding of the interaction between the robot and its environment. One of the purposes of this research is to enable the principled design of robot control programs. As a first step towards this aim we model the behaviour of the robot, as this emerges from its interaction with the environment, with the NARMAX modelling method (Nonlinear, Auto-Regressive, Moving Average models with eXogenous inputs). This method produces explicit polynomial functions that can be subsequently analysed using established mathematical methods. In this paper we demonstrate the fidelity of the obtained NARMAX models in the challenging task of robot route learning; we present a set of experiments in which a Magellan Pro mobile robot was taught to follow four different routes, always using the same mechanism to obtain the required control law.

Keywords: Mobile robotics, system identification, non-linear modelling, NARMAX.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1695
1734 Complex Network Approach to International Trade of Fossil Fuel

Authors: Semanur Soyyiğit Kaya, Ercan Eren

Abstract:

Energy has a prominent role for development of nations. Countries which have energy resources also have strategic power in the international trade of energy since it is essential for all stages of production in the economy. Thus, it is important for countries to analyze the weaknesses and strength of the system. On the other side, international trade is one of the fields that are analyzed as a complex network via network analysis. Complex network is one of the tools to analyze complex systems with heterogeneous agents and interaction between them. A complex network consists of nodes and the interactions between these nodes. Total properties which emerge as a result of these interactions are distinct from the sum of small parts (more or less) in complex systems. Thus, standard approaches to international trade are superficial to analyze these systems. Network analysis provides a new approach to analyze international trade as a network. In this network, countries constitute nodes and trade relations (export or import) constitute edges. It becomes possible to analyze international trade network in terms of high degree indicators which are specific to complex networks such as connectivity, clustering, assortativity/disassortativity, centrality, etc. In this analysis, international trade of crude oil and coal which are types of fossil fuel has been analyzed from 2005 to 2014 via network analysis. First, it has been analyzed in terms of some topological parameters such as density, transitivity, clustering etc. Afterwards, fitness to Pareto distribution has been analyzed via Kolmogorov-Smirnov test. Finally, weighted HITS algorithm has been applied to the data as a centrality measure to determine the real prominence of countries in these trade networks. Weighted HITS algorithm is a strong tool to analyze the network by ranking countries with regards to prominence of their trade partners. We have calculated both an export centrality and an import centrality by applying w-HITS algorithm to the data. As a result, impacts of the trading countries have been presented in terms of high-degree indicators.

Keywords: Complex network approach, fossil fuel, international trade, network theory.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2350
1733 Performance Evaluation and Cost Analysis of Standby Systems

Authors: M. A. Hajeeh

Abstract:

Pumping systems are an integral part of water desalination plants, their effective functioning is vital for the operation of a plant. In this research work, the reliability and availability of pressurized pumps in a reverse osmosis desalination plant are studied with the objective of finding configurations that provides optimal performance. Six configurations of a series system with different number of warm and cold standby components were examined. Closed form expressions for the mean time to failure (MTTF) and the long run availability are derived and compared under the assumption that the time between failures and repair times of the primary and standby components are exponentially distributed. Moreover, a cost/ benefit analysis is conducted in order to identify a configuration with the best performance and least cost. It is concluded that configurations with cold standby components are preferable especially when the pumps are of the size.

Keywords: Availability, Cost/ benefit, Mean time to failure, Pumps.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1686
1732 Direct Measurements of Wind Data over 100 Meters above the Ground in the Site of Lendinara, Italy

Authors: A. Dal Monte, M. Raciti Castelli, G. B. Bellato, L. Stevanato, E. Benini

Abstract:

The wind resource in the Italian site of Lendinara (RO) is analyzed through a systematic anemometric campaign performed on the top of the bell tower, at an altitude of over 100 m above the ground. Both the average wind speed and the Weibull distribution are computed. The resulting average wind velocity is in accordance with the numerical predictions of the Italian Wind Atlas, confirming the accuracy of the extrapolation of wind data adopted for the evaluation of wind potential at higher altitudes with respect to the commonly placed measurement stations.

Keywords: Anemometric campaign, wind resource, Weibull distribution, wind atlas

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1930
1731 Potential of Solar Energy in Zarqa Region

Authors: Ali M. Jawarneh, Ahmad S. AL-Shyyab

Abstract:

The purpose of this work is to present the potential of solar energy in Zarqa region. The solar radiation along year 2009 was obtained from Pyranometer which measures the global radiation over horizontal surfaces. Solar data in several different forms, over period of 5 minutes, hour-by-hour, daily and monthly data radiation have been presented. Briefly, the yearly global solar radiation in Zarqa is 7297.5 MJ/m2 (2027 kWh/m²) and the average annual solar radiation per day is 20 MJ/m2 (5.5 Kwh/m2). More specifically, the average annual solar radiation per day is 12.9 MJ/m2 (3.57 Kwh/m2) in winter and 25 MJ/m2 (7 Kwh/m2) in summer.

Keywords: Solar Energy, Pyranometer, Zarqa Region

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1885
1730 Fast Codevector Search Algorithm for 3-D Vector Quantized Codebook

Authors: H. B. Kekre, Tanuja K. Sarode

Abstract:

This paper presents a very simple and efficient algorithm for codebook search, which reduces a great deal of computation as compared to the full codebook search. The algorithm is based on sorting and centroid technique for search. The results table shows the effectiveness of the proposed algorithm in terms of computational complexity. In this paper we also introduce a new performance parameter named as Average fractional change in pixel value as we feel that it gives better understanding of the closeness of the image since it is related to the perception. This new performance parameter takes into consideration the average fractional change in each pixel value.

Keywords: Vector Quantization, Data Compression, Encoding, Searching.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1583
1729 A Design for Supply Chain Model by Integrated Evaluation of Design Value and Supply Chain Cost

Authors: Yuan-Jye Tseng, Jia-Shu Li

Abstract:

To design a product with the given product requirement and design objective, there can be alternative ways to propose the detailed design specifications of the product. In the design modeling stage, alternative design cases with detailed specifications can be modeled to fulfill the product requirement and design objective. Therefore, in the design evaluation stage, it is required to perform an evaluation of the alternative design cases for deciding the final design. The purpose of this research is to develop a product evaluation model for evaluating the alternative design cases by integrated evaluating the criteria of functional design, Kansei design, and design for supply chain. The criteria in the functional design group include primary function, expansion function, improved function, and new function. The criteria in the Kansei group include geometric shape, dimension, surface finish, and layout. The criteria in the design for supply chain group include material, manufacturing process, assembly, and supply chain operation. From the point of view of value and cost, the criteria in the functional design group and Kansei design group represent the design value of the product. The criteria in the design for supply chain group represent the supply chain and manufacturing cost of the product. It is required to evaluate the design value and the supply chain cost to determine the final design. For the purpose of evaluating the criteria in the three criteria groups, a fuzzy analytic network process (FANP) method is presented to evaluate a weighted index by calculating the total relational values among the three groups. A method using the technique for order preference by similarity to ideal solution (TOPSIS) is used to compare and rank the design alternative cases according to the weighted index using the total relational values of the criteria. The final decision of a design case can be determined by using the ordered ranking. For example, the design case with the top ranking can be selected as the final design case. Based on the criteria in the evaluation, the design objective can be achieved with a combined and weighted effect of the design value and manufacturing cost. An example product is demonstrated and illustrated in the presentation. It shows that the design evaluation model is useful for integrated evaluation of functional design, Kansei design, and design for supply chain to determine the best design case and achieve the design objective.

Keywords: Design evaluation, functional design, Kansei design, supply chain, design value, manufacturing cost, fuzzy analytic network process, technique for order preference by similarity to ideal solution.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 756
1728 DMC with Adaptive Weighted Output

Authors: Ahmed Abbas, M.R. M Rizk, Mohamed El-Sayed

Abstract:

This paper presents a new adaptive DMC controller that improves the controller performance in case of plant-model mismatch. The new controller monitors the plant measured output, compares it with the model output and calculates weights applied to the controller move. Simulations show that the new controller can help improve control performance and avoid instability in case of severe model mismatches.

Keywords: Adaptive control, dynamic matrix control, DMC, model predictive control

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2195
1727 Forecasting the Fluctuation of Currency Exchange Rate Using Random Forest

Authors: L. Basha, E. Gjika

Abstract:

The exchange rate is one of the most important economic variables, especially for a small, open economy such as Albania. Its effect is noticeable on one country's competitiveness, trade and current account, inflation, wages, domestic economic activity and bank stability. This study investigates the fluctuation of Albania’s exchange rates using monthly average foreign currency, Euro (Eur) to Albanian Lek (ALL) exchange rate with a time span from January 2008 to June 2021 and the macroeconomic factors that have a significant effect on the exchange rate. Initially, the Random Forest Regression algorithm is constructed to understand the impact of economic variables in the behavior of monthly average foreign currencies exchange rates. Then the forecast of macro-economic indicators for 12 months was performed using time series models. The predicted values received are placed in the random forest model in order to obtain the average monthly forecast of Euro to Albanian Lek (ALL) exchange rate for the period July 2021 to June 2022.

Keywords: Exchange rate, Random Forest, time series, Machine Learning, forecasting.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 588
1726 Energy Efficiency Testing of Fluorescent and WOLED (White Organic LED)

Authors: Hari Maghfiroh, Harry Prabowo

Abstract:

WOLED is widely used as lighting for high efficacy and little power consumption. In this research, power factor testing between WOLED and fluorescent lamp to see which one is more efficient in consuming energy. Since both lamps use semiconductor components, so calculation of the power factor need to consider the effects of harmonics. Harmonic make bigger losses. The study is conducted by comparing the value of the power factor regardless of harmonics (DPF) and also by included the harmonics (TPF). The average value of DPF of fluorescent is 0.953 while WOLED is 0.972. The average value of TPF of fluorescent is 0.717 whereas WOLED is 0.933. So from the review of power factor WOLED is more energy efficient than fluorescent lamp.

Keywords: Fluorescent, harmonic, power factor, WOLED.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1543
1725 Entropy Generation and Heat Transfer of Cu–Water Nanofluid Mixed Convection in a Cavity

Authors: Mliki Bouchmel, Belgacem Nabil, Abbassi Mohamed Ammar, Geudri Kamel, Omri Ahmed

Abstract:

In this numerical work, mixed convection and entropy generation of Cu–water nanofluid in a lid-driven square cavity have been investigated numerically using the Lattice Boltzmann Method. Horizontal walls of the cavity are adiabatic and vertical walls have constant temperature but different values. The top wall has been considered as moving from left to right at a constant speed, U0. The effects of different parameters such as nanoparticle volume concentration (0–0.05), Rayleigh number (104–106) and Reynolds numbers (1, 10 and 100) on the entropy generation, flow and temperature fields are studied. The results have shown that addition of nanoparticles to the base fluid affects the entropy generation, flow pattern and thermal behavior especially at higher Rayleigh and low Reynolds numbers. For pure fluid as well as nanofluid, the increase of Reynolds number increases the average Nusselt number and the total entropy generation, linearly. The maximum entropy generation occurs in nanofluid at low Rayleigh number and at high Reynolds number. The minimum entropy generation occurs in pure fluid at low Rayleigh and Reynolds numbers. Also at higher Reynolds number, the effect of Cu nanoparticles on enhancement of heat transfer was decreased because the effect of lid-driven cavity was increased. The present results are validated by favorable comparisons with previously published results. The results of the problem are presented in graphical and tabular forms and discussed.

Keywords: Entropy generation, mixed convection, nanofluid, lattice Boltzmann method.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1924
1724 Mathematics Anxiety among Male and Female Students

Authors: Wern Lin Yeo, Choo Kim Tan, Sook Ling Lew

Abstract:

The purpose of this study is to determine the relationship of anxiety level between male and female undergraduates at a private university in Malaysia. Convenient sampling method used in this study in which the students were selected based on the grouping assigned by the faculty. There were 214 undergraduates who registered the probability courses had participated in this study. Mathematics Anxiety Rating Scale (MARS) was the instrument used in study which used to determine students’ anxiety level towards probability. Reliability and validity of instrument was done before the major study was conducted. In the major study, students were given briefing about the study conducted. Participation of this study was voluntary. Students were given consent form to determine whether they agree to participate in the study. Duration of two weeks was given for students to complete the given online questionnaire. The data collected will be analyzed using Statistical Package for the Social Sciences (SPSS) to determine the level of anxiety. There were three anxiety level, i.e., low, average and high. Students’ anxiety level was determined based on their scores obtained compared with the mean and standard deviation. If the scores obtained were below mean and standard deviation, the anxiety level was low. If the scores were at below and above the mean and between one standard deviation, the anxiety level was average. If the scores were above the mean and greater than one standard deviation, the anxiety level was high. Results showed that both of genders were having average anxiety level. Among low, average and high anxiety level, frequency of males were found to be higher as compared to females. Hence, the mean values obtained for males (M = 3.62) was higher than females (M = 3.42). In order to be significant of anxiety level among the gender, the p-value should be less than .05. The p-value obtained in this study was .117. However, this value was greater than .05. Thus, there was no significant difference of anxiety level among the gender. In other words, there was no relationship of anxiety level with the gender.

Keywords: Anxiety level, gender, mathematics anxiety, probability and statistics.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4244
1723 Usage of Channel Coding Techniques for Peak-to-Average Power Ratio Reduction in Visible Light Communications Systems

Authors: P.L.D.N.M. de Silva, S.G. Edirisinghe, R. Weerasuriya

Abstract:

High Peak-to-Average Power Ratio (PAPR) is a concern of Orthogonal Frequency Division Multiplexing (OFDM) based Visible Light Communication (VLC) systems. Discrete Fourier Transform spread (DFT-s) OFDM is an alternative single carrier modulation scheme which would address this concern. Employing channel coding techniques is another mechanism to reduce the PAPR. In this study, the improvement which can be harnessed by hybridizing these two techniques for VLC system is being studied. Within the study, efficient techniques such as Hamming coding and Convolutional coding have been studied. Thus, we present the impact of the hybrid of DFT-s OFDM and Channel coding (Hamming coding and Convolutional coding) on PAPR in VLC systems, using MATLAB simulations.

Keywords: Convolutional Coding, Discrete Fourier Transform spread Orthogonal Frequency Division Multiplexing (DFT-s OFDM), Hamming Coding, Peak-to-Average Power Ratio (PAPR), Visible Light Communications (VLC).

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 441
1722 Tumble Flow Analysis in an Unfired Engine Using Particle Image Velocimetry

Authors: B. Murali Krishna, J. M. Mallikarjuna

Abstract:

This paper deals with the experimental investigations of the in-cylinder tumble flows in an unfired internal combustion engine with a flat piston at the engine speeds ranging from 400 to 1000 rev/min., and also with the dome and dome-cavity pistons at an engine speed of 1000 rev/min., using particle image velocimetry. From the two-dimensional in-cylinder flow measurements, tumble flow analysis is carried out in the combustion space on a vertical plane passing through cylinder axis. To analyze the tumble flows, ensemble average velocity vectors are used and to characterize it, tumble ratio is estimated. From the results, generally, we have found that tumble ratio varies mainly with crank angle position. Also, at the end of compression stroke, average turbulent kinetic energy is more at higher engine speeds. We have also found that, at 330 crank angle position, flat piston shows an improvement of about 85 and 23% in tumble ratio, and about 24 and 2.5% in average turbulent kinetic energy compared to dome and dome-cavity pistons respectively

Keywords: In-cylinder flow, Dome piston, Cavity, Tumble, PIV

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2244
1721 The Construction of Interactive Computer Multimedia Instruction on “Basic Japanese Vocabulary“

Authors: Kongrit Jittangthammagul, Sakesun Yampinij, Thapanee Endoo, Nattapong Kramwong

Abstract:

The study entitled “The Construction of Interactive Computer Multimedia Instruction on Basic Japanese Vocabulary" was aimed: 1) To construct the interactive computer multimedia instruction on Basic Japanese Vocabulary, 2) To find out multimedia-s quality, 3) To examine the student-s satisfaction and 4) To study the learning achievement in Basic Japanese vocabulary. The sampling group used in this study was composed of 40 1st year student in Educational Communications and Technology Department, Faculty of Industrial Education and Technology, King Mongkut-s University of Technology Thonburi, in the academic year 2553 B.E. (2010). According to research results, we found that 1). The quality assessment by 3 mass media experts was at 4.72 on average or at high level. 2) In terms of contents, the evaluation by 3 experts was at 4.81 on average or at high level. 3) In terms of achievement, there was a statistical significance between before and after the treatment at the .05 level. 4) The satisfaction of students towards the interactive computer multimedia Instruction on “Basic Japanese Vocabulary" was 4.35 on average, or at high level.

Keywords: Interactive Computer Multimedia on Basic Japanese Vocabulary, Learning Achievement, Quality

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1470