Search results for: average error.
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 2514

Search results for: average error.

1704 The use of a Bespoke Computer Game For Teaching Analogue Electronics

Authors: Olaf Hallan Graven, Dag Andreas Hals Samuelsen

Abstract:

An implementation of a design for a game based virtual learning environment is described. The game is developed for a course in analogue electronics, and the topic is the design of a power supply. This task can be solved in a number of different ways, with certain constraints, giving the students a certain amount of freedom, although the game is designed not to facilitate trial-and error approach. The use of storytelling and a virtual gaming environment provides the student with the learning material in a MMORPG environment. The game is tested on a group of second year electrical engineering students with good results.

Keywords: analogue electronics, e-learning, computer games for learning, virtual reality

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1518
1703 Study of Adaptive Filtering Algorithms and the Equalization of Radio Mobile Channel

Authors: Said Elkassimi, Said Safi, B. Manaut

Abstract:

This paper presented a study of three algorithms, the equalization algorithm to equalize the transmission channel with ZF and MMSE criteria, application of channel Bran A, and adaptive filtering algorithms LMS and RLS to estimate the parameters of the equalizer filter, i.e. move to the channel estimation and therefore reflect the temporal variations of the channel, and reduce the error in the transmitted signal. So far the performance of the algorithm equalizer with ZF and MMSE criteria both in the case without noise, a comparison of performance of the LMS and RLS algorithm.

Keywords: Adaptive filtering second equalizer, LMS, RLS Bran A, Proakis (B) MMSE, ZF.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2117
1702 The Impact of Revenue Gap on Economic Growth: A Case Study of Pakistan

Authors: M. Ilyas, M. W. Siddiqi

Abstract:

This study employs auto-regressive distributed lag (ARDL) bounds approach to cointegration for long run and errorcorrection modeling (ECM) for short run analysis to examine the relationship between revenue gap and economic growth for Pakistan using annual time series data over the period 1980 to 2008. The short and long run results indicate that revenue gap is statistical significant and negatively effect economic growth. The significant and negative coefficient of error correction term in ECM indicates that after a shock, the long rum equilibrium will again converge towards equilibrium about 10.406 percent within a year.

Keywords: ARDL cointegration, Economic Growth, RevenueGap, Pakistan.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2560
1701 pth Moment Exponential Synchronization of a Class of Chaotic Neural Networks with Mixed Delays

Authors: Zixin Liu, Shu Lü, Shouming Zhong, Mao Ye

Abstract:

This paper studies the pth moment exponential synchronization of a class of stochastic neural networks with mixed delays. Based on Lyapunov stability theory, by establishing a new integrodifferential inequality with mixed delays, several sufficient conditions have been derived to ensure the pth moment exponential stability for the error system. The criteria extend and improve some earlier results. One numerical example is presented to illustrate the validity of the main results.

Keywords: pth Moment Exponential synchronization, Stochastic, Neural networks, Mixed time delays

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1573
1700 A Study on Inference from Distance Variables in Hedonic Regression

Authors: Yan Wang, Yasushi Asami, Yukio Sadahiro

Abstract:

In urban area, several landmarks may affect housing price and rents, and hedonic analysis should employ distance variables corresponding to each landmarks. Unfortunately, the effects of distances to landmarks on housing prices are generally not consistent with the true price. These distance variables may cause magnitude error in regression, pointing a problem of spatial multicollinearity. In this paper, we provided some approaches for getting the samples with less bias and method on locating the specific sampling area to avoid the multicollinerity problem in two specific landmarks case.

Keywords: Landmarks, hedonic regression, distance variables, collinearity, multicollinerity.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1899
1699 The Application of Hybrid Orthonomal Bernstein and Block-Pulse Functions in Finding Numerical Solution of Fredholm Fuzzy Integral Equations

Authors: Mahmoud Zarrini, Sanaz Torkaman

Abstract:

In this paper, we have proposed a numerical method for solving fuzzy Fredholm integral equation of the second kind. In this method a combination of orthonormal Bernstein and Block-Pulse functions are used. In most cases, the proposed method leads to the exact solution. The advantages of this method are shown by an example and calculate the error analysis.

Keywords: Fuzzy Fredholm Integral Equation, Bernstein, Block-Pulse, Orthonormal.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2025
1698 A Cooperative Space-Time Transmission Scheme Based On Symbol Combinations

Authors: Keunhong Chae, Seokho Yoon

Abstract:

This paper proposes a cooperative Alamouti space time transmission scheme with low relay complexity for the cooperative communication systems. In the proposed scheme, the source node combines the data symbols to construct the Alamouti-coded form at the destination node, while the conventional scheme performs the corresponding operations at the relay nodes. In simulation results, it is shown that the proposed scheme achieves the second order cooperative diversity while maintaining the same bit error rate (BER) performance as that of the conventional scheme.

Keywords: Space-time transmission, cooperative communication system, MIMO.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1807
1697 New Wavelet Indices to Assess Muscle Fatigue during Dynamic Contractions

Authors: González-Izal M., Rodríguez-Carreño I, Mallor-Giménez F, Malanda A, Izquierdo M

Abstract:

The purpose of this study was to evaluate and compare new indices based on the discrete wavelet transform with another spectral parameters proposed in the literature as mean average voltage, median frequency and ratios between spectral moments applied to estimate acute exercise-induced changes in power output, i.e., to assess peripheral muscle fatigue during a dynamic fatiguing protocol. 15 trained subjects performed 5 sets consisting of 10 leg press, with 2 minutes rest between sets. Surface electromyography was recorded from vastus medialis (VM) muscle. Several surface electromyographic parameters were compared to detect peripheral muscle fatigue. These were: mean average voltage (MAV), median spectral frequency (Fmed), Dimitrov spectral index of muscle fatigue (FInsm5), as well as other five parameters obtained from the discrete wavelet transform (DWT) as ratios between different scales. The new wavelet indices achieved the best results in Pearson correlation coefficients with power output changes during acute dynamic contractions. Their regressions were significantly different from MAV and Fmed. On the other hand, they showed the highest robustness in presence of additive white gaussian noise for different signal to noise ratios (SNRs). Therefore, peripheral impairments assessed by sEMG wavelet indices may be a relevant factor involved in the loss of power output after dynamic high-loading fatiguing task.

Keywords: Median Frequency, EMG, wavelet transform, muscle fatigue

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1859
1696 On Constructing Approximate Convex Hull

Authors: M. Zahid Hossain, M. Ashraful Amin

Abstract:

The algorithms of convex hull have been extensively studied in literature, principally because of their wide range of applications in different areas. This article presents an efficient algorithm to construct approximate convex hull from a set of n points in the plane in O(n + k) time, where k is the approximation error control parameter. The proposed algorithm is suitable for applications preferred to reduce the computation time in exchange of accuracy level such as animation and interaction in computer graphics where rapid and real-time graphics rendering is indispensable.

Keywords: Convex hull, Approximation algorithm, Computational geometry, Linear time.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2295
1695 Joint Transmitter-Receiver Optimization for Bonded Wireline Communications

Authors: Mohammed H. Nafie, Ahmed F. Shalash

Abstract:

With the advent of DSL services, high data rates are now available over phone lines, yet higher rates are in demand. In this paper, we optimize the transmit filters that can be used over wireline channels. Results showing the bit error rates when optimized filters are used, and with a decision feedback equalizer (DFE) employed in the receiver, are given. We then show that significantly higher throughput can be achieved by modeling the channel as a multiple input multiple output (MIMO) channel. A receiver that employs a MIMO-DFE that deals jointly with several users is proposed and shown to provide significant improvement over the conventional DFE.

Keywords: DFE, MIMO Channels, Receiver Architectures, Transmit Filters.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1346
1694 Bitrate Reduction Using FMO for Video Streaming over Packet Networks

Authors: Le Thanh Ha, Hye-Soo Kim, Chun-Su Park, Seung-Won Jung, Sung-Jea Ko

Abstract:

Flexible macroblock ordering (FMO), adopted in the H.264 standard, allows to partition all macroblocks (MBs) in a frame into separate groups of MBs called Slice Groups (SGs). FMO can not only support error-resilience, but also control the size of video packets for different network types. However, it is well-known that the number of bits required for encoding the frame is increased by adopting FMO. In this paper, we propose a novel algorithm that can reduce the bitrate overhead caused by utilizing FMO. In the proposed algorithm, all MBs are grouped in SGs based on the similarity of the transform coefficients. Experimental results show that our algorithm can reduce the bitrate as compared with conventional FMO.

Keywords: Data Partition, Entropy Coding, Greedy Algorithm, H.264/AVC, Slice Group.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1301
1693 Laplace Decomposition Approximation Solution for a System of Multi-Pantograph Equations

Authors: M. A. Koroma, C. Zhan, A. F. Kamara, A. B. Sesay

Abstract:

In this work we adopt a combination of Laplace transform and the decomposition method to find numerical solutions of a system of multi-pantograph equations. The procedure leads to a rapid convergence of the series to the exact solution after computing a few terms. The effectiveness of the method is demonstrated in some examples by obtaining the exact solution and in others by computing the absolute error which decreases as the number of terms of the series increases.

Keywords: Laplace decomposition, pantograph equations, exact solution, numerical solution, approximate solution.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1639
1692 A Compact Pi Network for Reducing Bit Error Rate in Dispersive FIR Channel Noise Model

Authors: Kavita Burse, R.N. Yadav, S.C. Shrivastava, Vishnu Pratap Singh Kirar

Abstract:

During signal transmission, the combined effect of the transmitter filter, the transmission medium, and additive white Gaussian noise (AWGN) are included in the channel which distort and add noise to the signal. This causes the well defined signal constellation to spread causing errors in bit detection. A compact pi neural network with minimum number of nodes is proposed. The replacement of summation at each node by multiplication results in more powerful mapping. The resultant pi network is tested on six different channels.

Keywords: Additive white Gaussian noise, digitalcommunication system, multiplicative neuron, Pi neural network.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1660
1691 Algebraic Quantum Error Correction Codes

Authors: Ming-Chung Tsai, Kuan-Peng Chen, Zheng-Yao

Abstract:

A systematic and exhaustive method based on the group structure of a unitary Lie algebra is proposed to generate an enormous number of quantum codes. With respect to the algebraic structure, the orthogonality condition, which is the central rule of generating quantum codes, is proved to be fully equivalent to the distinguishability of the elements in this structure. In addition, four types of quantum codes are classified according to the relation of the codeword operators and some initial quantum state. By linking the unitary Lie algebra with the additive group, the classical correspondences of some of these quantum codes can be rendered.

Keywords: Quotient-Algebra Partition, Codeword Spinors, Basis Codewords, Syndrome Spinors

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1408
1690 Study of Temperature Changes in Fars Province

Authors: A. Gandomkar, R. Dehghani

Abstract:

Climate change is a phenomenon has been based on the available evidence from a very long time ago and now its existence is very probable. The speed and nature of climate parameters changes at the middle of twentieth century has been different and its quickness more than the before and its trend changed to some extent comparing to the past. Climate change issue now regarded as not only one of the most common scientific topic but also a social political one, is not a new issue. Climate change is a complicated atmospheric oceanic phenomenon on a global scale and long-term. Precipitation pattern change, fast decrease of snowcovered resources and its rapid melting, increased evaporation, the occurrence of destroying floods, water shortage crisis, severe reduction at the rate of harvesting agricultural products and, so on are all the significant of climate change. To cope with this phenomenon, its consequences and events in which public instruction is the most important but it may be climate that no significant cant and effective action has been done so far. The present article is included a part of one surrey about climate change in Fars. The study area having annually mean temperature 14 and precipitation 320 mm .23 stations inside the basin with a common 37 year statistical period have been applied to the meteorology data (1974-2010). Man-kendal and change factor methods are two statistical methods, applying them, the trend of changes and the annual mean average temperature and the annual minimum mean temperature were studied by using them. Based on time series for each parameter, the annual mean average temperature and the mean of annual maximum temperature have a rising trend so that this trend is clearer to the mean of annual maximum temperature.

Keywords: Climate change, Coefficient Variation, Fars province, Man-Kendal method.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1909
1689 Smart Power Scheduling to Reduce Peak Demand and Cost of Energy in Smart Grid

Authors: Hemant I. Joshi, Vivek J. Pandya

Abstract:

This paper discusses the simulation and experimental work of small Smart Grid containing ten consumers. Smart Grid is characterized by a two-way flow of real-time information and energy. RTP (Real Time Pricing) based tariff is implemented in this work to reduce peak demand, PAR (peak to average ratio) and cost of energy consumed. In the experimental work described here, working of Smart Plug, HEC (Home Energy Controller), HAN (Home Area Network) and communication link between consumers and utility server are explained. Algorithms for Smart Plug, HEC, and utility server are presented and explained in this work. After receiving the Real Time Price for different time slots of the day, HEC interacts automatically by running an algorithm which is based on Linear Programming Problem (LPP) method to find the optimal energy consumption schedule. Algorithm made for utility server can handle more than one off-peak time period during the day. Simulation and experimental work are carried out for different cases. At the end of this work, comparison between simulation results and experimental results are presented to show the effectiveness of the minimization method adopted.

Keywords: Smart Grid, Real Time Pricing, Peak to Average Ratio, Home Area Network, Home Energy Controller, Smart Plug, Utility Server, Linear Programming Problem.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1678
1688 More on Gaussian Quadratures for Fuzzy Functions

Authors: Shu-Xin Miao

Abstract:

In this paper, the Gaussian type quadrature rules for fuzzy functions are discussed. The errors representation and convergence theorems are given. Moreover, four kinds of Gaussian type quadrature rules with error terms for approximate of fuzzy integrals are presented. The present paper complements the theoretical results of the paper by T. Allahviranloo and M. Otadi [T. Allahviranloo, M. Otadi, Gaussian quadratures for approximate of fuzzy integrals, Applied Mathematics and Computation 170 (2005) 874-885]. The obtained results are illustrated by solving some numerical examples.

Keywords: Guassian quadrature rules, fuzzy number, fuzzy integral, fuzzy solution.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1430
1687 Performance Comparison of Different Regression Methods for a Polymerization Process with Adaptive Sampling

Authors: Florin Leon, Silvia Curteanu

Abstract:

Developing complete mechanistic models for polymerization reactors is not easy, because complex reactions occur simultaneously; there is a large number of kinetic parameters involved and sometimes the chemical and physical phenomena for mixtures involving polymers are poorly understood. To overcome these difficulties, empirical models based on sampled data can be used instead, namely regression methods typical of machine learning field. They have the ability to learn the trends of a process without any knowledge about its particular physical and chemical laws. Therefore, they are useful for modeling complex processes, such as the free radical polymerization of methyl methacrylate achieved in a batch bulk process. The goal is to generate accurate predictions of monomer conversion, numerical average molecular weight and gravimetrical average molecular weight. This process is associated with non-linear gel and glass effects. For this purpose, an adaptive sampling technique is presented, which can select more samples around the regions where the values have a higher variation. Several machine learning methods are used for the modeling and their performance is compared: support vector machines, k-nearest neighbor, k-nearest neighbor and random forest, as well as an original algorithm, large margin nearest neighbor regression. The suggested method provides very good results compared to the other well-known regression algorithms.

Keywords: Adaptive sampling, batch bulk methyl methacrylate polymerization, large margin nearest neighbor regression, machine learning.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1393
1686 Algorithm for Determining the Parameters of a Two-Layer Soil Model

Authors: Adekitan I. Aderibigbe, Fakolujo A. Olaosebikan

Abstract:

The parameters of a two-layer soil can be determined by processing resistivity data obtained from resistivity measurements carried out on the soil of interest. The processing usually entails applying the resistivity data as inputs to an optimisation function. This paper proposes an algorithm which utilises the square error as an optimisation function. Resistivity data from previous works were applied to test the accuracy of the new algorithm developed and the result obtained conforms significantly to results from previous works.

 

Keywords: Algorithm, earthing, resistivity, two-layer soil-model.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3327
1685 Integral Methods in the Determination of Temperature Fields of Cooled Blades of Gas Turbines

Authors: C. Ardil

Abstract:

A mathematical model and an effective numerical method for calculating the temperature field of the profile part of convection cooled blades have been developed. The theoretical substantiation of the method is proved by corresponding theorems. To this end, convergent quadrature processes were developed and error estimates were obtained in terms of the Zygmund continuity moduli.The boundary conditions for heat exchange are determined from the solution of the corresponding integral equations and empirical relations.The reliability of the developed methods is confirmed by the calculation-experimental studies of the thermohydraulic characteristics of the nozzle apparatus of the first stage of a gas turbine.

Keywords: Integral methods, determination of temperature fields, cooled blades, gas turbines.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 748
1684 Numerical Modeling of Temperature Fields in Aviation Gas Turbine Elements

Authors: A. M. Pashaev, R. A. Sadihov, A. S. Samedov, C. Ardil

Abstract:

A mathematical model and a numerical method for computing the temperature field of the profile part of convectionally cooled blades are developed. The theoretical substantiation of the method is proved by corresponding theorems. To this end, convergent quadrature processes were developed and error estimates were obtained in terms of the Zygmund continuity moduli. The boundary conditions for heat exchange are determined from the solution of the corresponding integral equations and empirical relations. The reliability of the developed methods is confirmed by calculation and experimental studies of the thermohydraulic characteristics of the nozzle apparatus of the first stage of the gas turbine.

Keywords: Aviation gas turbine, temperature field, cooled blades, numerical modeling.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 746
1683 A Kernel Classifier using Linearised Bregman Iteration

Authors: K. A. D. N. K Wimalawarne

Abstract:

In this paper we introduce a novel kernel classifier based on a iterative shrinkage algorithm developed for compressive sensing. We have adopted Bregman iteration with soft and hard shrinkage functions and generalized hinge loss for solving l1 norm minimization problem for classification. Our experimental results with face recognition and digit classification using SVM as the benchmark have shown that our method has a close error rate compared to SVM but do not perform better than SVM. We have found that the soft shrinkage method give more accuracy and in some situations more sparseness than hard shrinkage methods.

Keywords: Compressive sensing, Bregman iteration, Generalisedhinge loss, sparse, kernels, shrinkage functions

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1373
1682 Classification of Germinatable Mung Bean by Near Infrared Hyperspectral Imaging

Authors: Kaewkarn Phuangsombat, Arthit Phuangsombat, Anupun Terdwongworakul

Abstract:

Hard seeds will not grow and can cause mold in sprouting process. Thus, the hard seeds need to be separated from the normal seeds. Near infrared hyperspectral imaging in a range of 900 to 1700 nm was implemented to develop a model by partial least squares discriminant analysis to discriminate the hard seeds from the normal seeds. The orientation of the seeds was also studied to compare the performance of the models. The model based on hilum-up orientation achieved the best result giving the coefficient of determination of 0.98, and root mean square error of prediction of 0.07 with classification accuracy was equal to 100%.

Keywords: Mung bean, near infrared, germinatability, hard seed.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1154
1681 A Construction Management Tool: Determining a Project Schedule Typical Behaviors Using Cluster Analysis

Authors: Natalia Rudeli, Elisabeth Viles, Adrian Santilli

Abstract:

Delays in the construction industry are a global phenomenon. Many construction projects experience extensive delays exceeding the initially estimated completion time. The main purpose of this study is to identify construction projects typical behaviors in order to develop a prognosis and management tool. Being able to know a construction projects schedule tendency will enable evidence-based decision-making to allow resolutions to be made before delays occur. This study presents an innovative approach that uses Cluster Analysis Method to support predictions during Earned Value Analyses. A clustering analysis was used to predict future scheduling, Earned Value Management (EVM), and Earned Schedule (ES) principal Indexes behaviors in construction projects. The analysis was made using a database with 90 different construction projects. It was validated with additional data extracted from literature and with another 15 contrasting projects. For all projects, planned and executed schedules were collected and the EVM and ES principal indexes were calculated. A complete linkage classification method was used. In this way, the cluster analysis made considers that the distance (or similarity) between two clusters must be measured by its most disparate elements, i.e. that the distance is given by the maximum span among its components. Finally, through the use of EVM and ES Indexes and Tukey and Fisher Pairwise Comparisons, the statistical dissimilarity was verified and four clusters were obtained. It can be said that construction projects show an average delay of 35% of its planned completion time. Furthermore, four typical behaviors were found and for each of the obtained clusters, the interim milestones and the necessary rhythms of construction were identified. In general, detected typical behaviors are: (1) Projects that perform a 5% of work advance in the first two tenths and maintain a constant rhythm until completion (greater than 10% for each remaining tenth), being able to finish on the initially estimated time. (2) Projects that start with an adequate construction rate but suffer minor delays culminating with a total delay of almost 27% of the planned time. (3) Projects which start with a performance below the planned rate and end up with an average delay of 64%, and (4) projects that begin with a poor performance, suffer great delays and end up with an average delay of a 120% of the planned completion time. The obtained clusters compose a tool to identify the behavior of new construction projects by comparing their current work performance to the validated database, thus allowing the correction of initial estimations towards more accurate completion schedules.

Keywords: Cluster analysis, construction management, earned value, schedule.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1192
1680 Exploring Students’ Self-Evaluation on Their Learning Outcomes through an Integrated Cumulative Grade Point Average Reporting Mechanism

Authors: Suriyani Ariffin, Nor Aziah Alias, Khairil Iskandar Othman, Haslinda Yusoff

Abstract:

An Integrated Cumulative Grade Point Average (iCGPA) is a mechanism and strategy to ensure the curriculum of an academic programme is constructively aligned to the expected learning outcomes and student performance based on the attainment of those learning outcomes that is reported objectively in a spider web. Much effort and time has been spent to develop a viable mechanism and trains academics to utilize the platform for reporting. The question is: How well do learners conceive the idea of their achievement via iCGPA and whether quality learner attributes have been nurtured through the iCGPA mechanism? This paper presents the architecture of an integrated CGPA mechanism purported to address a holistic evaluation from the evaluation of courses learning outcomes to aligned programme learning outcomes attainment. The paper then discusses the students’ understanding of the mechanism and evaluation of their achievement from the generated spider web. A set of questionnaires were distributed to a group of students with iCGPA reporting and frequency analysis was used to compare the perspectives of students on their performance. In addition, the questionnaire also explored how they conceive the idea of an integrated, holistic reporting and how it generates their motivation to improve. The iCGPA group was found to be receptive to what they have achieved throughout their study period. They agreed that the achievement level generated from their spider web allows them to develop intervention and enhance the programme learning outcomes before they graduate.

Keywords: Learning outcomes attainment, iCGPA, programme learning outcomes, spider web, iCGPA reporting skills.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 767
1679 Determination of Measurement Uncertainty in Extracting of Forming Limit Diagrams

Authors: M. Mahboubkhah, H. Fayazfar

Abstract:

In this research, Forming Limit Diagrams for supertension sheet metals which are using in automobile industry have been obtained. The exerted strains to sheet metals have been measured with four different methods and the errors of each method have also been represented. These methods have been compared with together and the most efficient and economic way of extracting of the exerted strains to sheet metals has been introduced. In this paper total error and uncertainty of FLD extraction procedures have been derived. Determination of the measurement uncertainty in extracting of FLD has a great importance in design and analysis of the sheet metal forming process.

Keywords: Forming Limit Diagram, Major and Minor Strain, Measurement Uncertainty.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1992
1678 An Energy Aware Data Aggregation in Wireless Sensor Network Using Connected Dominant Set

Authors: M. Santhalakshmi, P Suganthi

Abstract:

Wireless Sensor Networks (WSNs) have many advantages. Their deployment is easier and faster than wired sensor networks or other wireless networks, as they do not need fixed infrastructure. Nodes are partitioned into many small groups named clusters to aggregate data through network organization. WSN clustering guarantees performance achievement of sensor nodes. Sensor nodes energy consumption is reduced by eliminating redundant energy use and balancing energy sensor nodes use over a network. The aim of such clustering protocols is to prolong network life. Low Energy Adaptive Clustering Hierarchy (LEACH) is a popular protocol in WSN. LEACH is a clustering protocol in which the random rotations of local cluster heads are utilized in order to distribute energy load among all sensor nodes in the network. This paper proposes Connected Dominant Set (CDS) based cluster formation. CDS aggregates data in a promising approach for reducing routing overhead since messages are transmitted only within virtual backbone by means of CDS and also data aggregating lowers the ratio of responding hosts to the hosts existing in virtual backbones. CDS tries to increase networks lifetime considering such parameters as sensors lifetime, remaining and consumption energies in order to have an almost optimal data aggregation within networks. Experimental results proved CDS outperformed LEACH regarding number of cluster formations, average packet loss rate, average end to end delay, life computation, and remaining energy computation.

Keywords: Wireless sensor network, connected dominant set, clustering, data aggregation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1126
1677 Irrigation Scheduling for Maize and Indian-mustard based on Daily Crop Water Requirement in a Semi- Arid Region

Authors: Vijay Shankar, C.S.P. Ojha, K.S. Hari Prasad

Abstract:

Maize and Indian mustard are significant crops in semi-arid climate zones of India. Improved water management requires precise scheduling of irrigation, which in turn requires an accurate computation of daily crop evapotranspiration (ETc). Daily crop evapotranspiration comes as a product of reference evapotranspiration (ET0) and the growth stage specific crop coefficients modified for daily variation. The first objective of present study is to develop crop coefficients Kc for Maize and Indian mustard. The estimated values of Kc for maize at the four crop growth stages (initial, development, mid-season, and late season) are 0.55, 1.08, 1.25, and 0.75, respectively, and for Indian mustard the Kc values at the four growth stages are 0.3, 0.6, 1.12, and 0.35, respectively. The second objective of the study is to compute daily crop evapotranspiration from ET0 and crop coefficients. Average daily ETc of maize varied from about 2.5 mm/d in the early growing period to > 6.5 mm/d at mid season. The peak ETc of maize is 8.3 mm/d and it occurred 64 days after sowing at the reproductive growth stage when leaf area index was 4.54. In the case of Indian mustard, average ETc is 1 mm/d at the initial stage, >1.8 mm/d at mid season and achieves a peak value of 2.12 mm/d on 56 days after sowing. Improved schedules of irrigation have been simulated based on daily crop evapo-transpiration and field measured data. Simulation shows a close match between modeled and field moisture status prevalent during crop season.

Keywords: Crop coefficient, Crop evapotranspiration, Field moisture, Irrigation Scheduling

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3323
1676 Approximating Fixed Points by a Two-Step Iterative Algorithm

Authors: Safeer Hussain Khan

Abstract:

In this paper, we introduce a two-step iterative algorithm to prove a strong convergence result for approximating common fixed points of three contractive-like operators. Our algorithm basically generalizes an existing algorithm..Our iterative algorithm also contains two famous iterative algorithms: Mann iterative algorithm and Ishikawa iterative algorithm. Thus our result generalizes the corresponding results proved for the above three iterative algorithms to a class of more general operators. At the end, we remark that nothing prevents us to extend our result to the case of the iterative algorithm with error terms.

Keywords: Contractive-like operator, iterative algorithm, fixed point, strong convergence.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2015
1675 Antenna Array Beamforming Using Neural Network

Authors: Maja Sarevska, Abdel-Badeeh M. Salem

Abstract:

This paper considers the problem of Null-Steering beamforming using Neural Network (NN) approach for antenna array system. Two cases are presented. First, unlike the other authors, the estimated Direction Of Arrivals (DOAs) are used for antenna array weights NN-based determination and the imprecise DOAs estimations are taken into account. Second, the blind null-steering beamforming is presented. In this case the antenna array outputs are presented at the input of the NN without DOAs estimation. The results of computer simulations will show much better relative mean error performances of the first NN approach compared to the NNbased blind beamforming.

Keywords: Beamforming, DOAs, neural network.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2411