Search results for: Continuous speech recognition
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 1705

Search results for: Continuous speech recognition

265 Comparison between Higher-Order SVD and Third-order Orthogonal Tensor Product Expansion

Authors: Chiharu Okuma, Jun Murakami, Naoki Yamamoto

Abstract:

In digital signal processing it is important to approximate multi-dimensional data by the method called rank reduction, in which we reduce the rank of multi-dimensional data from higher to lower. For 2-dimennsional data, singular value decomposition (SVD) is one of the most known rank reduction techniques. Additional, outer product expansion expanded from SVD was proposed and implemented for multi-dimensional data, which has been widely applied to image processing and pattern recognition. However, the multi-dimensional outer product expansion has behavior of great computation complex and has not orthogonally between the expansion terms. Therefore we have proposed an alterative method, Third-order Orthogonal Tensor Product Expansion short for 3-OTPE. 3-OTPE uses the power method instead of nonlinear optimization method for decreasing at computing time. At the same time the group of B. D. Lathauwer proposed Higher-Order SVD (HOSVD) that is also developed with SVD extensions for multi-dimensional data. 3-OTPE and HOSVD are similarly on the rank reduction of multi-dimensional data. Using these two methods we can obtain computation results respectively, some ones are the same while some ones are slight different. In this paper, we compare 3-OTPE to HOSVD in accuracy of calculation and computing time of resolution, and clarify the difference between these two methods.

Keywords: Singular value decomposition (SVD), higher-order SVD (HOSVD), higher-order tensor, outer product expansion, power method.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1532
264 From Separatism to Coalition: Variants in Language Politics and Leadership Pattern in Dravidian Movement

Authors: Subramaniam Chandran

Abstract:

This paper describes the evolution of language politics and the part played by political leaders with reference to the Dravidian parties in Tamil Nadu. It explores the interesting evolution from separatism to coalition in sustaining the values of parliamentary democracy and federalism. It seems that the appropriation of language politics is fully ascribed to the DMK leadership under Annadurai and Karunanidhi. For them, the Tamil language is a self-determining power, a terrain of nationhood, and a perennial source of social and political powers. The DMK remains a symbol of Tamil nationalist party playing language politics in the interest of the Tamils. Though electoral alliances largely determine the success, the language politics still has significant space in the politics of Tamil Nadu. Ironically, DMK moves from the periphery to centre for getting national recognition for the Tamils as well as for its own maximization of power. The evolution can be seen in two major phases as: language politics for party building; and language politics for state building with three successive political processes, namely, language politics in the process of separatism, representative politics and coalition. The much pronounced Dravidian Movement is radical enough to democratize the party ideology to survive the spirit of parliamentary democracy. This has secured its own rewards in terms of political power. The political power provides the means to achieve the social and political goal of the political party. Language politics and leadership pattern actualized this trend though the movement is shifted from separatism to coalition.

Keywords: Language politics, cultural nationalism, leadership, social justice

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1931
263 Successful Management of a Boy with Mild Persistent Asthma (A Longitudinal Case)

Authors: Lubis A., Setiawati L., Setyoningrum A. R., Suryawan A., Irwanto

Abstract:

Asthma is a condition that causing chronic health problems in children. In addition to basic therapy against disease, we must try to reduce the impact of chronic health problems and also optimize their medical aspect of growth and development. A boy with mild asthma attack frequent episode did not showed any improvement with medical treatment and his asthma control test was 11. From radiologic examination he got hyperaerated lung and billateral sinusitis maxillaris; skin test results were house dust, food and pet allergy; an overweight body; bad school grades; psychological and environmental problem. We followed and evaluated this boy in 6 months, treated holistically. Even we could not do much on environmental but no more psychological and school problems, his on a good bodyweight and his asthma control test was 22. A case of a child with mild asthma attack frequent episode was reported. Asthma clinical course show no significant improvement when other predisposing factor is not well-controlled and a child’s growth and development may be affected. Improving condition of the patient can be created with the help of loving and caring way of nurturing from the parents and supportive peer group. Therefore, continuous and consistent monitoring is required because prognosis of asthma is generally good when regularly and properly controlled.

Keywords: Asthma, chronic health problems, growth and development.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1616
262 Finding Pareto Optimal Front for the Multi-Mode Time, Cost Quality Trade-off in Project Scheduling

Authors: H. Iranmanesh, M. R. Skandari, M. Allahverdiloo

Abstract:

Project managers are the ultimate responsible for the overall characteristics of a project, i.e. they should deliver the project on time with minimum cost and with maximum quality. It is vital for any manager to decide a trade-off between these conflicting objectives and they will be benefited of any scientific decision support tool. Our work will try to determine optimal solutions (rather than a single optimal solution) from which the project manager will select his desirable choice to run the project. In this paper, the problem in project scheduling notated as (1,T|cpm,disc,mu|curve:quality,time,cost) will be studied. The problem is multi-objective and the purpose is finding the Pareto optimal front of time, cost and quality of a project (curve:quality,time,cost), whose activities belong to a start to finish activity relationship network (cpm) and they can be done in different possible modes (mu) which are non-continuous or discrete (disc), and each mode has a different cost, time and quality . The project is constrained to a non-renewable resource i.e. money (1,T). Because the problem is NP-Hard, to solve the problem, a meta-heuristic is developed based on a version of genetic algorithm specially adapted to solve multi-objective problems namely FastPGA. A sample project with 30 activities is generated and then solved by the proposed method.

Keywords: FastPGA, Multi-Execution Activity Mode, ParetoOptimality, Project Scheduling, Time-Cost-Quality Trade-Off.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1656
261 Reversible, Embedded and Highly Scalable Image Compression System

Authors: Federico Pérez González, Iñaki Goirizelaia Ordorika, Pedro Iriondo Bengoa

Abstract:

In this work a new method for low complexity image coding is presented, that permits different settings and great scalability in the generation of the final bit stream. This coding presents a continuous-tone still image compression system that groups loss and lossless compression making use of finite arithmetic reversible transforms. Both transformation in the space of color and wavelet transformation are reversible. The transformed coefficients are coded by means of a coding system in depending on a subdivision into smaller components (CFDS) similar to the bit importance codification. The subcomponents so obtained are reordered by means of a highly configure alignment system depending on the application that makes possible the re-configure of the elements of the image and obtaining different importance levels from which the bit stream will be generated. The subcomponents of each importance level are coded using a variable length entropy coding system (VBLm) that permits the generation of an embedded bit stream. This bit stream supposes itself a bit stream that codes a compressed still image. However, the use of a packing system on the bit stream after the VBLm allows the realization of a final highly scalable bit stream from a basic image level and one or several improvement levels.

Keywords: Image compression, wavelet transform, highly scalable, reversible transform, embedded, subcomponents.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1267
260 Reading against the Grain: Transcodifying Stimulus Meaning

Authors: Aba-Carina Pârlog

Abstract:

The paper shows that on transferring sense from the SL to the TL, the translator’s reading against the grain determines the creation of a faulty pattern of rendering the original meaning in the receiving culture which reflects the use of misleading transformative codes. In this case, the translator is a writer per se who decides what goes in and out of the book, how the style is to be ciphered and what elements of ideology are to be highlighted. The paper also proves that figurative language must not be flattened for the sake of clarity or naturalness. The missing figurative elements make the translated text less interesting, less challenging and less vivid which reflects poorly on the writer. There is a close connection between style and the writer’s person. If the writer’s style is very much altered in a translation, the translation is useless as the original writer and his / her imaginative world can no longer be discovered. The purpose of the paper is to prove that adaptation is a dangerous tool which leads to variants that sometimes reflect the original less than the reader would wish to. It contradicts the very essence of the process of translation which is that of making an original work available in a foreign language. If the adaptive transformative codes are so flexible that they encourage the translator to repeatedly leave out parts of the original work, then a subversive pattern emerges which changes the entire book. In conclusion, as a result of using adaptation, manipulative or subversive effects are created in the translated work. This is generally achieved by adding new words or connotations, creating new figures of speech or using explicitations. The additional meanings of the original work are neglected and the translator creates new meanings, implications, emphases and contexts. Again s/he turns into a new author who enjoys the freedom of expressing his / her own ideas without the constraints of the original text. Reading against the grain is unadvisable during the process of translation and consequently, following personal common sense becomes essential in the field of translation as well as everywhere else, so that translation should not become a source of fantasy.

Keywords: Speculative aesthetics, substance of expression, transformative code, translation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1625
259 Intelligent Recognition of Diabetes Disease via FCM Based Attribute Weighting

Authors: Kemal Polat

Abstract:

In this paper, an attribute weighting method called fuzzy C-means clustering based attribute weighting (FCMAW) for classification of Diabetes disease dataset has been used. The aims of this study are to reduce the variance within attributes of diabetes dataset and to improve the classification accuracy of classifier algorithm transforming from non-linear separable datasets to linearly separable datasets. Pima Indians Diabetes dataset has two classes including normal subjects (500 instances) and diabetes subjects (268 instances). Fuzzy C-means clustering is an improved version of K-means clustering method and is one of most used clustering methods in data mining and machine learning applications. In this study, as the first stage, fuzzy C-means clustering process has been used for finding the centers of attributes in Pima Indians diabetes dataset and then weighted the dataset according to the ratios of the means of attributes to centers of theirs. Secondly, after weighting process, the classifier algorithms including support vector machine (SVM) and k-NN (k- nearest neighbor) classifiers have been used for classifying weighted Pima Indians diabetes dataset. Experimental results show that the proposed attribute weighting method (FCMAW) has obtained very promising results in the classification of Pima Indians diabetes dataset.

Keywords: Fuzzy C-means clustering, Fuzzy C-means clustering based attribute weighting, Pima Indians diabetes dataset, SVM.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1735
258 The Kinetic of Biogas Production Rate from Cattle Manure in Batch Mode

Authors: Budiyono, I N. Widiasa, S. Johari, Sunarso

Abstract:

In this study, the kinetic of biogas production was studied by performing a series laboratory experiment using rumen fluid of animal ruminant as inoculums. Cattle manure as substrate was inoculated by rumen fluid to the anaerobic biodigester. Laboratory experiments using 400 ml biodigester were performed in batch operation mode. Given 100 grams of fresh cattle manure was fed to each biodigester and mixed with rumen fluid by manure : rumen weight ratio of 1:1 (MR11). The operating temperatures were varied at room temperature and 38.5 oC. The cumulative volume of biogas produced was used to measure the biodigester performance. The research showed that the rumen fluid inoculated to biodigester gave significant effect to biogas production (P<0.05). Rumen fluid inoculums caused biogas production rate and efficiency increase two to three times in compare to manure substrate without rumen fluid. With the rumen fluid inoculums, gave the kinetic parameters of biogas production i.e biogas production rate constants (U), maximum biogas production (A), and minimum time to produce biogas (λ) are 3.89 ml/(gVS.day); 172.51 (ml/gVS); dan 7.25 days, respectively. While the substrate without rumen fluid gave the kinetic parameters U, A, and λ are 1.74 ml/(gVS.day); 73.81 (ml/gVS); dan 14.75 days, respectively. The future work will be carried out to study the dynamics of biogas production if both the rumen inoculums and manure are fed in the continuous system.

Keywords: rumen fluid, inoculums, anaerobic digestion, biogasproduction.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2975
257 A Fuzzy-Rough Feature Selection Based on Binary Shuffled Frog Leaping Algorithm

Authors: Javad Rahimipour Anaraki, Saeed Samet, Mahdi Eftekhari, Chang Wook Ahn

Abstract:

Feature selection and attribute reduction are crucial problems, and widely used techniques in the field of machine learning, data mining and pattern recognition to overcome the well-known phenomenon of the Curse of Dimensionality. This paper presents a feature selection method that efficiently carries out attribute reduction, thereby selecting the most informative features of a dataset. It consists of two components: 1) a measure for feature subset evaluation, and 2) a search strategy. For the evaluation measure, we have employed the fuzzy-rough dependency degree (FRFDD) of the lower approximation-based fuzzy-rough feature selection (L-FRFS) due to its effectiveness in feature selection. As for the search strategy, a modified version of a binary shuffled frog leaping algorithm is proposed (B-SFLA). The proposed feature selection method is obtained by hybridizing the B-SFLA with the FRDD. Nine classifiers have been employed to compare the proposed approach with several existing methods over twenty two datasets, including nine high dimensional and large ones, from the UCI repository. The experimental results demonstrate that the B-SFLA approach significantly outperforms other metaheuristic methods in terms of the number of selected features and the classification accuracy.

Keywords: Binary shuffled frog leaping algorithm, feature selection, fuzzy-rough set, minimal reduct.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 702
256 An Anisotropic Model of Damage and Unilateral Effect for Brittle Materials

Authors: José Julio de C. Pituba

Abstract:

This work deals with the initial applications and formulation of an anisotropic plastic-damage constitutive model proposed for non-linear analysis of reinforced concrete structures submitted to a loading with change of the sign. The original constitutive model is based on the fundamental hypothesis of energy equivalence between real and continuous medium following the concepts of the Continuum Damage Mechanics. The concrete is assumed as an initial elastic isotropic medium presenting anisotropy, permanent strains and bimodularity (distinct elastic responses whether traction or compression stress states prevail) induced by damage evolution. In order to take into account the bimodularity, two damage tensors governing the rigidity in tension or compression regimes are introduced. Then, some conditions are introduced in the original version of the model in order to simulate the damage unilateral effect. The three-dimensional version of the proposed model is analyzed in order to validate its formulation when compared to micromechanical theory. The one-dimensional version of the model is applied in the analyses of a reinforced concrete beam submitted to a loading with change of the sign. Despite the parametric identification problems, the initial applications show the good performance of the model.

Keywords: Damage model, plastic strain, unilateral effect.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1799
255 Development and Initial Validation of the Social Competency Inventory for Tertiary Level Faculty Members

Authors: Glenn M. Calaguas, Carmela S. Dizon

Abstract:

This study aimed to develop and initially validate an instrument that measures social competency among tertiary level faculty members. A review of extant literature on social competence was done. The review of extant literature led to the writing of the items in the initial instrument which was evaluated by 11 Subject Matter Experts (SMEs). The SMEs were either educators or psychologists. The results of the evaluations done by the SMEs served as bases for the creation of the pre-try-out instrument used in the first trial-run. Insights from the first trial-run participants led to the development of the main try-out instrument used in the final test administration. One Hundred Forty-one participants from five private Higher Education Institutions (HEIs) in the National Capital Region (NCR) and five private HEIs in Central Luzon in the Philippines participated in the final test administration. The reliability of the instrument was evaluated using Cronbach-s Coefficient Alpha formula and had a Cronbach-s Alpha of 0.92. On the other hand, Factor Analysis was used to evaluate the validity of the instrument and six factors were identified. The development of the final instrument was based on the results of the evaluation of the instrument-s reliability and validity. For purposes of recognition, the instrument was named “Social Competency Inventory for Tertiary Level Faculty Members (SCI-TLFM)."

Keywords: development, initial validation, social competency, tertiary level faculty members

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2133
254 A Psychophysiological Evaluation of an Effective Recognition Technique Using Interactive Dynamic Virtual Environments

Authors: Mohammadhossein Moghimi, Robert Stone, Pia Rotshtein

Abstract:

Recording psychological and physiological correlates of human performance within virtual environments and interpreting their impacts on human engagement, ‘immersion’ and related emotional or ‘effective’ states is both academically and technologically challenging. By exposing participants to an effective, real-time (game-like) virtual environment, designed and evaluated in an earlier study, a psychophysiological database containing the EEG, GSR and Heart Rate of 30 male and female gamers, exposed to 10 games, was constructed. Some 174 features were subsequently identified and extracted from a number of windows, with 28 different timing lengths (e.g. 2, 3, 5, etc. seconds). After reducing the number of features to 30, using a feature selection technique, K-Nearest Neighbour (KNN) and Support Vector Machine (SVM) methods were subsequently employed for the classification process. The classifiers categorised the psychophysiological database into four effective clusters (defined based on a 3-dimensional space – valence, arousal and dominance) and eight emotion labels (relaxed, content, happy, excited, angry, afraid, sad, and bored). The KNN and SVM classifiers achieved average cross-validation accuracies of 97.01% (±1.3%) and 92.84% (±3.67%), respectively. However, no significant differences were found in the classification process based on effective clusters or emotion labels.

Keywords: Virtual Reality, effective computing, effective VR, emotion-based effective physiological database.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 967
253 A Novel SVM-Based OOK Detector in Low SNR Infrared Channels

Authors: J. P. Dubois, O. M. Abdul-Latif

Abstract:

Support Vector Machine (SVM) is a recent class of statistical classification and regression techniques playing an increasing role in applications to detection problems in various engineering problems, notably in statistical signal processing, pattern recognition, image analysis, and communication systems. In this paper, SVM is applied to an infrared (IR) binary communication system with different types of channel models including Ricean multipath fading and partially developed scattering channel with additive white Gaussian noise (AWGN) at the receiver. The structure and performance of SVM in terms of the bit error rate (BER) metric is derived and simulated for these channel stochastic models and the computational complexity of the implementation, in terms of average computational time per bit, is also presented. The performance of SVM is then compared to classical binary signal maximum likelihood detection using a matched filter driven by On-Off keying (OOK) modulation. We found that the performance of SVM is superior to that of the traditional optimal detection schemes used in statistical communication, especially for very low signal-to-noise ratio (SNR) ranges. For large SNR, the performance of the SVM is similar to that of the classical detectors. The implication of these results is that SVM can prove very beneficial to IR communication systems that notoriously suffer from low SNR at the cost of increased computational complexity.

Keywords: Least square-support vector machine, on-off keying, matched filter, maximum likelihood detector, wireless infrared communication.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1925
252 Embryo Transfer as an Assisted Reproductive Technology in Farm Animals

Authors: Diah Tri Widayati

Abstract:

Various assisted reproductive techniques have been developed and refined to obtain a large number of offspring from genetically superior animals or obtain offspring from infertile (or subfertile) animals. The embryo transfer is one assisted reproductive technique developed well, aimed at increased productivity of selected females, disease control, importation and exportation of livestock, rapid screening of AI sires for genetically recessive characteristics, treatment or circumvention of certain types of infertility. Embryo transfer also is a useful research tool for evaluating fetal and maternal interactions. This technique has been applied to nearly every species of domestic animal and many species of wildlife and exotic animals, including humans and non-human primates. The successful of embryo transfers have been limited to within-animal, homologous replacement of the embryos. There are several examples of interspecific and intergeneric embryo transfers in which embryos implanted but did not develop to term: sheep and goat, mouse and rat. An immunological rejections and placental incompatibility between the embryo and the surrogate mother appear to restrict interspecific embryo transfer/interspecific pregnancy. Recently, preimplantation embryo manipulation procedures have been applied, such as technique of inner cell mass transfer. This technique will possible to overcome the reproductive barrier interspecific embryo transfer/interspecific pregnancy, if there is a protective mechanism which prevents recognition of the foreign fetus by the mother of the other species

Keywords: Embryo Transfer, Assisted Reproductive Techology, Intraspesific-Interspesific Pregnancy, Inner cell mass.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4694
251 Finding Pareto Optimal Front for the Multi- Mode Time, Cost Quality Trade-off in Project Scheduling

Authors: H. Iranmanesh, M. R. Skandari, M. Allahverdiloo

Abstract:

Project managers are the ultimate responsible for the overall characteristics of a project, i.e. they should deliver the project on time with minimum cost and with maximum quality. It is vital for any manager to decide a trade-off between these conflicting objectives and they will be benefited of any scientific decision support tool. Our work will try to determine optimal solutions (rather than a single optimal solution) from which the project manager will select his desirable choice to run the project. In this paper, the problem in project scheduling notated as (1,T|cpm,disc,mu|curve:quality,time,cost) will be studied. The problem is multi-objective and the purpose is finding the Pareto optimal front of time, cost and quality of a project (curve:quality,time,cost), whose activities belong to a start to finish activity relationship network (cpm) and they can be done in different possible modes (mu) which are non-continuous or discrete (disc), and each mode has a different cost, time and quality . The project is constrained to a non-renewable resource i.e. money (1,T). Because the problem is NP-Hard, to solve the problem, a meta-heuristic is developed based on a version of genetic algorithm specially adapted to solve multi-objective problems namely FastPGA. A sample project with 30 activities is generated and then solved by the proposed method.

Keywords: FastPGA, Multi-Execution Activity Mode, Pareto Optimality, Project Scheduling, Time-Cost-Quality Trade-Off.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1779
250 On Pooling Different Levels of Data in Estimating Parameters of Continuous Meta-Analysis

Authors: N. R. N. Idris, S. Baharom

Abstract:

A meta-analysis may be performed using aggregate data (AD) or an individual patient data (IPD). In practice, studies may be available at both IPD and AD level. In this situation, both the IPD and AD should be utilised in order to maximize the available information. Statistical advantages of combining the studies from different level have not been fully explored. This study aims to quantify the statistical benefits of including available IPD when conducting a conventional summary-level meta-analysis. Simulated meta-analysis were used to assess the influence of the levels of data on overall meta-analysis estimates based on IPD-only, AD-only and the combination of IPD and AD (mixed data, MD), under different study scenario. The percentage relative bias (PRB), root mean-square-error (RMSE) and coverage probability were used to assess the efficiency of the overall estimates. The results demonstrate that available IPD should always be included in a conventional meta-analysis using summary level data as they would significantly increased the accuracy of the estimates.On the other hand, if more than 80% of the available data are at IPD level, including the AD does not provide significant differences in terms of accuracy of the estimates. Additionally, combining the IPD and AD has moderating effects on the biasness of the estimates of the treatment effects as the IPD tends to overestimate the treatment effects, while the AD has the tendency to produce underestimated effect estimates. These results may provide some guide in deciding if significant benefit is gained by pooling the two levels of data when conducting meta-analysis.

Keywords: Aggregate data, combined-level data, Individual patient data, meta analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1713
249 Effect of Chemical Pretreatments and Dehydration Methods on Quality Characteristics of Tomato Powder and Its Storage Stability

Authors: Reihaneh Ahmadzadeh Ghavidel, Mehdi Ghiafeh Davoodi

Abstract:

Dehydration process was carried out for tomato slices of var. Avinash after giving different pre-treatments such as calcium chloride (CaCl2), potassium metabisulphite (KMS), calcium chloride and potassium metabisulphite (CaCl2 +KMS), and sodium chloride (NaCl). Untreated samples served as control. Solar drier and continuous conveyor (tunnel) drier were used for dehydration. Quality characteristics of tomato slices viz. moisture content, sugar, titratable acidity, lycopene content, dehydration ratio, rehydration ratio and non-enzymatic browning as affected by dehydration process were studied. Storage study was also carried out for a period of six months for tomato powder packed into different types of packaging materials viz. metalized polyester (MP) film and low density poly ethylene (LDPE). Changes in lycopene content and non-enzymatic browning (NEB) were estimated during storage at room temperature. Pretreatment of 5 mm thickness of tomato slices with calcium chloride in combination with potassium metabisulphite and drying using a tunnel drier with subsequent storage of product in metalized polyester bags was selected as the best process.

Keywords: Drying pretreatments, Solar drying, Tomato powder, Tunnel drying

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2805
248 Active Segment Selection Method in EEG Classification Using Fractal Features

Authors: Samira Vafaye Eslahi

Abstract:

BCI (Brain Computer Interface) is a communication machine that translates brain massages to computer commands. These machines with the help of computer programs can recognize the tasks that are imagined. Feature extraction is an important stage of the process in EEG classification that can effect in accuracy and the computation time of processing the signals. In this study we process the signal in three steps of active segment selection, fractal feature extraction, and classification. One of the great challenges in BCI applications is to improve classification accuracy and computation time together. In this paper, we have used student’s 2D sample t-statistics on continuous wavelet transforms for active segment selection to reduce the computation time. In the next level, the features are extracted from some famous fractal dimension estimation of the signal. These fractal features are Katz and Higuchi. In the classification stage we used ANFIS (Adaptive Neuro-Fuzzy Inference System) classifier, FKNN (Fuzzy K-Nearest Neighbors), LDA (Linear Discriminate Analysis), and SVM (Support Vector Machines). We resulted that active segment selection method would reduce the computation time and Fractal dimension features with ANFIS analysis on selected active segments is the best among investigated methods in EEG classification.

Keywords: EEG, Student’s t- statistics, BCI, Fractal Features, ANFIS, FKNN.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2089
247 Improvement of Voltage Profile of Grid Integrated Wind Distributed Generation by SVC

Authors: Fariba Shavakhi Zavareh, Hadi Fotoohabadi, Reza Sedaghati

Abstract:

Due to the continuous increment of the load demand, identification of weaker buses, improvement of voltage profile and power losses in the context of the voltage stability problems has become one of the major concerns for the larger, complex, interconnected power systems. The objective of this paper is to review the impact of Flexible AC Transmission System (FACTS) controller in Wind generators connected electrical network for maintaining voltage stability. Wind energy could be the growing renewable energy due to several advantages. The influence of wind generators on power quality is a significant issue; non uniform power production causes variations in system voltage and frequency. Therefore, wind farm requires high reactive power compensation; the advances in high power semiconducting devices have led to the development of FACTS. The FACTS devices such as for example SVC inject reactive power into the system which helps in maintaining a better voltage profile. The performance is evaluated on an IEEE 14 bus system, two wind generators are connected at low voltage buses to meet the increased load demand and SVC devices are integrated at the buses with wind generators to keep voltage stability. Power flows, nodal voltage magnitudes and angles of the power network are obtained by iterative solutions using MIPOWER.

Keywords: Voltage Profile, FACTS Device, SVC, Distributed Generation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2635
246 Exploring Methods and Strategies for Sustainable Urban Development

Authors: Klio Monokrousou, Maria Giannopoulou

Abstract:

Urban areas, as they have been developed and operate today, are areas of accumulation of a significant amount of people and a large number of activities that generate desires and reasons for traveling. The territorial expansion of the cities as well as the need to preserve the importance of the central city areas lead to the continuous increase of transportation needs which in the limited urban space results in creating serious traffic and operational problems. The modern perception of urban planning is directed towards more holistic approaches and integrated policies that make it economically competitive, socially just and more environmentally friendly. Over the last 25 years, the goal of sustainable transport development has been central to the agenda of any plan or policy for the city. The modern planning of urban space takes into account the economic and social aspects of the city and the importance of the environment to sustainable urban development. In this context, the European Union promotes direct or indirect related interventions according to the cohesion and environmental policies; many countries even had the chance to actually test them. This paper explores the methods and processes that have been developed towards this direction and presents a review and systematic presentation of this work. The ultimate purpose of this research is to effectively use this review to create a decision making methodological framework which can be the basis of a useful operational tool for sustainable urban planning.

Keywords: Sustainable urban development, urban mobility, urban regeneration methods.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3475
245 Generation of Electro-Encephalography Readiness Potentials by Intention

Authors: Seokbeen Lim, Gilwon Yoon

Abstract:

The readiness potential in brain waves is a brain activity related with an intention whose potential arises even before its conscious intention. This study was carried out in order to understand the generation and mechanism of the readiness potential more. The experiment with two subjects was conducted in two ways following the Oddball task protocol. Firstly, auditory stimuli were randomly presented to the subjects. The subject was allowed to press the keyboard with the right index finger only when the subject heard the target stimulus but not the standard stimulus. Secondly, unlike the first one, the auditory stimuli were randomly presented, and the subjects pressed the keyboard in the same manner, but at the same time with grasping action of the left hand. The readiness potential showed up for both of these experiments. In the first Oddball experiment, the readiness potential was detected only when the target stimulus was presented. However, in the second Oddball experiment with the left hand action of grasping something, the readiness potential was detected at the presentation of for both standard and target stimuli. However, detected readiness potentials with the target stimuli were larger than those of the standard stimuli. We found an interesting phenomenon that the readiness potential was able to be detected even the standard stimulus. This indicates that motor-related readiness potentials can be generated only by the intention to move. These results present a new perspective in psychology and brain engineering since subconscious brain action may be prior to conscious recognition of the intention.

Keywords: Readiness potential, auditory stimuli, event-related potential, electroencephalography, oddball task.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 998
244 Continuous Fixed Bed Reactor Application for Decolourization of Textile Effluent by Adsorption on NaOH Treated Eggshell

Authors: M. Chafi, S. Akazdam, C. Asrir, L. Sebbahi, B. Gourich, N. Barka, M. Essahli

Abstract:

Fixed bed adsorption has become a frequently used industrial application in wastewater treatment processes. Various low cost adsorbents have been studied for their applicability in treatment of different types of effluents. In this work, the intention of the study was to explore the efficacy and feasibility for azo dye, Acid Orange 7 (AO7) adsorption onto fixed bed column of NaOH Treated eggshell (TES). The effect of various parameters like flow rate, initial dye concentration, and bed height were exploited in this study. The studies confirmed that the breakthrough curves were dependent on flow rate, initial dye concentration solution of AO7 and bed depth. The Thomas, Yoon–Nelson, and Adams and Bohart models were analysed to evaluate the column adsorption performance. The adsorption capacity, rate constant and correlation coefficient associated to each model for column adsorption was calculated and mentioned. The column experimental data were fitted well with Thomas model with coefficients of correlation R2 ≥0.93 at different conditions but the Yoon–Nelson, BDST and Bohart–Adams model (R2=0.911), predicted poor performance of fixed-bed column. The (TES) was shown to be suitable adsorbent for adsorption of AO7 using fixed-bed adsorption column.

Keywords: Adsorption models, acid orange 7, bed depth, breakthrough, dye adsorption, fixed-bed column, treated eggshell.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2233
243 Prioritising the TQM Enablers and IT Resources in the ICT Industry: An AHP Approach

Authors: Suby Khanam, Jamshed Siddiqui, Faisal Talib

Abstract:

Total Quality Management (TQM) is a managerial approach that improves the competitiveness of the industry, meanwhile Information technology (IT) was introduced with TQM for handling the technical issues which is supported by quality experts for fulfilling the customers’ requirement. Present paper aims to utilise AHP (Analytic Hierarchy Process) methodology to priorities and rank the hierarchy levels of TQM enablers and IT resource together for its successful implementation in the Information and Communication Technology (ICT) industry. A total of 17 TQM enablers (nine) and IT resources (eight) were identified and partitioned into 3 categories and were prioritised by AHP approach. The finding indicates that the 17 sub-criteria can be grouped into three main categories namely organizing, tools and techniques, and culture and people. Further, out of 17 sub-criteria, three sub-criteria: top management commitment and support, total employee involvement, and continuous improvement got highest priority whereas three sub-criteria such as structural equation modelling, culture change, and customer satisfaction got lowest priority. The result suggests a hierarchy model for ICT industry to prioritise the enablers and resources as well as to improve the TQM and IT performance in the ICT industry. This paper has some managerial implication which suggests the managers of ICT industry to implement TQM and IT together in their organizations to get maximum benefits and how to utilize available resources. At the end, conclusions, limitation, future scope of the study are presented.

Keywords: Analytic Hierarchy Process, Information Technology, Information and Communication Technology, Prioritization, Total Quality Management.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1934
242 Applying the Extreme-Based Teaching Model in Post-Secondary Online Classroom Setting: A Field Experiment

Authors: Leon Pan

Abstract:

The first programming course within post-secondary education has long been recognized as a challenging endeavor for both educators and students alike. Historically, these courses have exhibited high failure rates and a notable number of dropouts. Instructors often lament students' lack of effort on their coursework, and students often express frustration that the teaching methods employed are not effective. Drawing inspiration from the successful principles of Extreme Programming, this study introduces an approach—the Extremes-based teaching model—aimed at enhancing the teaching of introductory programming courses. To empirically determine the effectiveness of the model, a comparison was made between a section taught using the extreme-based model and another utilizing traditional teaching methods. Notably, the extreme-based teaching class required students to work collaboratively on projects, while also demanding continuous assessment and performance enhancement within groups. This paper details the application of the extreme-based model within the post-secondary online classroom context and presents the compelling results that emphasize its effectiveness in advancing the teaching and learning experiences. The extreme-based model led to a significant increase of 13.46 points in the weighted total average and a commendable 10% reduction in the failure rate.

Keywords: Extreme-based teaching model, innovative pedagogical methods, project-based learning, team-based learning.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 62
241 Service Quality and Consumer Behavior on Metered Taxi Services

Authors: Nattapong Techarattanased

Abstract:

The purposes of this research are to make comparisons in respect of the behaviors on the use of the services of metered taxi classified by the demographic factor and to study the influence of the recognition on service quality having the effect on usage behaviors of metered taxi services of consumers in Bangkok Metropolitan Areas. The samples used in this research were 400 metered taxi service users in Bangkok Metropolitan Areas and questionnaire was used as the tool for collecting the data. Analysis statistics are mean and multiple regression analysis. Results of the research revealed that the consumers recognize the overall quality of services in each aspect include tangible aspects of the service, responses to customers, assurance on the confidence, understanding and knowing of customers which is rated at the moderate level except the aspect of the assurance on the confidence and trustworthiness which are rated at a high level. For the result of hypothetical test, it is found that the quality in providing the services on the aspect of the assurance given to the customers has the effect on the usage behaviors of metered taxi services and the aspect of the frequency on the use of the services per month which in this connection. Such variable can forecast at one point nine percent (1.9%). In addition, quality in providing the services and the aspect of the responses to customers have the effect on the behaviors on the use of metered taxi services on the aspect of the expenses on the use of services per month which in this connection, such variable can forecast at two point one percent (2.1%).

Keywords: Consumer behavior, metered taxi, satisfaction, service quality.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3511
240 Variation of Streamwise and Vertical Turbulence Intensity in a Smooth and Rough Bed Open Channel Flow

Authors: Md Abdullah Al Faruque, Ram Balachandar

Abstract:

An experimental study with four different types of bed conditions was carried out to understand the effect of roughness in open channel flow at two different Reynolds numbers. The bed conditions include a smooth surface and three different roughness conditions, which were generated using sand grains with a median diameter of 2.46 mm. The three rough conditions include a surface with distributed roughness, a surface with continuously distributed roughness and a sand bed with a permeable interface. A commercial two-component fibre-optic LDA system was used to conduct the velocity measurements. The variables of interest include the mean velocity, turbulence intensity, correlation between the streamwise and the wall normal turbulence, Reynolds shear stress and velocity triple products. Quadrant decomposition was used to extract the magnitude of the Reynolds shear stress of the turbulent bursting events. The effect of roughness was evident throughout the flow depth. The results show that distributed roughness has the greatest roughness effect followed by the sand bed and the continuous roughness. Compared to the smooth bed, the streamwise turbulence intensity reduces but the vertical turbulence intensity increases at a location very close to the bed due to the introduction of roughness. Although the same sand grain is used to create the three different rough bed conditions, the difference in the turbulence intensity is an indication that the specific geometry of the roughness has an influence on turbulence structure.

Keywords: Open channel flow, smooth bed, rough bed, Reynolds number, turbulence.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2206
239 Targeting the Life Cycle Stages of the Diamond Back Moth (Plutella xylostella) with Three Different Parasitoid Wasps

Authors: F. O. Faithpraise, J. Idung, C. R. Chatwin, R. C. D. Young, P. Birch

Abstract:

A continuous time model of the interaction between crop insect pests and naturally beneficial pest enemies is created using a set of simultaneous, non-linear, ordinary differential equations incorporating natural death rates based on the Weibull distribution. The crop pest is present in all its life-cycle stages of: egg, larva, pupa and adult. The beneficial insects, parasitoid wasps, may be present in either or all parasitized: eggs, larva and pupa. Population modelling is used to estimate the quantity of the natural pest enemies that should be introduced into the pest infested environment to suppress the pest population density to an economically acceptable level within a prescribed number of days. The results obtained illustrate the effect of different combinations of parasitoid wasps, using the Pascal distribution to estimate their success in parasitizing different pest developmental stages, to deliver pest control to a sustainable level. Effective control, within a prescribed number of days, is established by the deployment of two or all three species of wasps, which partially destroy pest: egg, larvae and pupae stages. The selected scenarios demonstrate effective sustainable control of the pest in less than thirty days.

Keywords: Biological control, Diamondback moth, Parasitoid wasps, Population modeling.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3021
238 A Condition-Based Maintenance Policy for Multi-Unit Systems Subject to Deterioration

Authors: Nooshin Salari, Viliam Makis

Abstract:

In this paper, we propose a condition-based maintenance policy for multi-unit systems considering the existence of economic dependency among units. We consider a system composed of N identical units, where each unit deteriorates independently. Deterioration process of each unit is modeled as a three-state continuous time homogeneous Markov chain with two working states and a failure state. The average production rate of units varies in different working states and demand rate of the system is constant. Units are inspected at equidistant time epochs, and decision regarding performing maintenance is determined by the number of units in the failure state. If the total number of units in the failure state exceeds a critical level, maintenance is initiated, where units in failed state are replaced correctively and deteriorated state units are maintained preventively. Our objective is to determine the optimal number of failed units to initiate maintenance minimizing the long run expected average cost per unit time. The problem is formulated and solved in the semi-Markov decision process (SMDP) framework. A numerical example is developed to demonstrate the proposed policy and the comparison with the corrective maintenance policy is presented.

Keywords: Reliability, production, maintenance optimization, Semi-Markov Decision Process.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 767
237 ADA Tool for Satellite InSAR-Based Ground Displacement Analysis: The Granada Region

Authors: M. Cuevas-González, O. Monserrat, A. Barra, C. Reyes-Carmona, R. M. Mateos, J. P. Galve, R. Sarro, M. Cantalejo, E. Peña, M. Martínez-Corbella, J. A. Luque, J. M. Azañón, A. Millares, M. Béjar, J. A. Navarro, L. Solari

Abstract:

Geohazard prone areas require continuous monitoring to detect risks, understand the phenomena occurring in those regions and prevent disasters. Satellite interferometry (InSAR) has come to be a trustworthy technique for ground movement detection and monitoring in the last few years. InSAR based techniques allow to process large areas providing high number of displacement measurements at low cost. However, the results provided by such techniques are usually not easy to interpret by non-experienced users hampering its use for decision makers. This work presents a set of tools developed in the framework of different projects (Momit, Safety, U-Geohaz, Riskcoast) and an example of their use in the Granada Coastal area (Spain) is shown. The ADA (Active Displacement Areas) tool has been developed with the aim of easing the management, use and interpretation of InSAR based results. It provides a semi-automatic extraction of the most significant ADAs through the application ADAFinder tool. This tool aims to support the exploitation of the European Ground Motion Service (EU-GMS), which will offer reliable and systematic information on natural and anthropogenic ground motion phenomena across Europe.

Keywords: Ground displacements, InSAR, natural hazards, satellite imagery.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 360
236 Enhanced Particle Swarm Optimization Approach for Solving the Non-Convex Optimal Power Flow

Authors: M. R. AlRashidi, M. F. AlHajri, M. E. El-Hawary

Abstract:

An enhanced particle swarm optimization algorithm (PSO) is presented in this work to solve the non-convex OPF problem that has both discrete and continuous optimization variables. The objective functions considered are the conventional quadratic function and the augmented quadratic function. The latter model presents non-differentiable and non-convex regions that challenge most gradient-based optimization algorithms. The optimization variables to be optimized are the generator real power outputs and voltage magnitudes, discrete transformer tap settings, and discrete reactive power injections due to capacitor banks. The set of equality constraints taken into account are the power flow equations while the inequality ones are the limits of the real and reactive power of the generators, voltage magnitude at each bus, transformer tap settings, and capacitor banks reactive power injections. The proposed algorithm combines PSO with Newton-Raphson algorithm to minimize the fuel cost function. The IEEE 30-bus system with six generating units is used to test the proposed algorithm. Several cases were investigated to test and validate the consistency of detecting optimal or near optimal solution for each objective. Results are compared to solutions obtained using sequential quadratic programming and Genetic Algorithms.

Keywords: Particle Swarm Optimization, Optimal Power Flow, Economic Dispatch.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2339