Search results for: primary measures of music audiation
829 A New Analytical Approach to Reconstruct Residual Stresses Due to Turning Process
Authors: G.H. Farrahi, S.A. Faghidian, D.J. Smith
Abstract:
A thin layer on the component surface can be found with high tensile residual stresses, due to turning operations, which can dangerously affect the fatigue performance of the component. In this paper an analytical approach is presented to reconstruct the residual stress field from a limited incomplete set of measurements. Airy stress function is used as the primary unknown to directly solve the equilibrium equations and satisfying the boundary conditions. In this new method there exists the flexibility to impose the physical conditions that govern the behavior of residual stress to achieve a meaningful complete stress field. The analysis is also coupled to a least squares approximation and a regularization method to provide stability of the inverse problem. The power of this new method is then demonstrated by analyzing some experimental measurements and achieving a good agreement between the model prediction and the results obtained from residual stress measurement.Keywords: Residual stress, Limited measurements, Inverse problems, Turning process.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1432828 Experimental Results about the Dynamics of the Generalized Belief Propagation Used on LDPC Codes
Authors: Jean-Christophe Sibel, Sylvain Reynal, David Declercq
Abstract:
In the context of channel coding, the Generalized Belief Propagation (GBP) is an iterative algorithm used to recover the transmission bits sent through a noisy channel. To ensure a reliable transmission, we apply a map on the bits, that is called a code. This code induces artificial correlations between the bits to send, and it can be modeled by a graph whose nodes are the bits and the edges are the correlations. This graph, called Tanner graph, is used for most of the decoding algorithms like Belief Propagation or Gallager-B. The GBP is based on a non unic transformation of the Tanner graph into a so called region-graph. A clear advantage of the GBP over the other algorithms is the freedom in the construction of this graph. In this article, we explain a particular construction for specific graph topologies that involves relevant performance of the GBP. Moreover, we investigate the behavior of the GBP considered as a dynamic system in order to understand the way it evolves in terms of the time and in terms of the noise power of the channel. To this end we make use of classical measures and we introduce a new measure called the hyperspheres method that enables to know the size of the attractors.
Keywords: iterative decoder, LDPC, region-graph, chaos.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1648827 A New Objective Weight on Interval Type-2 Fuzzy Sets
Authors: Nurnadiah Z., Lazim A.
Abstract:
The design of weight is one of the important parts in fuzzy decision making, as it would have a deep effect on the evaluation results. Entropy is one of the weight measure based on objective evaluation. Non--probabilistic-type entropy measures for fuzzy set and interval type-2 fuzzy sets (IT2FS) have been developed and applied to weight measure. Since the entropy for (IT2FS) for decision making yet to be explored, this paper proposes a new objective weight method by using entropy weight method for multiple attribute decision making (MADM). This paper utilizes the nature of IT2FS concept in the evaluation process to assess the attribute weight based on the credibility of data. An example was presented to demonstrate the feasibility of the new method in decision making. The entropy measure of interval type-2 fuzzy sets yield flexible judgment and could be applied in decision making environment.Keywords: Objective weight, entropy weight, multiple attributedecision making, type-2 fuzzy sets, interval type-2 fuzzy sets
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1661826 Viability of Eggshells Ash Affecting the Setting Time of Cement
Authors: Fazeera Ujin, Kamran Shavarebi Ali, Zarina Yasmin Hanur Harith
Abstract:
This research paper reports on the feasibility and viability of eggshells ash and its effects on the water content and setting time of cement. An experiment was carried out to determine the quantity of water required in order to follow standard cement paste of normal consistency in accordance with MS EN 196-3:2007. The eggshells ash passing the 90µm sieve was used in the investigation. Eggshells ash with percentage of 0%, 0.1%, 0.5%, 1.0%, 1.5% and 2.0% were constituted to replace the cement. Chemical properties of both eggshells ash and cement are compared. From the results obtained, both eggshells ash and cement have the same chemical composition and primary composition which is the calcium compounds. Results from the setting time show that by adding the eggshells ash to the cement, the setting time of the cement decreases. In short, the higher amount of eggshells ash, the faster the rate of setting and apply to all percentage of eggshells ash that were used in this investigation. Both initial and final setting times fulfill the setting time requirements by Malaysian Standard. Hence, it is suggested that eggshells ash can be used as an admixture in concrete mix.
Keywords: Construction Materials, Eggshells Ash, Solid Waste, Setting Time.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2270825 Sizing the Protection Devices to Control Water Hammer Damage
Authors: I. Abuiziah, A. Oulhaj, K. Sebari, D. Ouazar
Abstract:
The primary objectives of transient analysis are to determine the values of transient pressures that can result from flow control operations and to establish the design criteria for system equipment and devices (such as control devices and pipe wall thickness) so as to provide an acceptable level of protection against system failure due to pipe collapse or bursting. Because of the complexity of the equations needed to describe transients, numerical computer models are used to analyze transient flow hydraulics. An effective numerical model allows the hydraulic engineer to analyze potential transient events and to identify and evaluate alternative solutions for controlling hydraulic transients, thereby protecting the integrity of the hydraulic system. This paper presents the influence of using the protection devices to control the adverse effects due to excessive and low pressure occurs in the transient.
Keywords: Flow Transient, Water hammer, Pipeline System, Surge Tank, Simulation Model, Protection Devices.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9495824 Hand Hygiene Habits of Ghanaian Youths in Accra, Ghana
Authors: Timothy B. Oppong, Haiyan Yang, Cecilia Amponsem-Boateng, Guangcai Duan
Abstract:
The human palm has been identified as one of the richest habitats for human microbial accommodation making hand hygiene essential to primary prevention of infection. Since the hand is in constant contact with fomites which have been proven to be mostly contaminated, building hand hygiene habits is essential for the prevention of infection. This research was conducted to assess the hand hygiene habits of Ghanaian youths in Accra. This study used a survey as a quantitative method of research. The findings of the study revealed that out of the 254 participants who fully answered the questionnaire, 22% had the habit of washing their hands after outings while only 51.6% had the habit of washing their hands after using the bathroom. However, about 60% of the participants said they sometimes ate with their hands while 28.9% had the habit of eating with the hand very often, a situation that put them at risk of infection from their hands since some participants had poor handwashing habits; prompting the need for continuous education on hand hygiene.
Keywords: Hand hygiene, hand hygiene habits, hand washing, hand sanitizer use.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 231823 Bayesian Networks for Earthquake Magnitude Classification in a Early Warning System
Authors: G. Zazzaro, F.M. Pisano, G. Romano
Abstract:
During last decades, worldwide researchers dedicated efforts to develop machine-based seismic Early Warning systems, aiming at reducing the huge human losses and economic damages. The elaboration time of seismic waveforms is to be reduced in order to increase the time interval available for the activation of safety measures. This paper suggests a Data Mining model able to correctly and quickly estimate dangerousness of the running seismic event. Several thousand seismic recordings of Japanese and Italian earthquakes were analyzed and a model was obtained by means of a Bayesian Network (BN), which was tested just over the first recordings of seismic events in order to reduce the decision time and the test results were very satisfactory. The model was integrated within an Early Warning System prototype able to collect and elaborate data from a seismic sensor network, estimate the dangerousness of the running earthquake and take the decision of activating the warning promptly.Keywords: Bayesian Networks, Decision Support System, Magnitude Classification, Seismic Early Warning System
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3598822 Response Quality Evaluation in Heterogeneous Question Answering System: A Black-box Approach
Authors: Goh Ong Sing, C. Ardil, Wilson Wong, Shahrin Sahib
Abstract:
The evaluation of the question answering system is a major research area that needs much attention. Before the rise of domain-oriented question answering systems based on natural language understanding and reasoning, evaluation is never a problem as information retrieval-based metrics are readily available for use. However, when question answering systems began to be more domains specific, evaluation becomes a real issue. This is especially true when understanding and reasoning is required to cater for a wider variety of questions and at the same time achieve higher quality responses The research in this paper discusses the inappropriateness of the existing measure for response quality evaluation and in a later part, the call for new standard measures and the related considerations are brought forward. As a short-term solution for evaluating response quality of heterogeneous systems, and to demonstrate the challenges in evaluating systems of different nature, this research presents a black-box approach using observation, classification scheme and a scoring mechanism to assess and rank three example systems (i.e. AnswerBus, START and NaLURI).
Keywords: Evaluation, question answering, response quality.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1859821 Optimizing Usability Testing with Collaborative Method in an E-Commerce Ecosystem
Authors: Markandeya Kunchi
Abstract:
Usability testing (UT) is one of the vital steps in the User-centred design (UCD) process when designing a product. In an e-commerce ecosystem, UT becomes primary as new products, features, and services are launched very frequently. And, there are losses attached to the company if an unusable and inefficient product is put out to market and is rejected by customers. This paper tries to answer why UT is important in the product life-cycle of an E-commerce ecosystem. Secondary user research was conducted to find out work patterns, development methods, type of stakeholders, and technology constraints, etc. of a typical E-commerce company. Qualitative user interviews were conducted with product managers and designers to find out the structure, project planning, product management method and role of the design team in a mid-level company. The paper tries to address the usual apprehensions of the company to inculcate UT within the team. As well, it stresses upon factors like monetary resources, lack of usability expert, narrow timelines, and lack of understanding of higher management as some primary reasons. Outsourcing UT to vendors is also very prevalent with mid-level e-commerce companies, but it has its own severe repercussions like very little team involvement, huge cost, misinterpretation of the findings, elongated timelines, and lack of empathy towards the customer, etc. The shortfalls of the unavailability of a UT process in place within the team and conducting UT through vendors are bad user experiences for customers while interacting with the product, badly designed products which are neither useful and nor utilitarian. As a result, companies see dipping conversions rates in apps and websites, huge bounce rates and increased uninstall rates. Thus, there was a need for a more lean UT system in place which could solve all these issues for the company. This paper highlights on optimizing the UT process with a collaborative method. The degree of optimization and structure of collaborative method is the highlight of this paper. Collaborative method of UT is one in which the centralised design team of the company takes for conducting and analysing the UT. The UT is usually a formative kind where designers take findings into account and uses in the ideation process. The success of collaborative method of UT is due to its ability to sync with the product management method employed by the company or team. The collaborative methods focus on engaging various teams (design, marketing, product, administration, IT, etc.) each with its own defined roles and responsibility in conducting a smooth UT with users In-house. The paper finally highlights the positive results of collaborative UT method after conducting more than 100 In-lab interviews with users across the different lines of businesses. Some of which are the improvement of interaction between stakeholders and the design team, empathy towards users, improved design iteration, better sanity check of design solutions, optimization of time and money, effective and efficient design solution. The future scope of collaborative UT is to make this method leaner, by reducing the number of days to complete the entire project starting from planning between teams to publishing the UT report.
Keywords: Usability testing, collaborative method, e-commerce, product management method.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 670820 Measurement of the Bipolarization Events
Authors: Stefan V. Stefanescu
Abstract:
We intend to point out the differences which exist between the classical Gini concentration coefficient and a proposed bipolarization index defined for an arbitrary random variable which have a finite support. In fact Gini's index measures only the "poverty degree" for the individuals from a given population taking into consideration their wages. The Gini coefficient is not so sensitive to the significant income variations in the "rich people class" . In practice there are multiple interdependent relations between the pauperization and the socio-economical polarization phenomena. The presence of a strong pauperization aspect inside the population induces often a polarization effect in this society. But the pauperization and the polarization phenomena are not identical. For this reason it isn't always adequate to use a Gini type coefficient, based on the Lorenz order, to estimate the bipolarization level of the individuals from the studied population. The present paper emphasizes these ideas by considering two families of random variables which have a linear or a triangular type distributions. In addition, the continuous variation, depending on the parameter "time" of the chosen distributions, could simulate a real dynamical evolution of the population.Keywords: Bipolarization phenomenon, Gini coefficient, income distribution, poverty measure.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1146819 Heat Exchanger Design
Authors: Su Thet Mon Than, Khin Aung Lin, Mi Sandar Mon
Abstract:
This paper is intended to assist anyone with some general technical experience, but perhaps limited specific knowledge of heat transfer equipment. A characteristic of heat exchanger design is the procedure of specifying a design, heat transfer area and pressure drops and checking whether the assumed design satisfies all requirements or not. The purpose of this paper is how to design the oil cooler (heat exchanger) especially for shell-and-tube heat exchanger which is the majority type of liquid-to-liquid heat exchanger. General design considerations and design procedure are also illustrated in this paper and a flow diagram is provided as an aid of design procedure. In design calculation, the MatLAB and AutoCAD software are used. Fundamental heat transfer concepts and complex relationships involved in such exchanger are also presented in this paper. The primary aim of this design is to obtain a high heat transfer rate without exceeding the allowable pressure drop. This computer program is highly useful to design the shell-and-tube type heat exchanger and to modify existing deign.
Keywords: Shell-and-Tube Heat Exchanger, MatLAB and AutoCAD
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 7932818 A Critical Review of the Adequacy of EIA Reports-Evidence from Pakistan
Authors: Obaidullah Nadeem, Rizwan Hameed
Abstract:
The preparation of good-quality Environmental Impact Assessment (EIA) reports contribute to enhancing overall effectiveness of EIA. This component of the EIA process becomes more important in situation where public participation is weak and there is lack of expertise on the part of the competent authority. In Pakistan, EIA became mandatory for every project likely to cause adverse environmental impacts from July 1994. The competent authority also formulated guidelines for preparation and review of EIA reports in 1997. However, EIA is yet to prove as a successful decision support tool to help in environmental protection. One of the several reasons of this ineffectiveness is the generally poor quality of EIA reports. This paper critically reviews EIA reports of some randomly selected projects. Interviews of EIA consultants, project proponents and concerned government officials have also been conducted to underpin the root causes of poor quality of EIA reports. The analysis reveals several inadequacies particularly in areas relating to identification, evaluation and mitigation of key impacts and consideration of alternatives. The paper identifies some opportunities and suggests measures for improving the quality of EIA reports and hence making EIA an effective tool to help in environmental protection.
Keywords: Environmental Impact Assessment, EIA Guidelines, EIA Reports, Pakistan.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3311817 Gabriel Mtsire’s "The Golden Spring" and Its Primary Sources, Textual and Content Changes Based on Cultural Development in the Context of the 4th-20th Centuries
Authors: Georgi Kalandadze
Abstract:
For studying the development of world civilizations, textual sources that have undergone textological and worldview changes are of great importance. The paper will discuss the collection of the XVIII century "The Golden Spring", compiled by Gabriel Mtsire, which includes texts of John Chrysostom. The teachings of John Chrysostom of the 4th century were translated into Georgian in the 10th-11th centuries by Euthymes of Athos. These texts correspond to the requirements of the Georgian society of the 10th-11th centuries. In the 18th century, Gabriel Mtsire collected and edited these texts to make them more understandable to his modern readers. In the 20th century, these texts were again adapted. Thus, the present study provides an opportunity to evaluate and outline the linguistic and content transformation process of the same work over 16 centuries.
Keywords: Gabriel Mtsire, John Chrysostom, Euthymius the Athonite, The Golden Spring.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 29816 Performance of Hybrid-MIMO Receiver Scheme in Cognitive Radio Network
Authors: Tanapong Khomyat, Peerapong Uthansakul, Monthippa Uthansakul
Abstract:
In this paper, we evaluate the performance of the Hybrid-MIMO Receiver Scheme (HMRS) in Cognitive Radio network (CR-network). We investigate the efficiency of the proposed scheme which the energy level and user number of primary user are varied according to the characteristic of CR-network. HMRS can allow users to transmit either Space-Time Block Code (STBC) or Spatial-Multiplexing (SM) streams simultaneously by using Successive Interference Cancellation (SIC) and Maximum Likelihood Detection (MLD). From simulation, the results indicate that the interference level effects to the performance of HMRS. Moreover, the exact closed-form capacity of the proposed scheme is derived and compared with STBC scheme.Keywords: Hybrid-MIMO, Cognitive radio network (CRnetwork), Symbol Error Rate (SER), Successive interference cancellation (SIC), Maximum likelihood detection (MLD).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1637815 Connectionist Approach to Generic Text Summarization
Authors: Rajesh S.Prasad, U. V. Kulkarni, Jayashree.R.Prasad
Abstract:
As the enormous amount of on-line text grows on the World-Wide Web, the development of methods for automatically summarizing this text becomes more important. The primary goal of this research is to create an efficient tool that is able to summarize large documents automatically. We propose an Evolving connectionist System that is adaptive, incremental learning and knowledge representation system that evolves its structure and functionality. In this paper, we propose a novel approach for Part of Speech disambiguation using a recurrent neural network, a paradigm capable of dealing with sequential data. We observed that connectionist approach to text summarization has a natural way of learning grammatical structures through experience. Experimental results show that our approach achieves acceptable performance. Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1591814 Adjustment and Compensation Techniques for the Rotary Axes of Five-axis CNC Machine Tools
Authors: Tung-Hui Hsu, Wen-Yuh Jywe
Abstract:
Five-axis computer numerical control (CNC) machine tools (three linear and two rotary axes) are ideally suited to the fabrication of complex work pieces, such as dies, turbo blades, and cams. The locations of the axis average line and centerline of the rotary axes strongly influence the performance of these machines; however, techniques to compensate for eccentric error in the rotary axes remain weak. This paper proposes optical (Non-Bar) techniques capable of calibrating five-axis CNC machine tools and compensating for eccentric error in the rotary axes. This approach employs the measurement path in ISO/CD 10791-6 to determine the eccentric error in two rotary axes, for which compensatory measures can be implemented. Experimental results demonstrate that the proposed techniques can improve the performance of various five-axis CNC machine tools by more than 90%. Finally, a result of the cutting test using a B-type five-axis CNC machine tool confirmed to the usefulness of this proposed compensation technique.
Keywords: Calibration, compensation, rotary axis, five-axis computer numerical control (CNC) machine tools, eccentric error, optical calibration system, ISO/CD 10791-6
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4168813 Bridging Quantitative and Qualitative of Glaucoma Detection
Authors: Noor Elaiza Abdul Khalid, Noorhayati Mohamed Noor, Zamalia Mahmud, Saadiah Yahya, and Norharyati Md Ariff
Abstract:
Glaucoma diagnosis involves extracting three features of the fundus image; optic cup, optic disc and vernacular. Present manual diagnosis is expensive, tedious and time consuming. A number of researches have been conducted to automate this process. However, the variability between the diagnostic capability of an automated system and ophthalmologist has yet to be established. This paper discusses the efficiency and variability between ophthalmologist opinion and digital technique; threshold. The efficiency and variability measures are based on image quality grading; poor, satisfactory or good. The images are separated into four channels; gray, red, green and blue. A scientific investigation was conducted on three ophthalmologists who graded the images based on the image quality. The images are threshold using multithresholding and graded as done by the ophthalmologist. A comparison of grade from the ophthalmologist and threshold is made. The results show there is a small variability between result of ophthalmologists and digital threshold.Keywords: Digital Fundus Image, Glaucoma Detection, Multithresholding, Segmentation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2044812 A Study on the Application of TRIZ to CAD/CAM System
Authors: Yuan L. Lai, Jian H. Chen, Jui P. Hung
Abstract:
This study created new graphical icons and operating functions in a CAD/CAM software system by analyzing icons in some of the popular systems, such as AutoCAD, AlphaCAM, Mastercam and the 1st edition of LiteCAM. These software systems all focused on geometric design and editing, thus how to transmit messages intuitively from icon itself to users is an important function of graphical icons. The primary purpose of this study is to design innovative icons and commands for new software. This study employed the TRIZ method, an innovative design method, to generate new concepts systematically. Through literature review, it then investigated and analyzed the relationship between TRIZ and idea development. Contradiction Matrix and 40 Principles were used to develop an assisting tool suitable for icon design in software development. We first gathered icon samples from the selected CAD/CAM systems. Then grouped these icons by meaningful functions, and compared useful and harmful properties. Finally, we developed new icons for new software systems in order to avoid intellectual property problem.Keywords: Icon, TRIZ, CAD/CAM.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1794811 Coastal Ecological Sensitivity and Risk Assessment: A Case Study of Sea Level Change in Apodi River (Atlantic Ocean), Northeast Brazil
Authors: Mukesh Singh Boori, Venerando Eustáquio Amaro, Helenice Vital
Abstract:
The present study has been carried out with a view to calculate the coastal vulnerability index (CVI) to know the high and low sensitive areas and area of inundation due to future SLR. Both conventional and remotely sensed data were used and analyzed through the modelling technique. Out of the total study area, 8.26% is very high risk, 14.21% high, 9.36% medium, 22.46% low and 7.35% in the very low vulnerable category, due to costal components. Results of the inundation analysis indicate that 225.2 km² and 397 km² of the land area will be submerged by flooding at 1m and 10m inundation levels. The most severely affected sectors are expected to be the residential, industrial and recreational areas. As this coast is planned for future coastal developmental activities, measures such as industrializations, building regulation, urban growth planning and agriculture, development of an integrated coastal zone management, strict enforcement of the Coastal Regulation Zone (CRZ) Act, monitoring of impacts and further research in this regard are recommended for the study area.
Keywords: Coastal planning, land use, satellite data, vulnerability.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1986810 Vetiver Oil Production from Root Culture of Vetiveria zizanioides
Authors: Rizkita R. Esyanti, Iriawati, Olga Mardisadora
Abstract:
Vetiver oil is secondary metabolite that accumulates in Vetiveria zizanioides roots. The aim of this study was to obtain best type of root culture which produce high amount of vetiver oil, and was similar to metabolites produce from its mother plant. Protein analysis was also conducted to detect protein, related to putative enzymes, which have a role in terpenoids synthesis in the root culture. The results showed that root culture derived from crown explant produced the best root growth. The root culture produced primary and lateral roots, ca. 40 branches. The vetiver oil produced from root culture was analyzed by using GC-MS., and the highest content of terpenoids from roots of crown explant attained 19.024%. The result of SDS PAGE showed proteins which were ±61 kD and ± 68 kD, each might be related to putative monoterpene synthase and sesquiterpenes complex, respectively.
Keywords: Protein, Root culture, Terpenoids, Vetiver oil.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3816809 Effects of Introducing Similarity Measures into Artificial Bee Colony Approach for Optimization of Vehicle Routing Problem
Authors: P. Shunmugapriya, S. Kanmani, P. Jude Fredieric, U. Vignesh, J. Reman Justin, K. Vivek
Abstract:
Vehicle Routing Problem (VRP) is a complex combinatorial optimization problem and it is quite difficult to find an optimal solution consisting of a set of routes for vehicles whose total cost is minimum. Evolutionary and swarm intelligent (SI) algorithms play a vital role in solving optimization problems. While the SI algorithms perform search, the diversity between the solutions they exploit is very important. This is because of the need to avoid early convergence and to get an appropriate balance between the exploration and exploitation. Therefore, it is important to check how far the solutions are diverse. In this paper, we measure the similarity between solutions, which ABC exploits while optimizing VRP. The similar solutions found are discarded at the end of the iteration and only unique solutions are passed on to the next iteration. The bees of discarded solutions become scouts and they start searching for new solutions. This process is continued and results show that the solution is optimized at lesser number of iterations but with the overhead of computing similarity in all the iterations. The problem instance from Solomon benchmarked dataset has been used for evaluating the presented methodology.
Keywords: ABC algorithm, vehicle routing problem, optimization, Jaccard’s similarity measure.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 846808 Using the Technology Acceptance Model to Examine Seniors’ Attitudes toward Facebook
Authors: Chien-Jen Liu, Shu Ching Yang
Abstract:
Using the technology acceptance model (TAM), this study examined the external variables of technological complexity (TC) to acquire a better understanding of the factors that influence the acceptance of computer application courses by learners at Active Aging Universities. After the learners in this study had completed a 27-hour Facebook course, 44 learners responded to a modified TAM survey. Data were collected to examine the path relationships among the variables that influence the acceptance of Facebook-mediated community learning. The partial least squares (PLS) method was used to test the measurement and the structural model. The study results demonstrated that attitudes toward Facebook use directly influence behavioral intentions (BI) with respect to Facebook use, evincing a high prediction rate of 58.3%. In addition to the perceived usefulness (PU) and perceived ease of use (PEOU) measures that are proposed in the TAM, other external variables, such as TC, also indirectly influence BI. These four variables can explain 88% of the variance in BI and demonstrate a high level of predictive ability. Finally, limitations of this investigation and implications for further research are discussed.
Keywords: Technology acceptance model (TAM), technological complexity, partial least squares (PLS), perceived usefulness.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3196807 Adaptive Filtering of Heart Rate Signals for an Improved Measure of Cardiac Autonomic Control
Authors: Desmond B. Keenan, Paul Grossman
Abstract:
In order to provide accurate heart rate variability indices of sympathetic and parasympathetic activity, the low frequency and high frequency components of an RR heart rate signal must be adequately separated. This is not always possible by just applying spectral analysis, as power from the high and low frequency components often leak into their adjacent bands. Furthermore, without the respiratory spectra it is not obvious that the low frequency component is not another respiratory component, which can appear in the lower band. This paper describes an adaptive filter, which aids the separation of the low frequency sympathetic and high frequency parasympathetic components from an ECG R-R interval signal, enabling the attainment of more accurate heart rate variability measures. The algorithm is applied to simulated signals and heart rate and respiratory signals acquired from an ambulatory monitor incorporating single lead ECG and inductive plethysmography sensors embedded in a garment. The results show an improvement over standard heart rate variability spectral measurements.Keywords: Heart rate variability, vagal tone, sympathetic, parasympathetic, spectral analysis, adaptive filter.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1755806 Combustion and Emission Characteristics in a Can-type Combustion Chamber
Authors: Selvakuma Kumaresh, Man Young Kim
Abstract:
Combustion phenomenon will be accomplished effectively by the development of low emission combustor. One of the significant factors influencing the entire Combustion process is the mixing between a swirling angular jet (Primary Air) and the non-swirling inner jet (fuel). To study this fundamental flow, the chamber had to be designed in such a manner that the combustion process to sustain itself in a continuous manner and the temperature of the products is sufficiently below the maximum working temperature in the turbine. This study is used to develop the effective combustion with low unburned combustion products by adopting the concept of high swirl flow and motility of holes in the secondary chamber. The proper selection of a swirler is needed to reduce emission which can be concluded from the emission of Nox and CO2. The capture of CO2 is necessary to mitigate CO2 emissions from natural gas. Thus the suppression of unburned gases is a meaningful objective for the development of high performance combustor without affecting turbine blade temperature.
Keywords: Combustion, Emission, Can-type Combustion Chamber, CFD, Motility of Holes, Swirl Flow.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3506805 Sustainable Urban Development of Slum Prone Area of Dhaka City
Authors: S. Ahmed Sinthia
Abstract:
Dhaka, the capital city of Bangladesh, is one of the densely populated cities in the world. Due to rapid urbanization 60% of its population lives in slum and squatter settlements. The reason behind this poverty is low economic growth, inequitable distribution of income, unequal distribution of productive assets, unemployment and underemployment, high rate of population growth, low level of human resource development, natural disasters, and limited access to public services. Along with poverty, creating pressure on urban land, shelter, plots, open spaces this creates environmental and ecological degradation. These constraints are mostly resulted from the failures of the government policies and measures and only Government can solve this problem. This is now prime time to establish planning and environmental management policy and sustainable urban development for the city and for the urban slum dwellers which are free from eviction, criminals, rent seekers and other miscreants.Keywords: Economic and resource constraints, environmental degradation and management, government policy, land management and policy, sustainable development.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5757804 Modeling Ambient Carbon Monoxide Pollutant Due to Road Traffic
Authors: Anjaneyulu M.V.L.R., Harikrishna M., Chenchuobulu S.
Abstract:
Rapid urbanization, industrialization and population growth have led to an increase in number of automobiles that cause air pollution. It is estimated that road traffic contributes 60% of air pollution in urban areas. A case by case assessment is required to predict the air quality in urban situations, so as to evolve certain traffic management measures to maintain the air quality levels with in the tolerable limits. Calicut city in the state of Kerala, India has been chosen as the study area. Carbon Monoxide (CO) concentration was monitored at 15 links in Calicut city and air quality performance was evaluated over each link. The CO pollutant concentration values were compared with the National Ambient Air Quality Standards (NAAQS), and the CO values were predicted by using CALINE4 and IITLS and Linear regression models. The study has revealed that linear regression model performs better than the CALINE4 and IITLS models. The possible association between CO pollutant concentration and traffic parameters like traffic flow, type of vehicle, and traffic stream speed was also evaluated.Keywords: CO pollution, Modelling, Traffic stream parameters.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2366803 A Meta-Heuristic Algorithm for Set Covering Problem Based on Gravity
Authors: S. Raja Balachandar, K. Kannan
Abstract:
A new Meta heuristic approach called "Randomized gravitational emulation search algorithm (RGES)" for solving large size set covering problems has been designed. This algorithm is found upon introducing randomization concept along with the two of the four primary parameters -velocity- and -gravity- in physics. A new heuristic operator is introduced in the domain of RGES to maintain feasibility specifically for the set covering problem to yield best solutions. The performance of this algorithm has been evaluated on a large set of benchmark problems from OR-library. Computational results showed that the randomized gravitational emulation search algorithm - based heuristic is capable of producing high quality solutions. The performance of this heuristic when compared with other existing heuristic algorithms is found to be excellent in terms of solution quality.
Keywords: Set covering problem, velocity, gravitational force, Newton's law, meta heuristic, combinatorial optimization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2235802 Some Design Issues in Designing of 50KW 50Krpm Permanent Magnet Synchronous Machine
Authors: Ali A. Mehna, Mohmed A. Ali, Ali S. Zayed
Abstract:
A numbers of important developments have led to an increasing attractiveness for very high speed electrical machines (either motor or generator). Specifically the increasing switching speed of power electronics, high energy magnets, high strength retaining materials, better high speed bearings and improvements in design analysis are the primary drivers in a move to higher speed. The design challenges come in the mechanical design both in terms of strength and resonant modes and in the electromagnetic design particularly in respect of iron losses and ac losses in the various conducting parts including the rotor. This paper describes detailed design work which has been done on a 50,000 rpm, 50kW permanent magnet( PM) synchronous machine. It describes work on electromagnetic and rotor eddy current losses using a variety of methods including both 2D finite element analysisKeywords: High speed, PM motor, rotor and stator losses, finiteelement analysis
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2647801 Predicting the Life Cycle of Complex Technical Systems (CTS)
Authors: Khalil A. Yaghi, Samer Barakat
Abstract:
Complex systems are composed of several plain interacting independent entities. Interaction between these entities creates a unified behavior at the global level that cannot be predicted by examining the behavior of any single individual component of the system. In this paper we consider a welded frame of an automobile trailer as a real example of Complex Technical Systems, The purpose of this paper is to introduce a Statistical method for predicting the life cycle of complex technical systems. To organize gathering of primary data for modeling the life cycle of complex technical systems an “Automobile Trailer Frame" were used as a prototype in this research. The prototype represents a welded structure of several pieces. Both information flows underwent a computerized analysis and classification for the acquisition of final results to reach final recommendations for improving the trailers structure and their operational conditions.
Keywords: Complex Technical System (CTS), AutomobileTrailer Frame, Automobile Service.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1235800 Spatial Analysis of Park and Ride Users’ Dynamic Accessibility to Train Station: A Case Study in Perth
Authors: Ting (Grace) Lin, Jianhong (Cecilia) Xia, Todd Robinson
Abstract:
Accessibility analysis, examining people’s ability to access facilities and destinations, is a fundamental assessment for transport planning, policy making, and social exclusion research. Dynamic accessibility which measures accessibility in real-time traffic environment has been an advanced accessibility indicator in transport research. It is also a useful indicator to help travelers to understand travel time daily variability, assists traffic engineers to monitor traffic congestions, and finally develop effective strategies in order to mitigate traffic congestions. This research involved real-time traffic information by collecting travel time data with 15-minute interval via the TomTom® API. A framework for measuring dynamic accessibility was then developed based on the gravity theory and accessibility dichotomy theory through space and time interpolation. Finally, the dynamic accessibility can be derived at any given time and location under dynamic accessibility spatial analysis framework.
Keywords: Dynamic accessibility, space-time continuum, transport research, TomTom® API.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1084