Search results for: Control Approach.
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 8145

Search results for: Control Approach.

5505 Performance Evaluation of Universities as Groups of Decision Making Units

Authors: Ali Payan, Bijan Rahmani Parchicolaie

Abstract:

Universities have different offices such as educational, research, student, administrative, and financial offices. This paper considers universities as groups of decision making units (DMUs) in which DMUs are their offices. This approach gives us with a more just evaluation of universities instead of separate evaluation of the offices of universities. The proposed approach to evaluate group performance of universities is based on common set of weights method in DEA. The suggested method not only can compare groups and measure their efficiencies, but also can calculate the efficiency of units within group and efficiency spread of groups. At last, the suggested method is applied for the analysis of the performance of universities in 14th district of Islamic Azad University as groups under evaluation.

Keywords: Common set of weights, group efficiency, performance analysis, spread efficiency.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1463
5504 A New Biologically Inspired Pattern Recognition Spproach for Face Recognition

Authors: V. Kabeer, N.K.Narayanan

Abstract:

This paper reports a new pattern recognition approach for face recognition. The biological model of light receptors - cones and rods in human eyes and the way they are associated with pattern vision in human vision forms the basis of this approach. The functional model is simulated using CWD and WPD. The paper also discusses the experiments performed for face recognition using the features extracted from images in the AT & T face database. Artificial Neural Network and k- Nearest Neighbour classifier algorithms are employed for the recognition purpose. A feature vector is formed for each of the face images in the database and recognition accuracies are computed and compared using the classifiers. Simulation results show that the proposed method outperforms traditional way of feature extraction methods prevailing for pattern recognition in terms of recognition accuracy for face images with pose and illumination variations.

Keywords: Face recognition, Image analysis, Wavelet feature extraction, Pattern recognition, Classifier algorithms

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1672
5503 Influence of Radio Frequency Identification Technology in Logistic, Inventory Control and Supply Chain Optimization

Authors: H. Amoozad-khalili, R. Tavakkoli-Moghaddam, N.Shahab-Dehkordi

Abstract:

The main aim of Supply Chain Management (SCM) is to produce, distribute, logistics and deliver goods and equipment in right location, right time, right amount to satisfy costumers, with minimum time and cost waste. So implementing techniques that reduce project time and cost, and improve productivity and performance is very important. Emerging technologies such as the Radio Frequency Identification (RFID) are now making it possible to automate supply chains in a real time manner and making them more efficient than the simple supply chain of the past for tracing and monitoring goods and products and capturing data on movements of goods and other events. This paper considers concepts, components and RFID technology characteristics by concentration of warehouse and inventories management. Additionally, utilization of RFID in the role of improving information management in supply chain is discussed. Finally, the facts of installation and this technology-s results in direction with warehouse and inventory management and business development will be presented.

Keywords: Logistics, Supply Chain Management, RFIDTechnology, Inventory Control.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1828
5502 Evolutionary Approach for Automated Discovery of Censored Production Rules

Authors: Kamal K. Bharadwaj, Basheer M. Al-Maqaleh

Abstract:

In the recent past, there has been an increasing interest in applying evolutionary methods to Knowledge Discovery in Databases (KDD) and a number of successful applications of Genetic Algorithms (GA) and Genetic Programming (GP) to KDD have been demonstrated. The most predominant representation of the discovered knowledge is the standard Production Rules (PRs) in the form If P Then D. The PRs, however, are unable to handle exceptions and do not exhibit variable precision. The Censored Production Rules (CPRs), an extension of PRs, were proposed by Michalski & Winston that exhibit variable precision and supports an efficient mechanism for handling exceptions. A CPR is an augmented production rule of the form: If P Then D Unless C, where C (Censor) is an exception to the rule. Such rules are employed in situations, in which the conditional statement 'If P Then D' holds frequently and the assertion C holds rarely. By using a rule of this type we are free to ignore the exception conditions, when the resources needed to establish its presence are tight or there is simply no information available as to whether it holds or not. Thus, the 'If P Then D' part of the CPR expresses important information, while the Unless C part acts only as a switch and changes the polarity of D to ~D. This paper presents a classification algorithm based on evolutionary approach that discovers comprehensible rules with exceptions in the form of CPRs. The proposed approach has flexible chromosome encoding, where each chromosome corresponds to a CPR. Appropriate genetic operators are suggested and a fitness function is proposed that incorporates the basic constraints on CPRs. Experimental results are presented to demonstrate the performance of the proposed algorithm.

Keywords: Censored Production Rule, Data Mining, MachineLearning, Evolutionary Algorithms.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1876
5501 Hi-Fi Traffic Clearance Technique for Life Saving Vehicles using Differential GPS System

Authors: N. Yuvaraj, V. B. Prakash, D. Venkatraj

Abstract:

This paper may be considered as combination of both pervasive computing and Differential GPS (global positioning satellite) which relates to control automatic traffic signals in such a way as to pre-empt normal signal operation and permit lifesaving vehicles. Before knowing the arrival of the lifesaving vehicles from the signal there is a chance of clearing the traffic. Traffic signal preemption system includes a vehicle equipped with onboard computer system capable of capturing diagnostic information and estimated location of the lifesaving vehicle using the information provided by GPS receiver connected to the onboard computer system and transmitting the information-s using a wireless transmitter via a wireless network. The fleet management system connected to a wireless receiver is capable of receiving the information transmitted by the lifesaving vehicle .A computer is also located at the intersection uses corrected vehicle position, speed & direction measurements, in conjunction with previously recorded data defining approach routes to the intersection, to determine the optimum time to switch a traffic light controller to preemption mode so that lifesaving vehicles can pass safely. In case when the ambulance need to take a “U" turn in a heavy traffic area we suggest a solution. Now we are going to make use of computerized median which uses LINKED BLOCKS (removable) to solve the above problem.

Keywords: Ubiquitous computing, differential GPS, fleet management system, wireless transmitter and receiver computerized median i.e. linked blocks (removable).

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1986
5500 A Multilevel Comparative Assessment Approach to International Services Trade Competitiveness: The Case of Romania and Bulgaria

Authors: Ana Bobirca, Paul-Gabriel Miclaus

Abstract:

International competitiveness receives much attention nowadays, but up to now its assessment has been heavily based on manufacturing industry statistics. This paper addresses the need for competitiveness indicators that cover the service sector and sets out a multilevel framework for measuring international services trade competitiveness. The approach undertaken here aims at comparatively examining the international competitiveness of the EU-25 (the twenty-five European Union member states before the 1st of January 2007), Romanian and Bulgarian services trade, as well as the last two countries- structure of specialization on the EU-25 services market. The primary changes in the international competitiveness of three major services sectors – transportation, travel and other services - are analyzed. This research attempts to determine the ability of the two recent European Union (EU) member states to contend with the challenges that might arise from the hard competition within the enlarged EU, in the field of services trade.

Keywords: Bulgaria, EU-25, international competitiveness, international services trade, Romania.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2010
5499 Model-Based Small Area Estimation with Application to Unemployment Estimates

Authors: Hichem Omrani, Philippe Gerber, Patrick Bousch

Abstract:

The problem of Small Area Estimation (SAE) is complex because of various information sources and insufficient data. In this paper, an approach for SAE is presented for decision-making at national, regional and local level. We propose an Empirical Best Linear Unbiased Predictor (EBLUP) as an estimator in order to combine several information sources to evaluate various indicators. First, we present the urban audit project and its environmental, social and economic indicators. Secondly, we propose an approach for decision making in order to estimate indicators. An application is used to validate the theoretical proposal. Finally, a decision support system is presented based on open-source environment.

Keywords: Small area estimation, statistical method, sampling, empirical best linear unbiased predictor (EBLUP), decision-making.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1702
5498 Feature Reduction of Nearest Neighbor Classifiers using Genetic Algorithm

Authors: M. Analoui, M. Fadavi Amiri

Abstract:

The design of a pattern classifier includes an attempt to select, among a set of possible features, a minimum subset of weakly correlated features that better discriminate the pattern classes. This is usually a difficult task in practice, normally requiring the application of heuristic knowledge about the specific problem domain. The selection and quality of the features representing each pattern have a considerable bearing on the success of subsequent pattern classification. Feature extraction is the process of deriving new features from the original features in order to reduce the cost of feature measurement, increase classifier efficiency, and allow higher classification accuracy. Many current feature extraction techniques involve linear transformations of the original pattern vectors to new vectors of lower dimensionality. While this is useful for data visualization and increasing classification efficiency, it does not necessarily reduce the number of features that must be measured since each new feature may be a linear combination of all of the features in the original pattern vector. In this paper a new approach is presented to feature extraction in which feature selection, feature extraction, and classifier training are performed simultaneously using a genetic algorithm. In this approach each feature value is first normalized by a linear equation, then scaled by the associated weight prior to training, testing, and classification. A knn classifier is used to evaluate each set of feature weights. The genetic algorithm optimizes a vector of feature weights, which are used to scale the individual features in the original pattern vectors in either a linear or a nonlinear fashion. By this approach, the number of features used in classifying can be finely reduced.

Keywords: Feature reduction, genetic algorithm, pattern classification, nearest neighbor rule classifiers (k-NNR).

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1759
5497 Split-Pipe Design of Water Distribution Networks Using a Combination of Tabu Search and Genetic Algorithm

Authors: J. Tospornsampan, I. Kita, M. Ishii, Y. Kitamura

Abstract:

In this paper a combination approach of two heuristic-based algorithms: genetic algorithm and tabu search is proposed. It has been developed to obtain the least cost based on the split-pipe design of looped water distribution network. The proposed combination algorithm has been applied to solve the three well-known water distribution networks taken from the literature. The development of the combination of these two heuristic-based algorithms for optimization is aimed at enhancing their strengths and compensating their weaknesses. Tabu search is rather systematic and deterministic that uses adaptive memory in search process, while genetic algorithm is probabilistic and stochastic optimization technique in which the solution space is explored by generating candidate solutions. Split-pipe design may not be realistic in practice but in optimization purpose, optimal solutions are always achieved with split-pipe design. The solutions obtained in this study have proved that the least cost solutions obtained from the split-pipe design are always better than those obtained from the single pipe design. The results obtained from the combination approach show its ability and effectiveness to solve combinatorial optimization problems. The solutions obtained are very satisfactory and high quality in which the solutions of two networks are found to be the lowest-cost solutions yet presented in the literature. The concept of combination approach proposed in this study is expected to contribute some useful benefits in diverse problems.

Keywords: GAs, Heuristics, Looped network, Least-cost design, Pipe network, Optimization, TS

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1782
5496 Embedding the Dimensions of Sustainability into City Information Modelling

Authors: Ali M. Al-Shaery

Abstract:

The purpose of this paper is to address the functions of sustainability dimensions in city information modelling and to present the required sustainability criteria that support establishing a sustainable planning framework for enhancing existing cities and developing future smart cities. The paper is divided into two sections. The first section is based on the examination of a wide and extensive array of cross-disciplinary literature in the last decade and a half to conceptualize the terms ‘sustainable’ and ‘smart city’, and map their associated criteria to city information modelling. The second section is based on analyzing two approaches relating to city information modelling, namely statistical and dynamic approaches, and their suitability in the development of cities’ action plans. The paper argues that the use of statistical approaches to embed sustainability dimensions in city information modelling have limited value. Despite the popularity of such approaches in addressing other dimensions like utility and service management in development and action plans of the world cities, these approaches are unable to address the dynamics across various city sectors with regards to economic, environmental and social criteria. The paper suggests an integrative dynamic and cross-disciplinary planning approach to embedding sustainability dimensions in city information modelling frameworks. Such an approach will pave the way towards optimal planning and implementation of priority actions of projects and investments. The approach can be used to achieve three main goals: (1) better development and action plans for world cities (2) serve the development of an integrative dynamic and cross-disciplinary framework that incorporates economic, environmental and social sustainability criteria and (3) address areas that require further attention in the development of future sustainable and smart cities. The paper presents an innovative approach for city information modelling and a well-argued, balanced hierarchy of sustainability criteria that can contribute to an area of research which is still in its infancy in terms of development and management.

Keywords: Information modelling, smart city, sustainable city, sustainability dimensions, sustainability criteria, city development planning.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1165
5495 Long-Term Economic-Ecological Assessment of Optimal Local Heat-Generating Technologies for the German Unrefurbished Residential Building Stock on the Quarter Level

Authors: M. A. Spielmann, L. Schebek

Abstract:

In order to reach the long-term national climate goals of the German government for the building sector, substantial energetic measures have to be executed. Historically, those measures were primarily energetic efficiency measures at the buildings’ shells. Advanced technologies for the on-site generation of heat (or other types of energy) often are not feasible at this small spatial scale of a single building. Therefore, the present approach uses the spatially larger dimension of a quarter. The main focus of the present paper is the long-term economic-ecological assessment of available decentralized heat-generating (CHP power plants and electrical heat pumps) technologies at the quarter level for the German unrefurbished residential buildings. Three distinct terms have to be described methodologically: i) Quarter approach, ii) Economic assessment, iii) Ecological assessment. The quarter approach is used to enable synergies and scaling effects over a single-building. For the present study, generic quarters that are differentiated according to significant parameters concerning their heat demand are used. The core differentiation of those quarters is made by the construction time period of the buildings. The economic assessment as the second crucial parameter is executed with the following structure: Full costs are quantized for each technology combination and quarter. The investment costs are analyzed on an annual basis and are modeled with the acquisition of debt. Annuity loans are assumed. Consequently, for each generic quarter, an optimal technology combination for decentralized heat generation is provided in each year of the temporal boundaries (2016-2050). The ecological assessment elaborates for each technology combination and each quarter a Life Cycle assessment. The measured impact category hereby is GWP 100. The technology combinations for heat production can be therefore compared against each other concerning their long-term climatic impacts. Core results of the approach can be differentiated to an economic and ecological dimension. With an annual resolution, the investment and running costs of different energetic technology combinations are quantified. For each quarter an optimal technology combination for local heat supply and/or energetic refurbishment of the buildings within the quarter is provided. Coherently to the economic assessment, the climatic impacts of the technology combinations are quantized and compared against each other.

Keywords: Building sector, heat, LCA, quarter level, systemic approach.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 941
5494 A Hybrid Approach for Color Image Quantization Using K-means and Firefly Algorithms

Authors: Parisut Jitpakdee, Pakinee Aimmanee, Bunyarit Uyyanonvara

Abstract:

Color Image quantization (CQ) is an important problem in computer graphics, image and processing. The aim of quantization is to reduce colors in an image with minimum distortion. Clustering is a widely used technique for color quantization; all colors in an image are grouped to small clusters. In this paper, we proposed a new hybrid approach for color quantization using firefly algorithm (FA) and K-means algorithm. Firefly algorithm is a swarmbased algorithm that can be used for solving optimization problems. The proposed method can overcome the drawbacks of both algorithms such as the local optima converge problem in K-means and the early converge of firefly algorithm. Experiments on three commonly used images and the comparison results shows that the proposed algorithm surpasses both the base-line technique k-means clustering and original firefly algorithm.

Keywords: Clustering, Color quantization, Firefly algorithm, Kmeans.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2215
5493 Fault-Tolerant Optimal Broadcast Algorithm for the Hypercube Topology

Authors: Lokendra Singh Umrao, Ravi Shankar Singh

Abstract:

This paper presents an optimal broadcast algorithm for the hypercube networks. The main focus of the paper is the effectiveness of the algorithm in the presence of many node faults. For the optimal solution, our algorithm builds with spanning tree connecting the all nodes of the networks, through which messages are propagated from source node to remaining nodes. At any given time, maximum n − 1 nodes may fail due to crashing. We show that the hypercube networks are strongly fault-tolerant. Simulation results analyze to accomplish algorithm characteristics under many node faults. We have compared our simulation results between our proposed method and the Fu’s method. Fu’s approach cannot tolerate n − 1 faulty nodes in the worst case, but our approach can tolerate n − 1 faulty nodes.

Keywords: Fault tolerance, hypercube, broadcasting, link/node faults, routing.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1874
5492 Communicative and Artistic Machines: A Survey of Models and Experiments on Artificial Agents

Authors: Artur Matuck, Guilherme F. Nobre

Abstract:

Machines can be either tool, media, or social agents. Advances in technology have been delivering machines capable of autonomous expression, both through communication and art. This paper deals with models (theoretical approach) and experiments (applied approach) related to artificial agents. On one hand it traces how social sciences' scholars have worked with topics such as text automatization, man-machine writing cooperation, and communication. On the other hand it covers how computer sciences' scholars have built communicative and artistic machines, including the programming of creativity. The aim is to present a brief survey on artificially intelligent communicators and artificially creative writers, and provide the basis to understand the meta-authorship and also to new and further man-machine co-authorship.

Keywords: Artificial communication, artificial creativity, artificial writers, meta-authorship, robotic art.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1307
5491 Probability Distribution of Rainfall Depth at Hourly Time-Scale

Authors: S. Dan'azumi, S. Shamsudin, A. A. Rahman

Abstract:

Rainfall data at fine resolution and knowledge of its characteristics plays a major role in the efficient design and operation of agricultural, telecommunication, runoff and erosion control as well as water quality control systems. The paper is aimed to study the statistical distribution of hourly rainfall depth for 12 representative stations spread across Peninsular Malaysia. Hourly rainfall data of 10 to 22 years period were collected and its statistical characteristics were estimated. Three probability distributions namely, Generalized Pareto, Exponential and Gamma distributions were proposed to model the hourly rainfall depth, and three goodness-of-fit tests, namely, Kolmogorov-Sminov, Anderson-Darling and Chi-Squared tests were used to evaluate their fitness. Result indicates that the east cost of the Peninsular receives higher depth of rainfall as compared to west coast. However, the rainfall frequency is found to be irregular. Also result from the goodness-of-fit tests show that all the three models fit the rainfall data at 1% level of significance. However, Generalized Pareto fits better than Exponential and Gamma distributions and is therefore recommended as the best fit.

Keywords: Goodness-of-fit test, Hourly rainfall, Malaysia, Probability distribution.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2913
5490 Reduction of False Positives in Head-Shoulder Detection Based on Multi-Part Color Segmentation

Authors: Lae-Jeong Park

Abstract:

The paper presents a method that utilizes figure-ground color segmentation to extract effective global feature in terms of false positive reduction in the head-shoulder detection. Conventional detectors that rely on local features such as HOG due to real-time operation suffer from false positives. Color cue in an input image provides salient information on a global characteristic which is necessary to alleviate the false positives of the local feature based detectors. An effective approach that uses figure-ground color segmentation has been presented in an effort to reduce the false positives in object detection. In this paper, an extended version of the approach is presented that adopts separate multipart foregrounds instead of a single prior foreground and performs the figure-ground color segmentation with each of the foregrounds. The multipart foregrounds include the parts of the head-shoulder shape and additional auxiliary foregrounds being optimized by a search algorithm. A classifier is constructed with the feature that consists of a set of the multiple resulting segmentations. Experimental results show that the presented method can discriminate more false positive than the single prior shape-based classifier as well as detectors with the local features. The improvement is possible because the presented approach can reduce the false positives that have the same colors in the head and shoulder foregrounds.

Keywords: Pedestrian detection, color segmentation, false positives, feature extraction.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1139
5489 An Interactive Ontology Visualization Approach for the Networked Home Environment

Authors: Ilkka Niskanen, Jarmo Kalaoja, Julia Kantorovitch, Toni Piirainen

Abstract:

Ontologies are broadly used in the context of networked home environments. With ontologies it is possible to define and store context information, as well as to model different kinds of physical environments. Ontologies are central to networked home environments as they carry the meaning. However, ontologies and the OWL language is complex. Several ontology visualization approaches have been developed to enhance the understanding of ontologies. The domain of networked home environments sets some special requirements for the ontology visualization approach. The visualization tool presented here, visualizes ontologies in a domain-specific way. It represents effectively the physical structures and spatial relationships of networked home environments. In addition, it provides extensive interaction possibilities for editing and manipulating the visualization. The tool shortens the gap from beginner to intermediate OWL ontology reader by visualizing instances in their actual locations and making OWL ontologies more interesting and concrete, and above all easier to comprehend.

Keywords: Ontologies, visualization, interaction.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1370
5488 Hearing Aids Maintenance Training for Hearing-Impaired Preschool Children with the Help of Motion Graphic Tools

Authors: M. Mokhtarzadeh, M. Taheri Qomi, M. Nikafrooz, A. Atashafrooz

Abstract:

The purpose of the present study was to investigate the effectiveness of using motion graphics as a learning medium on training hearing aids maintenance skills to hearing-impaired children. The statistical population of this study consisted of all children with hearing loss in Ahvaz city, at age 4 to 7 years old. As the sample, 60, whom were selected by multistage random sampling, were randomly assigned to two groups; experimental (30 children) and control (30 children) groups. The research method was experimental and the design was pretest-posttest with the control group. The intervention consisted of a 2-minute motion graphics clip to train hearing aids maintenance skills. Data were collected using a 9-question researcher-made questionnaire. The data were analyzed by using one-way analysis of covariance. Results showed that the training of hearing aids maintenance skills with motion graphics was significantly effective for those children. The results of this study can be used by educators, teachers, professionals, and parents to train children with disabilities or normal students.

Keywords: Hearing-impaired children, hearing aids, hearing aids maintenance skill, and motion graphics.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 562
5487 Index t-SNE: Tracking Dynamics of High-Dimensional Datasets with Coherent Embeddings

Authors: G. Candel, D. Naccache

Abstract:

t-SNE is an embedding method that the data science community has widely used. It helps two main tasks: to display results by coloring items according to the item class or feature value; and for forensic, giving a first overview of the dataset distribution. Two interesting characteristics of t-SNE are the structure preservation property and the answer to the crowding problem, where all neighbors in high dimensional space cannot be represented correctly in low dimensional space. t-SNE preserves the local neighborhood, and similar items are nicely spaced by adjusting to the local density. These two characteristics produce a meaningful representation, where the cluster area is proportional to its size in number, and relationships between clusters are materialized by closeness on the embedding. This algorithm is non-parametric. The transformation from a high to low dimensional space is described but not learned. Two initializations of the algorithm would lead to two different embedding. In a forensic approach, analysts would like to compare two or more datasets using their embedding. A naive approach would be to embed all datasets together. However, this process is costly as the complexity of t-SNE is quadratic, and would be infeasible for too many datasets. Another approach would be to learn a parametric model over an embedding built with a subset of data. While this approach is highly scalable, points could be mapped at the same exact position, making them indistinguishable. This type of model would be unable to adapt to new outliers nor concept drift. This paper presents a methodology to reuse an embedding to create a new one, where cluster positions are preserved. The optimization process minimizes two costs, one relative to the embedding shape and the second relative to the support embedding’ match. The embedding with the support process can be repeated more than once, with the newly obtained embedding. The successive embedding can be used to study the impact of one variable over the dataset distribution or monitor changes over time. This method has the same complexity as t-SNE per embedding, and memory requirements are only doubled. For a dataset of n elements sorted and split into k subsets, the total embedding complexity would be reduced from O(n2) to O(n2/k), and the memory requirement from n2 to 2(n/k)2 which enables computation on recent laptops. The method showed promising results on a real-world dataset, allowing to observe the birth, evolution and death of clusters. The proposed approach facilitates identifying significant trends and changes, which empowers the monitoring high dimensional datasets’ dynamics.

Keywords: Concept drift, data visualization, dimension reduction, embedding, monitoring, reusability, t-SNE, unsupervised learning.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 483
5486 Designing Creative Events with Deconstructivism Approach

Authors: Maryam Memarian, Mahmood Naghizadeh

Abstract:

Deconstruction is an approach that is entirely incompatible with the traditional prevalent architecture. Considering the fact that this approach attempts to put architecture in sharp contrast with its opposite events and transpires with attending to the neglected and missing aspects of architecture and deconstructing its stable structures. It also recklessly proceeds beyond the existing frameworks and intends to create a different and more efficient prospect for space. The aim of deconstruction architecture is to satisfy both the prospective and retrospective visions as well as takes into account all tastes of the present in order to transcend time. Likewise, it ventures to fragment the facts and symbols of the past and extract new concepts from within their heart, which coincide with today’s circumstances. Since this approach is an attempt to surpass the limits of the prevalent architecture, it can be employed to design places in which creative events occur and imagination and ambition flourish. Thought-provoking artistic events can grow and mature in such places and be represented in the best way possible to all people. The concept of event proposed in the plan grows out of the interaction between space and creation. In addition to triggering surprise and high impressions, it is also considered as a bold journey into the suspended realms of the traditional conflicts in architecture such as architecture-landscape, interior-exterior, center-margin, product-process, and stability-instability. In this project, at first, through interpretive-historical research method and examining the inputs and data collection, recognition and organizing takes place. After evaluating the obtained data using deductive reasoning, the data is eventually interpreted. Given the fact that the research topic is in its infancy and there is not a similar case in Iran with limited number of corresponding instances across the world, the selected topic helps to shed lights on the unrevealed and neglected parts in architecture. Similarly, criticizing, investigating and comparing specific and highly prized cases in other countries with the project under study can serve as an introduction into this architecture style.

Keywords: Creativity, deconstruction, event.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1894
5485 Effect of Gibberellic Acid and 2,4- Dichlorophenoxyacetic Acid on Fruit Development and Fruit Quality of Wax Apple

Authors: Nguyen Minh Tuan, Yen Chung–Ruey

Abstract:

This study was conducted to evaluate the effects of gibberellic acid and 2,4- dichlorophenoxyacetic acid on flower number, fruit growth and fruit quality of wax apple. GA3 and 2,4-D were applied at small bud and petal fall stage. Number of flower, fruit set, fruit drop, fruit crack, fruit growth and fruit quality were recorded. Results indicated that spraying with 10 ppm GA3 had the best results in number of flower. GA3 spray at 30 ppm gave the faster rate of fruit growth than the other treatments. Fruit set, fruit size as well as fruit weight markedly improved by spraying 30 ppm GA3, followed by 10 ppm GA3 compared to untreated control. Moreover, spray GA3 at 30 ppm was the most effective and increased total soluble solids, reduced titratable acidity and fruit drop. On the other hand, it was noticed that with 10 ppm 2,4-D application also enhanced the fruit growth rate, improved physiological and biochemical characters of fruit compared to untreated control. It was concluded that both GA3 and 2,4-D spray have positive effects on fruit development, reduced fruit drop, fruit crack and improved fruit quality of wax apple under field conditions.

Keywords: Wax apple, GA3, 2, 4-D, fruit growth, fruit quality.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5439
5484 Attention-Based Spatio-Temporal Approach for Fire and Smoke Detection

Authors: A. Mirrashid, M. Khoshbin, A. Atghaei, H. Shahbazi

Abstract:

In various industries, smoke and fire are two of the most important threats in the workplace. One of the common methods for detecting smoke and fire is the use of infrared thermal and smoke sensors, which cannot be used in outdoor applications. Therefore, the use of vision-based methods seems necessary. The problem of smoke and fire detection is spatiotemporal and requires spatiotemporal solutions. This paper presents a method that uses spatial features along with temporal-based features to detect smoke and fire in the scene. It consists of three main parts; the task of each part is to reduce the error of the previous part so that the final model has a robust performance. This method also uses transformer modules to increase the accuracy of the model. The results of our model show the proper performance of the proposed approach in solving the problem of smoke and fire detection and can be used to increase workplace safety.

Keywords: Attention, fire detection, smoke detection, spatiotemporal.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 337
5483 Verification of a Locked CFD Approach to Cool Down Modeling

Authors: P. Bárta

Abstract:

Increasing demand on the performance of Subsea Production Systems (SPS) suggests a need for more detailed investigation of fluid behavior taking place in subsea equipment. Complete CFD cool down analyses of subsea equipment are very time demanding. The objective of this paper is to investigate a Locked CFD approach, which enables significant reduction of the computational time and at the same time maintains sufficient accuracy during thermal cool down simulations. The result comparison of a dead leg simulation using the Full CFD and the three LCFD-methods confirms the validity of the locked flow field assumption for the selected case. For the tested case the LCFD simulation speed up by factor of 200 results in the absolute thermal error of 0.5 °C (3% relative error), speed up by factor of 10 keeps the LCFD results within 0.1 °C (0.5 % relative error) comparing to the Full CFD.

Keywords: CFD, Locked Flow Field, Speed up of CFD simulation time, Subsea

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1669
5482 On Measuring the Reusability Proneness of Mobile Applications

Authors: Fathi Taibi

Abstract:

The abnormal increase in the number of applications available for download in Android markets is a good indication that they are being reused. However, little is known about their real reusability potential. A considerable amount of these applications is reported as having a poor quality or being malicious. Hence, in this paper, an approach to measure the reusability potential of classes in Android applications is proposed. The approach is not meant specifically for this particular type of applications. Rather, it is intended for Object-Oriented (OO) software systems in general and aims also to provide means to discard the classes of low quality and defect prone applications from being reused directly through inheritance and instantiation. An empirical investigation is conducted to measure and rank the reusability potential of the classes of randomly selected Android applications. The results obtained are thoroughly analyzed in order to understand the extent of this potential and the factors influencing it.

Keywords: Reusability, Software Quality Factors, Software Metrics, Empirical Investigation, Object-Oriented Software, Android Applications.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1793
5481 Calibrations and Effect of Different Operating Conditions on the Performance of a Fluid Power Control System with Servo Solenoid Valve

Authors: Tahany W. Sadak, Fouly, A. Anwer, M. Rizk

Abstract:

The current investigation presents a study on the hydraulic performance of an electro-hydraulic servo solenoid valve controlled linear piston used in hydraulic systems. Advanced methods have been used to measure and record laboratory experiments, to ensure accurate analysis and evaluation. Experiments have been conducted under different values of temperature (28, 40 and 50 °C), supply pressure (10, 20, 30, 40 and 50 bar), system stiffness (32 N/mm), and load (0.0 & 5560 N). It is concluded that increasing temperature of hydraulic oil increases the quantity of flow rate, so it achieves an increase of the quantity of flow by 5.75 % up to 48.8 % depending on operating conditions. The values of pressure decay at low temperature are less than the values at high temperature. The frequency increases with the increase of the temperature. When we connect the springs to the system, it decreases system frequency. These results are very useful in the process of packing and manufacturing of fluid products, where the properties are not affected by 50 °C, so energy and time are saved.

Keywords: Electro Hydraulic Servo Valve, fluid power control system, system stiffness, static and dynamic performance.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 676
5480 Low-Cost Space-Based Geoengineering: An Assessment Based on Self-Replicating Manufacturing of in-Situ Resources on the Moon

Authors: Alex Ellery

Abstract:

Geoengineering approaches to climate change mitigation are unpopular and regarded with suspicion. Of these, space-based approaches are regarded as unworkable and enormously costly. Here, a space-based approach is presented that is modest in cost, fully controllable and reversible, and acts as a natural spur to the development of solar power satellites over the longer term as a clean source of energy. The low-cost approach exploits self-replication technology which it is proposed may be enabled by 3D printing technology. Self-replication of 3D printing platforms will enable mass production of simple spacecraft units. Key elements being developed are 3D-printable electric motors and 3D-printable vacuum tube-based electronics. The power of such technologies will open up enormous possibilities at low cost including space-based geoengineering.

Keywords: 3D printing, in-situ resource utilization, self-replication technology, space-based geoengineering.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1094
5479 Loop Back Connected Component Labeling Algorithm and Its Implementation in Detecting Face

Authors: A. Rakhmadi, M. S. M. Rahim, A. Bade, H. Haron, I. M. Amin

Abstract:

In this study, a Loop Back Algorithm for component connected labeling for detecting objects in a digital image is presented. The approach is using loop back connected component labeling algorithm that helps the system to distinguish the object detected according to their label. Deferent than whole window scanning technique, this technique reduces the searching time for locating the object by focusing on the suspected object based on certain features defined. In this study, the approach was also implemented for a face detection system. Face detection system is becoming interesting research since there are many devices or systems that require detecting the face for certain purposes. The input can be from still image or videos, therefore the sub process of this system has to be simple, efficient and accurate to give a good result.

Keywords: Image processing, connected components labeling, face detection.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2291
5478 Performance Evaluation of Intelligent Controllers for AGC in Thermal Systems

Authors: Muhammad Muhsin, Abhishek Mishra, Shreyansh Vishwakarma, K. Dasaratha Babu, Anudevi Samuel

Abstract:

In an interconnected power system, any sudden small load perturbation in any of the interconnected areas causes the deviation of the area frequencies, the tie line power and voltage deviation at the generator terminals. This paper deals with the study of performance of intelligent Fuzzy Logic controllers coupled with Conventional Controllers (PI and PID) for Load Frequency Control. For analysis, an isolated single area and interconnected two area thermal power systems with and without generation rate constraints (GRC) have been considered. The studies have been performed with conventional PI and PID controllers and their performance has been compared with intelligent fuzzy controllers. It can be demonstrated that these controllers can successfully bring back the excursions in area frequencies and tie line powers within acceptable limits in smaller time periods and with lesser transients as compared to the performance of conventional controllers under same load disturbance conditions. The simulations in MATLAB have been used for comparative studies.

Keywords: Area Control Error, Fuzzy Logic, Generation rate constraint, Load Frequency, Tie line Power.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2454
5477 Preliminary Design of Frozen Soil Simulation System Based on Finite Element Simulation

Authors: Wenyu Song, Bingxi Li, Zhongbin Fu, Baocheng Jiang

Abstract:

Full - Scale Accelerated Loading System, one part of “the Eleventh - Five - Year National Grand Technology Infrastructure Program" is a facility to evaluate the performance and service life of different kinds of pavements subjected to traffic loading under full - controlled environment. While simulating the environments of frigid zone and permafrost zone, the accurate control of air temperature, road temperature and roadbed temperature are the key points and also aporias for the designment. In this paper, numerical simulations are used to determine the design parameters of the frozen soil simulation system. At first, a brief introduction of the Full - Scale Accelerate Loading System was given. Then, the temperature control method of frozen soil simulation system was proposed. Finally, by using finite element simulations, the optimal design of frozen soil simulation system was obtained. This proposed design, which was obtained by finite element simulations, provided significant referents to the ultimate design of the environment simulation system.

Keywords: China, finite element simulation, frozen soilsimulation system, preliminary design.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1576
5476 Increase Energy Savings with Lighting Automation Using Light Pipes and Power LEDs

Authors: İ. Kıyak, G. Gökmen

Abstract:

Using of natural lighting has come into prominence in constructed buildings, especially in last ten years, under scope of energy efficiency. Natural lighting methods are one of the methods that aim to take advantage of day light in maximum level and decrease using of artificial lighting. Increasing of day light amount in buildings by using suitable methods will give optimum result in terms of comfort and energy saving when the daylight-artificial light integration is ensured with a suitable control system. Using of natural light in places that require lighting will ensure energy saving in great extent. With this study, it is aimed to save energy used for purpose of lighting. Under this scope, lighting of a scanning laboratory of a hospital was realized by using a lighting automation containing natural and artificial lighting. In natural lighting, light pipes were used and in artificial lighting, dimmable power LED modules were used. Necessity of lighting was followed with motion sensors. The lighting automation containing natural and artificial light was ensured with fuzzy logic control. At the scanning laboratory where this application was realized, energy saving in lighting was obtained.

Keywords: Daylight transfer, fuzzy logic controller, light pipe, Power LED.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2148