Search results for: N priority classes
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 587

Search results for: N priority classes

167 Implementation of Student-Centered Learning Approach in Building Surveying Course

Authors: Amal A. Abdel-Sattar

Abstract:

The curriculum of architecture department in Prince Sultan University includes ‘Building Surveying’ course which is usually a part of civil engineering courses. As a fundamental requirement of the course, it requires a strong background in mathematics and physics, which are not usually preferred subjects to the architecture students and many of them are not giving the required and necessary attention to these courses during their preparation year before commencing their architectural study. This paper introduces the concept and the methodology of the student-centered learning approach in the course of building surveying for architects. One of the major outcomes is the improvement in the students’ involvement in the course and how this will cover and strength their analytical weak points and improve their mathematical skills. The study is conducted through three semesters with a total number of 99 students. The effectiveness of the student-centered learning approach is studied using the student survey at the end of each semester and teacher observations. This survey showed great acceptance of the students for these methods. Also, the teachers observed a great improvement in the students’ mathematical abilities and how keener they became in attending the classes which were clearly reflected on the low absence record.

Keywords: Architecture, building surveying, student-centered learning, teaching, and learning.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1283
166 Swarmed Discriminant Analysis for Multifunction Prosthesis Control

Authors: Rami N. Khushaba, Ahmed Al-Ani, Adel Al-Jumaily

Abstract:

One of the approaches enabling people with amputated limbs to establish some sort of interface with the real world includes the utilization of the myoelectric signal (MES) from the remaining muscles of those limbs. The MES can be used as a control input to a multifunction prosthetic device. In this control scheme, known as the myoelectric control, a pattern recognition approach is usually utilized to discriminate between the MES signals that belong to different classes of the forearm movements. Since the MES is recorded using multiple channels, the feature vector size can become very large. In order to reduce the computational cost and enhance the generalization capability of the classifier, a dimensionality reduction method is needed to identify an informative yet moderate size feature set. This paper proposes a new fuzzy version of the well known Fisher-s Linear Discriminant Analysis (LDA) feature projection technique. Furthermore, based on the fact that certain muscles might contribute more to the discrimination process, a novel feature weighting scheme is also presented by employing Particle Swarm Optimization (PSO) for estimating the weight of each feature. The new method, called PSOFLDA, is tested on real MES datasets and compared with other techniques to prove its superiority.

Keywords: Discriminant Analysis, Pattern Recognition, SignalProcessing.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1556
165 Platform-as-a-Service Sticky Policies for Privacy Classification in the Cloud

Authors: Maha Shamseddine, Amjad Nusayr, Wassim Itani

Abstract:

In this paper, we present a Platform-as-a-Service (PaaS) model for controlling the privacy enforcement mechanisms applied on user data when stored and processed in Cloud data centers. The proposed architecture consists of establishing user configurable ‘sticky’ policies on the Graphical User Interface (GUI) data-bound components during the application development phase to specify the details of privacy enforcement on the contents of these components. Various privacy classification classes on the data components are formally defined to give the user full control on the degree and scope of privacy enforcement including the type of execution containers to process the data in the Cloud. This not only enhances the privacy-awareness of the developed Cloud services, but also results in major savings in performance and energy efficiency due to the fact that the privacy mechanisms are solely applied on sensitive data units and not on all the user content. The proposed design is implemented in a real PaaS cloud computing environment on the Microsoft Azure platform.

Keywords: Privacy enforcement, Platform-as-a-Service privacy awareness, cloud computing privacy.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 759
164 Enhancing Performance of Bluetooth Piconets Using Priority Scheduling and Exponential Back-Off Mechanism

Authors: Dharmendra Chourishi “Maitraya”, Sridevi Seshadri

Abstract:

Bluetooth is a personal wireless communication technology and is being applied in many scenarios. It is an emerging standard for short range, low cost, low power wireless access technology. Current existing MAC (Medium Access Control) scheduling schemes only provide best-effort service for all masterslave connections. It is very challenging to provide QoS (Quality of Service) support for different connections due to the feature of Master Driven TDD (Time Division Duplex). However, there is no solution available to support both delay and bandwidth guarantees required by real time applications. This paper addresses the issue of how to enhance QoS support in a Bluetooth piconet. The Bluetooth specification proposes a Round Robin scheduler as possible solution for scheduling the transmissions in a Bluetooth Piconet. We propose an algorithm which will reduce the bandwidth waste and enhance the efficiency of network. We define token counters to estimate traffic of real-time slaves. To increase bandwidth utilization, a back-off mechanism is then presented for best-effort slaves to decrease the frequency of polling idle slaves. Simulation results demonstrate that our scheme achieves better performance over the Round Robin scheduling.

Keywords: Piconet, Medium Access Control, Polling algorithm, Scheduling, QoS, Time Division Duplex (TDD).

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1700
163 Dialogue Journals as an EFL Learning Strategy in the Preparatory Year Program: Learners' Attitudes and Perceptions

Authors: Asma Alyahya

Abstract:

This study attempts to elicit the perceptions and attitudes of EFL learners of the Preparatory Year Program at KSU towards dialogue journal writing as an EFL learning strategy. The descriptive research design used incorporated both qualitative and quantitative instruments to accomplish the objectives of the study. A learners’ attitude questionnaire and follow-up interviewswith learners from a randomly selected representative sample of the participants were employed. The participants were 55 female Saudi university students in the Preparatory Year Program at King Saud University. The analysis of the results indicated that the PYP learners had highly positive attitudes towards dialogue journal writing in their EFL classes and positive perceptions of the benefits of the use of dialogue journal writing as an EFL learning strategy. The results also revealed that dialogue journals are considered an effective EFL learning strategy since they fulfill various needs for both learners and instructors. Interestingly, the analysis of the results also revealed that Saudi university level students tend to write about personal topics in their dialogue journals more than academic ones.

Keywords: Dialogue journals, EFL, learning strategy, writing.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2066
162 Market Feasibility for New Brand Coffee House: The Case Study of Thailand

Authors: Pongsiri K.

Abstract:

This research aimed to study the market feasibility for new brand coffee house, the case study of Thailand.. This study is a mixed methods research combining quantitative research and the qualitative research. Primary data 350 sets of questionnaires were distributed, and the high quality completed questionnaires of 320 sets returned. Research samples are identified as customers’ of Hi-end department stores in Thailand. The sources of secondary data were critical selected from highly reliable sources, both from public and private sectors. The results were used to classify the customer group into two main groups, the younger than 25 and the older than 25years old. Results of the younger group, are give priority to the dimension of coffee house and its services dimension more than others, then branding dimension and the product dimension respectively. On the other hand, the older group give the difference result as they rate the important of the branding, coffee house and its services, then the product respectively. Coffee consuming is not just the trend but it has become part of people lifestyle. And the new cultures also created by the wise businessman. Coffee was long produced and consumed in Thailand. But it is surprisingly the hi-end brand coffee houses in Thai market are mostly imported brands. The café business possibility for Thai brand coffee house in Thai market were discussed in the paper.

Keywords: Coffee House, Café, Coffee Consuming and new entry branding, market feasibility

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14044
161 An Algorithm of Finite Capacity Material Requirement Planning System for Multi-stage Assembly Flow Shop

Authors: T. Wuttipornpun, U. Wangrakdiskul, W. Songserm

Abstract:

This paper aims to develop an algorithm of finite capacity material requirement planning (FCMRP) system for a multistage assembly flow shop. The developed FCMRP system has two main stages. The first stage is to allocate operations to the first and second priority work centers and also determine the sequence of the operations on each work center. The second stage is to determine the optimal start time of each operation by using a linear programming model. Real data from a factory is used to analyze and evaluate the effectiveness of the proposed FCMRP system and also to guarantee a practical solution to the user. There are five performance measures, namely, the total tardiness, the number of tardy orders, the total earliness, the number of early orders, and the average flow-time. The proposed FCMRP system offers an adjustable solution which is a compromised solution among the conflicting performance measures. The user can adjust the weight of each performance measure to obtain the desired performance. The result shows that the combination of FCMRP NP3 and EDD outperforms other combinations in term of overall performance index. The calculation time for the proposed FCMRP system is about 10 minutes which is practical for the planners of the factory.

Keywords: Material requirement planning, Finite capacity, Linear programming, Permutation, Application in industry.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2301
160 Attribute Weighted Class Complexity: A New Metric for Measuring Cognitive Complexity of OO Systems

Authors: Dr. L. Arockiam, A. Aloysius

Abstract:

In general, class complexity is measured based on any one of these factors such as Line of Codes (LOC), Functional points (FP), Number of Methods (NOM), Number of Attributes (NOA) and so on. There are several new techniques, methods and metrics with the different factors that are to be developed by the researchers for calculating the complexity of the class in Object Oriented (OO) software. Earlier, Arockiam et.al has proposed a new complexity measure namely Extended Weighted Class Complexity (EWCC) which is an extension of Weighted Class Complexity which is proposed by Mishra et.al. EWCC is the sum of cognitive weights of attributes and methods of the class and that of the classes derived. In EWCC, a cognitive weight of each attribute is considered to be 1. The main problem in EWCC metric is that, every attribute holds the same value but in general, cognitive load in understanding the different types of attributes cannot be the same. So here, we are proposing a new metric namely Attribute Weighted Class Complexity (AWCC). In AWCC, the cognitive weights have to be assigned for the attributes which are derived from the effort needed to understand their data types. The proposed metric has been proved to be a better measure of complexity of class with attributes through the case studies and experiments

Keywords: Software Complexity, Attribute Weighted Class Complexity, Weighted Class Complexity, Data Type

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2121
159 Many-Sided Self Risk Analysis Model for Information Asset to Secure Stability of the Information and Communication Service

Authors: Jin-Tae Lee, Jung-Hoon Suh, Sang-Soo Jang, Jae-Il Lee

Abstract:

Information and communication service providers (ICSP) that are significant in size and provide Internet-based services take administrative, technical, and physical protection measures via the information security check service (ISCS). These protection measures are the minimum action necessary to secure the stability and continuity of the information and communication services (ICS) that they provide. Thus, information assets are essential to providing ICS, and deciding the relative importance of target assets for protection is a critical procedure. The risk analysis model designed to decide the relative importance of information assets, which is described in this study, evaluates information assets from many angles, in order to choose which ones should be given priority when it comes to protection. Many-sided risk analysis (MSRS) grades the importance of information assets, based on evaluation of major security check items, evaluation of the dependency on the information and communication facility (ICF) and influence on potential incidents, and evaluation of major items according to their service classification, in order to identify the ISCS target. MSRS could be an efficient risk analysis model to help ICSPs to identify their core information assets and take information protection measures first, so that stability of the ICS can be ensured.

Keywords: Information Asset, Information CommunicationFacility, Evaluation, ISCS (Information Security Check Service), Evaluation, Grade.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1448
158 New Chances of Reforming Pedagogical Approach in Secondary English Class in China under the New English Curriculum and National College Entrance Examination Reform

Authors: Yue Wang

Abstract:

Five years after the newest English curriculum, reform policy was enacted in China and hand-wringing spread among teachers who accused that this is another “wearing new shoes to walk the old road” policy. This paper provides a thoroughly philosophical policy analysis of serious efforts that had been made to support this reform and revealed the hindrances that bridled the reform to yield the desired effect. Blame could be easily put on teachers for their insufficient pedagogical content knowledge, conservative resistance, and the handicaps of large class sizes and limited teaching times and so on. However, the underlying causes for this implementation failure are the interrelated factors in the NCEE-centred education system, such as the reluctance from students, the lack of school and education bureau support and insufficient teacher training. A further discussion of the 2017 to 2020’s NCEE reform on English prompts new possibilities for the authentic pedagogical approach reform in secondary English classes. In all, the pedagogical approach reform at the secondary level is heading towards a brighter future with the initiation of new NCEE reform.

Keywords: English curriculum, failure, NCEE, new possibilities, pedagogical, policy analysis, reform.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 527
157 Hiding Data in Images Using PCP

Authors: Souvik Bhattacharyya, Gautam Sanyal

Abstract:

In recent years, everything is trending toward digitalization and with the rapid development of the Internet technologies, digital media needs to be transmitted conveniently over the network. Attacks, misuse or unauthorized access of information is of great concern today which makes the protection of documents through digital media a priority problem. This urges us to devise new data hiding techniques to protect and secure the data of vital significance. In this respect, steganography often comes to the fore as a tool for hiding information. Steganography is a process that involves hiding a message in an appropriate carrier like image or audio. It is of Greek origin and means "covered or hidden writing". The goal of steganography is covert communication. Here the carrier can be sent to a receiver without any one except the authenticated receiver only knows existence of the information. Considerable amount of work has been carried out by different researchers on steganography. In this work the authors propose a novel Steganographic method for hiding information within the spatial domain of the gray scale image. The proposed approach works by selecting the embedding pixels using some mathematical function and then finds the 8 neighborhood of the each selected pixel and map each bit of the secret message in each of the neighbor pixel coordinate position in a specified manner. Before embedding a checking has been done to find out whether the selected pixel or its neighbor lies at the boundary of the image or not. This solution is independent of the nature of the data to be hidden and produces a stego image with minimum degradation.

Keywords: Cover Image, LSB, Pixel Coordinate Position (PCP), Stego Image.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1821
156 Comparison between Deterministic and Probabilistic Stability Analysis, Featuring Consequent Risk Assessment

Authors: Isabela Moreira Queiroz

Abstract:

Slope stability analyses are largely carried out by deterministic methods and evaluated through a single security factor. Although it is known that the geotechnical parameters can present great dispersal, such analyses are considered fixed and known. The probabilistic methods, in turn, incorporate the variability of input key parameters (random variables), resulting in a range of values of safety factors, thus enabling the determination of the probability of failure, which is an essential parameter in the calculation of the risk (probability multiplied by the consequence of the event). Among the probabilistic methods, there are three frequently used methods in geotechnical society: FOSM (First-Order, Second-Moment), Rosenblueth (Point Estimates) and Monte Carlo. This paper presents a comparison between the results from deterministic and probabilistic analyses (FOSM method, Monte Carlo and Rosenblueth) applied to a hypothetical slope. The end was held to evaluate the behavior of the slope and consequent risk analysis, which is used to calculate the risk and analyze their mitigation and control solutions. It can be observed that the results obtained by the three probabilistic methods were quite close. It should be noticed that the calculation of the risk makes it possible to list the priority to the implementation of mitigation measures. Therefore, it is recommended to do a good assessment of the geological-geotechnical model incorporating the uncertainty in viability, design, construction, operation and closure by means of risk management. 

Keywords: Probabilistic methods, risk assessment, risk management, slope stability.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1739
155 A Study of Combined Mechanical and Chemical Stabilisation of Fine Grained Dredge Soil of River Jhelum

Authors: Adnan F. Sheikh, Fayaz A. Mir

Abstract:

After the recent devastating flood in Kashmir in 2014, dredging of the local water bodies, especially Jhelum River has become a priority for the government. Local government under the project name of 'Comprehensive Flood Management Programme' plans to undertake an increase in discharge of existing flood channels by removal of encroachments and acquisition of additional land, dredging and other works of the water bodies. The total quantity of soil to be dredged will be 16.15 lac cumecs. Dredged soil is a major component that would result from the project which requires disposal/utilization. This study analyses the effect of cement and sand on the engineering properties of soil. The tests were conducted with variable additions of sand (10%, 20% and 30%), whereas cement was added at 12%. Samples with following compositions: soil-cement (12%) and soil-sand (30%) were tested as well. Laboratory experiments were conducted to determine the engineering characteristics of soil, i.e., compaction, strength, and CBR characteristics. The strength characteristics of the soil were determined by unconfined compressive strength test and direct shear test. Unconfined compressive strength of the soil was tested immediately and for a curing period of seven days. CBR test was performed for unsoaked, soaked (worst condition- 4 days) and cured (4 days) samples.

Keywords: Comprehensive flood management programme, dredge soil, strength characteristics, flood.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 887
154 Artificial Intelligence Techniques Applications for Power Disturbances Classification

Authors: K.Manimala, Dr.K.Selvi, R.Ahila

Abstract:

Artificial Intelligence (AI) methods are increasingly being used for problem solving. This paper concerns using AI-type learning machines for power quality problem, which is a problem of general interest to power system to provide quality power to all appliances. Electrical power of good quality is essential for proper operation of electronic equipments such as computers and PLCs. Malfunction of such equipment may lead to loss of production or disruption of critical services resulting in huge financial and other losses. It is therefore necessary that critical loads be supplied with electricity of acceptable quality. Recognition of the presence of any disturbance and classifying any existing disturbance into a particular type is the first step in combating the problem. In this work two classes of AI methods for Power quality data mining are studied: Artificial Neural Networks (ANNs) and Support Vector Machines (SVMs). We show that SVMs are superior to ANNs in two critical respects: SVMs train and run an order of magnitude faster; and SVMs give higher classification accuracy.

Keywords: back propagation network, power quality, probabilistic neural network, radial basis function support vector machine

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1556
153 Methodology for Developing an Intelligent Tutoring System Based on Marzano’s Taxonomy

Authors: Joaquin Navarro Perales, Ana Lidia Franzoni Velázquez, Francisco Cervantes Pérez

Abstract:

The Mexican educational system faces diverse challenges related with the quality and coverage of education. The development of Intelligent Tutoring Systems (ITS) may help to solve some of them by helping teachers to customize their classes according to the performance of the students in online courses. In this work, we propose the adaptation of a functional ITS based on Bloom’s taxonomy called Sistema de Apoyo Generalizado para la Enseñanza Individualizada (SAGE), to measure student’s metacognition and their emotional response based on Marzano’s taxonomy. The students and the system will share the control over the advance in the course, so they can improve their metacognitive skills. The system will not allow students to get access to subjects not mastered yet. The interaction between the system and the student will be implemented through Natural Language Processing techniques, thus avoiding the use of sensors to evaluate student’s response. The teacher will evaluate student’s knowledge utilization, which is equivalent to the last cognitive level in Marzano’s taxonomy.

Keywords: Intelligent tutoring systems, student modelling, metacognition, affective computing, natural language processing.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1010
152 Antibiotic Resistance Profile of Bacterial Isolates from Animal Farming Aquatic Environments and Meats in a Peri-Urban Community in South Korea

Authors: Hyunjin Rho, Bongjin Shin, Okbok Lee, Yu-Hyun Choi, Jiyoung Lee, Jaerang Rho

Abstract:

The increasing usage of antibiotics in the animal farming industry is an emerging worldwide problem contributing to the development of antibiotic resistance. The purpose of this work was to investigate the prevalence and antibiotic resistance profile of bacterial isolates collected from aquatic environments and meats in a peri-urban community in Daejeon, Korea. In an antibacterial susceptibility test, the bacterial isolates showed a high incidence of resistance (~ 26.04 %) to cefazolin, tetracycline, gentamycin, norfloxacin, erythromycin and vancomycin. The results from a test for multiple antibiotic resistance indicated that the isolates were displaying an approximately 5-fold increase in the incidence of multiple antibiotic resistance to combinations of two different antibiotics compared to combinations of three or more antibiotics. Most of the isolates showed multi-antibiotic resistance, and the resistance patterns were similar among the sampling groups. Sequencing data analysis of 16S rRNA showed that most of the resistant isolates appeared to be dominated by the classes Betaproteobacteria and Gammaproteobacteria in the phylum Proteobacteria.

Keywords: Antibiotics, Antibiotic resistance, Antimicrobial resistance, Multi-resistance

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2229
151 Early Recognition and Grading of Cataract Using a Combined Log Gabor/Discrete Wavelet Transform with ANN and SVM

Authors: Hadeer R. M. Tawfik, Rania A. K. Birry, Amani A. Saad

Abstract:

Eyes are considered to be the most sensitive and important organ for human being. Thus, any eye disorder will affect the patient in all aspects of life. Cataract is one of those eye disorders that lead to blindness if not treated correctly and quickly. This paper demonstrates a model for automatic detection, classification, and grading of cataracts based on image processing techniques and artificial intelligence. The proposed system is developed to ease the cataract diagnosis process for both ophthalmologists and patients. The wavelet transform combined with 2D Log Gabor Wavelet transform was used as feature extraction techniques for a dataset of 120 eye images followed by a classification process that classified the image set into three classes; normal, early, and advanced stage. A comparison between the two used classifiers, the support vector machine SVM and the artificial neural network ANN were done for the same dataset of 120 eye images. It was concluded that SVM gave better results than ANN. SVM success rate result was 96.8% accuracy where ANN success rate result was 92.3% accuracy.

Keywords: Cataract, classification, detection, feature extraction, grading, log-gabor, neural networks, support vector machines, wavelet.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 993
150 Detecting HCC Tumor in Three Phasic CT Liver Images with Optimization of Neural Network

Authors: Mahdieh Khalilinezhad, Silvana Dellepiane, Gianni Vernazza

Abstract:

The aim of this work is to build a model based on tissue characterization that is able to discriminate pathological and non-pathological regions from three-phasic CT images. With our research and based on a feature selection in different phases, we are trying to design a neural network system with an optimal neuron number in a hidden layer. Our approach consists of three steps: feature selection, feature reduction, and classification. For each region of interest (ROI), 6 distinct sets of texture features are extracted such as: first order histogram parameters, absolute gradient, run-length matrix, co-occurrence matrix, autoregressive model, and wavelet, for a total of 270 texture features. When analyzing more phases, we show that the injection of liquid cause changes to the high relevant features in each region. Our results demonstrate that for detecting HCC tumor phase 3 is the best one in most of the features that we apply to the classification algorithm. The percentage of detection between pathology and healthy classes, according to our method, relates to first order histogram parameters with accuracy of 85% in phase 1, 95% in phase 2, and 95% in phase 3.

Keywords: Feature selection, Multi-phasic liver images, Neural network, Texture analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2535
149 Object Negotiation Mechanism for an Intelligent Environment Using Event Agents

Authors: Chiung-Hui Chen

Abstract:

With advancements in science and technology, the concept of the Internet of Things (IoT) has gradually developed. The development of the intelligent environment adds intelligence to objects in the living space by using the IoT. In the smart environment, when multiple users share the living space, if different service requirements from different users arise, then the context-aware system will have conflicting situations for making decisions about providing services. Therefore, the purpose of establishing a communication and negotiation mechanism among objects in the intelligent environment is to resolve those service conflicts among users. This study proposes developing a decision-making methodology that uses “Event Agents” as its core. When the sensor system receives information, it evaluates a user’s current events and conditions; analyses object, location, time, and environmental information; calculates the priority of the object; and provides the user services based on the event. Moreover, when the event is not single but overlaps with another, conflicts arise. This study adopts the “Multiple Events Correlation Matrix” in order to calculate the degree values of incidents and support values for each object. The matrix uses these values as the basis for making inferences for system service, and to further determine appropriate services when there is a conflict.

Keywords: Internet of things, intelligent object, event agents, negotiation mechanism, degree of similarity.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1190
148 Reclaiming Pedestrian Space from Car Dominated Neighborhoods

Authors: Andreas L. Savvides

Abstract:

For a long time as a result of accommodating car traffic, planning ideologies in the past put a low priority on public space, pedestrianism and the role of city space as a meeting place for urban dwellers. In addition, according to authors such as Jan Gehl, market forces and changing architectural perceptions began to shift the focus of planning practice from the integration of public space in various pockets around the contemporary city to individual buildings. Eventually, these buildings have become increasingly more isolated and introverted and have turned their backs to the realm of the public space adjoining them. As a result of this practice, the traditional function of public space as a social forum for city dwellers has in many cases been reduced or even phased out. Author Jane Jacobs published her seminal book “The Death and Life of Great American Cities" more than fifty years ago, but her observations and predictions at the time still ring true today, where she pointed out how the dramatic increase in car traffic and its accommodation by the urban planning ideology that was brought about by the Modern movement has prompted a separation of the uses of the city. At the same time it emphasizes free standing buildings that threaten urban space and city life and result in underutilized and lifeless urban cores. In this discussion context, the aim of this paper is to showcase a reversal of just such a situation in the case of the Dasoupolis neighborhood in Strovolos, Cyprus, where enlightened urban design practice has see the reclamation of pedestrian space in a car dominated area.

Keywords: Urban Design, Public Space, Right to the City, Accessibility, Mobility

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1915
147 Refinement of Object-Z Specifications Using Morgan-s Refinement Calculus

Authors: Mehrnaz Najafi, Hassan Haghighi

Abstract:

Morgan-s refinement calculus (MRC) is one of the well-known methods allowing the formality presented in the program specification to be continued all the way to code. On the other hand, Object-Z (OZ) is an extension of Z adding support for classes and objects. There are a number of methods for obtaining code from OZ specifications that can be categorized into refinement and animation methods. As far as we know, only one refinement method exists which refines OZ specifications into code. However, this method does not have fine-grained refinement rules and thus cannot be automated. On the other hand, existing animation methods do not present mapping rules formally and do not support the mapping of several important constructs of OZ, such as all cases of operation expressions and most of constructs in global paragraph. In this paper, with the aim of providing an automatic path from OZ specifications to code, we propose an approach to map OZ specifications into their counterparts in MRC in order to use fine-grained refinement rules of MRC. In this way, having counterparts of our specifications in MRC, we can refine them into code automatically using MRC tools such as RED. Other advantages of our work pertain to proposing mapping rules formally, supporting the mapping of all important constructs of Object-Z, and considering dynamic instantiation of objects while OZ itself does not cover this facility.

Keywords: Formal method, Formal specification, Formalprogram development, Morgan's Refinement Calculus, Object-Z

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1320
146 Modelling a Hospital as a Queueing Network: Analysis for Improving Performance

Authors: Emad Alenany, M. Adel El-Baz

Abstract:

In this paper, the flow of different classes of patients into a hospital is modelled and analyzed by using the queueing network analyzer (QNA) algorithm and discrete event simulation. Input data for QNA are the rate and variability parameters of the arrival and service times in addition to the number of servers in each facility. Patient flows mostly match real flow for a hospital in Egypt. Based on the analysis of the waiting times, two approaches are suggested for improving performance: Separating patients into service groups, and adopting different service policies for sequencing patients through hospital units. The separation of a specific group of patients, with higher performance target, to be served separately from the rest of patients requiring lower performance target, requires the same capacity while improves performance for the selected group of patients with higher target. Besides, it is shown that adopting the shortest processing time and shortest remaining processing time service policies among other tested policies would results in, respectively, 11.47% and 13.75% reduction in average waiting time relative to first come first served policy.

Keywords: Queueing network, discrete-event simulation, health applications, SPT.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1529
145 A Robust Adaptive Congestion Control Strategy for Large Scale Networks with Differentiated Services Traffic

Authors: R. R. Chen, K. Khorasani

Abstract:

In this paper, a robust decentralized congestion control strategy is developed for a large scale network with Differentiated Services (Diff-Serv) traffic. The network is modeled by a nonlinear fluid flow model corresponding to two classes of traffic, namely the premium traffic and the ordinary traffic. The proposed congestion controller does take into account the associated physical network resource limitations and is shown to be robust to the unknown and time-varying delays. Our proposed decentralized congestion control strategy is developed on the basis of Diff-Serv architecture by utilizing a robust adaptive technique. A Linear Matrix Inequality (LMI) condition is obtained to guarantee the ultimate boundedness of the closed-loop system. Numerical simulation implementations are presented by utilizing the QualNet and Matlab software tools to illustrate the effectiveness and capabilities of our proposed decentralized congestion control strategy.

Keywords: Congestion control, Large scale networks, Decentralized control, Differentiated services traffic, Time-delay systems.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1989
144 Investigating the Relation between Student Engagement and Attainment in a Flexible Learning Environment

Authors: Y. Bi, T. Anderson, M. Huang

Abstract:

The use of technology is increasingly adopted to support flexible learning in Higher Education institutions. The adoption of more sophisticated technologies offers a broad range of facilities for communication and resource sharing, thereby creating a flexible learning environment that facilitates and even encourages students not to physically attend classes. However this emerging trend seems to contradict class attendance requirements within universities, inevitably leading to a dilemma between amending traditional regulations and creating new policies for the higher education institutions. This study presents an investigation into student engagement in a technology enhanced/driven flexible environment along with its relationship to attainment. We propose an approach to modelling engagement from different perspectives in terms of indicators and then consider what impact these indicators have on student academic performance. We have carried out a case study on the relation between attendance and attainment in a flexible environment. Although our preliminary results show attendance is quantitatively correlated with successful student development and learning outcomes, our results also indicate there is a cohort that did not follow such a pattern. Nevertheless the preliminary results could provide an insight into pilot studies in the wider deployment of new technology to support flexible learning.

Keywords: Engagement, flexible leaning, attendance and attainment.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1795
143 Characteristics of Wall Thickness Increase in Pipe Reduction Process using Planetary Rolls

Authors: Yuji Kotani, Shunsuke Kanai, Hisaki Watari

Abstract:

In recent years, global warming has become a worldwide problem. The reduction of carbon dioxide emissions is a top priority for many companies in the manufacturing industry. In the automobile industry as well, the reduction of carbon dioxide emissions is one of the most important issues. Technology to reduce the weight of automotive parts improves the fuel economy of automobiles, and is an important technology for reducing carbon dioxide. Also, even if this weight reduction technology is applied to electric automobiles rather than gasoline automobiles, reducing energy consumption remains an important issue. Plastic processing of hollow pipes is one important technology for realizing the weight reduction of automotive parts. Ohashi et al. [1],[2] present an example of research on pipe formation in which a process was carried out to enlarge a pipe diameter using a lost core, achieving the suppression of wall thickness reduction and greater pipe expansion than hydroforming. In this study, we investigated a method to increase the wall thickness of a pipe through pipe compression using planetary rolls. The establishment of a technology whereby the wall thickness of a pipe can be controlled without buckling the pipe is an important technology for the weight reduction of products. Using the finite element analysis method, we predicted that it would be possible to increase the compression of an aluminum pipe with a 3mm wall thickness by approximately 20%, and wall thickness by approximately 20% by pressing the hollow pipe with planetary rolls.

Keywords: Pipe-Forming, Wall Thickness, Finite-element-method

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2985
142 Transient Stability Assessment Using Fuzzy SVM and Modified Preventive Control

Authors: B. Dora Arul Selvi, .N. Kamaraj

Abstract:

Transient Stability is an important issue in power systems planning, operation and extension. The objective of transient stability analysis problem is not satisfied with mere transient instability detection or evaluation and it is most important to complement it by defining fast and efficient control measures in order to ensure system security. This paper presents a new Fuzzy Support Vector Machines (FSVM) to investigate the stability status of power systems and a modified generation rescheduling scheme to bring back the identified unstable cases to a more economical and stable operating point. FSVM improves the traditional SVM (Support Vector Machines) by adding fuzzy membership to each training sample to indicate the degree of membership of this sample to different classes. The preventive control based on economic generator rescheduling avoids the instability of the power systems with minimum change in operating cost under disturbed conditions. Numerical results on the New England 39 bus test system show the effectiveness of the proposed method.

Keywords: Fuzzy Support Vector Machine (FSVM), Incremental Cost, Preventive Control, Transient stability

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1491
141 An Investigation of Performance versus Security in Cognitive Radio Networks with Supporting Cloud Platforms

Authors: Kurniawan D. Irianto, Demetres D. Kouvatsos

Abstract:

The growth of wireless devices affects the availability of limited frequencies or spectrum bands as it has been known that spectrum bands are a natural resource that cannot be added. Meanwhile, the licensed frequencies are idle most of the time. Cognitive radio is one of the solutions to solve those problems. Cognitive radio is a promising technology that allows the unlicensed users known as secondary users (SUs) to access licensed bands without making interference to licensed users or primary users (PUs). As cloud computing has become popular in recent years, cognitive radio networks (CRNs) can be integrated with cloud platform. One of the important issues in CRNs is security. It becomes a problem since CRNs use radio frequencies as a medium for transmitting and CRNs share the same issues with wireless communication systems. Another critical issue in CRNs is performance. Security has adverse effect to performance and there are trade-offs between them. The goal of this paper is to investigate the performance related to security trade-off in CRNs with supporting cloud platforms. Furthermore, Queuing Network Models with preemptive resume and preemptive repeat identical priority are applied in this project to measure the impact of security to performance in CRNs with or without cloud platform. The generalized exponential (GE) type distribution is used to reflect the bursty inter-arrival and service times at the servers. The results show that the best performance is obtained when security is disabled and cloud platform is enabled.

Keywords: Cloud Platforms, Cognitive Radio Networks, GEtype Distribution, Performance Vs Security.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2521
140 Water Resources Crisis in Saudi Arabia, Challenges and Possible Management Options: An Analytic Review

Authors: A. A. Ghanim

Abstract:

The Kingdom of Saudi Arabia (KSA) is heading towards a severe and rapidly expanding water crisis, which can have negative impacts on the country’s environment and economy. Of the total water consumption in KSA, the agricultural sector accounts for nearly 87% of the total water use and, therefore, any attempt that overlooks this sector will not help in improving the sustainability of the country’s water resources. KSA Vision 2030 gives priority of water use in the agriculture sector for the regions that have natural renewable water resources. It means that there is little concern for making reuse of municipal wastewater for irrigation purposes in any region in general and in water-scarce regions in particular. The use of treated wastewater is very limited in Saudi Arabia, but it has very considerable potential for future expansion due its numerous beneficial uses. This study reviews the current situation of water resources in Saudi Arabia, providing more highlights on agriculture and wastewater reuse. The reviewed study is proposing some corrective measures for development and better management of water resources in the Kingdom. Suggestions also include consideration of treated water as an alternative source for irrigation in some regions of the country. The study concluded that a sustainable solution for the water crisis in KSA requires implementation of multiple measures in an integrated manner. The integrated solution plan should focus on two main directions: first, improving the current management practices of the existing water resources; second, developing new water supplies from both conventional and non-conventional sources.

Keywords: Saudi Arabia, water resources, water crisis, treated wastewater.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1938
139 A Review and Comparative Analysis on Cluster Ensemble Methods

Authors: S. Sarumathi, P. Ranjetha, C. Saraswathy, M. Vaishnavi, S. Geetha

Abstract:

Clustering is an unsupervised learning technique for aggregating data objects into meaningful classes so that intra cluster similarity is maximized and inter cluster similarity is minimized in data mining. However, no single clustering algorithm proves to be the most effective in producing the best result. As a result, a new challenging technique known as the cluster ensemble approach has blossomed in order to determine the solution to this problem. For the cluster analysis issue, this new technique is a successful approach. The cluster ensemble's main goal is to combine similar clustering solutions in a way that achieves the precision while also improving the quality of individual data clustering. Because of the massive and rapid creation of new approaches in the field of data mining, the ongoing interest in inventing novel algorithms necessitates a thorough examination of current techniques and future innovation. This paper presents a comparative analysis of various cluster ensemble approaches, including their methodologies, formal working process, and standard accuracy and error rates. As a result, the society of clustering practitioners will benefit from this exploratory and clear research, which will aid in determining the most appropriate solution to the problem at hand.

Keywords: Clustering, cluster ensemble methods, consensus function, data mining, unsupervised learning.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 820
138 An Approach for Reducing the Computational Complexity of LAMSTAR Intrusion Detection System using Principal Component Analysis

Authors: V. Venkatachalam, S. Selvan

Abstract:

The security of computer networks plays a strategic role in modern computer systems. Intrusion Detection Systems (IDS) act as the 'second line of defense' placed inside a protected network, looking for known or potential threats in network traffic and/or audit data recorded by hosts. We developed an Intrusion Detection System using LAMSTAR neural network to learn patterns of normal and intrusive activities, to classify observed system activities and compared the performance of LAMSTAR IDS with other classification techniques using 5 classes of KDDCup99 data. LAMSAR IDS gives better performance at the cost of high Computational complexity, Training time and Testing time, when compared to other classification techniques (Binary Tree classifier, RBF classifier, Gaussian Mixture classifier). we further reduced the Computational Complexity of LAMSTAR IDS by reducing the dimension of the data using principal component analysis which in turn reduces the training and testing time with almost the same performance.

Keywords: Binary Tree Classifier, Gaussian Mixture, IntrusionDetection System, LAMSTAR, Radial Basis Function.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1747