Search results for: Time Series Prediction
5145 From Mother Tongue Education to Multilingual Higher Education
Authors: Mario R. Acevedo Amaya, Fernanda M. Martinez Reyes
Abstract:
Through the time, the higher education has changed the learning system since mother tongue to bilingual, and in this new century has been coming develop a multilingual education. All as part of globalization process of the countries and the education. Nevertheless, this change only has been effectively in countries of the first world, the rest have been lagging. Therefore, these countries require strengthen their higher education systems through models that give way to multilingual and bilingual education. In this way, shows a new model adapted from a systemic form to allow a higher bilingual and multilingual education in Latin America. This systematization aims to increase the skills and competencies student’s, decrease the time learning of a second tongue, add to multilingualism in the American Latin Universities, also, contribute to position the region´s countries in a better global status, and stimulate the development of new research in this area.Keywords: Bilingual Education, Higher Education, Multilingual Education, Multilingual Education Model
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19335144 Border Limited Adaptive Subdivision Based On Triangle Meshes
Authors: Pichayut Peerasathien, Hiroshi Nagahashi
Abstract:
Subdivision is a method to create a smooth surface from a coarse mesh by subdividing the entire mesh. The conventional ways to compute and render surfaces are inconvenient both in terms of memory and computational time as the number of meshes will increase exponentially. An adaptive subdivision is the way to reduce the computational time and memory by subdividing only certain selected areas. In this paper, a new adaptive subdivision method for triangle meshes is introduced. This method defines a new adaptive subdivision rules by considering the properties of each triangle's neighbors and is embedded in a traditional Loop's subdivision. It prevents some undesirable side effects that appear in the conventional adaptive ways. Models that were subdivided by our method are compared with other adaptive subdivision methods
Keywords: Subdivision, loop subdivision, handle cracks, smooth surface.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16525143 Reachable Set Bounding Estimation for Distributed Delay Systems with Disturbances
Authors: Li Xu, Shouming Zhong
Abstract:
The reachable set bounding estimation for distributed delay systems with disturbances is a new problem. In this paper,we consider this problem subject to not only time varying delay and polytopic uncertainties but also distributed delay systems which is not studied fully untill now. we can obtain improved non-ellipsoidal reachable set estimation for neural networks with time-varying delay by the maximal Lyapunov-Krasovskii fuctional which is constructed as the pointwise maximum of a family of Lyapunov-Krasovskii fuctionals corresponds to vertexes of uncertain polytope.On the other hand,matrix inequalities containing only one scalar and Matlabs LMI Toolbox is utilized to give a non-ellipsoidal description of the reachable set.finally,numerical examples are given to illustrate the existing results.
Keywords: Reachable set, Distributed delay, Lyapunov-Krasovskii function, Polytopic uncertainties.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18565142 Efficient Dimensionality Reduction of Directional Overcurrent Relays Optimal Coordination Problem
Authors: Fouad Salha , X. Guillaud
Abstract:
Directional over current relays (DOCR) are commonly used in power system protection as a primary protection in distribution and sub-transmission electrical systems and as a secondary protection in transmission systems. Coordination of protective relays is necessary to obtain selective tripping. In this paper, an approach for efficiency reduction of DOCRs nonlinear optimum coordination (OC) is proposed. This was achieved by modifying the objective function and relaxing several constraints depending on the four constraints classification, non-valid, redundant, pre-obtained and valid constraints. According to this classification, the far end fault effect on the objective function and constraints, and in consequently on relay operating time, was studied. The study was carried out, firstly by taking into account the near-end and far-end faults in DOCRs coordination problem formulation; and then faults very close to the primary relays (nearend faults). The optimal coordination (OC) was achieved by simultaneously optimizing all variables (TDS and Ip) in nonlinear environment by using of Genetic algorithm nonlinear programming techniques. The results application of the above two approaches on 6-bus and 26-bus system verify that the far-end faults consideration on OC problem formulation don-t lose the optimality.
Keywords: Backup/Primary relay, Coordination time interval (CTI), directional over current relays, Genetic algorithm, time dial setting (TDS), pickup current setting (Ip), nonlinear programming.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15845141 Identification of Flexographic-printed Newspapers with NIR Spectral Imaging
Authors: Raimund Leitner, Susanne Rosskopf
Abstract:
Near-infrared (NIR) spectroscopy is a widely used method for material identification for laboratory and industrial applications. While standard spectrometers only allow measurements at one sampling point at a time, NIR Spectral Imaging techniques can measure, in real-time, both the size and shape of an object as well as identify the material the object is made of. The online classification and sorting of recovered paper with NIR Spectral Imaging (SI) is used with success in the paper recycling industry throughout Europe. Recently, the globalisation of the recycling material streams caused that water-based flexographic-printed newspapers mainly from UK and Italy appear also in central Europe. These flexo-printed newspapers are not sufficiently de-inkable with the standard de-inking process originally developed for offset-printed paper. This de-inking process removes the ink from recovered paper and is the fundamental processing step to produce high-quality paper from recovered paper. Thus, the flexo-printed newspapers are a growing problem for the recycling industry as they reduce the quality of the produced paper if their amount exceeds a certain limit within the recovered paper material. This paper presents the results of a research project for the development of an automated entry inspection system for recovered paper that was jointly conducted by CTR AG (Austria) and PTS Papiertechnische Stiftung (Germany). Within the project an NIR SI prototype for the identification of flexo-printed newspaper has been developed. The prototype can identify and sort out flexoprinted newspapers in real-time and achieves a detection accuracy for flexo-printed newspaper of over 95%. NIR SI, the technology the prototype is based on, allows the development of inspection systems for incoming goods in a paper production facility as well as industrial sorting systems for recovered paper in the recycling industry in the near future.Keywords: spectral imaging, imaging spectroscopy, NIR, waterbasedflexographic, flexo-printed, recovered paper, real-time classification.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15455140 The Importance of Issues for the Youth in Voter Decision Making: A Case Study among University Students in Malaysia
Authors: Sivamurugan Pandian
Abstract:
In the 13th Malaysia’s General Elections held in 2013, it was observed that large numbers of urban constituencies saw strongly decisive young voters (between 21-39 age group) determine the outcome in their favour. Also, the Elections Commission had approximated that 70% of some 4.2 million unregistered voters at the time were citizens aged between 21 and 40 years old. If they are not already considered an important form of political leverage, 450,000 young Malaysians turn 21 years old each year. Further compounding this fact were the 2.4 million new voters registered in 2012, which at the time constituted almost 30% of the entire voting population. This article discusses the importance of issues for the youth, with reference to the university students in Malaysia in their decision making on polling day.
Keywords: Malaysia, Youth, Issues, Voting Patterns, Elections.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 35485139 A Review on Technology Forecasting Methods and Their Application Area
Authors: Daekook Kang, Wooseok Jang, Hyeonjeong Lee, Hyun Joung No
Abstract:
Technology changes have been acknowledged as a critical factor in determining competitiveness of organization. Under such environment, the right anticipation of technology change has been of huge importance in strategic planning. To monitor technology change, technology forecasting (TF) is frequently utilized. In academic perspective, TF has received great attention for a long time. However, few researches have been conducted to provide overview of the TF literature. Even though some studies deals with review of TF research, they generally focused on type and characteristics of various TF, so hardly provides information about patterns of TF research and which TF method is used in certain technology industry. Accordingly, this study profile developments in and patterns of scholarly research in TF over time. Also, this study investigates which technology industries have used certain TF method and identifies their relationships. This study will help in understanding TF research trend and their application area.Keywords: Technology forecasting, technology industry, TF trend, technology trajectory.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 29555138 NewPerceptual Organization within Temporal Displacement
Authors: Michele Sinico
Abstract:
The psychological present has an actual extension. When a sequence of instantaneous stimuli falls in this short interval of time, observers perceive a compresence of events in succession and the temporal order depends on the qualitative relationships between the perceptual properties of the events. Two experiments were carried out to study the influence of perceptual grouping, with and without temporal displacement, on the duration of auditory sequences. The psychophysical method of adjustment was adopted. The first experiment investigated the effect of temporal displacement of a white noise on sequence duration. The second experiment investigated the effect of temporal displacement, along the pitch dimension, on temporal shortening of sequence. The results suggest that the temporal order of sounds, in the case of temporal displacement, is organized along the pitch dimension.Keywords: Time perception, perceptual present, temporal displacement, gestalt laws of perceptual organization
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 8085137 Towards an Enhanced Quality of IPTV Media Server Architecture over Software Defined Networking
Authors: Esmeralda Hysenbelliu
Abstract:
The aim of this paper is to present the QoE (Quality of Experience) IPTV SDN-based media streaming server enhanced architecture for configuring, controlling, management and provisioning the improved delivery of IPTV service application with low cost, low bandwidth, and high security. Furthermore, it is given a virtual QoE IPTV SDN-based topology to provide an improved IPTV service based on QoE Control and Management of multimedia services functionalities. Inside OpenFlow SDN Controller there are enabled in high flexibility and efficiency Service Load-Balancing Systems; based on the Loading-Balance module and based on GeoIP Service. This two Load-balancing system improve IPTV end-users Quality of Experience (QoE) with optimal management of resources greatly. Through the key functionalities of OpenFlow SDN controller, this approach produced several important features, opportunities for overcoming the critical QoE metrics for IPTV Service like achieving incredible Fast Zapping time (Channel Switching time) < 0.1 seconds. This approach enabled Easy and Powerful Transcoding system via FFMPEG encoder. It has the ability to customize streaming dimensions bitrates, latency management and maximum transfer rates ensuring delivering of IPTV streaming services (Audio and Video) in high flexibility, low bandwidth and required performance. This QoE IPTV SDN-based media streaming architecture unlike other architectures provides the possibility of Channel Exchanging between several IPTV service providers all over the word. This new functionality brings many benefits as increasing the number of TV channels received by end –users with low cost, decreasing stream failure time (Channel Failure time < 0.1 seconds) and improving the quality of streaming services.
Keywords: Improved QoE, OpenFlow SDN controller, IPTV service application, softwarization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10305136 Entrepreneurial Orientation and Customers Satisfaction: Evidences nearby Khao San Road
Authors: Vichada Chokesikarin
Abstract:
The study aims to determine which factors account for customer satisfaction and to investigate the relationship between entrepreneurial orientation and business success, in particular, context of the information understanding of hostel business in Pranakorn district, Bangkok and the significant element of entrepreneurship in tourism industry. This study covers 352 hostels customers and 61 hostel owners/managers nearby Khao San road. Data collection methods were used by survey questionnaire and a series of hypotheses were developed from services marketing literature. The findings suggest the customer satisfaction most influenced by image, service quality, room quality and price accordingly. Furthermore the findings revealed that significant relationships exist between entrepreneurial orientation and business success; while competitive aggressiveness was found unrelated. The ECSI model’s generic measuring customer satisfaction was found partially mediate the business success. A reconsideration of other variables applicable should be supported with the model of hostel business. The study provides context and overall view of hostel business while discussing from the entrepreneurial orientation to customer satisfaction, thereby reducing decision risk on hostel investment.
Keywords: Customer satisfaction, ECSI Model, entrepreneurial orientation, small hotels, hostel, business performance.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 34525135 Dynamic Analysis of the Dome with Arches and Rings from Romexpo Bucharest
Authors: V. Precupas, A. Ivan, M. Ivan
Abstract:
The dome with ribs and rings, which covers the ROMEXPO pavilion from Bucharest, was designed after the collapse of the single layer reticulated dome. In this paper, it was made the checking of the structure, under the dynamic loads with three recorded accelerograms calibrated according to Romanian seismic design code P100-1/2006. Under the action the dynamic loadings, it was made a time-history analysis to determine the zones where the plastic hinges appear, at what accelerations and their position on the structure. The studied dome is formed by 32 spatial semi arches and three rings: one circular ring located at the top of the dome and another two rings, design as trusses, the first near the supports and the second as an intermediate rings above the skylights. Above the skylights up to the top, the dome is tight together with purlins and bracings.Keywords: dome, dynamic analysis, plastic hinges, time-history
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 24505134 Intelligent Transport System: Classification of Traffic Signs Using Deep Neural Networks in Real Time
Authors: Anukriti Kumar, Tanmay Singh, Dinesh Kumar Vishwakarma
Abstract:
Traffic control has been one of the most common and irritating problems since the time automobiles have hit the roads. Problems like traffic congestion have led to a significant time burden around the world and one significant solution to these problems can be the proper implementation of the Intelligent Transport System (ITS). It involves the integration of various tools like smart sensors, artificial intelligence, position technologies and mobile data services to manage traffic flow, reduce congestion and enhance driver's ability to avoid accidents during adverse weather. Road and traffic signs’ recognition is an emerging field of research in ITS. Classification problem of traffic signs needs to be solved as it is a major step in our journey towards building semi-autonomous/autonomous driving systems. The purpose of this work focuses on implementing an approach to solve the problem of traffic sign classification by developing a Convolutional Neural Network (CNN) classifier using the GTSRB (German Traffic Sign Recognition Benchmark) dataset. Rather than using hand-crafted features, our model addresses the concern of exploding huge parameters and data method augmentations. Our model achieved an accuracy of around 97.6% which is comparable to various state-of-the-art architectures.
Keywords: Multiclass classification, convolution neural network, OpenCV, Data Augmentation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 8165133 The Risk and Value Engineering Structures and their Integration with Industrial Projects Management (A Case Study on I. K.Corporation)
Authors: Lida Haghnegahdar, Ezzatollah Asgharizadeh
Abstract:
Value engineering is an efficacious contraption for administrators to make up their minds. Value perusals proffer the gaffers a suitable instrument to decrease the expenditures of the life span, quality amelioration, structural improvement, curtailment of the construction schedule, longevity prolongation or a merging of the aforementioned cases. Subjecting organizers to pressures on one hand and their accountability towards their pertinent fields together with inherent risks and ambiguities of other options on the other hand set some comptrollers in a dilemma utilization of risk management and the value engineering in projects manipulation with regard to complexities of implementing projects can be wielded as a contraption to identify and efface each item which wreaks unnecessary expenses and time squandering sans inflicting any damages upon the essential project applications. Of course It should be noted that implementation of risk management and value engineering with regard to the betterment of efficiency and functions may lead to the project implementation timing elongation. Here time revamping does not refer to time diminishing in the whole cases. his article deals with risk and value engineering conceptualizations at first. The germane reverberations effectuated due to its execution in Iran Khodro Corporation are regarded together with the joint features and amalgamation of the aforesaid entia; hence the proposed blueprint is submitted to be taken advantage of in engineering and industrial projects including Iran Khodro Corporation.Keywords: Management, risk engineering, value engineering, project manipulation, Iran Khodro.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 23525132 Development of Manufacturing Simulation Model for Semiconductor Fabrication
Authors: Syahril Ridzuan Ab Rahim, Ibrahim Ahmad, Mohd Azizi Chik, Ahmad Zafir Md. Rejab, and U. Hashim
Abstract:
This research presents the development of simulation modeling for WIP management in semiconductor fabrication. Manufacturing simulation modeling is needed for productivity optimization analysis due to the complex process flows involved more than 35 percent re-entrance processing steps more than 15 times at same equipment. Furthermore, semiconductor fabrication required to produce high product mixed with total processing steps varies from 300 to 800 steps and cycle time between 30 to 70 days. Besides the complexity, expansive wafer cost that potentially impact the company profits margin once miss due date is another motivation to explore options to experiment any analysis using simulation modeling. In this paper, the simulation model is developed using existing commercial software platform AutoSched AP, with customized integration with Manufacturing Execution Systems (MES) and Advanced Productivity Family (APF) for data collections used to configure the model parameters and data source. Model parameters such as processing steps cycle time, equipment performance, handling time, efficiency of operator are collected through this customization. Once the parameters are validated, few customizations are made to ensure the prior model is executed. The accuracy for the simulation model is validated with the actual output per day for all equipments. The comparison analysis from result of the simulation model compared to actual for achieved 95 percent accuracy for 30 days. This model later was used to perform various what if analysis to understand impacts on cycle time and overall output. By using this simulation model, complex manufacturing environment like semiconductor fabrication (fab) now have alternative source of validation for any new requirements impact analysis.Keywords: Advanced Productivity Family (APF), Complementary Metal Oxide Semiconductor (CMOS), Manufacturing Execution Systems (MES), Work In Progress (WIP).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 32205131 Unit Testing with Déjà-Vu Objects
Authors: Sharareh Afsharian, Andrea Bei, Marco Bianchi
Abstract:
In this paper we introduce a new unit test technique called déjà-vu object. Déjà-vu objects replace real objects used by classes under test, allowing the execution of isolated unit tests. A déjà-vu object is able to observe and record the behaviour of a real object during real sessions, and to replace it during unit tests, returning previously recorded results. Consequently déjà-vu object technique can be useful when a bottom-up development and testing strategy is adopted. In this case déjà-vu objects can increase test portability and test source code readability. At the same time they can reduce the time spent by programmers to develop test code and the risk of incompatibility during the switching between déjà-vu and production code.Keywords: Bottom-up testing approach, integration test, testportability, unit test.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 25955130 Spreading of Swirling Double–Concentric Jets at Low and High Pulsation Intensities
Authors: Shiferaw R. Jufar, Rong F. Huang, Ching M. Hsu
Abstract:
The spreading characteristics of acoustically excited swirling double-concentric jets were studied experimentally. The central jet was acoustically excited at low and high pulsation intensities. A smoke wire flow visualization and a hot-wire anemometer velocity measurement results show that excitation forces a vortex ring to roll-up from the edge of the central tube during each excitation period. At low pulsation intensities, the vortex ring evolves downstream, and eventually breaks up into turbulent eddies. At high pulsation intensities, the primary vortex ring evolves and a series of trailing vortex rings form during the same period of excitation. The trailing vortex rings accelerate while evolving downstream and overtake the primary vortex ring within the same cycle. In the process, the primary vortex ring becomes unstable and breaks up early. The effect of the fast traveling trailing vortex rings combined with the swirl motion of the annular flow improve jet spreading compared with the naturally evolving jets.Keywords: Acoustic excitation, double–concentric jets, flow control, swirling jet.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19815129 QoS Expectations in IP Networks: A Practical View
Authors: S. Arrizabalaga, A. Salterain, M. Domínguez, I. Alvaro
Abstract:
Traditionally, Internet has provided best-effort service to every user regardless of its requirements. However, as Internet becomes universally available, users demand more bandwidth and applications require more and more resources, and interest has developed in having the Internet provide some degree of Quality of Service. Although QoS is an important issue, the question of how it will be brought into the Internet has not been solved yet. Researches, due to the rapid advances in technology are proposing new and more desirable capabilities for the next generation of IP infrastructures. But neither all applications demand the same amount of resources, nor all users are service providers. In this way, this paper is the first of a series of papers that presents an architecture as a first step to the optimization of QoS in the Internet environment as a solution to a SMSE's problem whose objective is to provide public service to internet with certain Quality of Service expectations. The service provides new business opportunities, but also presents new challenges. We have designed and implemented a scalable service framework that supports adaptive bandwidth based on user demands, and the billing based on usage and on QoS. The developed application has been evaluated and the results show that traffic limiting works at optimum and so it does exceeding bandwidth distribution. However, some considerations are done and currently research is under way in two basic areas: (i) development and testing new transfer protocols, and (ii) developing new strategies for traffic improvements based on service differentiation.
Keywords: Differentiated Services, Linux, Quality of Service, queueing disciplines, web application.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19235128 Estimation of Thermal Conductivity of Nanofluids Using MD-Stochastic Simulation Based Approach
Authors: Sujoy Das, M. M. Ghosh
Abstract:
The thermal conductivity of a fluid can be significantly enhanced by dispersing nano-sized particles in it, and the resultant fluid is termed as "nanofluid". A theoretical model for estimating the thermal conductivity of a nanofluid has been proposed here. It is based on the mechanism that evenly dispersed nanoparticles within a nanofluid undergo Brownian motion in course of which the nanoparticles repeatedly collide with the heat source. During each collision a rapid heat transfer occurs owing to the solidsolid contact. Molecular dynamics (MD) simulation of the collision of nanoparticles with the heat source has shown that there is a pulselike pick up of heat by the nanoparticles within 20-100 ps, the extent of which depends not only on thermal conductivity of the nanoparticles, but also on the elastic and other physical properties of the nanoparticle. After the collision the nanoparticles undergo Brownian motion in the base fluid and release the excess heat to the surrounding base fluid within 2-10 ms. The Brownian motion and associated temperature variation of the nanoparticles have been modeled by stochastic analysis. Repeated occurrence of these events by the suspended nanoparticles significantly contributes to the characteristic thermal conductivity of the nanofluids, which has been estimated by the present model for a ethylene glycol based nanofluid containing Cu-nanoparticles of size ranging from 8 to 20 nm, with Gaussian size distribution. The prediction of the present model has shown a reasonable agreement with the experimental data available in literature.
Keywords: Brownian dynamics, Molecular dynamics, Nanofluid, Thermal conductivity.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22645127 A Model for Estimation of Efforts in Development of Software Systems
Authors: Parvinder S. Sandhu, Manisha Prashar, Pourush Bassi, Atul Bisht
Abstract:
Software effort estimation is the process of predicting the most realistic use of effort required to develop or maintain software based on incomplete, uncertain and/or noisy input. Effort estimates may be used as input to project plans, iteration plans, budgets. There are various models like Halstead, Walston-Felix, Bailey-Basili, Doty and GA Based models which have already used to estimate the software effort for projects. In this study Statistical Models, Fuzzy-GA and Neuro-Fuzzy (NF) Inference Systems are experimented to estimate the software effort for projects. The performances of the developed models were tested on NASA software project datasets and results are compared with the Halstead, Walston-Felix, Bailey-Basili, Doty and Genetic Algorithm Based models mentioned in the literature. The result shows that the NF Model has the lowest MMRE and RMSE values. The NF Model shows the best results as compared with the Fuzzy-GA based hybrid Inference System and other existing Models that are being used for the Effort Prediction with lowest MMRE and RMSE values.Keywords: Neuro-Fuzzy Model, Halstead Model, Walston-Felix Model, Bailey-Basili Model, Doty Model, GA Based Model, Genetic Algorithm.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 32275126 The Visualizer for Real-Time Analysis of Internet Trends
Authors: Radek Malinský, Ivan Jelínek
Abstract:
The current web has become a modern encyclopedia, where people share their thoughts and ideas on various topics around them. This kind of encyclopedia is very useful for other people who are looking for answers to their questions. However, with the growing popularity of social networking and blogging and ever expanding network services, there has also been a growing diversity of technologies along with a different structure of individual web sites. It is therefore difficult to directly find a relevant answer for a common Internet user. This paper presents a web application for the real-time end-to-end analysis of selected Internet trends where the trend can be whatever the people post online. The application integrates fully configurable tools for data collection and analysis using selected webometric algorithms, and for its chronological visualization to user. It can be assumed that the application facilitates the users to evaluate the quality of various products that are mentioned online.Keywords: Trend, visualizer, web analysis, web 2.0.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22415125 Methodology: A Review in Modelling and Predictability of Embankment in Soft Ground
Authors: Bhim Kumar Dahal
Abstract:
Transportation network development in the developing country is in rapid pace. The majority of the network belongs to railway and expressway which passes through diverse topography, landform and geological conditions despite the avoidance principle during route selection. Construction of such networks demand many low to high embankment which required improvement in the foundation soil. This paper is mainly focused on the various advanced ground improvement techniques used to improve the soft soil, modelling approach and its predictability for embankments construction. The ground improvement techniques can be broadly classified in to three groups i.e. densification group, drainage and consolidation group and reinforcement group which are discussed with some case studies. Various methods were used in modelling of the embankments from simple 1-dimensional to complex 3-dimensional model using variety of constitutive models. However, the reliability of the predictions is not found systematically improved with the level of sophistication. And sometimes the predictions are deviated more than 60% to the monitored value besides using same level of erudition. This deviation is found mainly due to the selection of constitutive model, assumptions made during different stages, deviation in the selection of model parameters and simplification during physical modelling of the ground condition. This deviation can be reduced by using optimization process, optimization tools and sensitivity analysis of the model parameters which will guide to select the appropriate model parameters.
Keywords: Embankment, ground improvement, modelling, model prediction.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9525124 Monitoring Blood Pressure Using Regression Techniques
Authors: Qasem Qananwah, Ahmad Dagamseh, Hiam AlQuran, Khalid Shaker Ibrahim
Abstract:
Blood pressure helps the physicians greatly to have a deep insight into the cardiovascular system. The determination of individual blood pressure is a standard clinical procedure considered for cardiovascular system problems. The conventional techniques to measure blood pressure (e.g. cuff method) allows a limited number of readings for a certain period (e.g. every 5-10 minutes). Additionally, these systems cause turbulence to blood flow; impeding continuous blood pressure monitoring, especially in emergency cases or critically ill persons. In this paper, the most important statistical features in the photoplethysmogram (PPG) signals were extracted to estimate the blood pressure noninvasively. PPG signals from more than 40 subjects were measured and analyzed and 12 features were extracted. The features were fed to principal component analysis (PCA) to find the most important independent features that have the highest correlation with blood pressure. The results show that the stiffness index means and standard deviation for the beat-to-beat heart rate were the most important features. A model representing both features for Systolic Blood Pressure (SBP) and Diastolic Blood Pressure (DBP) was obtained using a statistical regression technique. Surface fitting is used to best fit the series of data and the results show that the error value in estimating the SBP is 4.95% and in estimating the DBP is 3.99%.
Keywords: Blood pressure, noninvasive optical system, PCA, continuous monitoring.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6875123 Weighted Data Replication Strategy for Data Grid Considering Economic Approach
Authors: N. Mansouri, A. Asadi
Abstract:
Data Grid is a geographically distributed environment that deals with data intensive application in scientific and enterprise computing. Data replication is a common method used to achieve efficient and fault-tolerant data access in Grids. In this paper, a dynamic data replication strategy, called Enhanced Latest Access Largest Weight (ELALW) is proposed. This strategy is an enhanced version of Latest Access Largest Weight strategy. However, replication should be used wisely because the storage capacity of each Grid site is limited. Thus, it is important to design an effective strategy for the replication replacement task. ELALW replaces replicas based on the number of requests in future, the size of the replica, and the number of copies of the file. It also improves access latency by selecting the best replica when various sites hold replicas. The proposed replica selection selects the best replica location from among the many replicas based on response time that can be determined by considering the data transfer time, the storage access latency, the replica requests that waiting in the storage queue and the distance between nodes. Simulation results utilizing the OptorSim show our replication strategy achieve better performance overall than other strategies in terms of job execution time, effective network usage and storage resource usage.
Keywords: Data grid, data replication, simulation, replica selection, replica placement.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21105122 Expectation-Confirmation Model of Information System Continuance: A Meta-Analysis
Authors: Hui-Min Lai, Chin-Pin Chen, Yung-Fu Chang
Abstract:
The expectation-confirmation model (ECM) is one of the most widely used models for evaluating information system continuance, and this model has been extended to other study backgrounds, or expanded with other theoretical perspectives. However, combining ECM with other theories or investigating the background problem may produce some disparities, thus generating inaccurate conclusions. Habit is considered to be an important factor that influences the user’s continuance behavior. This paper thus critically examines seven pairs of relationships from the original ECM and the habit variable. A meta-analysis was used to tackle the development of ECM research over the last 10 years from a range of journals and conference papers published in 2005–2014. Forty-six journal articles and 19 conference papers were selected for analysis. The results confirm our prediction that a high effect size for the seven pairs of relationships was obtained (ranging from r=0.386 to r=0.588). Furthermore, a meta-analytic structural equation modeling was performed to simultaneously test all relationships. The results show that habit had a significant positive effect on continuance intention at p<=0.05 and that the six other pairs of relationships were significant at p<0.10. Based on the findings, we refined our original research model and an alternative model was proposed for understanding and predicting information system continuance. Some theoretical implications are also discussed.Keywords: Expectation-confirmation theory, expectation- confirmation model, meta-analysis, meta-analytic structural equation modeling.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 27325121 Experimental Studies on Treated Sub-base Soil with Fly Ash and Cement for Sustainable Design Recommendations
Authors: M. Jayakumar, Lau Chee Sing
Abstract:
The pavement constructions on soft and expansive soils are not durable and unable to sustain heavy traffic loading. As a result, pavement failures and settlement problems will occur very often even under light traffic loading due to cyclic and rolling effects. Geotechnical engineers have dwelled deeply into this matter, and adopt various methods to improve the engineering characteristics of soft fine-grained soils and expansive soils. The problematic soils are either replaced by good and better quality material or treated by using chemical stabilization with various binding materials. Increased the strength and durability are also the part of the sustainability drive to reduce the environment footprint of the built environment by the efficient use of resources and waste recycle materials. This paper presents a series of laboratory tests and evaluates the effect of cement and fly ash on the strength and drainage characteristics of soil in Miri. The tests were performed at different percentages of cement and fly ash by dry weight of soil. Additional tests were also performed on soils treated with the combinations of fly ash with cement and lime. The results of this study indicate an increase in unconfined compression strength and a decrease in hydraulic conductivity of the treated soil.
Keywords: Pavement failure, soft soil, sustainability, stabilization, fly ash, strength and permeability.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 34395120 Developing of Fragility Curve for Two-Span Simply Supported Concrete Bridge in Near-Fault Area
Authors: S. Shirazian, M.R. Ghayamghamian, G.R. Nouri
Abstract:
Bridges are one of the main components of transportation networks. They should be functional before and after earthquake for emergency services. Therefore we need to assess seismic performance of bridges under different seismic loadings. Fragility curve is one of the popular tools in seismic evaluations. The fragility curves are conditional probability statements, which give the probability of a bridge reaching or exceeding a particular damage level for a given intensity level. In this study, the seismic performance of a two-span simply supported concrete bridge is assessed. Due to usual lack of empirical data, the analytical fragility curve was developed by results of the dynamic analysis of bridge subjected to the different time histories in near-fault area.Keywords: Fragility curve, Seismic behavior, Time historyanalysis, Transportation Network.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 27965119 Motion Prediction and Motion Vector Cost Reduction during Fast Block Motion Estimation in MCTF
Authors: Karunakar A K, Manohara Pai M M
Abstract:
In 3D-wavelet video coding framework temporal filtering is done along the trajectory of motion using Motion Compensated Temporal Filtering (MCTF). Hence computationally efficient motion estimation technique is the need of MCTF. In this paper a predictive technique is proposed in order to reduce the computational complexity of the MCTF framework, by exploiting the high correlation among the frames in a Group Of Picture (GOP). The proposed technique applies coarse and fine searches of any fast block based motion estimation, only to the first pair of frames in a GOP. The generated motion vectors are supplied to the next consecutive frames, even to subsequent temporal levels and only fine search is carried out around those predicted motion vectors. Hence coarse search is skipped for all the motion estimation in a GOP except for the first pair of frames. The technique has been tested for different fast block based motion estimation algorithms over different standard test sequences using MC-EZBC, a state-of-the-art scalable video coder. The simulation result reveals substantial reduction (i.e. 20.75% to 38.24%) in the number of search points during motion estimation, without compromising the quality of the reconstructed video compared to non-predictive techniques. Since the motion vectors of all the pair of frames in a GOP except the first pair will have value ±1 around the motion vectors of the previous pair of frames, the number of bits required for motion vectors is also reduced by 50%.Keywords: Motion Compensated Temporal Filtering, predictivemotion estimation, lifted wavelet transform, motion vector
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16195118 Bi-Lateral Comparison between NIS-Egypt and NMISA-South Africa for the Calibration of an Optical Time Domain Reflectometer
Authors: Osama Terra, Mariesa Nel, Hatem Hussein
Abstract:
Calibration of Optical Time Domain Reflectometer (OTDR) has a crucial role for the accurate determination of fault locations and the accurate calculation of loss budget of long-haul optical fibre links during installation and repair. A comparison has been made between the Egyptian National Institute for Standards (NIS-Egypt) and the National Metrology institute of South Africa (NMISA-South Africa) for the calibration of an OTDR. The distance and the attenuation scales of a transfer OTDR have been calibrated by both institutes using their standards according to the standard IEC 61746-1 (2009). The results of this comparison have been compiled in this report.
Keywords: OTDR calibration, recirculating loop, concatenated method, standard fibre.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9355117 Scheduling of Bus Fleet Departure Time Based on Mathematical Model of Number of Bus Stops for Municipality Bus Organization
Authors: Ali Abdi Kordani, Hamid Bigdelirad, Sid Mohammad Boroomandrad
Abstract:
Operating Urban Bus Transit System is a phenomenon that has a major role in transporting passengers in cities. There are many factors involved in planning and operating an Urban Bus Transit System, one of which is selecting optimized number of stops and scheduling of bus fleet departure. In this paper, we tried to introduce desirable methodology to select number of stops and schedule properly. Selecting the right number of stops causes convenience in accessibility and reduction in travel time and finally increase in public preference of this transportation mode. The achieved results revealed that number of stops must reduce from 33 to 25. Also according to scheduling and conducted economic analysis, the number of buses must decrease from 17 to 11 to have the most appropriate status for the Bus Organization.
Keywords: Number of optimized stops, organizing bus system, scheduling, urban transit.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 7055116 Leaching of Flotation Concentrate of Oxide Copper Ore from Sepon Mine, Lao PDR
Authors: C. Rattanakawin, S. Vasailor
Abstract:
Acid leaching of flotation concentrate of oxide copper ore containing mainly of malachite was performed in a standard agitation tank with various parameters. The effects of solid to liquid ratio, sulfuric acid concentration, agitation speed, leaching temperature and time were examined to get proper conditions. The best conditions are 1:8 solid to liquid ratio, 10% concentration by weight, 250 rev/min, 30 oC and 5-min leaching time in respect. About 20% Cu grade assayed by atomic absorption technique with 98% copper recovery was obtained from these combined optimum conditions. Dissolution kinetics of the concentrate was approximated as a logarithmic function. As a result, the first-order reaction rate is suggested from this leaching study.Keywords: Agitation leaching, dissolution kinetics, flotation concentrate, oxide copper ore, sulfuric acid.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 692