Search results for: Continuous Factory Planning
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 1853

Search results for: Continuous Factory Planning

533 Finding Pareto Optimal Front for the Multi-Mode Time, Cost Quality Trade-off in Project Scheduling

Authors: H. Iranmanesh, M. R. Skandari, M. Allahverdiloo

Abstract:

Project managers are the ultimate responsible for the overall characteristics of a project, i.e. they should deliver the project on time with minimum cost and with maximum quality. It is vital for any manager to decide a trade-off between these conflicting objectives and they will be benefited of any scientific decision support tool. Our work will try to determine optimal solutions (rather than a single optimal solution) from which the project manager will select his desirable choice to run the project. In this paper, the problem in project scheduling notated as (1,T|cpm,disc,mu|curve:quality,time,cost) will be studied. The problem is multi-objective and the purpose is finding the Pareto optimal front of time, cost and quality of a project (curve:quality,time,cost), whose activities belong to a start to finish activity relationship network (cpm) and they can be done in different possible modes (mu) which are non-continuous or discrete (disc), and each mode has a different cost, time and quality . The project is constrained to a non-renewable resource i.e. money (1,T). Because the problem is NP-Hard, to solve the problem, a meta-heuristic is developed based on a version of genetic algorithm specially adapted to solve multi-objective problems namely FastPGA. A sample project with 30 activities is generated and then solved by the proposed method.

Keywords: FastPGA, Multi-Execution Activity Mode, ParetoOptimality, Project Scheduling, Time-Cost-Quality Trade-Off.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1675
532 Reversible, Embedded and Highly Scalable Image Compression System

Authors: Federico Pérez González, Iñaki Goirizelaia Ordorika, Pedro Iriondo Bengoa

Abstract:

In this work a new method for low complexity image coding is presented, that permits different settings and great scalability in the generation of the final bit stream. This coding presents a continuous-tone still image compression system that groups loss and lossless compression making use of finite arithmetic reversible transforms. Both transformation in the space of color and wavelet transformation are reversible. The transformed coefficients are coded by means of a coding system in depending on a subdivision into smaller components (CFDS) similar to the bit importance codification. The subcomponents so obtained are reordered by means of a highly configure alignment system depending on the application that makes possible the re-configure of the elements of the image and obtaining different importance levels from which the bit stream will be generated. The subcomponents of each importance level are coded using a variable length entropy coding system (VBLm) that permits the generation of an embedded bit stream. This bit stream supposes itself a bit stream that codes a compressed still image. However, the use of a packing system on the bit stream after the VBLm allows the realization of a final highly scalable bit stream from a basic image level and one or several improvement levels.

Keywords: Image compression, wavelet transform, highly scalable, reversible transform, embedded, subcomponents.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1291
531 Urban Growth Prediction in Athens, Greece, Using Artificial Neural Networks

Authors: D. Triantakonstantis, D. Stathakis

Abstract:

Urban areas have been expanded throughout the globe. Monitoring and modelling urban growth have become a necessity for a sustainable urban planning and decision making. Urban prediction models are important tools for analyzing the causes and consequences of urban land use dynamics. The objective of this research paper is to analyze and model the urban change, which has been occurred from 1990 to 2000 using CORINE land cover maps. The model was developed using drivers of urban changes (such as road distance, slope, etc.) under an Artificial Neural Network modelling approach. Validation was achieved using a prediction map for 2006 which was compared with a real map of Urban Atlas of 2006. The accuracy produced a Kappa index of agreement of 0,639 and a value of Cramer's V of 0,648. These encouraging results indicate the importance of the developed urban growth prediction model which using a set of available common biophysical drivers could serve as a management tool for the assessment of urban change.

Keywords: Artificial Neural Networks, CORINE, Urban Atlas, Urban Growth Prediction.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3441
530 A Dose Distribution Approach Using Monte Carlo Simulation in Dosimetric Accuracy Calculation for Treating the Lung Tumor

Authors: Md Abdullah Al Mashud, M. Tariquzzaman, M. Jahangir Alam, Tapan Kumar Godder, M. Mahbubur Rahman

Abstract:

This paper presents a Monte Carlo (MC) method-based dose distributions on lung tumor for 6 MV photon beam to improve the dosimetric accuracy for cancer treatment. The polystyrene which is tissue equivalent material to the lung tumor density is used in this research. In the empirical calculations, TRS-398 formalism of IAEA has been used, and the setup was made according to the ICRU recommendations. The research outcomes were compared with the state-of-the-art experimental results. From the experimental results, it is observed that the proposed based approach provides more accurate results and improves the accuracy than the existing approaches. The average %variation between measured and TPS simulated values was obtained 1.337±0.531, which shows a substantial improvement comparing with the state-of-the-art technology.

Keywords: Lung tumor, Monte Carlo, polystyrene, elekta synergy, Monaco Planning System.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1226
529 The Kinetic of Biogas Production Rate from Cattle Manure in Batch Mode

Authors: Budiyono, I N. Widiasa, S. Johari, Sunarso

Abstract:

In this study, the kinetic of biogas production was studied by performing a series laboratory experiment using rumen fluid of animal ruminant as inoculums. Cattle manure as substrate was inoculated by rumen fluid to the anaerobic biodigester. Laboratory experiments using 400 ml biodigester were performed in batch operation mode. Given 100 grams of fresh cattle manure was fed to each biodigester and mixed with rumen fluid by manure : rumen weight ratio of 1:1 (MR11). The operating temperatures were varied at room temperature and 38.5 oC. The cumulative volume of biogas produced was used to measure the biodigester performance. The research showed that the rumen fluid inoculated to biodigester gave significant effect to biogas production (P<0.05). Rumen fluid inoculums caused biogas production rate and efficiency increase two to three times in compare to manure substrate without rumen fluid. With the rumen fluid inoculums, gave the kinetic parameters of biogas production i.e biogas production rate constants (U), maximum biogas production (A), and minimum time to produce biogas (λ) are 3.89 ml/(gVS.day); 172.51 (ml/gVS); dan 7.25 days, respectively. While the substrate without rumen fluid gave the kinetic parameters U, A, and λ are 1.74 ml/(gVS.day); 73.81 (ml/gVS); dan 14.75 days, respectively. The future work will be carried out to study the dynamics of biogas production if both the rumen inoculums and manure are fed in the continuous system.

Keywords: rumen fluid, inoculums, anaerobic digestion, biogasproduction.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3019
528 An Anisotropic Model of Damage and Unilateral Effect for Brittle Materials

Authors: José Julio de C. Pituba

Abstract:

This work deals with the initial applications and formulation of an anisotropic plastic-damage constitutive model proposed for non-linear analysis of reinforced concrete structures submitted to a loading with change of the sign. The original constitutive model is based on the fundamental hypothesis of energy equivalence between real and continuous medium following the concepts of the Continuum Damage Mechanics. The concrete is assumed as an initial elastic isotropic medium presenting anisotropy, permanent strains and bimodularity (distinct elastic responses whether traction or compression stress states prevail) induced by damage evolution. In order to take into account the bimodularity, two damage tensors governing the rigidity in tension or compression regimes are introduced. Then, some conditions are introduced in the original version of the model in order to simulate the damage unilateral effect. The three-dimensional version of the proposed model is analyzed in order to validate its formulation when compared to micromechanical theory. The one-dimensional version of the model is applied in the analyses of a reinforced concrete beam submitted to a loading with change of the sign. Despite the parametric identification problems, the initial applications show the good performance of the model.

Keywords: Damage model, plastic strain, unilateral effect.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1820
527 Rural Tourism Planning from the Perspective of Water Resource Protection and Regional Integration: Taking Villages along Tongji Lake as an Example

Authors: Pianpian Zhang, Qingping Luo

Abstract:

Currently, there is a great tendency that more and more villages in China are trying to increase income by development of tourism. Especially in Zhejiang Province, 'Beautiful Rural Construction' provides an excellent opportunity for the development of tourism. In this context, development orientation, transportation routes and tourism service facilities are analyzed under the perspective of water resources protection and regional integration based on the development tourism industry of the six villages in Pujiang County, Zhejiang Province as a research object. In the program, the biggest issue is the contradiction between the ecological protection of the water and the development of economy. How to deal with the relationship between protection and development is the key to the design of this case. Furthermore, the six villages are regarded as a whole, connecting to each other by the system of five-path and the landscape along the lake. Every village has its own features, but cannot develop without one another. The article is actively exploring for suggestions and countermeasures to promote the development premised on protection and based on a regional view.

Keywords: Development, integration, protection, rural tourism.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 872
526 Solving the Set Covering Problem Using the Binary Cat Swarm Optimization Metaheuristic

Authors: Broderick Crawford, Ricardo Soto, Natalia Berrios, Eduardo Olguin

Abstract:

In this paper, we present a binary cat swarm optimization for solving the Set covering problem. The set covering problem is a well-known NP-hard problem with many practical applications, including those involving scheduling, production planning and location problems. Binary cat swarm optimization is a recent swarm metaheuristic technique based on the behavior of discrete cats. Domestic cats show the ability to hunt and are curious about moving objects. The cats have two modes of behavior: seeking mode and tracing mode. We illustrate this approach with 65 instances of the problem from the OR-Library. Moreover, we solve this problem with 40 new binarization techniques and we select the technical with the best results obtained. Finally, we make a comparison between results obtained in previous studies and the new binarization technique, that is, with roulette wheel as transfer function and V3 as discretization technique.

Keywords: Binary cat swarm optimization, set covering problem, metaheuristic, binarization methods.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2321
525 Remote Sensing, GIS, and AHP for Assessing Physical Vulnerability to Tsunami Hazard

Authors: Abu Bakar Sambah, Fusanori Miura

Abstract:

Remote sensing image processing, spatial data analysis through GIS approach, and analytical hierarchy process were introduced in this study for assessing the vulnerability area and inundation area due to tsunami hazard in the area of Rikuzentakata, Iwate Prefecture, Japan. Appropriate input parameters were derived from GSI DEM data, ALOS AVNIR-2, and field data. We used the parameters of elevation, slope, shoreline distance, and vegetation density. Five classes of vulnerability were defined and weighted via pairwise comparison matrix. The assessment results described that 14.35km2 of the study area was under tsunami vulnerability zone. Inundation areas are those of high and slightly high vulnerability. The farthest area reached by a tsunami was about 7.50km from the shoreline and shows that rivers act as flooding strips that transport tsunami waves into the hinterland. This study can be used for determining a priority for land-use planning in the scope of tsunami hazard risk management.

Keywords: AHP, GIS, remote sensing, tsunami vulnerability.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3326
524 Optimal Placement of DG in Distribution System to Mitigate Power Quality Disturbances

Authors: G.V.K Murthy, S. Sivanagaraju, S. Satyanarayana, B. Hanumantha Rao

Abstract:

Distributed Generation (DG) systems are considered an integral part in future distribution system planning. Appropriate size and location of distributed generation plays a significant role in minimizing power losses in distribution systems. Among the benefits of distributed generation is the reduction in active power losses, which can improve the system performance, reliability and power quality. In this paper, Artificial Bee Colony (ABC) algorithm is proposed to determine the optimal DG-unit size and location by loss sensitivity index in order to minimize the real power loss, total harmonic distortion (THD) and voltage sag index improvement. Simulation study is conducted on 69-bus radial test system to verify the efficacy of the proposed method.

Keywords: Distributed generation, artificial bee colony method, loss reduction, radial distribution network.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2848
523 Study on Changes of Land Use impacting the Process of Urbanization, by Using Landsat Data in African Regions: A Case Study in Kigali, Rwanda

Authors: Delphine Mukaneza, Lin Qiao, Wang Pengxin, Li Yan, Chen Yingyi

Abstract:

Human activities on land use make the land-cover gradually change or transit. In this study, we examined the use of Landsat TM data to detect the land use change of Kigali between 1987 and 2009 using remote sensing techniques and analysis of data using ENVI and ArcGIS, a GIS software. Six different categories of land use were distinguished: bare soil, built up land, wetland, water, vegetation, and others. With remote sensing techniques, we analyzed land use data in 1987, 1999 and 2009, changed areas were found and a dynamic situation of land use in Kigali city was found during the 22 years studied. According to relevant Landsat data, the research focused on land use change in accordance with the role of remote sensing in the process of urbanization. The result of the work has shown the rapid increase of built up land between 1987 and 1999 and a big decrease of vegetation caused by the rebuild of the city after the 1994 genocide, while in the period of 1999 to 2009 there was a reduction in built up land and vegetation, after the authority of Kigali city established, a Master Plan where all constructions which were not in the range of the master Plan were destroyed. Rwanda's capital, Kigali City, through the expansion of the urban area, it is increasing the internal employment rate and attracts business investors and the service sector to improve their economy, which will increase the population growth and provide a better life. The overall planning of the city of Kigali considers the environment, land use, infrastructure, cultural and socio-economic factors, the economic development and population forecast, urban development, and constraints specification. To achieve the above purpose, the Government has set for the overall planning of city Kigali, different stages of the detailed description of the design, strategy and action plan that would guide Kigali planners and members of the public in the future to have more detailed regional plans and practical measures. Thus, land use change is significantly the performance of Kigali active human area, which plays an important role for the country to take certain decisions. Another area to take into account is the natural situation of Kigali city. Agriculture in the region does not occupy a dominant position, and with the population growth and socio-economic development, the construction area will gradually rise and speed up the process of urbanization. Thus, as a developing country, Rwanda's population continues to grow and there is low rate of utilization of land, where urbanization remains low. As mentioned earlier, the 1994 genocide massacres, population growth and urbanization processes, have been the factors driving the dramatic changes in land use. The focus on further research would be on analysis of Rwanda’s natural resources, social and economic factors that could be, the driving force of land use change.

Keywords: Land use change, urbanization, Kigali City, Landsat.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1048
522 ERP Implementation in Iran: (A Successful Experience in DGC)

Authors: Mohammad Reza Ostad Ali Naghi Kashani

Abstract:

Nowadays, the amounts of companies which tend to have an Enterprise Resource Planning (ERP) application are increasing. Although ERP projects are expensive, time consuming, and complex, there are some successful experiences. These days, developing countries are striving to implement ERP projects successfully; however, there are many obstacles. Therefore, these projects would be failed or partially failed. This paper concerns the implementation of a successful ERP implementation, IFS, in Iran at Dana Geophysics Company (DGC). After a short review of ERP and ERP market in Iran, we propose a three phases deployment methodology (phase 1: Preparation and Business Process Management (BPM) phase 2: implementation and phase 3: testing, golive-1 (pilot) and golive-2 (final)). Then, we present five guidelines (Project Management, Change Management, Business Process Management (BPM), Training& Knowledge Management, and Technical Management), which were chose as work streams. In this case study we present lessons learned in Project management and Business process Management.

Keywords: Business Process Management, Critical Success Factors, ERP, Project Management.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2937
521 Finding Pareto Optimal Front for the Multi- Mode Time, Cost Quality Trade-off in Project Scheduling

Authors: H. Iranmanesh, M. R. Skandari, M. Allahverdiloo

Abstract:

Project managers are the ultimate responsible for the overall characteristics of a project, i.e. they should deliver the project on time with minimum cost and with maximum quality. It is vital for any manager to decide a trade-off between these conflicting objectives and they will be benefited of any scientific decision support tool. Our work will try to determine optimal solutions (rather than a single optimal solution) from which the project manager will select his desirable choice to run the project. In this paper, the problem in project scheduling notated as (1,T|cpm,disc,mu|curve:quality,time,cost) will be studied. The problem is multi-objective and the purpose is finding the Pareto optimal front of time, cost and quality of a project (curve:quality,time,cost), whose activities belong to a start to finish activity relationship network (cpm) and they can be done in different possible modes (mu) which are non-continuous or discrete (disc), and each mode has a different cost, time and quality . The project is constrained to a non-renewable resource i.e. money (1,T). Because the problem is NP-Hard, to solve the problem, a meta-heuristic is developed based on a version of genetic algorithm specially adapted to solve multi-objective problems namely FastPGA. A sample project with 30 activities is generated and then solved by the proposed method.

Keywords: FastPGA, Multi-Execution Activity Mode, Pareto Optimality, Project Scheduling, Time-Cost-Quality Trade-Off.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1798
520 On Pooling Different Levels of Data in Estimating Parameters of Continuous Meta-Analysis

Authors: N. R. N. Idris, S. Baharom

Abstract:

A meta-analysis may be performed using aggregate data (AD) or an individual patient data (IPD). In practice, studies may be available at both IPD and AD level. In this situation, both the IPD and AD should be utilised in order to maximize the available information. Statistical advantages of combining the studies from different level have not been fully explored. This study aims to quantify the statistical benefits of including available IPD when conducting a conventional summary-level meta-analysis. Simulated meta-analysis were used to assess the influence of the levels of data on overall meta-analysis estimates based on IPD-only, AD-only and the combination of IPD and AD (mixed data, MD), under different study scenario. The percentage relative bias (PRB), root mean-square-error (RMSE) and coverage probability were used to assess the efficiency of the overall estimates. The results demonstrate that available IPD should always be included in a conventional meta-analysis using summary level data as they would significantly increased the accuracy of the estimates.On the other hand, if more than 80% of the available data are at IPD level, including the AD does not provide significant differences in terms of accuracy of the estimates. Additionally, combining the IPD and AD has moderating effects on the biasness of the estimates of the treatment effects as the IPD tends to overestimate the treatment effects, while the AD has the tendency to produce underestimated effect estimates. These results may provide some guide in deciding if significant benefit is gained by pooling the two levels of data when conducting meta-analysis.

Keywords: Aggregate data, combined-level data, Individual patient data, meta analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1734
519 The Aspect of the Human Bias in Decision Making within Quality Management Systems & LEAN Theory

Authors: Adriana Ávila Zúñiga Nordfjeld

Abstract:

This paper provides a literature review to document the state of the art with respect to handling “human bias” in decision making within the established quality management systems (QMS) and LEAN theory, in the context of shipbuilding. Previous research shows that in shipbuilding there is a huge deviation from the planned man-hours under the project management to the actual man-hours used because of errors in planning and reworks caused by human bias in the information flows, among others. This reduces the efficiency, and increases operational costs. Thus, the research question is how QMS and LEAN handle biases. The findings show the gap in studying the integration of methods to handle human bias in decision making into QMS and lean, not only within shipbuilding, but in general. Theoretical and practical implications are discussed for researchers and practitioners in the areas of decision making, QMS and LEAN, and future research is suggested.

Keywords: Human bias, decision making, LEAN Shipbuilding, quality management systems.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2952
518 Effect of Chemical Pretreatments and Dehydration Methods on Quality Characteristics of Tomato Powder and Its Storage Stability

Authors: Reihaneh Ahmadzadeh Ghavidel, Mehdi Ghiafeh Davoodi

Abstract:

Dehydration process was carried out for tomato slices of var. Avinash after giving different pre-treatments such as calcium chloride (CaCl2), potassium metabisulphite (KMS), calcium chloride and potassium metabisulphite (CaCl2 +KMS), and sodium chloride (NaCl). Untreated samples served as control. Solar drier and continuous conveyor (tunnel) drier were used for dehydration. Quality characteristics of tomato slices viz. moisture content, sugar, titratable acidity, lycopene content, dehydration ratio, rehydration ratio and non-enzymatic browning as affected by dehydration process were studied. Storage study was also carried out for a period of six months for tomato powder packed into different types of packaging materials viz. metalized polyester (MP) film and low density poly ethylene (LDPE). Changes in lycopene content and non-enzymatic browning (NEB) were estimated during storage at room temperature. Pretreatment of 5 mm thickness of tomato slices with calcium chloride in combination with potassium metabisulphite and drying using a tunnel drier with subsequent storage of product in metalized polyester bags was selected as the best process.

Keywords: Drying pretreatments, Solar drying, Tomato powder, Tunnel drying

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2825
517 Impact of Revenue Gap on Budget Deficit, Debt Burden and Economic Growth: An Evidence from Pakistan

Authors: M. W. Siddiqi, M. Ilyas

Abstract:

Availability and mobilization of revenue is the main essential with which an economy is managed and run. While planning or while making the budgets nations set revenue targets to be achieved. But later when the accounts are closed the actual collections of revenue through taxes or even the non-tax revenue collection would invariably be different as compared to the initial estimates and targets set to be achieved. This revenue-gap distorts the whole system and the economy disturbing all the major macroeconomic indicators. This study is aimed to find out short and long term impact of revenue gap on budget deficit, debt burden and economic growth on the economy of Pakistan. For this purpose the study uses autoregressive distributed lag approach to cointegration and error correction mechanism on three different models for the period 1980 to 2009. The empirical results show that revenue gap has a short and long run relationship with economic growth and budget deficit. However, revenue gap has no impact on debt burden.

Keywords: Revenue Gap, Economic Growth, Budget Deficit, Debt Burden

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2809
516 Modeling and Simulation of Flow Shop Scheduling Problem through Petri Net Tools

Authors: Joselito Medina Marin, Norberto Hernández Romero, Juan Carlos Seck Tuoh Mora, Erick S. Martinez Gomez

Abstract:

The Flow Shop Scheduling Problem (FSSP) is a typical problem that is faced by production planning managers in Flexible Manufacturing Systems (FMS). This problem consists in finding the optimal scheduling to carry out a set of jobs, which are processed in a set of machines or shared resources. Moreover, all the jobs are processed in the same machine sequence. As in all the scheduling problems, the makespan can be obtained by drawing the Gantt chart according to the operations order, among other alternatives. On this way, an FMS presenting the FSSP can be modeled by Petri nets (PNs), which are a powerful tool that has been used to model and analyze discrete event systems. Then, the makespan can be obtained by simulating the PN through the token game animation and incidence matrix. In this work, we present an adaptive PN to obtain the makespan of FSSP by applying PN analytical tools.

Keywords: Flow-shop scheduling problem, makespan, Petri nets, state equation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1731
515 Minimizing Energy Consumption in Wireless Sensor Networks using Binary Integer Linear Programming

Authors: Chompunut Jantarasorn, Chutima Prommak

Abstract:

The important issue considered in the widespread deployment of Wireless Sensor Networks (WSNs) is an efficiency of the energy consumption. In this paper, we present a study of the optimal relay station planning problems using Binary Integer Linear Programming (BILP) model to minimize the energy consumption in WSNs. Our key contribution is that the proposed model not only ensures the required network lifetime but also guarantees the radio connectivity at high level of communication quality. Specially, we take into account effects of noise, signal quality limitation and bit error rate characteristics. Numerical experiments were conducted in various network scenarios. We analyzed the effects of different sensor node densities and distribution on the energy consumption.

Keywords: Binary Integer Linear Programming, BILP, Energy consumption, Optimal node placement and Wireless sensor networks.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2193
514 Active Segment Selection Method in EEG Classification Using Fractal Features

Authors: Samira Vafaye Eslahi

Abstract:

BCI (Brain Computer Interface) is a communication machine that translates brain massages to computer commands. These machines with the help of computer programs can recognize the tasks that are imagined. Feature extraction is an important stage of the process in EEG classification that can effect in accuracy and the computation time of processing the signals. In this study we process the signal in three steps of active segment selection, fractal feature extraction, and classification. One of the great challenges in BCI applications is to improve classification accuracy and computation time together. In this paper, we have used student’s 2D sample t-statistics on continuous wavelet transforms for active segment selection to reduce the computation time. In the next level, the features are extracted from some famous fractal dimension estimation of the signal. These fractal features are Katz and Higuchi. In the classification stage we used ANFIS (Adaptive Neuro-Fuzzy Inference System) classifier, FKNN (Fuzzy K-Nearest Neighbors), LDA (Linear Discriminate Analysis), and SVM (Support Vector Machines). We resulted that active segment selection method would reduce the computation time and Fractal dimension features with ANFIS analysis on selected active segments is the best among investigated methods in EEG classification.

Keywords: EEG, Student’s t- statistics, BCI, Fractal Features, ANFIS, FKNN.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2109
513 Forecasting Direct Normal Irradiation at Djibouti Using Artificial Neural Network

Authors: Ahmed Kayad Abdourazak, Abderafi Souad, Zejli Driss, Idriss Abdoulkader Ibrahim

Abstract:

In this paper Artificial Neural Network (ANN) is used to predict the solar irradiation in Djibouti for the first Time that is useful to the integration of Concentrating Solar Power (CSP) and sites selections for new or future solar plants as part of solar energy development. An ANN algorithm was developed to establish a forward/reverse correspondence between the latitude, longitude, altitude and monthly solar irradiation. For this purpose the German Aerospace Centre (DLR) data of eight Djibouti sites were used as training and testing in a standard three layers network with the back propagation algorithm of Lavenber-Marquardt. Results have shown a very good agreement for the solar irradiation prediction in Djibouti and proves that the proposed approach can be well used as an efficient tool for prediction of solar irradiation by providing so helpful information concerning sites selection, design and planning of solar plants.

Keywords: Artificial neural network, solar irradiation, concentrated solar power, Lavenberg-Marquardt.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1071
512 Optimizing Usability Testing with Collaborative Method in an E-Commerce Ecosystem

Authors: Markandeya Kunchi

Abstract:

Usability testing (UT) is one of the vital steps in the User-centred design (UCD) process when designing a product. In an e-commerce ecosystem, UT becomes primary as new products, features, and services are launched very frequently. And, there are losses attached to the company if an unusable and inefficient product is put out to market and is rejected by customers. This paper tries to answer why UT is important in the product life-cycle of an E-commerce ecosystem. Secondary user research was conducted to find out work patterns, development methods, type of stakeholders, and technology constraints, etc. of a typical E-commerce company. Qualitative user interviews were conducted with product managers and designers to find out the structure, project planning, product management method and role of the design team in a mid-level company. The paper tries to address the usual apprehensions of the company to inculcate UT within the team. As well, it stresses upon factors like monetary resources, lack of usability expert, narrow timelines, and lack of understanding of higher management as some primary reasons. Outsourcing UT to vendors is also very prevalent with mid-level e-commerce companies, but it has its own severe repercussions like very little team involvement, huge cost, misinterpretation of the findings, elongated timelines, and lack of empathy towards the customer, etc. The shortfalls of the unavailability of a UT process in place within the team and conducting UT through vendors are bad user experiences for customers while interacting with the product, badly designed products which are neither useful and nor utilitarian. As a result, companies see dipping conversions rates in apps and websites, huge bounce rates and increased uninstall rates. Thus, there was a need for a more lean UT system in place which could solve all these issues for the company. This paper highlights on optimizing the UT process with a collaborative method. The degree of optimization and structure of collaborative method is the highlight of this paper. Collaborative method of UT is one in which the centralised design team of the company takes for conducting and analysing the UT. The UT is usually a formative kind where designers take findings into account and uses in the ideation process. The success of collaborative method of UT is due to its ability to sync with the product management method employed by the company or team. The collaborative methods focus on engaging various teams (design, marketing, product, administration, IT, etc.) each with its own defined roles and responsibility in conducting a smooth UT with users In-house. The paper finally highlights the positive results of collaborative UT method after conducting more than 100 In-lab interviews with users across the different lines of businesses. Some of which are the improvement of interaction between stakeholders and the design team, empathy towards users, improved design iteration, better sanity check of design solutions, optimization of time and money, effective and efficient design solution. The future scope of collaborative UT is to make this method leaner, by reducing the number of days to complete the entire project starting from planning between teams to publishing the UT report.

Keywords: Usability testing, collaborative method, e-commerce, product management method.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 657
511 Case Studies of CSAMT Method Applied to Study of Complex Rock Mass Structure and Hidden Tectonic

Authors: Yuxin Chen, Qingyun Di, C. Dinis da Gama

Abstract:

In projects like waterpower, transportation and mining, etc., proving up the rock-mass structure and hidden tectonic to estimate the geological body-s activity is very important. Integrating the seismic results, drilling and trenching data, CSAMT method was carried out at a planning dame site in southwest China to evaluate the stability of a deformation. 2D and imitated 3D inversion resistivity results of CSAMT method were analyzed. The results indicated that CSAMT was an effective method for defining an outline of deformation body to several hundred meters deep; the Lung Pan Deformation was stable in natural conditions; but uncertain after the future reservoir was impounded. This research presents a good case study of the fine surveying and research on complex geological structure and hidden tectonic in engineering project.

Keywords: CSAMT Surveying, Deformation Stability.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2434
510 A Systematic Review on the Integration of Project Management with Organizational Flows

Authors: Maurício Covolan Rosito, Ricardo Melo Bastos

Abstract:

Software projects are very dynamic and require recurring adjustments of their project plans. These settings can be understood as reconfigurations in the schedule, in the resources allocation and other design elements. Yet, during the planning and execution of a software project, the integration of specific activities in the projects with the activities that take part in the organization-s common activity flow should be considered. This article presents the results from a systematic review of aspects related to software projects- dynamic reconfiguration emphasizing the integration of project management with the organizational flows. A series of studies was analyzed from the year 2000 to the present. The results of this work show that there is a diversity of techniques and strategies for dynamic reconfiguration of software projects-. However, few approaches consider the integration of software project activities with the activities that take part in the organization-s common workflow.

Keywords: Dynamic Reconfiguration, Organizational workflows, Project Management, Systematic Review

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1727
509 Obstacle Classification Method Based On 2D LIDAR Database

Authors: Moohyun Lee, Soojung Hur, Yongwan Park

Abstract:

We propose obstacle classification method based on 2D LIDAR Database. The existing obstacle classification method based on 2D LIDAR, has an advantage in terms of accuracy and shorter calculation time. However, it was difficult to classifier the type of obstacle and therefore accurate path planning was not possible. In order to overcome this problem, a method of classifying obstacle type based on width data of obstacle was proposed. However, width data was not sufficient to improve accuracy. In this paper, database was established by width and intensity data; the first classification was processed by the width data; the second classification was processed by the intensity data; classification was processed by comparing to database; result of obstacle classification was determined by finding the one with highest similarity values. An experiment using an actual autonomous vehicle under real environment shows that calculation time declined in comparison to 3D LIDAR and it was possible to classify obstacle using single 2D LIDAR.

Keywords: Obstacle, Classification, LIDAR, Segmentation, Width, Intensity, Database.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3438
508 Improvement of Voltage Profile of Grid Integrated Wind Distributed Generation by SVC

Authors: Fariba Shavakhi Zavareh, Hadi Fotoohabadi, Reza Sedaghati

Abstract:

Due to the continuous increment of the load demand, identification of weaker buses, improvement of voltage profile and power losses in the context of the voltage stability problems has become one of the major concerns for the larger, complex, interconnected power systems. The objective of this paper is to review the impact of Flexible AC Transmission System (FACTS) controller in Wind generators connected electrical network for maintaining voltage stability. Wind energy could be the growing renewable energy due to several advantages. The influence of wind generators on power quality is a significant issue; non uniform power production causes variations in system voltage and frequency. Therefore, wind farm requires high reactive power compensation; the advances in high power semiconducting devices have led to the development of FACTS. The FACTS devices such as for example SVC inject reactive power into the system which helps in maintaining a better voltage profile. The performance is evaluated on an IEEE 14 bus system, two wind generators are connected at low voltage buses to meet the increased load demand and SVC devices are integrated at the buses with wind generators to keep voltage stability. Power flows, nodal voltage magnitudes and angles of the power network are obtained by iterative solutions using MIPOWER.

Keywords: Voltage Profile, FACTS Device, SVC, Distributed Generation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2656
507 Numerical Modeling of Waves and Currents by Using a Hydro-Sedimentary Model

Authors: Mustapha Kamel Mihoubi, Hocine Dahmani

Abstract:

Over recent years much progress has been achieved in the fields of numerical modeling shoreline processes: waves, currents, waves and current. However, there are still some problems in the existing models to link the on the first, the hydrodynamics of waves and currents and secondly, the sediment transport processes and due to the variability in time, space and interaction and the simultaneous action of wave-current near the shore. This paper is the establishment of a numerical modeling to forecast the sediment transport from development scenarios of harbor structure. It is established on the basis of a numerical simulation of a water-sediment model via a 2D model using a set of codes calculation MIKE 21-DHI software. This is to examine the effect of the sediment transport drivers following the dominant incident wave in the direction to pass input harbor work under different variants planning studies to find the technical and economic limitations to the sediment transport and protection of the harbor structure optimum solution.

Keywords: Swell, current, radiation, stress, mesh, MIKE21, sediment.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1342
506 Continuous Fixed Bed Reactor Application for Decolourization of Textile Effluent by Adsorption on NaOH Treated Eggshell

Authors: M. Chafi, S. Akazdam, C. Asrir, L. Sebbahi, B. Gourich, N. Barka, M. Essahli

Abstract:

Fixed bed adsorption has become a frequently used industrial application in wastewater treatment processes. Various low cost adsorbents have been studied for their applicability in treatment of different types of effluents. In this work, the intention of the study was to explore the efficacy and feasibility for azo dye, Acid Orange 7 (AO7) adsorption onto fixed bed column of NaOH Treated eggshell (TES). The effect of various parameters like flow rate, initial dye concentration, and bed height were exploited in this study. The studies confirmed that the breakthrough curves were dependent on flow rate, initial dye concentration solution of AO7 and bed depth. The Thomas, Yoon–Nelson, and Adams and Bohart models were analysed to evaluate the column adsorption performance. The adsorption capacity, rate constant and correlation coefficient associated to each model for column adsorption was calculated and mentioned. The column experimental data were fitted well with Thomas model with coefficients of correlation R2 ≥0.93 at different conditions but the Yoon–Nelson, BDST and Bohart–Adams model (R2=0.911), predicted poor performance of fixed-bed column. The (TES) was shown to be suitable adsorbent for adsorption of AO7 using fixed-bed adsorption column.

Keywords: Adsorption models, acid orange 7, bed depth, breakthrough, dye adsorption, fixed-bed column, treated eggshell.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2262
505 Prioritising the TQM Enablers and IT Resources in the ICT Industry: An AHP Approach

Authors: Suby Khanam, Jamshed Siddiqui, Faisal Talib

Abstract:

Total Quality Management (TQM) is a managerial approach that improves the competitiveness of the industry, meanwhile Information technology (IT) was introduced with TQM for handling the technical issues which is supported by quality experts for fulfilling the customers’ requirement. Present paper aims to utilise AHP (Analytic Hierarchy Process) methodology to priorities and rank the hierarchy levels of TQM enablers and IT resource together for its successful implementation in the Information and Communication Technology (ICT) industry. A total of 17 TQM enablers (nine) and IT resources (eight) were identified and partitioned into 3 categories and were prioritised by AHP approach. The finding indicates that the 17 sub-criteria can be grouped into three main categories namely organizing, tools and techniques, and culture and people. Further, out of 17 sub-criteria, three sub-criteria: top management commitment and support, total employee involvement, and continuous improvement got highest priority whereas three sub-criteria such as structural equation modelling, culture change, and customer satisfaction got lowest priority. The result suggests a hierarchy model for ICT industry to prioritise the enablers and resources as well as to improve the TQM and IT performance in the ICT industry. This paper has some managerial implication which suggests the managers of ICT industry to implement TQM and IT together in their organizations to get maximum benefits and how to utilize available resources. At the end, conclusions, limitation, future scope of the study are presented.

Keywords: Analytic Hierarchy Process, Information Technology, Information and Communication Technology, Prioritization, Total Quality Management.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1950
504 Applying the Extreme-Based Teaching Model in Post-Secondary Online Classroom Setting: A Field Experiment

Authors: Leon Pan

Abstract:

The first programming course within post-secondary education has long been recognized as a challenging endeavor for both educators and students alike. Historically, these courses have exhibited high failure rates and a notable number of dropouts. Instructors often lament students' lack of effort on their coursework, and students often express frustration that the teaching methods employed are not effective. Drawing inspiration from the successful principles of Extreme Programming, this study introduces an approach—the Extremes-based teaching model—aimed at enhancing the teaching of introductory programming courses. To empirically determine the effectiveness of the model, a comparison was made between a section taught using the extreme-based model and another utilizing traditional teaching methods. Notably, the extreme-based teaching class required students to work collaboratively on projects, while also demanding continuous assessment and performance enhancement within groups. This paper details the application of the extreme-based model within the post-secondary online classroom context and presents the compelling results that emphasize its effectiveness in advancing the teaching and learning experiences. The extreme-based model led to a significant increase of 13.46 points in the weighted total average and a commendable 10% reduction in the failure rate.

Keywords: Extreme-based teaching model, innovative pedagogical methods, project-based learning, team-based learning.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 99