Search results for: Imperfect channel state information
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 6341

Search results for: Imperfect channel state information

2411 Determining Food Habits in Süleymanpasa Town of Tekirdag City, Turkey

Authors: Emine Yilmaz, Ismail Yilmaz, Harun Uran

Abstract:

Food-borne problems have been placed among the most leading problems of the society especially in recent years. This state arises as a problem which affects the society wholly such as the supply of food stuffs that are necessary for an individual to perform his physiological and biological functions, their amount, compound, their effects on health and distribution by individuals. This study was conducted in order to determine the sensitivities and criteria of people, who have different socio-economic backgrounds and live in Süleymanpasa Town of Tekirdag City, in their preference of food stuffs. The research data were collected by means of Interview Technique with individuals within the scope of the study (300) and applying surveys with convenience sampling. According to the research results, quality appears in the first rank among the factors by which consumers are affected while buying food stuffs. Consumers stated that they try to be careful with not buying food sold outdoors. The most preferred food among the ones being sold outdoor were found to be breakfast food. Also, food stuff which consumers become the most selective for while buying was determined to be meat and meat products. Due to general knowledge about the food stuff consumed in human nutrition may affect their health negatively; consumers expressed that they are very relevant with their diets and this circumstances affects their purchase preferences.  

Keywords: Consumption, food safety, consumer behavior, purchase preferences.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2796
2410 Deployment of Service Quality Characteristics

Authors: Shuki Dror

Abstract:

This work discusses an innovative methodology for deployment of service quality characteristics. Four groups of organizational features that may influence the quality of services are identified: human resource, technology, planning, and organizational relationships. A House of Service Quality (HOSQ) matrix is built to extract the desired improvement in the service quality characteristics and to translate them into a hierarchy of important organizational features. The Mean Square Error (MSE) criterion enables the pinpointing of the few essential service quality characteristics to be improved as well as selection of the vital organizational features. The method was implemented in an engineering supply enterprise and provides useful information on its vital service dimensions.

Keywords: HOQ, organizational features, service quality.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1865
2409 Study of the Effect of Seismic Behavior of Twin Tunnels Position on Each Other

Authors: M. Azadi, M. Kalhor

Abstract:

Excavation of shallow tunnels such as subways in urban areas plays a significant role as a life line and investigation of the soil behavior against tunnel construction is one of the vital subjects studied in the geotechnical scope. Nowadays, urban tunnels are mostly drilled by T.B.Ms and changing the applied forces to tunnel lining is one of the most risky matters while drilling tunnels by these machines. Variation of soil cementation can change the behavior of these forces in the tunnel lining. Therefore, this article is designed to assess the impact of tunnel excavation in different soils and several amounts of cementation on applied loads to tunnel lining under static and dynamic loads. According to the obtained results, changing the cementation of soil will affect the applied loadings to the tunnel envelope significantly. It can be determined that axial force in tunnel lining decreases considerably when soil cementation increases. Also, bending moment and shear force in tunnel lining decreases as the soil cementation increases and causes bending and shear behavior of the segments to improve. Based on the dynamic analyses, as cohesion factor in soil increases, bending moment, axial and shear forces of segments decrease but lining behavior of the tunnel is the same as static state. The results show that decreasing the overburden applied to lining caused by cementation is different in two static and dynamic states.

Keywords: Tunnel, Soil cementation, Static, Dynamic.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2223
2408 Assessing Water Quality Using GIS: The Case of Northern Lebanon Miocene Aquifer

Authors: M. Saba, A. Iaaly, E. Carlier, N. Georges

Abstract:

This research focuses on assessing the ground water quality of Northern Lebanon affected by saline water intrusion. The chemical, physical and microbiological parameters were collected in various seasons spanning over the period of two years. Results were assessed using Geographic Information System (GIS) due to its visual capabilities in presenting the pollution extent in the studied region. Future projections of the excessive pumping were also simulated using GIS in order to assess the extent of the problem of saline intrusion in the near future.

Keywords: GIS, saline water, quality control, drinkable water quality standards, pumping.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1566
2407 Decision Tree Based Scheduling for Flexible Job Shops with Multiple Process Plans

Authors: H.-H. Doh, J.-M. Yu, Y.-J. Kwon, J.-H. Shin, H.-W. Kim, S.-H. Nam, D.-H. Lee

Abstract:

This paper suggests a decision tree based approach for flexible job shop scheduling with multiple process plans, i.e. each job can be processed through alternative operations, each of which can be processed on alternative machines. The main decision variables are: (a) selecting operation/machine pair; and (b) sequencing the jobs assigned to each machine. As an extension of the priority scheduling approach that selects the best priority rule combination after many simulation runs, this study suggests a decision tree based approach in which a decision tree is used to select a priority rule combination adequate for a specific system state and hence the burdens required for developing simulation models and carrying out simulation runs can be eliminated. The decision tree based scheduling approach consists of construction and scheduling modules. In the construction module, a decision tree is constructed using a four-stage algorithm, and in the scheduling module, a priority rule combination is selected using the decision tree. To show the performance of the decision tree based approach suggested in this study, a case study was done on a flexible job shop with reconfigurable manufacturing cells and a conventional job shop, and the results are reported by comparing it with individual priority rule combinations for the objectives of minimizing total flow time and total tardiness.

Keywords: Flexible job shop scheduling, Decision tree, Priority rules, Case study.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3317
2406 A Two-Stage Expert System for Diagnosis of Leukemia Based on Type-2 Fuzzy Logic

Authors: Ali Akbar Sadat Asl

Abstract:

Diagnosis and deciding about diseases in medical fields is facing innate uncertainty which can affect the whole process of treatment. This decision is made based on expert knowledge and the way in which an expert interprets the patient's condition, and the interpretation of the various experts from the patient's condition may be different. Fuzzy logic can provide mathematical modeling for many concepts, variables, and systems that are unclear and ambiguous and also it can provide a framework for reasoning, inference, control, and decision making in conditions of uncertainty. In systems with high uncertainty and high complexity, fuzzy logic is a suitable method for modeling. In this paper, we use type-2 fuzzy logic for uncertainty modeling that is in diagnosis of leukemia. The proposed system uses an indirect-direct approach and consists of two stages: In the first stage, the inference of blood test state is determined. In this step, we use an indirect approach where the rules are extracted automatically by implementing a clustering approach. In the second stage, signs of leukemia, duration of disease until its progress and the output of the first stage are combined and the final diagnosis of the system is obtained. In this stage, the system uses a direct approach and final diagnosis is determined by the expert. The obtained results show that the type-2 fuzzy expert system can diagnose leukemia with the average accuracy about 97%.

Keywords: Expert system, leukemia, medical diagnosis, type-2 fuzzy logic.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1052
2405 MEAL Project: Modifying Eating Attitudes and Actions through Learning

Authors: E. Oliver, A. Cebolla, A. Dominguez, A. Gonzalez-Segura, E. de la Cruz, S. Albertini, L. Ferrini, K. Kronika, T. Nilsen, R. Baños

Abstract:

The main objective of MEAL is to develop a pedagogical tool aimed to help teachers and nutritionists (students and professionals) to acquire, train, promote and deliver to children basic nutritional education and healthy eating behaviours competencies. MEAL is focused on eating behaviours and not only in nutritional literacy, and will use new technologies like Information and Communication Technologies (ICTs) and serious games (SG) platforms to consolidate the nutritional competences and habits.

Keywords: Nutritional Education, Pedagogical ICT Platform, Serious Games, Teachers and Nutritionists, Training Course.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2274
2404 Using Automatic Ontology Learning Methods in Human Plausible Reasoning Based Systems

Authors: A. R. Vazifedoost, M. Rahgozar, F. Oroumchian

Abstract:

Knowledge discovery from text and ontology learning are relatively new fields. However their usage is extended in many fields like Information Retrieval (IR) and its related domains. Human Plausible Reasoning based (HPR) IR systems for example need a knowledge base as their underlying system which is currently made by hand. In this paper we propose an architecture based on ontology learning methods to automatically generate the needed HPR knowledge base.

Keywords: Ontology Learning, Human Plausible Reasoning, knowledge extraction, knowledge representation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1600
2403 Automatic Sleep Stage Scoring with Wavelet Packets Based on Single EEG Recording

Authors: Luay A. Fraiwan, Natheer Y. Khaswaneh, Khaldon Y. Lweesy

Abstract:

Sleep stage scoring is the process of classifying the stage of the sleep in which the subject is in. Sleep is classified into two states based on the constellation of physiological parameters. The two states are the non-rapid eye movement (NREM) and the rapid eye movement (REM). The NREM sleep is also classified into four stages (1-4). These states and the state wakefulness are distinguished from each other based on the brain activity. In this work, a classification method for automated sleep stage scoring based on a single EEG recording using wavelet packet decomposition was implemented. Thirty two ploysomnographic recording from the MIT-BIH database were used for training and validation of the proposed method. A single EEG recording was extracted and smoothed using Savitzky-Golay filter. Wavelet packets decomposition up to the fourth level based on 20th order Daubechies filter was used to extract features from the EEG signal. A features vector of 54 features was formed. It was reduced to a size of 25 using the gain ratio method and fed into a classifier of regression trees. The regression trees were trained using 67% of the records available. The records for training were selected based on cross validation of the records. The remaining of the records was used for testing the classifier. The overall correct rate of the proposed method was found to be around 75%, which is acceptable compared to the techniques in the literature.

Keywords: Features selection, regression trees, sleep stagescoring, wavelet packets.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2328
2402 Combined Feature Based Hyperspectral Image Classification Technique Using Support Vector Machines

Authors: Mrs.K.Kavitha, S.Arivazhagan

Abstract:

A spatial classification technique incorporating a State of Art Feature Extraction algorithm is proposed in this paper for classifying a heterogeneous classes present in hyper spectral images. The classification accuracy can be improved if and only if both the feature extraction and classifier selection are proper. As the classes in the hyper spectral images are assumed to have different textures, textural classification is entertained. Run Length feature extraction is entailed along with the Principal Components and Independent Components. A Hyperspectral Image of Indiana Site taken by AVIRIS is inducted for the experiment. Among the original 220 bands, a subset of 120 bands is selected. Gray Level Run Length Matrix (GLRLM) is calculated for the selected forty bands. From GLRLMs the Run Length features for individual pixels are calculated. The Principle Components are calculated for other forty bands. Independent Components are calculated for next forty bands. As Principal & Independent Components have the ability to represent the textural content of pixels, they are treated as features. The summation of Run Length features, Principal Components, and Independent Components forms the Combined Features which are used for classification. SVM with Binary Hierarchical Tree is used to classify the hyper spectral image. Results are validated with ground truth and accuracies are calculated.

Keywords: Multi-class, Run Length features, PCA, ICA, classification and Support Vector Machines.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1521
2401 A Real Time Development Study for Automated Centralized Remote Monitoring System at Royal Belum Forest

Authors: Amri Yusoff, Shahrizuan Shafiril, Ashardi Abas, Norma Che Yusoff

Abstract:

Nowadays, illegal logging has been causing many effects including flash flood, avalanche, global warming, and etc. The purpose of this study was to maintain the earth ecosystem by keeping and regulate Malaysia’s treasurable rainforest by utilizing a new technology that will assist in real-time alert and give faster response to the authority to act on these illegal activities. The methodology of this research consisted of design stages that have been conducted as well as the system model and system architecture of the prototype in addition to the proposed hardware and software that have been mainly used such as microcontroller, sensor with the implementation of GSM, and GPS integrated system. This prototype was deployed at Royal Belum forest in December 2014 for phase 1 and April 2015 for phase 2 at 21 pinpoint locations. The findings of this research were the capture of data in real-time such as temperature, humidity, gaseous, fire, and rain detection which indicate the current natural state and habitat in the forest. Besides, this device location can be detected via GPS of its current location and then transmitted by SMS via GSM system. All of its readings were sent in real-time for further analysis. The data that were compared to meteorological department showed that the precision of this device was about 95% and these findings proved that the system is acceptable and suitable to be used in the field.

Keywords: Remote monitoring system, forest data, GSM, GPS, wireless sensor.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1621
2400 Testing Object-Oriented Framework Applications Using FIST2 Tool: A Case Study

Authors: Jehad Al Dallal

Abstract:

An application framework provides a reusable design and implementation for a family of software systems. Frameworks are introduced to reduce the cost of a product line (i.e., a family of products that shares the common features). Software testing is a timeconsuming and costly ongoing activity during the application software development process. Generating reusable test cases for the framework applications during the framework development stage, and providing and using the test cases to test part of the framework application whenever the framework is used reduces the application development time and cost considerably. This paper introduces the Framework Interface State Transition Tester (FIST2), a tool for automated unit testing of Java framework applications. During the framework development stage, given the formal descriptions of the framework hooks, the specifications of the methods of the framework-s extensible classes, and the illegal behavior description of the Framework Interface Classes (FICs), FIST2 generates unitlevel test cases for the classes. At the framework application development stage, given the customized method specifications of the implemented FICs, FIST2 automates the use, execution, and evaluation of the already generated test cases to test the implemented FICs. The paper illustrates the use of the FIST2 tool for testing several applications that use the SalesPoint framework.

Keywords: Automated testing, class testing, FICs, FIST2, object-oriented framework, object-oriented testing.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1617
2399 Fast Extraction of Edge Histogram in DCT Domain based on MPEG7

Authors: Minyoung Eom, Yoonsik Choe

Abstract:

In these days, multimedia data is transmitted and processed in compressed format. Due to the decoding procedure and filtering for edge detection, the feature extraction process of MPEG-7 Edge Histogram Descriptor is time-consuming as well as computationally expensive. To improve efficiency of compressed image retrieval, we propose a new edge histogram generation algorithm in DCT domain in this paper. Using the edge information provided by only two AC coefficients of DCT coefficients, we can get edge directions and strengths directly in DCT domain. The experimental results demonstrate that our system has good performance in terms of retrieval efficiency and effectiveness.

Keywords: DCT, Descriptor, EHD, MPEG7.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2124
2398 Semi-Automatic Approach for Semantic Annotation

Authors: Mohammad Yasrebi, Mehran Mohsenzadeh

Abstract:

The third phase of web means semantic web requires many web pages which are annotated with metadata. Thus, a crucial question is where to acquire these metadata. In this paper we propose our approach, a semi-automatic method to annotate the texts of documents and web pages and employs with a quite comprehensive knowledge base to categorize instances with regard to ontology. The approach is evaluated against the manual annotations and one of the most popular annotation tools which works the same as our tool. The approach is implemented in .net framework and uses the WordNet for knowledge base, an annotation tool for the Semantic Web.

Keywords: Semantic Annotation, Metadata, Information Extraction, Semantic Web, knowledge base.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1865
2397 Intelligent Transport System: Classification of Traffic Signs Using Deep Neural Networks in Real Time

Authors: Anukriti Kumar, Tanmay Singh, Dinesh Kumar Vishwakarma

Abstract:

Traffic control has been one of the most common and irritating problems since the time automobiles have hit the roads. Problems like traffic congestion have led to a significant time burden around the world and one significant solution to these problems can be the proper implementation of the Intelligent Transport System (ITS). It involves the integration of various tools like smart sensors, artificial intelligence, position technologies and mobile data services to manage traffic flow, reduce congestion and enhance driver's ability to avoid accidents during adverse weather. Road and traffic signs’ recognition is an emerging field of research in ITS. Classification problem of traffic signs needs to be solved as it is a major step in our journey towards building semi-autonomous/autonomous driving systems. The purpose of this work focuses on implementing an approach to solve the problem of traffic sign classification by developing a Convolutional Neural Network (CNN) classifier using the GTSRB (German Traffic Sign Recognition Benchmark) dataset. Rather than using hand-crafted features, our model addresses the concern of exploding huge parameters and data method augmentations. Our model achieved an accuracy of around 97.6% which is comparable to various state-of-the-art architectures.

Keywords: Multiclass classification, convolution neural network, OpenCV, Data Augmentation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 813
2396 How Do Politicians Recover Their Costs? The Political Economy of Representative Democracy in India

Authors: Subramaniam Chandran

Abstract:

This paper explores the features of political economy in the dynamics of representative politics in India. Politics is seen as enhancing economic benefits through acquiring and maintenance of power in the realm of democratic set up. The system of representation is riddled with competitive populism. Emerging leaders and parties are forced to accommodate their ideologies in coping with competitive politics. Electoral politics and voting behaviour reflect series of influences mooted by the politicians. Voters are accustomed to expect benefits outs of state exchequer. The electoral competitors show a changing phase of investment and return policy. Every elector has to spend and realize his costs in his tenure. In the case of defeated electors, even the cost recovery is not possible directly; there are indirect means to recover their costs. The series of case studies show the method of party funding, campaign financing, electoral expenditure, and cost recovery. Regulations could not restrict the level of spending. Several cases of disproportionate accumulation of wealth by the politicians reveal that money played a major part in electoral process. The political economy of representative politics hitherto ignores how a politician spends and recovers his cost and multiples his wealth. To be sure, the acquiring and maintenance of power is to enhance the wealth of the electors.

Keywords: Political economy, representative politics, costrecovery, electoral politics

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1550
2395 Resistance to Sulfuric Acid Attacks of Self-Consolidating Concrete: Effect Metakaolin and Various Cements Types

Authors: Kianoosh Samimi, Farhad Estakhr, Mahdi Mahdikhani, Faramaz Moodi

Abstract:

Due to their fluidity and simplicity of use, self-compacting concretes (SCCs) have undeniable advantages. In recent years, the role of metakaolin as a one of pozzolanic materials in concrete has been considered by researchers. It can modify various properties of concrete, due to high pozzolanic reactions and also makes a denser microstructure. The objective of this paper is to examine the influence of three type of Portland cement and metakaolin on fresh state, compressive strength and sulfuric acid attacks in self- consolidating concrete at early age up to 90 days of curing in lime water. Six concrete mixtures were prepared with three types of different cement as Portland cement type II, Portland Slag Cement (PSC), Pozzolanic Portland Cement (PPC) and 15% substitution of metakaolin by every cement. The results show that the metakaolin admixture increases the viscosity and the demand amount of superplasticizer. According to the compressive strength results, the highest value of compressive strength was achieved for PSC and without any metakaolin at age of 90 days. Conversely, the lowest level of compressive strength at all ages of conservation was obtained for PPC and containing 15% metakaolin. According to this study, the total substitution of PSC and PPC by Portland cement type II is beneficial to the increasing in the chemical resistance of the SCC with respect to the sulfuric acid attack. On the other hand, this increase is more noticeable by the use of 15% of metakaolin. Therefore, it can be concluded that metakaolin has a positive effect on the chemical resistance of SCC containing of Portland cement type II, PSC, and PPC.

Keywords: SCC, metakaolin, cement type, durability, compressive strength, sulfuric acid attacks.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 894
2394 A Microcontroller Implementation of Constrained Model Predictive Control

Authors: Amira Kheriji Abbes, Faouzi Bouani, Mekki Ksouri

Abstract:

Model Predictive Control (MPC) is an established control technique in a wide range of process industries. The reason for this success is its ability to handle multivariable systems and systems having input, output or state constraints. Neverthless comparing to PID controller, the implementation of the MPC in miniaturized devices like Field Programmable Gate Arrays (FPGA) and microcontrollers has historically been very small scale due to its complexity in implementation and its computation time requirement. At the same time, such embedded technologies have become an enabler for future manufacturing enterprisers as well as a transformer of organizations and markets. In this work, we take advantage of these recent advances in this area in the deployment of one of the most studied and applied control technique in the industrial engineering. In this paper, we propose an efficient firmware for the implementation of constrained MPC in the performed STM32 microcontroller using interior point method. Indeed, performances study shows good execution speed and low computational burden. These results encourage to develop predictive control algorithms to be programmed in industrial standard processes. The PID anti windup controller was also implemented in the STM32 in order to make a performance comparison with the MPC. The main features of the proposed constrained MPC framework are illustrated through two examples.

Keywords: Embedded software, microcontroller, constrainedModel Predictive Control, interior point method, PID antiwindup, Keil tool, C/Cµ language.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2797
2393 Analysis of Security Vulnerabilities for Mobile Health Applications

Authors: Y. Cifuentes, L. Beltrán, L. Ramírez

Abstract:

The availability to deploy mobile applications for health care is increasing daily thru different mobile app stores. But within these capabilities the number of hacking attacks has also increased, in particular into medical mobile applications. The security vulnerabilities in medical mobile apps can be triggered by errors in code, incorrect logic, poor design, among other parameters. This is usually used by malicious attackers to steal or modify the users’ information. The aim of this research is to analyze the vulnerabilities detected in mobile medical apps according to risk factor standards defined by OWASP in 2014.

Keywords: mHealth apps, OWASP, protocols, security vulnerabilities, risk factors.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4409
2392 Deep Learning for Renewable Power Forecasting: An Approach Using LSTM Neural Networks

Authors: Fazıl Gökgöz, Fahrettin Filiz

Abstract:

Load forecasting has become crucial in recent years and become popular in forecasting area. Many different power forecasting models have been tried out for this purpose. Electricity load forecasting is necessary for energy policies, healthy and reliable grid systems. Effective power forecasting of renewable energy load leads the decision makers to minimize the costs of electric utilities and power plants. Forecasting tools are required that can be used to predict how much renewable energy can be utilized. The purpose of this study is to explore the effectiveness of LSTM-based neural networks for estimating renewable energy loads. In this study, we present models for predicting renewable energy loads based on deep neural networks, especially the Long Term Memory (LSTM) algorithms. Deep learning allows multiple layers of models to learn representation of data. LSTM algorithms are able to store information for long periods of time. Deep learning models have recently been used to forecast the renewable energy sources such as predicting wind and solar energy power. Historical load and weather information represent the most important variables for the inputs within the power forecasting models. The dataset contained power consumption measurements are gathered between January 2016 and December 2017 with one-hour resolution. Models use publicly available data from the Turkish Renewable Energy Resources Support Mechanism. Forecasting studies have been carried out with these data via deep neural networks approach including LSTM technique for Turkish electricity markets. 432 different models are created by changing layers cell count and dropout. The adaptive moment estimation (ADAM) algorithm is used for training as a gradient-based optimizer instead of SGD (stochastic gradient). ADAM performed better than SGD in terms of faster convergence and lower error rates. Models performance is compared according to MAE (Mean Absolute Error) and MSE (Mean Squared Error). Best five MAE results out of 432 tested models are 0.66, 0.74, 0.85 and 1.09. The forecasting performance of the proposed LSTM models gives successful results compared to literature searches.

Keywords: Deep learning, long-short-term memory, energy, renewable energy load forecasting.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1594
2391 Characteristics of Regional Issues in Local Municipalities of Japan in Consideration of Socio-Economic Condition

Authors: Akiko Kondo, Akio Kondo

Abstract:

We are facing serious problems related to long-term depopulation and an aging society with a falling birth rate in Japan. In this situation, we are suffering from a shortfall in human resources as well as a shortage of workforce in rural regions. In addition, we are struggling with a protracted economic slump and excess concentration of population in the Tokyo Metropolitan area. It is an urgent national issue to consider how to live in this country and what kind of structure of society and administration policy is needed. It is necessary to clarify people’s desire for their way of living and social assistance to be provided. The aim of this study is to clarify the characteristics of regional issues and the degree of their seriousness in local municipalities of Japan. We conducted a questionnaire survey about regional agenda in all local municipalities in Japan. We obtained responses concerning the degree of seriousness of regional issues and degree of importance of policies. Based on the data gathered from the survey, it is apparent that many local municipalities are facing an aging population and declining population. We constructed a model to analyze factors for declining population. Using the model, it was clarified that a population’s age structure, job opportunities and income level affect the decline of population. In addition, we showed the way of the evaluation of state of local municipality.

Keywords: Evaluation, Local municipality, Regional analysis, Regional issue.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1861
2390 Motion Prediction and Motion Vector Cost Reduction during Fast Block Motion Estimation in MCTF

Authors: Karunakar A K, Manohara Pai M M

Abstract:

In 3D-wavelet video coding framework temporal filtering is done along the trajectory of motion using Motion Compensated Temporal Filtering (MCTF). Hence computationally efficient motion estimation technique is the need of MCTF. In this paper a predictive technique is proposed in order to reduce the computational complexity of the MCTF framework, by exploiting the high correlation among the frames in a Group Of Picture (GOP). The proposed technique applies coarse and fine searches of any fast block based motion estimation, only to the first pair of frames in a GOP. The generated motion vectors are supplied to the next consecutive frames, even to subsequent temporal levels and only fine search is carried out around those predicted motion vectors. Hence coarse search is skipped for all the motion estimation in a GOP except for the first pair of frames. The technique has been tested for different fast block based motion estimation algorithms over different standard test sequences using MC-EZBC, a state-of-the-art scalable video coder. The simulation result reveals substantial reduction (i.e. 20.75% to 38.24%) in the number of search points during motion estimation, without compromising the quality of the reconstructed video compared to non-predictive techniques. Since the motion vectors of all the pair of frames in a GOP except the first pair will have value ±1 around the motion vectors of the previous pair of frames, the number of bits required for motion vectors is also reduced by 50%.

Keywords: Motion Compensated Temporal Filtering, predictivemotion estimation, lifted wavelet transform, motion vector

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1618
2389 Preliminary Evaluation of Decommissioning Wastes for the First Commercial Nuclear Power Reactor in South Korea

Authors: Kyomin Lee, Joohee Kim, Sangho Kang

Abstract:

The commercial nuclear power reactor in South Korea, Kori Unit 1, which was a 587 MWe pressurized water reactor that started operation since 1978, was permanently shut down in June 2017 without an additional operating license extension. The Kori 1 Unit is scheduled to become the nuclear power unit to enter the decommissioning phase. In this study, the preliminary evaluation of the decommissioning wastes for the Kori Unit 1 was performed based on the following series of process: firstly, the plant inventory is investigated based on various documents (i.e., equipment/ component list, construction records, general arrangement drawings). Secondly, the radiological conditions of systems, structures and components (SSCs) are established to estimate the amount of radioactive waste by waste classification. Third, the waste management strategies for Kori Unit 1 including waste packaging are established. Forth, selection of the proper decontamination and dismantling (D&D) technologies is made considering the various factors. Finally, the amount of decommissioning waste by classification for Kori 1 is estimated using the DeCAT program, which was developed by KEPCO-E&C for a decommissioning cost estimation. The preliminary evaluation results have shown that the expected amounts of decommissioning wastes were less than about 2% and 8% of the total wastes generated (i.e., sum of clean wastes and radwastes) before/after waste processing, respectively, and it was found that the majority of contaminated material was carbon or alloy steel and stainless steel. In addition, within the range of availability of information, the results of the evaluation were compared with the results from the various decommissioning experiences data or international/national decommissioning study. The comparison results have shown that the radioactive waste amount from Kori Unit 1 decommissioning were much less than those from the plants decommissioned in U.S. and were comparable to those from the plants in Europe. This result comes from the difference of disposal cost and clearance criteria (i.e., free release level) between U.S. and non-U.S. The preliminary evaluation performed using the methodology established in this study will be useful as a important information in establishing the decommissioning planning for the decommissioning schedule and waste management strategy establishment including the transportation, packaging, handling, and disposal of radioactive wastes.

Keywords: Characterization, classification, decommissioning, decontamination and dismantling, Kori 1, radioactive waste.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1480
2388 Feasibility Study on Designing a Flat Loop Heat Pipe (LHP) to Recover the Heat from Exhaust of a Gas Turbine

Authors: M.H.Ghaffari

Abstract:

A theoretical study is conducted to design and explore the effect of different parameters such as heat loads, the tube size of piping system, wick thickness, porosity and hole size on the performance and capability of a Loop Heat Pipe(LHP). This paper presents a steady state model that describes the different phenomena inside a LHP. Loop Heat Pipes(LHPs) are two-phase heat transfer devices with capillary pumping of a working fluid. By their original design comparing with heat pipes and special properties of the capillary structure, they-re capable of transferring heat efficiency for distances up to several meters at any orientation in the gravity field, or to several meters in a horizontal position. This theoretical model is described by different relations to satisfy important limits such as capillary and nucleate boiling. An algorithm is developed to predict the size of the LHP satisfying the limitations mentioned above for a wide range of applied loads. Finally, to assess and evaluate the algorithm and all the relations considered, we have used to design a new kind of LHP to recover the heat from the exhaust of an actual Gas Turbine. By finding the results, it showed that we can use the LHP as a very high efficient device to recover the heat even in high amount of loads(exhaust of a gas turbine). The sizes of all parts of the LHP were obtained using the developed algorithm.

Keywords: Loop Heat Pipe, Head Load, Liquid-Vapor Interface, Heat Transfer, Design Algorithm

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2071
2387 A Robust Audio Fingerprinting Algorithm in MP3 Compressed Domain

Authors: Ruili Zhou, Yuesheng Zhu

Abstract:

In this paper, a new robust audio fingerprinting algorithm in MP3 compressed domain is proposed with high robustness to time scale modification (TSM). Instead of simply employing short-term information of the MP3 stream, the new algorithm extracts the long-term features in MP3 compressed domain by using the modulation frequency analysis. Our experiment has demonstrated that the proposed method can achieve a hit rate of above 95% in audio retrieval and resist the attack of 20% TSM. It has lower bit error rate (BER) performance compared to the other algorithms. The proposed algorithm can also be used in other compressed domains, such as AAC.

Keywords: Audio Fingerprinting, MP3, Modulation Frequency, TSM

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2195
2386 Performance and Availability Analyses of PV Generation Systems in Taiwan

Authors: H. S. Huang, J. C. Jao, K. L. Yen, C. T. Tsai

Abstract:

The purpose of this article applies the monthly final energy yield and failure data of 202 PV systems installed in Taiwan to analyze the PV operational performance and system availability. This data is collected by Industrial Technology Research Institute through manual records. Bad data detection and failure data estimation approaches are proposed to guarantee the quality of the received information. The performance ratio value and system availability are then calculated and compared with those of other countries. It is indicated that the average performance ratio of Taiwan-s PV systems is 0.74 and the availability is 95.7%. These results are similar with those of Germany, Switzerland, Italy and Japan.

Keywords: availability, performance ratio, PV system, Taiwan

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4436
2385 Environmental Assessment Methods in Abu Dhabi

Authors: Huda Al Salmi, Hisham Al Kadi, Simone Leao

Abstract:

Abu Dhabi is one of the fastest developed cities in the region. On top of all the current and future environmental challenges, Abu Dhabi aims to be among the top governments in the world in sustainable development. Abu Dhabi plans to create an attractive, livable and sustainable managed urban environment in which all necessary services and infrastructure are provided in a sustainable and timely manner. Abu Dhabi is engaged in a difficult challenge to develop credible environmental indicators that would assess the ambitious environmental targets. The aim of those indicators is to provide reliable guidance to decision makers and the public concerning key factors that determine the state of urban environment and identify major areas for policy intervention. In order to ensure sustainable development in UAE in general, and of Abu Dhabi City in particular, relevant and contextual environmental indicators need to be carefully considered. These indicators provide a gauge at a national government scale of how close countries are to establish environmental policy goals. The environment indicators assist city decision-making in such areas as identification of significant environmental aspects and observation of environmental performance trends. Those can help to find ways of reducing environmental pollution and in improving eco-efficiency. This paper outlines recent strategies implemented in Abu Dhabi that aims to improve the sustainable performance of the city-s built environment. The paper explores the variety of current and possible indicators at different levels and their roles in the development of the city.

Keywords: Abu Dhabi, Sustainable Development, Indicators, Assessment Methods.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2736
2384 Evaluating the Role of Multisensory Elements in Foreign Language Acquisition

Authors: Sari Myréen

Abstract:

The aim of this study was to evaluate the role of multisensory elements in enhancing and facilitating foreign language acquisition among adult students in a language classroom. The use of multisensory elements enables the creation of a student-centered classroom, where the focus is on individual learner’s language learning process, perceptions and motivation. Multisensory language learning is a pedagogical approach where the language learner uses all the senses more effectively than in a traditional in-class environment. Language learning is facilitated due to multisensory stimuli which increase the number of cognitive connections in the learner and take into consideration different types of learners. A living lab called Multisensory Space creates a relaxed and receptive state in the learners through various multisensory stimuli, and thus promotes their natural foreign language acquisition. Qualitative and quantitative data were collected in two questionnaire inquiries among the Finnish students of a higher education institute at the end of their basic French courses in December 2014 and 2016. The inquiries discussed the effects of multisensory elements on the students’ motivation to study French as well as their learning outcomes. The results show that the French classes in the Multisensory Space provide the students with an encouraging and pleasant learning environment, which has a positive impact on their motivation to study the foreign language as well as their language learning outcomes.

Keywords: Foreign language acquisition, foreign language learning, higher education, multisensory learning, pedagogical approach, transcultural learning.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1378
2383 Use of Carica papaya as a Bio-Sorbent for Removal of Heavy Metals in Wastewater

Authors: W. E. Igwegbe, B. C. Okoro, J. C. Osuagwu

Abstract:

The study assessed the effectiveness of Pawpaw (Carica papaya) wood in reducing the concentrations of heavy metals in wastewater acting as a bio-sorbent. The following heavy metals were considered; Zinc, Cadmium, Lead, Copper, Iron, Selenium, Nickel and Manganese. The physiochemical properties of Carica papaya stem were studied. The experimental sample was sourced from the trunk of a felled matured pawpaw tree. Wastewater for experimental use was prepared by dissolving soil samples collected from a dump site at Owerri, Imo state of Nigeria in water. The concentration of each metal remaining in solution as residual metal after bio-sorption was determined using Atomic absorption Spectrometer. The effects of pH and initial heavy metal concentration were studied in a batch reactor. The results of Spectrometer test showed that there were different functional groups detected in the Carica papaya stem biomass. There was increase in metal removal as the pH increased for all the metals considered except for Nickel and Manganese. Optimum bio-sorption occurred at pH 5.9 with 5g/100ml solution of bio-sorbent. The results of the study showed that the treated wastewater is fit for irrigation purpose based on Canada wastewater quality guideline for the protection of Agricultural standard. This approach thus provides a cost effective and environmentally friendly option for treating wastewater.

Keywords: Biomass, bio-sorption, Carica papaya, heavy metal, wastewater.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2818
2382 Using the Polynomial Approximation Algorithm in the Algorithm 2 for Manipulator's Control in an Unknown Environment

Authors: Pavel K. Lopatin, Artyom S. Yegorov

Abstract:

The Algorithm 2 for a n-link manipulator movement amidst arbitrary unknown static obstacles for a case when a sensor system supplies information about local neighborhoods of different points in the configuration space is presented. The Algorithm 2 guarantees the reaching of a target position in a finite number of steps. The Algorithm 2 is reduced to a finite number of calls of a subroutine for planning a trajectory in the presence of known forbidden states. The polynomial approximation algorithm which is used as the subroutine is presented. The results of the Algorithm2 implementation are given.

Keywords: Manipulator, trajectory planning, unknown obstacles.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1238