Search results for: mean square process
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 16628

Search results for: mean square process

14708 The Essence and Attribution of Intellectual Property Rights Generated in the Digitization of Intangible Cultural Heritage

Authors: Jiarong Zhang

Abstract:

Digitizing intangible cultural heritage is a complex and comprehensive process from which sorts of intellectual property rights may be generated. Digitizing may be a repacking process of cultural heritage, which creates copyrights; recording folk songs and indigenous performances can create 'related rights'. At the same time, digitizing intangible cultural heritage may infringe the intellectual property rights of others unintentionally. Recording religious rituals of indigenous communities without authorization can violate the moral right of the ceremony participants of the community; making digital copies of rock paintings may infringe the right of reproduction. In addition, several parties are involved in the digitization process: indigenous peoples, museums, and archives can be holders of cultural heritage; companies and research institutions can be technology providers; internet platforms can be promoters and sellers; the public and groups above can be beneficiaries. When diverse intellectual property rights versus various parties, problems and disputes can arise easily. What are the types of intellectual property rights generated in the digitization process? What is the essence of these rights? Who should these rights belong to? How to use intellectual property to protect the digitalization of cultural heritage? How to avoid infringing on the intellectual property rights of others? While the digitization has been regarded as an effective approach to preserve intangible cultural heritage, related intellectual property issues have not received the attention and full discussion. Thus, parties involving in the digitization process may face intellectual property infringement lawsuits. The article will explore those problems from the intersection perspective of intellectual property law and cultural heritage. From a comparative approach, the paper will analysis related legal documents and cases, and shed some lights of those questions listed. The findings show, although there are no intellectual property laws targeting the cultural heritage in most countries, the involved stakeholders can seek protection from existing intellectual property rights following the suggestions of the article. The research will contribute to the digitization of intangible cultural heritage from a legal and policy aspect.

Keywords: copyright, digitization, intangible cultural heritage, intellectual property, Internet platforms

Procedia PDF Downloads 145
14707 Reverse Logistics Information Management Using Ontological Approach

Authors: F. Lhafiane, A. Elbyed, M. Bouchoum

Abstract:

Reverse Logistics (RL) Process is considered as complex and dynamic network that involves many stakeholders such as: suppliers, manufactures, warehouse, retails, and costumers, this complexity is inherent in such process due to lack of perfect knowledge or conflicting information. Ontologies, on the other hand, can be considered as an approach to overcome the problem of sharing knowledge and communication among the various reverse logistics partners. In this paper, we propose a semantic representation based on hybrid architecture for building the Ontologies in an ascendant way, this method facilitates the semantic reconciliation between the heterogeneous information systems (ICT) that support reverse logistics Processes and product data.

Keywords: Reverse Logistics, information management, heterogeneity, ontologies, semantic web

Procedia PDF Downloads 492
14706 Perceived Influence of Information Communication Technology on Empowerment Amongst the College of Education Physical and Health Education Students in Oyo State

Authors: I. O. Oladipo, Olusegun Adewale Ajayi, Omoniyi Oladipupo Adigun

Abstract:

Information Communication Technology (ICT) have the potential to contribute to different facets of educational development and effective learning; expanding access, promoting efficiency, improve the quality of learning, enhancing the quality of teaching and provide important mechanism for the economic crisis. Considering the prevalence of unemployment among the higher institution graduates in this nation, in which much seems not to have been achieved in this direction. In view of this, the purpose of this study is to create an awareness and enlightenment of ICT for empowerment opportunities after school. A self-developed modified 4-likert scale questionnaire was used for data collection among Colleges of Education, Physical and Health Education students in Oyo State. Inferential statistical analysis of chi-square set at 0.05 alpha levels was used to analyze the stated hypotheses. The study concludes that awareness and enlightenment of ICT significantly influence empowerment opportunities and recommended that college of education students should be encouraged on the application of ICT for job opportunity after school.

Keywords: employment, empowerment, information communication technology, physical education

Procedia PDF Downloads 390
14705 Data Refinement Enhances The Accuracy of Short-Term Traffic Latency Prediction

Authors: Man Fung Ho, Lap So, Jiaqi Zhang, Yuheng Zhao, Huiyang Lu, Tat Shing Choi, K. Y. Michael Wong

Abstract:

Nowadays, a tremendous amount of data is available in the transportation system, enabling the development of various machine learning approaches to make short-term latency predictions. A natural question is then the choice of relevant information to enable accurate predictions. Using traffic data collected from the Taiwan Freeway System, we consider the prediction of short-term latency of a freeway segment with a length of 17 km covering 5 measurement points, each collecting vehicle-by-vehicle data through the electronic toll collection system. The processed data include the past latencies of the freeway segment with different time lags, the traffic conditions of the individual segments (the accumulations, the traffic fluxes, the entrance and exit rates), the total accumulations, and the weekday latency profiles obtained by Gaussian process regression of past data. We arrive at several important conclusions about how data should be refined to obtain accurate predictions, which have implications for future system-wide latency predictions. (1) We find that the prediction of median latency is much more accurate and meaningful than the prediction of average latency, as the latter is plagued by outliers. This is verified by machine-learning prediction using XGBoost that yields a 35% improvement in the mean square error of the 5-minute averaged latencies. (2) We find that the median latency of the segment 15 minutes ago is a very good baseline for performance comparison, and we have evidence that further improvement is achieved by machine learning approaches such as XGBoost and Long Short-Term Memory (LSTM). (3) By analyzing the feature importance score in XGBoost and calculating the mutual information between the inputs and the latencies to be predicted, we identify a sequence of inputs ranked in importance. It confirms that the past latencies are most informative of the predicted latencies, followed by the total accumulation, whereas inputs such as the entrance and exit rates are uninformative. It also confirms that the inputs are much less informative of the average latencies than the median latencies. (4) For predicting the latencies of segments composed of two or three sub-segments, summing up the predicted latencies of each sub-segment is more accurate than the one-step prediction of the whole segment, especially with the latency prediction of the downstream sub-segments trained to anticipate latencies several minutes ahead. The duration of the anticipation time is an increasing function of the traveling time of the upstream segment. The above findings have important implications to predicting the full set of latencies among the various locations in the freeway system.

Keywords: data refinement, machine learning, mutual information, short-term latency prediction

Procedia PDF Downloads 169
14704 Comparative Study of Different Enhancement Techniques for Computed Tomography Images

Authors: C. G. Jinimole, A. Harsha

Abstract:

One of the key problems facing in the analysis of Computed Tomography (CT) images is the poor contrast of the images. Image enhancement can be used to improve the visual clarity and quality of the images or to provide a better transformation representation for further processing. Contrast enhancement of images is one of the acceptable methods used for image enhancement in various applications in the medical field. This will be helpful to visualize and extract details of brain infarctions, tumors, and cancers from the CT image. This paper presents a comparison study of five contrast enhancement techniques suitable for the contrast enhancement of CT images. The types of techniques include Power Law Transformation, Logarithmic Transformation, Histogram Equalization, Contrast Stretching, and Laplacian Transformation. All these techniques are compared with each other to find out which enhancement provides better contrast of CT image. For the comparison of the techniques, the parameters Peak Signal to Noise Ratio (PSNR) and Mean Square Error (MSE) are used. Logarithmic Transformation provided the clearer and best quality image compared to all other techniques studied and has got the highest value of PSNR. Comparison concludes with better approach for its future research especially for mapping abnormalities from CT images resulting from Brain Injuries.

Keywords: computed tomography, enhancement techniques, increasing contrast, PSNR and MSE

Procedia PDF Downloads 314
14703 Polymer Recycling by Biomaterial and Its Application in Grease Formulation

Authors: Amitkumar Barot, Vijaykumar Sinha

Abstract:

There is growing interest in the development of new materials based on recycled polymers from plastic waste, and also in the field of lubricants much effort has been spent on substitution of petro-based raw materials by natural-based renewable ones. This is due to the facts of depleting fossil fuels and due to strict environmental laws. In relevance to this, new technique for the formulation of grease that combines the chemical recycling of poly (ethylene terephthalate) PET with the use of castor oil (CO) has been developed. Comparison to diols used in chemical recycling of PET, castor oil is renewable, easily available, environmentally friendly, economically cheaper and hence sustainability indeed. The process parameters like CO concentration and temperature were altered, and further, the influences of the process parameters have been studied in order to establish technically and commercially viable process. Further thereby formed depolymerized product find an application as base oil in the formulation of grease. A depolymerized product has been characterized by various chemical and instrumental methods, while formulated greases have been evaluated for its tribological properties. The grease formulated using this new environmentally friendly approach presents applicative properties similar, and in some cases superior, compared to those of a commercial grease obtained from non-renewable resources.

Keywords: castor oil, grease formulation, recycling, sustainability

Procedia PDF Downloads 219
14702 Personnel Selection Based on Step-Wise Weight Assessment Ratio Analysis and Multi-Objective Optimization on the Basis of Ratio Analysis Methods

Authors: Emre Ipekci Cetin, Ebru Tarcan Icigen

Abstract:

Personnel selection process is considered as one of the most important and most difficult issues in human resources management. At the stage of personnel selection, the applicants are handled according to certain criteria, the candidates are dealt with, and efforts are made to select the most appropriate candidate. However, this process can be more complicated in terms of the managers who will carry out the staff selection process. Candidates should be evaluated according to different criteria such as work experience, education, foreign language level etc. It is crucial that a rational selection process is carried out by considering all the criteria in an integrated structure. In this study, the problem of choosing the front office manager of a 5 star accommodation enterprise operating in Antalya is addressed by using multi-criteria decision-making methods. In this context, SWARA (Step-wise weight assessment ratio analysis) and MOORA (Multi-Objective Optimization on the basis of ratio analysis) methods, which have relatively few applications when compared with other methods, have been used together. Firstly SWARA method was used to calculate the weights of the criteria and subcriteria that were determined by the business. After the weights of the criteria were obtained, the MOORA method was used to rank the candidates using the ratio system and the reference point approach. Recruitment processes differ from sector to sector, from operation to operation. There are a number of criteria that must be taken into consideration by businesses in accordance with the structure of each sector. It is of utmost importance that all candidates are evaluated objectively in the framework of these criteria, after these criteria have been carefully selected in the selection of suitable candidates for employment. In the study, staff selection process was handled by using SWARA and MOORA methods together.

Keywords: accommodation establishments, human resource management, multi-objective optimization on the basis of ratio analysis, multi-criteria decision making, step-wise weight assessment ratio analysis

Procedia PDF Downloads 343
14701 Decision-Making Process Based on Game Theory in the Process of Urban Transformation

Authors: Cemil Akcay, Goksun Yerlikaya

Abstract:

Buildings are the living spaces of people with an active role in every aspect of life in today's world. While some structures have survived from the early ages, most of the buildings that completed their lifetime have not transported to the present day. Nowadays, buildings that do not meet the social, economic, and safety requirements of the age return to life with a transformation process. This transformation is called urban transformation. Urban transformation is the renewal of the areas with a risk of disaster and the technological infrastructure required by the structure. The transformation aims to prevent damage to earthquakes and other disasters by rebuilding buildings that have completed their non-earthquake-resistant economic life. It is essential to decide on other issues related to conversion and transformation in places where most of the building stock should transform into the first-degree earthquake belt, such as Istanbul. In urban transformation, property owners, local authority, and contractor must deal at a common point. Considering that hundreds of thousands of property owners are sometimes in the areas of transformation, it is evident how difficult it is to make the deal and decide. For the optimization of these decisions, the use of game theory is foreseeing. The main problem in this study is that the urban transformation is carried out in place, or the building or buildings are transport to a different location. There are many stakeholders in the Istanbul University Cerrahpaşa Medical Faculty Campus, which is planned to be carried out in the process of urban transformation, was tried to solve the game theory applications. An analysis of the decisions given on a real urban transformation project and the logical suitability of decisions taken without the use of game theory were also supervised using game theory. In each step of this study, many decision-makers are classifying according to a specific logical sequence, and in the game trees that emerged as a result of this classification, Nash balances were tried to observe, and optimum decisions were determined. All decisions taken for this project have been subjected to two significant differentiated comparisons using game theory, and as decisions are taken without the use of game theory, and according to the results, solutions for the decision phase of the urban transformation process introduced. The game theory model developed from beginning to the end of the urban transformation process, particularly as a solution to the difficulty of making rational decisions in large-scale projects with many participants in the decision-making process. The use of a decision-making mechanism can provide an optimum answer to the demands of the stakeholders. In today's world for the construction sector, it is also seeing that the game theory is a non-surprising consequence of the fact that it is the most critical issues of planning and making the right decision in future years.

Keywords: urban transformation, the game theory, decision making, multi-actor project

Procedia PDF Downloads 140
14700 Optimization of Titanium Leaching Process Using Experimental Design

Authors: Arash Rafiei, Carroll Moore

Abstract:

Leaching process as the first stage of hydrometallurgy is a multidisciplinary system including material properties, chemistry, reactor design, mechanics and fluid dynamics. Therefore, doing leaching system optimization by pure scientific methods need lots of times and expenses. In this work, a mixture of two titanium ores and one titanium slag are used for extracting titanium for leaching stage of TiO2 pigment production procedure. Optimum titanium extraction can be obtained from following strategies: i) Maximizing titanium extraction without selective digestion; and ii) Optimizing selective titanium extraction by balancing between maximum titanium extraction and minimum impurity digestion. The main difference between two strategies is due to process optimization framework. For the first strategy, the most important stage of production process is concerned as the main stage and rest of stages would be adopted with respect to the main stage. The second strategy optimizes performance of more than one stage at once. The second strategy has more technical complexity compared to the first one but it brings more economical and technical advantages for the leaching system. Obviously, each strategy has its own optimum operational zone that is not as same as the other one and the best operational zone is chosen due to complexity, economical and practical aspects of the leaching system. Experimental design has been carried out by using Taguchi method. The most important advantages of this methodology are involving different technical aspects of leaching process; minimizing the number of needed experiments as well as time and expense; and concerning the role of parameter interactions due to principles of multifactor-at-time optimization. Leaching tests have been done at batch scale on lab with appropriate control on temperature. The leaching tank geometry has been concerned as an important factor to provide comparable agitation conditions. Data analysis has been done by using reactor design and mass balancing principles. Finally, optimum zone for operational parameters are determined for each leaching strategy and discussed due to their economical and practical aspects.

Keywords: titanium leaching, optimization, experimental design, performance analysis

Procedia PDF Downloads 372
14699 Random Vertical Seismic Vibrations of the Long Span Cantilever Beams

Authors: Sergo Esadze

Abstract:

Seismic resistance norms require calculation of cantilevers on vertical components of the base seismic acceleration. Long span cantilevers, as a rule, must be calculated as a separate construction element. According to the architectural-planning solution, functional purposes and environmental condition of a designing buildings/structures, long span cantilever construction may be of very different types: both by main bearing element (beam, truss, slab), and by material (reinforced concrete, steel). A choice from these is always linked with bearing construction system of the building. Research of vertical seismic vibration of these constructions requires individual approach for each (which is not specified in the norms) in correlation with model of seismic load. The latest may be given both as deterministic load and as a random process. Loading model as a random process is more adequate to this problem. In presented paper, two types of long span (from 6m – up to 12m) reinforcement concrete cantilever beams have been considered: a) bearing elements of cantilevers, i.e., elements in which they fixed, have cross-sections with large sizes and cantilevers are made with haunch; b) cantilever beam with load-bearing rod element. Calculation models are suggested, separately for a) and b) types. They are presented as systems with finite quantity degree (concentrated masses) of freedom. Conditions for fixing ends are corresponding with its types. Vertical acceleration and vertical component of the angular acceleration affect masses. Model is based on assumption translator-rotational motion of the building in the vertical plane, caused by vertical seismic acceleration. Seismic accelerations are considered as random processes and presented by multiplication of the deterministic envelope function on stationary random process. Problem is solved within the framework of the correlation theory of random process. Solved numerical examples are given. The method is effective for solving the specific problems.

Keywords: cantilever, random process, seismic load, vertical acceleration

Procedia PDF Downloads 189
14698 Optimization of Monascus Orange Pigments Production Using pH-Controlled Fed-Batch Fermentation

Authors: Young Min Kim, Deokyeong Choe, Chul Soo Shin

Abstract:

Monascus pigments, commonly used as a natural colorant in Asia, have many biological activities, such as cholesterol level control, anti-obesity, anti-cancer, and anti-oxidant, that have recently been elucidated. Especially, amino acid derivatives of Monascus pigments are receiving much attention because they have higher biological activities than original Monascus pigments. Previously, there have been two ways to produce amino acid derivatives: one-step production and two-step production. However, the one-step production has low purity, and the two-step production—precursor(orange pigments) fermentation and derivatives synthesis—has low productivity and growth rate during its precursor fermentation step. In this study, it was verified that pH is a key factor that affects the stability of orange pigments and the growth rate of Monascus. With an optimal pH profile obtained by pH-stat fermentation, we designed a process of precursor(orange pigments) fermentation that is a pH-controlled fed-batch fermentation. The final concentration of orange pigments in this process increased to 5.5g/L which is about 30% higher than the concentration produced from the previously used precursor fermentation step.

Keywords: cultivation process, fed-batch fermentation, monascus pigments, pH stability

Procedia PDF Downloads 298
14697 Energy-Led Sustainability Assessment Approach for Energy-Efficient Manufacturing

Authors: Aldona Kluczek

Abstract:

In recent years, manufacturing processes have interacted with sustainability issues realized in the cost-effective ways that minimalize energy, decrease negative impacts on the environment and are safe for society. However, the attention has been on separate sustainability assessment methods considering energy and material flow, energy consumption, and emission release or process control. In this paper, the energy-led sustainability assessment approach combining the methods: energy Life Cycle Assessment to assess environmental impact, Life Cycle Cost to analyze costs, and Social Life Cycle Assessment through ‘energy LCA-based value stream map’, is used to assess the energy sustainability of the hardwood lumber manufacturing process in terms of technologies. The approach integrating environmental, economic and social issues can be visualized in the considered energy-efficient technologies on the map of an energy LCA-related (input and output) inventory data. It will enable the identification of efficient technology of a given process to be reached, through the effective analysis of energy flow. It is also indicated that interventions in the considered technology should focus on environmental, economic improvements to achieve energy sustainability. The results have indicated that the most intense energy losses are caused by a cogeneration technology. The environmental impact analysis shows that a substantial reduction by 34% can be achieved with the improvement of it. From the LCC point of view, the result seems to be cost-effective, when done at that plant where the improvement is used. By demonstrating the social dimension, every component of the energy of plant labor use in the life-cycle process of the lumber production has positive energy benefits. The energy required to install the energy-efficient technology amounts to 30.32 kJ compared to others components of the energy of plant labor and it has the highest value in terms of energy-related social indicators. The paper depicts an example of hardwood lumber production in order to prove the applicability of a sustainability assessment method.

Keywords: energy efficiency, energy life cycle assessment, life cycle cost, social life cycle analysis, manufacturing process, sustainability assessment

Procedia PDF Downloads 247
14696 Re-Engineering Management Process in IRAN’s Smart Schools

Authors: M. R. Babaei, S. M. Hosseini, S. Rahmani, L. Moradi

Abstract:

Today, the quality of education and training systems and the effectiveness of the education systems of most concern to stakeholders and decision-makers of our country's development in each country. In Iran this is a double issue of concern to numerous reasons; So that governments, over the past decade have hardly even paid the running costs of education. ICT is claiming it has the power to change the structure of a program for training, reduce costs and increase quality, and do education systems and products consistent with the needs of the community and take steps to practice education. Own of the areas that the introduction of information technology has fundamentally changed is the field of education. The aim of this research is process reengineering management in schools simultaneously has been using field studies to collect data in the form of interviews and a questionnaire survey. The statistical community of this research has been the country of Iran and smart schools under the education. Sampling was targeted. The data collection tool was a questionnaire composed of two parts. The questionnaire consists of 36 questions that each question designates one of effective factors on the management of smart schools. Also each question consists of two parts. The first part designates the operating position in the management process, which represents the domain's belonging to the management agent (planning, organizing, leading, controlling). According to the classification of Dabryn and in second part the factors affect the process of managing the smart schools were examined, that Likert scale is used to classify. Questions the validity of the group of experts and prominent university professors in the fields of information technology, management and reengineering of approved and Cronbach's alpha reliability and also with the use of the formula is evaluated and approved. To analyse the data, descriptive and inferential statistics were used to analyse the factors contributing to the rating of (Linkert scale) descriptive statistics (frequency table data, mean, median, mode) was used. To analyse the data using analysis of variance and nonparametric tests and Friedman test, the assumption was evaluated. The research conclusions show that the factors influencing the management process re-engineering smart schools in school performance is affected.

Keywords: re-engineering, management process, smart school, Iran's school

Procedia PDF Downloads 244
14695 An Automatic Model Transformation Methodology Based on Semantic and Syntactic Comparisons and the Granularity Issue Involved

Authors: Tiexin Wang, Sebastien Truptil, Frederick Benaben

Abstract:

Model transformation, as a pivotal aspect of Model-driven engineering, attracts more and more attentions both from researchers and practitioners. Many domains (enterprise engineering, software engineering, knowledge engineering, etc.) use model transformation principles and practices to serve to their domain specific problems; furthermore, model transformation could also be used to fulfill the gap between different domains: by sharing and exchanging knowledge. Since model transformation has been widely used, there comes new requirement on it: effectively and efficiently define the transformation process and reduce manual effort that involved in. This paper presents an automatic model transformation methodology based on semantic and syntactic comparisons, and focuses particularly on granularity issue that existed in transformation process. Comparing to the traditional model transformation methodologies, this methodology serves to a general purpose: cross-domain methodology. Semantic and syntactic checking measurements are combined into a refined transformation process, which solves the granularity issue. Moreover, semantic and syntactic comparisons are supported by software tool; manual effort is replaced in this way.

Keywords: automatic model transformation, granularity issue, model-driven engineering, semantic and syntactic comparisons

Procedia PDF Downloads 395
14694 Hydrodynamic Study and Sizing of a Distillation Column by HYSYS Software

Authors: Derrouazin Mohammed Redhouane, Souakri Mohammed Lotfi, Henini Ghania

Abstract:

This work consists, first of all, of mastering one of the powerful process simulation tools currently used in the industrial processes, which is the HYSYS sizing software, and second, of simulating a petroleum distillation column. This study is divided into two parts; where the first one consists of a dimensioning of the column with a fast approximating method using state equations, iterative calculations, and then a precise simulation method with the HYSYS software. The second part of this study is a hydrodynamic study in order to verify by obtained results the proper functioning of the plates.

Keywords: industry process engineering, water distillation, environment, HYSYS simulation tool

Procedia PDF Downloads 129
14693 Road Maintenance Management Decision System Using Multi-Criteria and Geographical Information System for Takoradi Roads, Ghana

Authors: Eric Mensah, Carlos Mensah

Abstract:

The road maintenance backlogs created as a result of deferred maintenance especially in developing countries has caused considerable deterioration of many road assets. This is usually due to difficulties encountered in selecting and prioritising maintainable roads based on objective criteria rather than some political or other less important criteria. In order to ensure judicious use of limited resources for road maintenance, five factors were identified as the most important criteria for road management within the study area. This was based on the judgements of 40 experts. The results were further used to develop weightings using the Multi-Criteria Decision Process (MCDP) to analyse and select road alternatives according to maintenance goal. Using Geographical Information Systems (GIS), maintainable roads were grouped using the Jenk’s natural breaks to allow for further prioritised in order of importance for display on a dashboard of maps, charts, and tables. This reduces the problems of subjective maintenance and road selections, thereby reducing wastage of resources and easing the maintenance process through an object organised spatial decision support system.

Keywords: decision support, geographical information systems, multi-criteria decision process, weighted sum

Procedia PDF Downloads 376
14692 Automatic Classification of Periodic Heart Sounds Using Convolutional Neural Network

Authors: Jia Xin Low, Keng Wah Choo

Abstract:

This paper presents an automatic normal and abnormal heart sound classification model developed based on deep learning algorithm. MITHSDB heart sounds datasets obtained from the 2016 PhysioNet/Computing in Cardiology Challenge database were used in this research with the assumption that the electrocardiograms (ECG) were recorded simultaneously with the heart sounds (phonocardiogram, PCG). The PCG time series are segmented per heart beat, and each sub-segment is converted to form a square intensity matrix, and classified using convolutional neural network (CNN) models. This approach removes the need to provide classification features for the supervised machine learning algorithm. Instead, the features are determined automatically through training, from the time series provided. The result proves that the prediction model is able to provide reasonable and comparable classification accuracy despite simple implementation. This approach can be used for real-time classification of heart sounds in Internet of Medical Things (IoMT), e.g. remote monitoring applications of PCG signal.

Keywords: convolutional neural network, discrete wavelet transform, deep learning, heart sound classification

Procedia PDF Downloads 348
14691 Dynamic Process Model for Designing Smart Spaces Based on Context-Awareness and Computational Methods Principles

Authors: Heba M. Jahin, Ali F. Bakr, Zeyad T. Elsayad

Abstract:

As smart spaces can be defined as any working environment which integrates embedded computers, information appliances and multi-modal sensors to remain focused on the interaction between the users, their activity, and their behavior in the space; hence, smart space must be aware of their contexts and automatically adapt to their changing context-awareness, by interacting with their physical environment through natural and multimodal interfaces. Also, by serving the information used proactively. This paper suggests a dynamic framework through the architectural design process of the space based on the principles of computational methods and context-awareness principles to help in creating a field of changes and modifications. It generates possibilities, concerns about the physical, structural and user contexts. This framework is concerned with five main processes: gathering and analyzing data to generate smart design scenarios, parameters, and attributes; which will be transformed by coding into four types of models. Furthmore, connecting those models together in the interaction model which will represent the context-awareness system. Then, transforming that model into a virtual and ambient environment which represents the physical and real environments, to act as a linkage phase between the users and their activities taking place in that smart space . Finally, the feedback phase from users of that environment to be sure that the design of that smart space fulfill their needs. Therefore, the generated design process will help in designing smarts spaces that can be adapted and controlled to answer the users’ defined goals, needs, and activity.

Keywords: computational methods, context-awareness, design process, smart spaces

Procedia PDF Downloads 331
14690 The Implementation of the European Landscape Convention in Turkey: Opportunities and Constraints

Authors: Tutku Ak, Abdullah Kelkit, Cihad Öztürk

Abstract:

An increase has been witnessed with the number of multinational environmental agreements in the past decade, particularly in Europe. Success with implementation, however, shows variation. While many countries are willing to join these agreements, they do not always fully honor their obligations to put their commitments into practice. One reason for this is that countries have different legal and administrative systems. One example of an international multilateral environmental agreement is the European Landscape Convention (ELC). ELC expresses a concern to achieve sustainable development based on a balanced and harmonious relationship between social needs, economic activity, and the environment. Member states are required to implement the convention in accordance with their own administrative structure, respecting subsidiarity. In particular, the importance of cooperation in the protection, management, and planning of the resources is expressed through the convention. In this paper, it is intended to give a broad view of ELC’s implementation process in Turkey and what factors have influenced by the process. Under this context, the paper will focus on the objectives of the convention for addressing the issue of the loss of European landscapes, and the justification and tools used to accomplish these objectives. The degree to which these objectives have been implemented in Turkey and the opportunities and constraints that have been faced during this process have been discussed.

Keywords: European landscape convention, implementation, multinational environmental agreements, policy tools

Procedia PDF Downloads 298
14689 Effect of Equal Channel Angular Pressing Process on Impact Property of Pure Copper

Authors: Fahad Al-Mufadi, F. Djavanroodi

Abstract:

Ultrafine grained (UFG) and nanostructured (NS) materials have experienced a rapid development during the last decade and made profound impact on every field of materials science and engineering. The present work has been undertaken to develop ultra-fine grained pure copper by severe plastic deformation method and to examine the impact property by different characterizing tools. For this aim, equal channel angular pressing die with the channel angle, outer corner angle and channel diameter of 90°, 17° and 20 mm had been designed and manufactured. Commercial pure copper billets were ECAPed up to four passes by route BC at the ambient temperature. The results indicated that there is a great improvement at the hardness measurement, yield strength and ultimate tensile strength after ECAP process. It is found that the magnitudes of HV reach 136HV from 52HV after the final pass. Also, about 285% and 125% enhancement at the YS and UTS values have been obtained after the fourth pass as compared to the as-received conditions, respectively. On the other hand, the elongation to failure and impact energy have been reduced by imposing ECAP process and pass numbers. It is needed to say that about 56% reduction in the impact energy have been attained for the samples as contrasted to annealed specimens.

Keywords: SPD, ECAP, pure cu, impact property

Procedia PDF Downloads 259
14688 Comparing Xbar Charts: Conventional versus Reweighted Robust Estimation Methods for Univariate Data Sets

Authors: Ece Cigdem Mutlu, Burak Alakent

Abstract:

Maintaining the quality of manufactured products at a desired level depends on the stability of process dispersion and location parameters and detection of perturbations in these parameters as promptly as possible. Shewhart control chart is the most widely used technique in statistical process monitoring to monitor the quality of products and control process mean and variability. In the application of Xbar control charts, sample standard deviation and sample mean are known to be the most efficient conventional estimators in determining process dispersion and location parameters, respectively, based on the assumption of independent and normally distributed datasets. On the other hand, there is no guarantee that the real-world data would be normally distributed. In the cases of estimated process parameters from Phase I data clouded with outliers, efficiency of traditional estimators is significantly reduced, and performance of Xbar charts are undesirably low, e.g. occasional outliers in the rational subgroups in Phase I data set may considerably affect the sample mean and standard deviation, resulting a serious delay in detection of inferior products in Phase II. For more efficient application of control charts, it is required to use robust estimators against contaminations, which may exist in Phase I. In the current study, we present a simple approach to construct robust Xbar control charts using average distance to the median, Qn-estimator of scale, M-estimator of scale with logistic psi-function in the estimation of process dispersion parameter, and Harrell-Davis qth quantile estimator, Hodge-Lehmann estimator and M-estimator of location with Huber psi-function and logistic psi-function in the estimation of process location parameter. Phase I efficiency of proposed estimators and Phase II performance of Xbar charts constructed from these estimators are compared with the conventional mean and standard deviation statistics both under normality and against diffuse-localized and symmetric-asymmetric contaminations using 50,000 Monte Carlo simulations on MATLAB. Consequently, it is found that robust estimators yield parameter estimates with higher efficiency against all types of contaminations, and Xbar charts constructed using robust estimators have higher power in detecting disturbances, compared to conventional methods. Additionally, utilizing individuals charts to screen outlier subgroups and employing different combination of dispersion and location estimators on subgroups and individual observations are found to improve the performance of Xbar charts.

Keywords: average run length, M-estimators, quality control, robust estimators

Procedia PDF Downloads 190
14687 Nanda Ways of Knowing, Being and Doing: Our Process of Research Engagement and Research Impacts

Authors: Steven Kelly

Abstract:

A fundament role of the researcher is research engagement, that is, the interaction between researchers and research end-users outside of academia for the mutually beneficial transfer of knowledge, technologies, methods, or resources. While research impact is the contribution that research makes to the economy, society, environment, or culture beyond the contribution to academic research. Ironically, traditional impact metrics in the academy are designed to focus on the outputs; it dismisses the important role engagement plays in fostering a collaborative process that leads to meaningful, ethical, and useful impacts. Dr. Kelly, aNanda (First Nations) man himself, has worked closely with the Nanda community over the past decade, ensuring cultural protocols are upheld and implemented while doing research engagement. The focus was on the process, which was essential to foster a positive research impact culture. The contributions that flowed from this process were the naming of a new species of squat lobster in the Nanda language, a poster design in collaboration with The University of Melbourne, Museums Victoria and Bundiyarra - IrraWanga language centre, media coverage, and the formation of the “Nanda language, Nanda country project”. The Nanda language, Nanda country project is a language revitalization project that focused on reconnecting Nanda people with the language & culture on Nanda Country. Such outcomes are imperative on the eve of the United Nations International Decade of Indigenous Languages. In this paperDr, Kellywill discuss howNanda cultural practicesinformed research engagement to foster a collaborative processthat, in turn, ledto meaningful, ethical, and useful impacts within and outside of the academy.

Keywords: community collaboration, indigenous, nanda, research engagement, research impacts

Procedia PDF Downloads 114
14686 The Automatic Transliteration Model of Images of the Book Hamong Tani Using Statistical Approach

Authors: Agustinus Rudatyo Himamunanto, Anastasia Rita Widiarti

Abstract:

Transliteration using Javanese manuscripts is one of methods to preserve and legate the wealth of literature in the past for the present generation in Indonesia. The transliteration manual process commonly requires philologists and takes a relatively long time. The automatic transliteration process is expected to shorten the time so as to help the works of philologists. The preprocessing and segmentation stage firstly done is used to manage the document images, thus obtaining image script units that will compile input document images free from noise and have the similarity in properties in the thickness, size, and slope. The next stage of characteristic extraction is used to find unique characteristics that will distinguish each Javanese script image. One of characteristics that is used in this research is the number of black pixels in each image units. Each image of Java scripts contained in the data training will undergo the same process similar to the input characters. The system testing was performed with the data of the book Hamong Tani. The book Hamong Tani was selected due to its content, age and number of pages. Those were considered sufficient as a model experimental input. Based on the results of random page automatic transliteration process testing, it was determined that the maximum percentage correctness obtained was 81.53%. The percentage of success was obtained in 32x32 pixel input image size with the 5x5 image window. With regard to the results, it can be concluded that the automatic transliteration model offered is relatively good.

Keywords: Javanese script, character recognition, statistical, automatic transliteration

Procedia PDF Downloads 339
14685 Imputation Technique for Feature Selection in Microarray Data Set

Authors: Younies Saeed Hassan Mahmoud, Mai Mabrouk, Elsayed Sallam

Abstract:

Analysing DNA microarray data sets is a great challenge, which faces the bioinformaticians due to the complication of using statistical and machine learning techniques. The challenge will be doubled if the microarray data sets contain missing data, which happens regularly because these techniques cannot deal with missing data. One of the most important data analysis process on the microarray data set is feature selection. This process finds the most important genes that affect certain disease. In this paper, we introduce a technique for imputing the missing data in microarray data sets while performing feature selection.

Keywords: DNA microarray, feature selection, missing data, bioinformatics

Procedia PDF Downloads 574
14684 Towards a Comprehensive Framework on Civic Competence Development of Teachers: A Systematic Review of Literature

Authors: Emilie Vandevelde, Ellen Claes

Abstract:

This study aims to develop a comprehensive model for the civic socialization process of teachers. Citizenship has become one of the main objectives for the European education systems. It is expected that teachers are well prepared and equipped with the necessary knowledge, skills, and attitudes to also engage students in democratic citizenship. While a lot is known about young peoples’ civic competence development and how schools and teachers (don’t) support this process, less is known about how teachers themselves engage with (the teaching of) civics. Other than the civic socialization process of young adolescents that focuses on personal competence development, the civic socialization process of teachers includes the development of professional, civic competences. These professional competences make that they are able to prepare pupils to carry out their civic responsibilities in thoughtful ways. Existing models for the civic socialization process of young adolescents do not take this dual purpose into account. Based on these observations, this paper will investigate (1)What personal and professional civic competences teachers need to effectively teach civic education and (2) how teachers acquire these personal and professional civic competences. To answer the first research question, a systematic review of literature of existing civic education frameworks was carried out and linked to literature on teacher training. The second research question was addressed by adapting the Octagon model, developed by the International Association for the Evaluation of Educational Achievement (IEA), to the context of teachers. This was done by carrying out a systematic review of the recent literature linking three theoretical topics involved in teachers’ civic competence development: theories about the civic socialization process of young adolescents, Schulmans (1987) theoretical assumptions on pedagogical content knowledge (PCK), and Nogueira & Moreira’s (2012) framework for civic education teachers’ knowledge and literature on teachers’ professional development. This resulted in a comprehensive conceptual framework describing the personal and professional civic competences of civic education teachers. In addition, this framework is linked to the OctagonT model: a model that describes the processes through which teachers acquire these personal and professional civic competences. This model recognizes that teachers’ civic socialization process is influenced by interconnected variables located at different levels in a multi-level structure (the individual teacher (e.g., civic beliefs), everyday contacts (e.g., teacher educators, the intended, informal and hidden curriculum of the teacher training program, internship contacts, participation opportunities in teacher training, etc.) and the influence of the national educational context (e.g., vision on civic education)). Furthermore, implications for teacher education programs are described.

Keywords: civic education, civic competences, civic socialization, octagon model, teacher training

Procedia PDF Downloads 303
14683 Cluster Analysis and Benchmarking for Performance Optimization of a Pyrochlore Processing Unit

Authors: Ana C. R. P. Ferreira, Adriano H. P. Pereira

Abstract:

Given the frequent variation of mineral properties throughout the Araxá pyrochlore deposit, even if a good homogenization work has been carried out before feeding the processing plants, an operation with quality and performance’s high variety standard is expected. These results could be improved and standardized if the blend composition parameters that most influence the processing route are determined, and then the types of raw materials are grouped by them, finally presenting a great reference with operational settings for each group. Associating the physical and chemical parameters of a unit operation through benchmarking or even an optimal reference of metallurgical recovery and product quality reflects in the reduction of the production costs, optimization of the mineral resource, and guarantee of greater stability in the subsequent processes of the production chain that uses the mineral of interest. Conducting a comprehensive exploratory data analysis to identify which characteristics of the ore are most relevant to the process route, associated with the use of Machine Learning algorithms for grouping the raw material (ore) and associating these with reference variables in the process’ benchmark is a reasonable alternative for the standardization and improvement of mineral processing units. Clustering methods through Decision Tree and K-Means were employed, associated with algorithms based on the theory of benchmarking, with criteria defined by the process team in order to reference the best adjustments for processing the ore piles of each cluster. A clean user interface was created to obtain the outputs of the created algorithm. The results were measured through the average time of adjustment and stabilization of the process after a new pile of homogenized ore enters the plant, as well as the average time needed to achieve the best processing result. Direct gains from the metallurgical recovery of the process were also measured. The results were promising, with a reduction in the adjustment time and stabilization when starting the processing of a new ore pile, as well as reaching the benchmark. Also noteworthy are the gains in metallurgical recovery, which reflect a significant saving in ore consumption and a consequent reduction in production costs, hence a more rational use of the tailings dams and life optimization of the mineral deposit.

Keywords: mineral clustering, machine learning, process optimization, pyrochlore processing

Procedia PDF Downloads 143
14682 Three Dimensional Simulation of the Transient Modeling and Simulation of Different Gas Flows Velocity and Flow Distribution in Catalytic Converter with Porous Media

Authors: Amir Reza Radmanesh, Sina Farajzadeh Khosroshahi, Hani Sadr

Abstract:

The transient catalytic converter performance is governed by complex interactions between exhaust gas flow and the monolithic structure of the catalytic converter. Stringent emission regulations around the world necessitate the use of highly-efficient catalytic converters in vehicle exhaust systems. Computational fluid dynamics (CFD) is a powerful tool for calculating the flow field inside the catalytic converter. Radial velocity profiles, obtained by a commercial CFD code, present very good agreement with respective experimental results published in the literature. However the applicability of CFD for transient simulations is limited by the high CPU demands. In the present work, Geometric modeling ceramic monolith substrate is done with square shaped channel type of Catalytic converter and it is coated platinum and palladium. This example illustrates the effect of flow distribution on thermal response of a catalytic converter and different gas flow velocities, during the critical phase of catalytic converter warm up.

Keywords: catalytic converter, computational fluid dynamic, porous media, velocity distribution

Procedia PDF Downloads 858
14681 The Role of the Founding Ideals of a University in Higher Education

Authors: Masanori Kimura

Abstract:

The “founding spirit” of a Japanese university is similar to the mission statement of a Western university, but the difference between the two is that the former more closely reflects the founder’s inner world because it usually originates from the strong personal beliefs that the founder held when establishing the university. To find how much this ideology actually valued in today’s higher education, this paper surveys 2091 job openings for foreign language full-time faculty positions, posted by Japanese private universities from 2012 to 2016. The results suggest that women’s universities and universities with religious affiliations have a tendency to request that successful candidates observe their founding spirit, or at least demonstrate some understanding of the ideology after being hired. On the other hand, co-ed universities and universities with no religious affiliations do not show such a distinct tendency. A chi-square test revealed that this difference between the two groups was statistically significant (p<.01). Furthermore, the paper argues that it is not always appropriate for universities to evaluate themselves based on a single evaluation scale such as university rankings, and that both faculty and staff members need to be more aware of the founding spirit to improve the quality of the education the university provides.

Keywords: founding spirit, higher education, university administrative management, university evaluation

Procedia PDF Downloads 277
14680 Preparation Non-Woven Nanofiber Structures for Uniform and Rapid Drug Releasing Applications Using an Electrospinning Process

Authors: Cho-Liang Chung

Abstract:

Uniform and rapid drug release are important for trauma dressing application. Low glass transition polymer system and non-woven nanofiber structures as the designs conduct rapid-release characteristics. In this study, polyvinylpyrrolidone, polysulfone, and polystyrene were dissolved in dimethylformamide to form precursor solution. These solutions were blended with vitamin C to form the electrospinning solutions. The non-woven nanofibers structures were successfully prepared using an electrospinning process. The following instruments were used to analyze the characteristics of non-woven nanofibers structures: Atomic force microscopy (AFM), Field Emission Scanning Electron Microscope (FE-SEM), and X-ray Diffraction (XRD). The AFM was used to scan the nanofibers. 3D Graphics were applied to explore the surface morphology of nanofibers. FE-SEM was used to explore the morphology of non-woven structures. XRD was used to identify crystal structures in the non-woven structures. The evolution of morphology of non-woven structures was changed dramatically in different durations, because of the moisture absorption and decreasing glass transition temperature; the non-woven nanofiber structures can be applied to uniform and rapid drug release for trauma dressing application.

Keywords: nanofibers, non-woven, electrospinning process, rapid drug releasing

Procedia PDF Downloads 140
14679 Synthesis and Characterization of PVDF, FG, PTFE, and PES Membrane Distillation Modified with Silver Nanoparticles

Authors: Lopez J., Mehrvar M., Quinones E., Suarez A., RomeroC.

Abstract:

The Silver Nanoparticles (AgNP) are used as deliver of heat on surface of Membrane Distillation in order to fight against Thermal Polarization and improving the Desalination Process. In this study AgNPwere deposited by dip coating process over PVDF, FG hydrophilic, and PTFE hydrophobic commercial membranes as substrate. Membranes were characterized by SEM, EDS, contact angle, Pore size distributionand using a UV lamp and a thermal camera were measured the performance of heat deliver. The presence of AgNP 50 – 150 nm and the increase in absorption of energy over membrane were verified.

Keywords: silver nanoparticles, membrane distillation, plasmon effect, heat deliver

Procedia PDF Downloads 125