Search results for: data quality
8855 An Exhaustive Investigation of Green Building Certification on the Productivity and Mental and Physical Health of Buildings Occupants in Tehran, Iran
Authors: Armin Samarghandi, Amirreza Jafari, Mohamad Ghiasi
Abstract:
Since individuals spend the majority of their time indoors, this immediately affects their productivity and health. An essential factor in occupational health and public health is the influence of indoor environmental quality (IEQ) in buildings on occupant welfare and productivity. Since then, empirical data have been equivocal, indicating either that the studies are inaccurate or that the research has just scratched the surface of green buildings in offices, accommodation, and hospital settings and not taken the aforementioned holistically. This study compared three green-certified buildings — a residential green building, a green hospital, and a green school — with conventional structures in Tehran, Iran by means of a questionnaire spread among those utilizing these buildings, and assessing their productivity and health rate as opposed to the time they resided, worked in conventional buildings. The results demonstrated higher scores pertaining to productivity, physical and mental wellness as a consequence of better indoor environmental quality (IEQ), natural lighting, design, and sustainability of these building against non-green buildings. In addition, ancillary matters — environmental, financial, intellectual, emotional, social, spiritual dimensions of participants — were indirectly evaluated, and the same results are produced.
Keywords: Green building, Iran, productivity, physical and mental health, Tehran.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5538854 The Application of Six Sigma to Integration of Computer Based Systems
Authors: Zenon Chaczko, Essam Rahali, Rizwan Tariq
Abstract:
This paper introduces a process for the module level integration of computer based systems. It is based on the Six Sigma Process Improvement Model, where the goal of the process is to improve the overall quality of the system under development. We also present a conceptual framework that shows how this process can be implemented as an integration solution. Finally, we provide a partial implementation of key components in the conceptual framework.
Keywords: Software Quality, Six Sigma, System Integration, 3SI Process, 3SI Conceptual Framework.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16688853 A New History Based Method to Handle the Recurring Concept Shifts in Data Streams
Authors: Hossein Morshedlou, Ahmad Abdollahzade Barforoush
Abstract:
Recent developments in storage technology and networking architectures have made it possible for broad areas of applications to rely on data streams for quick response and accurate decision making. Data streams are generated from events of real world so existence of associations, which are among the occurrence of these events in real world, among concepts of data streams is logical. Extraction of these hidden associations can be useful for prediction of subsequent concepts in concept shifting data streams. In this paper we present a new method for learning association among concepts of data stream and prediction of what the next concept will be. Knowing the next concept, an informed update of data model will be possible. The results of conducted experiments show that the proposed method is proper for classification of concept shifting data streams.Keywords: Data Stream, Classification, Concept Shift, History.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 12778852 Incremental Learning of Independent Topic Analysis
Authors: Takahiro Nishigaki, Katsumi Nitta, Takashi Onoda
Abstract:
In this paper, we present a method of applying Independent Topic Analysis (ITA) to increasing the number of document data. The number of document data has been increasing since the spread of the Internet. ITA was presented as one method to analyze the document data. ITA is a method for extracting the independent topics from the document data by using the Independent Component Analysis (ICA). ICA is a technique in the signal processing; however, it is difficult to apply the ITA to increasing number of document data. Because ITA must use the all document data so temporal and spatial cost is very high. Therefore, we present Incremental ITA which extracts the independent topics from increasing number of document data. Incremental ITA is a method of updating the independent topics when the document data is added after extracted the independent topics from a just previous the data. In addition, Incremental ITA updates the independent topics when the document data is added. And we show the result applied Incremental ITA to benchmark datasets.Keywords: Text mining, topic extraction, independent, incremental, independent component analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10578851 Mitigation of Sag in Real Time
Authors: Vijay Gajanan Neve, Pallavi V. Pullawar, G. M. Dhole
Abstract:
Modern industrial processes are based on a large amount of electronic devices such as programmable logic controllers and adjustable speed drives. Unfortunately, electronic devices are sensitive to disturbances, and thus, industrial loads become less tolerant to power quality problems such as sags, swells, and harmonics. Voltage sags are an important power quality problem. In this paper proposed a new configuration of Static Var Compensator (SVC) considering three different conditions named as topologies and Booster transformer with fuzzy logic based controller, capable of compensating for power quality problems associated with voltage sags and maintaining a prescribed level of voltage profile. Fuzzy logic controller is designed to achieve the firing angles for SVC such that it maintains voltage profile. The online monitoring system for voltage sag mitigation in the laboratory using the hardware is used. The results are presented from the performance of each topology and Booster transformer considered in this paper.
Keywords: Booster Transformer, Fuzzy logic, Static Var Compensator, Voltage sag.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 25808850 Valorization of Beer Brewing Wastes by Composting
Authors: M. E. Silva, I. Brás
Abstract:
The aim of this work was to study the viability of recycling the residual yeast and diatomaceous earth (RYDE) slurry generated by the beer brewing industry by composting with animal manures, as well as to evaluate the quality of the composts obtained. Two pilot composting trials were carried out with different mixes: cow manure/RYDE slurry (Pile CM) and sheep manure/RYDE slurry (Pile SM). For all piles, wood chips were applied as bulking agent. The process was monitored by evaluating standard physical and chemical parameters. The compost quality was assessed by the heavy metals content and phytotoxicity. Both piles reached a thermophilic phase in the first day, however having different trends. The pH showed a slight alkaline character. The C/N reached values lower than 19 at the end of composting process. Generally, all the piles exhibited absence of heavy metals. However, the pile SM exhibited phytotoxicity. This study showed that RYDE slurry can be valorized by composting with cow manure.Keywords: Beer brewing wastes, compost; quality, valorization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10398849 Quad Tree Decomposition Based Analysis of Compressed Image Data Communication for Lossy and Lossless Using WSN
Authors: N. Muthukumaran, R. Ravi
Abstract:
The Quad Tree Decomposition based performance analysis of compressed image data communication for lossy and lossless through wireless sensor network is presented. Images have considerably higher storage requirement than text. While transmitting a multimedia content there is chance of the packets being dropped due to noise and interference. At the receiver end the packets that carry valuable information might be damaged or lost due to noise, interference and congestion. In order to avoid the valuable information from being dropped various retransmission schemes have been proposed. In this proposed scheme QTD is used. QTD is an image segmentation method that divides the image into homogeneous areas. In this proposed scheme involves analysis of parameters such as compression ratio, peak signal to noise ratio, mean square error, bits per pixel in compressed image and analysis of difficulties during data packet communication in Wireless Sensor Networks. By considering the above, this paper is to use the QTD to improve the compression ratio as well as visual quality and the algorithm in MATLAB 7.1 and NS2 Simulator software tool.
Keywords: Image compression, Compression Ratio, Quad tree decomposition, Wireless sensor networks, NS2 simulator.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 23908848 A Machine Learning Approach for Anomaly Detection in Environmental IoT-Driven Wastewater Purification Systems
Authors: Giovanni Cicceri, Roberta Maisano, Nathalie Morey, Salvatore Distefano
Abstract:
The main goal of this paper is to present a solution for a water purification system based on an Environmental Internet of Things (EIoT) platform to monitor and control water quality and machine learning (ML) models to support decision making and speed up the processes of purification of water. A real case study has been implemented by deploying an EIoT platform and a network of devices, called Gramb meters and belonging to the Gramb project, on wastewater purification systems located in Calabria, south of Italy. The data thus collected are used to control the wastewater quality, detect anomalies and predict the behaviour of the purification system. To this extent, three different statistical and machine learning models have been adopted and thus compared: Autoregressive Integrated Moving Average (ARIMA), Long Short Term Memory (LSTM) autoencoder, and Facebook Prophet (FP). The results demonstrated that the ML solution (LSTM) out-perform classical statistical approaches (ARIMA, FP), in terms of both accuracy, efficiency and effectiveness in monitoring and controlling the wastewater purification processes.Keywords: EIoT, machine learning, anomaly detection, environment monitoring.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10258847 A Study of Distinctive Models for Pre-hospital EMS in Thailand: Knowledge Capture
Authors: R. Sinthavalai, N. Memongkol, N. Patthanaprechawong, J. Viriyanantavong, C. Choosuk
Abstract:
In Thailand, the practice of pre-hospital Emergency Medical Service (EMS) in each area reveals the different growth rates and effectiveness of the practices. Those can be found as the diverse quality and quantity. To shorten the learning curve prior to speed-up the practices in other areas, story telling and lessons learnt from the effective practices are valued as meaningful knowledge. To this paper, it was to ascertain the factors, lessons learnt and best practices that have impact as contributing to the success of prehospital EMS system. Those were formulized as model prior to speedup the practice in other areas. To develop the model, Malcolm Baldrige National Quality Award (MBNQA), which is widely recognized as a framework for organizational quality assessment and improvement, was chosen as the discussion framework. Remarkably, this study was based on the consideration of knowledge capture; however it was not to complete the loop of knowledge activities. Nevertheless, it was to highlight the recognition of knowledge capture, which is the initiation of knowledge management.Keywords: Emergency Medical Service, Modeling, MBNQA, Thailand.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15578846 A Framework for Data Mining Based Multi-Agent: An Application to Spatial Data
Authors: H. Baazaoui Zghal, S. Faiz, H. Ben Ghezala
Abstract:
Data mining is an extraordinarily demanding field referring to extraction of implicit knowledge and relationships, which are not explicitly stored in databases. A wide variety of methods of data mining have been introduced (classification, characterization, generalization...). Each one of these methods includes more than algorithm. A system of data mining implies different user categories,, which mean that the user-s behavior must be a component of the system. The problem at this level is to know which algorithm of which method to employ for an exploratory end, which one for a decisional end, and how can they collaborate and communicate. Agent paradigm presents a new way of conception and realizing of data mining system. The purpose is to combine different algorithms of data mining to prepare elements for decision-makers, benefiting from the possibilities offered by the multi-agent systems. In this paper the agent framework for data mining is introduced, and its overall architecture and functionality are presented. The validation is made on spatial data. Principal results will be presented.
Keywords: Databases, data mining, multi-agent, spatial datamart.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20448845 Dynamic Admission Control Based on Effective Demand for Next Generation Wireless Networks
Authors: Somenath Mukherjee, Rajdeep Ray, Raj Kumar Samanta, Mofazzal H. Khondekar, Gautam Sanyal
Abstract:
In next generation wireless networks (i.e., 4G and beyond), one of the main objectives is to ensure highest level of customer satisfaction in terms of data transfer speed, decrease in cost and delay, non-rejection and no drop of calls, availability of ‘always-on’ connectivity and services, continuity of connected services, hastle-free roaming in addition to the convenience of use of network services from anywhere and anytime. To take care of these requirements effectively, internet service providers (ISPs) and network planners have to go for major capacity enhancement of network resources and at the same time these resources are to be used effectively and efficiently to reduce cost and to increase revenue. In this work, the effective bandwidth available in a Mobile Switching Center (MSC) of a wireless network providing multi-class multimedia services is analyzed. Bandwidth requirement of the users for a customized Quality of Service (QoS) is estimated. The findings of the QoS estimation are applied for the capacity planning and admission control of the multi-class traffic flows coming into the MSC.
Keywords: Next generation wireless network, mobile switching center, multi-class traffic, quality of service, admission control, effective bandwidth.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 8408844 Latent Topic Based Medical Data Classification
Authors: Jian-hua Yeh, Shi-yi Kuo
Abstract:
This paper discusses the classification process for medical data. In this paper, we use the data from ACM KDDCup 2008 to demonstrate our classification process based on latent topic discovery. In this data set, the target set and outliers are quite different in their nature: target set is only 0.6% size in total, while the outliers consist of 99.4% of the data set. We use this data set as an example to show how we dealt with this extremely biased data set with latent topic discovery and noise reduction techniques. Our experiment faces two major challenge: (1) extremely distributed outliers, and (2) positive samples are far smaller than negative ones. We try to propose a suitable process flow to deal with these issues and get a best AUC result of 0.98.
Keywords: classification, latent topics, outlier adjustment, feature scaling
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16418843 Contamination in Industrial Areas and Environmental Management in Latvia
Authors: Juris Burlakovs, Maris Klavins, Raimonds Ernsteins, Armands Ruskulis
Abstract:
Environmental contamination is a common problem in ex-industrial and industrial sites. This article gives a brief description of general applied environmental investigation methodologies and possible remediation applications in Latvia. Most of contaminated areas are situated in former and active industrial, military areas and ports. Industrial and logistic activities very often have been with great impact for more than hundred years thus the contamination level with heavy metals, hydrocarbons, pesticides, persistent organic pollutants is high and is threatening health and environment in general. 242 territories now are numbered as contaminated and fixed in the National Register of contaminated territories in Latvia. Research and remediation of contamination in densely populated areas are of important environmental policy domain. Four different investigation case studies of contaminated areas are given describing the history of use, environmental quality assessment as well as planned environmental management actions. All four case study locations are situated in Riga - the capital of the Republic of Latvia. The aim of this paper is to analyze the situation and problems with management of contaminated areas in Latvia, give description of field research methods and recommendations for remediation industry based on scientific data and innovations.
Keywords: Remediation technology, environmental quality assessment, heavy metals, hydrocarbon contamination, environmental management.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22998842 Data Transformation Services (DTS): Creating Data Mart by Consolidating Multi-Source Enterprise Operational Data
Authors: J. D. D. Daniel, K. N. Goh, S. M. Yusop
Abstract:
Trends in business intelligence, e-commerce and remote access make it necessary and practical to store data in different ways on multiple systems with different operating systems. As business evolve and grow, they require efficient computerized solution to perform data update and to access data from diverse enterprise business applications. The objective of this paper is to demonstrate the capability of DTS [1] as a database solution for automatic data transfer and update in solving business problem. This DTS package is developed for the sales of variety of plants and eventually expanded into commercial supply and landscaping business. Dimension data modeling is used in DTS package to extract, transform and load data from heterogeneous database systems such as MySQL, Microsoft Access and Oracle that consolidates into a Data Mart residing in SQL Server. Hence, the data transfer from various databases is scheduled to run automatically every quarter of the year to review the efficient sales analysis. Therefore, DTS is absolutely an attractive solution for automatic data transfer and update which meeting today-s business needs.Keywords: Data Transformation Services (DTS), ObjectLinking and Embedding Database (OLEDB), Data Mart, OnlineAnalytical Processing (OLAP), Online Transactional Processing(OLTP).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20358841 Extraction of Data from Web Pages: A Vision Based Approach
Authors: P. S. Hiremath, Siddu P. Algur
Abstract:
With the explosive growth of information sources available on the World Wide Web, it has become increasingly difficult to identify the relevant pieces of information, since web pages are often cluttered with irrelevant content like advertisements, navigation-panels, copyright notices etc., surrounding the main content of the web page. Hence, tools for the mining of data regions, data records and data items need to be developed in order to provide value-added services. Currently available automatic techniques to mine data regions from web pages are still unsatisfactory because of their poor performance and tag-dependence. In this paper a novel method to extract data items from the web pages automatically is proposed. It comprises of two steps: (1) Identification and Extraction of the data regions based on visual clues information. (2) Identification of data records and extraction of data items from a data region. For step1, a novel and more effective method is proposed based on visual clues, which finds the data regions formed by all types of tags using visual clues. For step2 a more effective method namely, Extraction of Data Items from web Pages (EDIP), is adopted to mine data items. The EDIP technique is a list-based approach in which the list is a linear data structure. The proposed technique is able to mine the non-contiguous data records and can correctly identify data regions, irrespective of the type of tag in which it is bound. Our experimental results show that the proposed technique performs better than the existing techniques.
Keywords: Web data records, web data regions, web mining.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18998840 Visual-Graphical Methods for Exploring Longitudinal Data
Authors: H. W. Ker
Abstract:
Longitudinal data typically have the characteristics of changes over time, nonlinear growth patterns, between-subjects variability, and the within errors exhibiting heteroscedasticity and dependence. The data exploration is more complicated than that of cross-sectional data. The purpose of this paper is to organize/integrate of various visual-graphical techniques to explore longitudinal data. From the application of the proposed methods, investigators can answer the research questions include characterizing or describing the growth patterns at both group and individual level, identifying the time points where important changes occur and unusual subjects, selecting suitable statistical models, and suggesting possible within-error variance.Keywords: Data exploration, exploratory analysis, HLMs/LMEs, longitudinal data, visual-graphical methods.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20938839 A Materialized Approach to the Integration of XML Documents: the OSIX System
Authors: H. Ahmad, S. Kermanshahani, A. Simonet, M. Simonet
Abstract:
The data exchanged on the Web are of different nature from those treated by the classical database management systems; these data are called semi-structured data since they do not have a regular and static structure like data found in a relational database; their schema is dynamic and may contain missing data or types. Therefore, the needs for developing further techniques and algorithms to exploit and integrate such data, and extract relevant information for the user have been raised. In this paper we present the system OSIX (Osiris based System for Integration of XML Sources). This system has a Data Warehouse model designed for the integration of semi-structured data and more precisely for the integration of XML documents. The architecture of OSIX relies on the Osiris system, a DL-based model designed for the representation and management of databases and knowledge bases. Osiris is a viewbased data model whose indexing system supports semantic query optimization. We show that the problem of query processing on a XML source is optimized by the indexing approach proposed by Osiris.Keywords: Data integration, semi-structured data, views, XML.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15898838 Web Content Mining: A Solution to Consumer's Product Hunt
Authors: Syed Salman Ahmed, Zahid Halim, Rauf Baig, Shariq Bashir
Abstract:
With the rapid growth in business size, today's businesses orient towards electronic technologies. Amazon.com and e-bay.com are some of the major stakeholders in this regard. Unfortunately the enormous size and hugely unstructured data on the web, even for a single commodity, has become a cause of ambiguity for consumers. Extracting valuable information from such an everincreasing data is an extremely tedious task and is fast becoming critical towards the success of businesses. Web content mining can play a major role in solving these issues. It involves using efficient algorithmic techniques to search and retrieve the desired information from a seemingly impossible to search unstructured data on the Internet. Application of web content mining can be very encouraging in the areas of Customer Relations Modeling, billing records, logistics investigations, product cataloguing and quality management. In this paper we present a review of some very interesting, efficient yet implementable techniques from the field of web content mining and study their impact in the area specific to business user needs focusing both on the customer as well as the producer. The techniques we would be reviewing include, mining by developing a knowledge-base repository of the domain, iterative refinement of user queries for personalized search, using a graphbased approach for the development of a web-crawler and filtering information for personalized search using website captions. These techniques have been analyzed and compared on the basis of their execution time and relevance of the result they produced against a particular search.
Keywords: Data mining, web mining, search engines, knowledge discovery.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20528837 Quality Attributes of Various Spray Dried Pulp Powder Prepared from Low Temperature Stored Calcium Salts Pretreated Guava Fruits
Authors: Renu Rahel, A. S. Chauhan, K. Srinivasulu, R. Ravi, V. B. Kudachikar
Abstract:
The effect of calcium salts on the storage stability and on the quality attributes of both fresh and processed product (guava powder) of white flesh guavas (var ‘Allahabad safeda’) was studied. The pulp behavioral studies of fully ripened guava fruits indicated that fruits pretreated with 3% and 4.5% calcium chloride had the least viscosity. The guava pulp powder using spray drying technique was developed and its storage stability and the moisture sorption studies were carried out for product quality evaluation at normal storage condition (27°C; 65%RH). Results revealed that powder obtained from 3% calcium chloride pretreated guavas was found to be at par with the powder obtained from control guavas after 90 days of normal storage. Studies on microbiological quality of guava pulp powder indicated that among the treatments powder obtained from guava fruit pretreated with 3% calcium chloride to be the most effective through restricting microbial counts of total plate count, yeast, mold, Staphylococcus and E. coli below their permissible limit. Moisture sorption studies of guava powder revealed that foil laminate 12μm PET/9 μm foil/38-40 μm is the most suitable packaging material recommended.Keywords: White flesh guava, calcium salts, spray drying, powder, storage stability.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20638836 Identifying and Prioritizing Factors Affecting Consumer Behavior Based on Product Value
Authors: Houshang Taghizadeh, Gholamreza Soltani Fesghandis
Abstract:
Nowadays, without the awareness of consumer behavior and correct understanding of it, it is not possible for organizations to take appropriate measures to meet the consumer needs and demands. The aim of this paper is the identification and prioritization of the factors affecting the consumer behavior based on the product value. The population of the study includes all the consumers of furniture producing firms in East Azarbaijan province, Iran. The research sample includes 93 people selected by the sampling formula in unlimited population. The data collection instrument was a questionnaire, the validity of which was confirmed through face validity and the reliability of which was determined, using Cronbach's alpha coefficient. The Kolmogorov-Smironov test was used to test data normality, the t-test for identification of factors affecting the product value, and Friedman test for prioritizing the factors. The results show that quality, satisfaction, styling, price, finishing operation, performance, safety, worth, shape, use, and excellence are placed from 1 to 11 priorities, respectively.Keywords: Consumer Behavior, Consumer Satisfaction, Product, Value
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 29028835 Tools for Analysis and Optimization of Standalone Green Microgrids
Authors: William Anderson, Kyle Kobold, Oleg Yakimenko
Abstract:
Green microgrids using mostly renewable energy (RE) for generation, are complex systems with inherent nonlinear dynamics. Among a variety of different optimization tools there are only a few ones that adequately consider this complexity. This paper evaluates applicability of two somewhat similar optimization tools tailored for standalone RE microgrids and also assesses a machine learning tool for performance prediction that can enhance the reliability of any chosen optimization tool. It shows that one of these microgrid optimization tools has certain advantages over another and presents a detailed routine of preparing input data to simulate RE microgrid behavior. The paper also shows how neural-network-based predictive modeling can be used to validate and forecast solar power generation based on weather time series data, which improves the overall quality of standalone RE microgrid analysis.Keywords: Microgrid, renewable energy, complex systems, optimization, predictive modeling, neural networks.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10588834 Balance of Rural and Urban Structures
Authors: Ehrenstorfer Barbara, Peherstorfer Tanja, Nový Jan
Abstract:
Urbanization and regionalization are two different approaches when it comes to economical structures and development, infrastructure and mobility, quality of life and living, education, social cohesion and many other topics. At first glance, the structures associated with urbanization and regionalization seems to be contradicting. This paper discusses possibilities of transfer and cooperation between rural and urban structures. An empirical investigation contributed to reveal scenarios of supposable forms of exchange and cooperation of remote rural areas and big cities.Keywords: Learning Regions, Quality of Life and Living, Regional and Rural Development, Social Innovation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17378833 Effect of Visual Speech in Sign Speech Synthesis
Authors: Zdenek Krnoul
Abstract:
This article investigates a contribution of synthesized visual speech. Synthesis of visual speech expressed by a computer consists in an animation in particular movements of lips. Visual speech is also necessary part of the non-manual component of a sign language. Appropriate methodology is proposed to determine the quality and the accuracy of synthesized visual speech. Proposed methodology is inspected on Czech speech. Hence, this article presents a procedure of recording of speech data in order to set a synthesis system as well as to evaluate synthesized speech. Furthermore, one option of the evaluation process is elaborated in the form of a perceptual test. This test procedure is verified on the measured data with two settings of the synthesis system. The results of the perceptual test are presented as a statistically significant increase of intelligibility evoked by real and synthesized visual speech. Now, the aim is to show one part of evaluation process which leads to more comprehensive evaluation of the sign speech synthesis system.
Keywords: Perception test, Sign speech synthesis, Talking head, Visual speech.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14758832 Education Quality Development for Excellence Performance with Higher Education by Using COBIT 5
Authors: Kemkanit Sanyanunthana
Abstract:
The purpose of this research is to study the management system of information technology which supports the education of five private universities in Thailand, according to the case studies which have been developing their qualities and standards of management and education by service provision of information technology to support the excellence performance. The concept to connect information technology with a suitable system has been created by information technology administrators for development, as a system that can be used throughout the organizations to help reach the utmost benefits of using all resources. Hence, the researcher as a person who has been performing these duties within higher education is interested to do this research by selecting the Control Objective for Information and Related Technology 5 (COBIT 5) for the Malcolm Baldrige National Quality Award (MBNQA) of America, or the National Award which applies the concept of Total Quality Management (TQM) to the organization evaluation. Such evaluation is called the Education Criteria for Performance Excellence (EdPEx) focuses on studying and comparing education quality development for excellent performance using COBIT 5 in terms of information technology to study the problems and obstacles of the investigation process for an information technology system, which is considered as an instrument to drive all organizations to reach the excellence performance of the information technology, and to be the model of evaluation and analysis of the process to be in accordance with the strategic plans of the information technology in the universities. This research is conducted in the form of descriptive and survey research according to the case studies. The data collection were carried out by using questionnaires through the administrators working related to the information technology field, and the research documents related to the change management as the main study. The research can be concluded that the performance based on the APO domain process (ALIGN, PLAN AND ORGANISE) of the COBIT 5 standard frame, which emphasizes concordant governance and management of strategic plans for the organizations, could reach only 95%. This might be because of some restrictions such as organizational cultures; therefore, the researcher has studied and analyzed the management of information technology in universities as a whole, under the organizational structures, to reach the performance in accordance with the overall APO domain which would affect the determined strategic plans to be able to develop based on the excellence performance of information technology, and to apply the risk management system at the organizational level into every performance process which would develop the work effectiveness for the resources management of information technology to reach the utmost benefits.
Keywords: COBIT 5, APO, EdPEx Criteria, MBNQA.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14878831 Fuzzy Hyperbolization Image Enhancement and Artificial Neural Network for Anomaly Detection
Authors: Sri Hartati, 1Agus Harjoko, Brad G. Nickerson
Abstract:
A prototype of an anomaly detection system was developed to automate process of recognizing an anomaly of roentgen image by utilizing fuzzy histogram hyperbolization image enhancement and back propagation artificial neural network. The system consists of image acquisition, pre-processor, feature extractor, response selector and output. Fuzzy Histogram Hyperbolization is chosen to improve the quality of the roentgen image. The fuzzy histogram hyperbolization steps consist of fuzzyfication, modification of values of membership functions and defuzzyfication. Image features are extracted after the the quality of the image is improved. The extracted image features are input to the artificial neural network for detecting anomaly. The number of nodes in the proposed ANN layers was made small. Experimental results indicate that the fuzzy histogram hyperbolization method can be used to improve the quality of the image. The system is capable to detect the anomaly in the roentgen image.Keywords: Image processing, artificial neural network, anomaly detection.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21118830 Multivariate Analytical Insights into Spatial and Temporal Variation in Water Quality of a Major Drinking Water Reservoir
Authors: Azadeh Golshan, Craig Evans, Phillip Geary, Abigail Morrow, Zoe Rogers, Marcel Maeder
Abstract:
22 physicochemical variables have been determined in water samples collected weekly from January to December in 2013 from three sampling stations located within a major drinking water reservoir. Classical Multivariate Curve Resolution Alternating Least Squares (MCR-ALS) analysis was used to investigate the environmental factors associated with the physico-chemical variability of the water samples at each of the sampling stations. Matrix augmentation MCR-ALS (MA-MCR-ALS) was also applied, and the two sets of results were compared for interpretative clarity. Links between these factors, reservoir inflows and catchment land-uses were investigated and interpreted in relation to chemical composition of the water and their resolved geographical distribution profiles. The results suggested that the major factors affecting reservoir water quality were those associated with agricultural runoff, with evidence of influence on algal photosynthesis within the water column. Water quality variability within the reservoir was also found to be strongly linked to physical parameters such as water temperature and the occurrence of thermal stratification. The two methods applied (MCR-ALS and MA-MCR-ALS) led to similar conclusions; however, MA-MCR-ALS appeared to provide results more amenable to interpretation of temporal and geological variation than those obtained through classical MCR-ALS.
Keywords: Catchment management, drinking water reservoir, multivariate curve resolution alternating least squares, thermal stratification, water quality.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9208829 Comparison of Neural Network and Logistic Regression Methods to Predict Xerostomia after Radiotherapy
Authors: Hui-Min Ting, Tsair-Fwu Lee, Ming-Yuan Cho, Pei-Ju Chao, Chun-Ming Chang, Long-Chang Chen, Fu-Min Fang
Abstract:
To evaluate the ability to predict xerostomia after radiotherapy, we constructed and compared neural network and logistic regression models. In this study, 61 patients who completed a questionnaire about their quality of life (QoL) before and after a full course of radiation therapy were included. Based on this questionnaire, some statistical data about the condition of the patients’ salivary glands were obtained, and these subjects were included as the inputs of the neural network and logistic regression models in order to predict the probability of xerostomia. Seven variables were then selected from the statistical data according to Cramer’s V and point-biserial correlation values and were trained by each model to obtain the respective outputs which were 0.88 and 0.89 for AUC, 9.20 and 7.65 for SSE, and 13.7% and 19.0% for MAPE, respectively. These parameters demonstrate that both neural network and logistic regression methods are effective for predicting conditions of parotid glands.
Keywords: NPC, ANN, logistic regression, xerostomia.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16358828 Data-Driven Decision-Making in Digital Entrepreneurship
Authors: Abeba Nigussie Turi, Xiangming Samuel Li
Abstract:
Data-driven business models are more typical for established businesses than early-stage startups that strive to penetrate a market. This paper provided an extensive discussion on the principles of data analytics for early-stage digital entrepreneurial businesses. Here, we developed data-driven decision-making (DDDM) framework that applies to startups prone to multifaceted barriers in the form of poor data access, technical and financial constraints, to state some. The startup DDDM framework proposed in this paper is novel in its form encompassing startup data analytics enablers and metrics aligning with startups' business models ranging from customer-centric product development to servitization which is the future of modern digital entrepreneurship.
Keywords: Startup data analytics, data-driven decision-making, data acquisition, data generation, digital entrepreneurship.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 8268827 Design and Analysis of Gauge R&R Studies: Making Decisions Based on ANOVA Method
Authors: Afrooz Moatari Kazerouni
Abstract:
In a competitive production environment, critical decision making are based on data resulted by random sampling of product units. Efficiency of these decisions depends on data quality and also their reliability scale. This point leads to the necessity of a reliable measurement system. Therefore, the conjecture process and analysing the errors contributes to a measurement system known as Measurement System Analysis (MSA). The aim of this research is on determining the necessity and assurance of extensive development in analysing measurement systems, particularly with the use of Repeatability and Reproducibility Gages (GR&R) to improve physical measurements. Nowadays in productive industries, repeatability and reproducibility gages released so well but they are not applicable as well as other measurement system analysis methods. To get familiar with this method and gain a feedback in improving measurement systems, this survey would be on “ANOVA" method as the most widespread way of calculating Repeatability and Reproducibility (R&R).Keywords: Analysis of Variance (ANOVA), MeasurementSystem Analysis (MSA), Part-Operator interaction effect, Repeatability and Reproducibility.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 46678826 A Review on Process Parameters of Ti/Al Dissimilar Joint Using Laser Beam Welding
Authors: K. Kalaiselvan, K. Sekar, S. Elavarasi
Abstract:
The use of laser beam welding for joining titanium and aluminum offers more advantages compared with conventional joining processes. Dissimilar metal combination is very much needed for aircraft structural industries and research activities. The quality of a weld joint is directly influenced by the welding input parameters. The common problem that is faced by the manufactures is the control of the process parameters to obtain a good weld joint with minimal detrimental. To overcome this issue, various parameters can be preferred to obtain quality of weld joint. In this present study an overall literature review on processing parameters such as offset distance, welding speed, laser power, shielding gas and filler metals are discussed with the effects on quality weldment. Additionally, mechanical properties of welds joint are discussed. The aim of the report is to review the recent progress in the welding of dissimilar titanium (Ti) and aluminum (Al) alloys to provide a basis for follow up research.
Keywords: Laser beam welding, titanium, aluminum, process parameters.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 604