Search results for: sensor-based methods
915 A New Suburb Renovation Concept
Authors: A. Soikkeli, L. Sorri
Abstract:
Finnish national research project, User- and Business-oriented Suburb Renovation Concept (KLIKK), was started in January 2012 and will end in June 2014. The perspective of energy efficiency is emphasised in the project, but also it addresses what improving the energy efficiency of suburban apartment buildings means from the standpoint of architecturally valuable buildings representing different periods. The project will also test the impacts of stricter energy efficiency requirements on renovation projects.
The primary goal of the project is to develop a user-oriented, industrial, economic renovation concept for suburban apartment building renovation, extension and construction of additional storeys. The concept will make it possible to change from performance- and cost-based operation to novel service- and user-oriented, site-specifically tailored renovation methods utilizing integrated order and delivery chains.
The present project is collaborating with Ministry of the Environment and participating cities in developing a new type of lighter town planning model for suburban renovations and in-fill construction. To support this, the project will simultaneously develop practices for environmental impact assessment tools in renovation and suburban supplementary and in-fill construction.
Keywords: Energy efficiency, Prefabrication, Renovation concept, Suburbs, Sustainability, User-Orientated.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2037914 Chemical Composition of Variety 'Nante' Hybrid Carrots Cultivated in Latvia
Authors: Tatjana Rakcejeva, Ingrida Augspole, Lija Dukalska, Fredijs Dimins
Abstract:
carrot is one of the important root vegetable crops, and it is highly nutritious as it contains appreciable amount of vitamins, minerals and β-carotene. The major objective of current research was to evaluate the chemical composition of carrot variety 'Nante' hybrids in general and to select the best samples for fresh-cut salad production. The research was accomplished on fresh in Latvia cultivated carrots harvested in Zemgale region in the first part of October, 2011 and immediately used for experiments. Late-bearing variety 'Nante' hybrid carrots were used for analysis: 'Nante/Berlikum', 'Nante/Maestro', 'Nante/Forto', 'Nante/Bolero' and 'Nante/Champion'. The quality parameters as moisture, soluble solid, firmness, b-carotene, carotenoid, color, polyphenols, total phenolic compounds and total antioxidant capacity were analyzed using standard methods. For fresh-cut salad production as more applicable could be recommended hybrids 'Nante/Forto' and 'Nante/Berlikum' - mainly because it-s higher nutritive value, as higher total phenolic compounds, polyphenols and pronounced antioxidant capacity.Keywords: carrots, chemical composition, evaluation
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2170913 The Emoji Method: An Approach for Identifying and Formulating Problem Ideas
Authors: Thorsten Herrmann, Alexander Laukemann, Hansgeorg Binz, Daniel Roth
Abstract:
For the analysis of already identified and existing problems, the pertinent literature provides a comprehensive collection of approaches as well as methods in order to analyze the problems in detail. But coming up with problems, which are assets worth pursuing further, is often challenging. However, the importance of well-formulated problem ideas and their influence of subsequent creative processes are incontestable and proven. In order to meet the covered challenges, the Institute for Engineering Design and Industrial Design (IKTD) developed the Emoji Method. This paper presents the Emoji Method, which support designers to generate problem ideas in a structured way. Considering research findings from knowledge management and innovation management, research into emojis and emoticons reveal insights by means of identifying and formulating problem ideas within the early design phase. The simple application and the huge supporting potential of the Emoji Method within the early design phase are only few of the many successful results of the conducted evaluation. The Emoji Method encourages designers to identify problem ideas and describe them in a structured way in order to start focused with generating solution ideas for the revealed problem ideas.
Keywords: Emojis, problem ideas, innovation management, knowledge management.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1081912 An Integrative Bayesian Approach to Supporting the Prediction of Protein-Protein Interactions: A Case Study in Human Heart Failure
Authors: Fiona Browne, Huiru Zheng, Haiying Wang, Francisco Azuaje
Abstract:
Recent years have seen a growing trend towards the integration of multiple information sources to support large-scale prediction of protein-protein interaction (PPI) networks in model organisms. Despite advances in computational approaches, the combination of multiple “omic" datasets representing the same type of data, e.g. different gene expression datasets, has not been rigorously studied. Furthermore, there is a need to further investigate the inference capability of powerful approaches, such as fullyconnected Bayesian networks, in the context of the prediction of PPI networks. This paper addresses these limitations by proposing a Bayesian approach to integrate multiple datasets, some of which encode the same type of “omic" data to support the identification of PPI networks. The case study reported involved the combination of three gene expression datasets relevant to human heart failure (HF). In comparison with two traditional methods, Naive Bayesian and maximum likelihood ratio approaches, the proposed technique can accurately identify known PPI and can be applied to infer potentially novel interactions.Keywords: Bayesian network, Classification, Data integration, Protein interaction networks.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1616911 Efficient High Fidelity Signal Reconstruction Based on Level Crossing Sampling
Authors: Negar Riazifar, Nigel G. Stocks
Abstract:
This paper proposes strategies in level crossing (LC) sampling and reconstruction that provide high fidelity signal reconstruction for speech signals; these strategies circumvent the problem of exponentially increasing number of samples as the bit-depth is increased and hence are highly efficient. Specifically, the results indicate that the distribution of the intervals between samples is one of the key factors in the quality of signal reconstruction; including samples with short intervals does not improve the accuracy of the signal reconstruction, whilst samples with large intervals lead to numerical instability. The proposed sampling method, termed reduced conventional level crossing (RCLC) sampling, exploits redundancy between samples to improve the efficiency of the sampling without compromising performance. A reconstruction technique is also proposed that enhances the numerical stability through linear interpolation of samples separated by large intervals. Interpolation is demonstrated to improve the accuracy of the signal reconstruction in addition to the numerical stability. We further demonstrate that the RCLC and interpolation methods can give useful levels of signal recovery even if the average sampling rate is less than the Nyquist rate.
Keywords: Level crossing sampling, numerical stability, speech processing, trigonometric polynomial.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 430910 Anonymous Editing Prevention Technique Using Gradient Method for High-Quality Video
Authors: Jiwon Lee, Chanho Jung, Si-Hwan Jang, Kyung-Ill Kim, Sanghyun Joo, Wook-Ho Son
Abstract:
Since the advances in digital imaging technologies have led to development of high quality digital devices, there are a lot of illegal copies of copyrighted video content on the Internet. Also, unauthorized editing is occurred frequently. Thus, we propose an editing prevention technique for high-quality (HQ) video that can prevent these illegally edited copies from spreading out. The proposed technique is applied spatial and temporal gradient methods to improve the fidelity and detection performance. Also, the scheme duplicates the embedding signal temporally to alleviate the signal reduction caused by geometric and signal-processing distortions. Experimental results show that the proposed scheme achieves better performance than previously proposed schemes and it has high fidelity. The proposed scheme can be used in unauthorized access prevention method of visual communication or traitor tracking applications which need fast detection process to prevent illegally edited video content from spreading out.Keywords: Editing prevention technique, gradient method, high-quality video, luminance change, visual communication.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1930909 Medical Image Registration by Minimizing Divergence Measure Based on Tsallis Entropy
Authors: Shaoyan Sun, Liwei Zhang, Chonghui Guo
Abstract:
As the use of registration packages spreads, the number of the aligned image pairs in image databases (either by manual or automatic methods) increases dramatically. These image pairs can serve as a set of training data. Correspondingly, the images that are to be registered serve as testing data. In this paper, a novel medical image registration method is proposed which is based on the a priori knowledge of the expected joint intensity distribution estimated from pre-aligned training images. The goal of the registration is to find the optimal transformation such that the distance between the observed joint intensity distribution obtained from the testing image pair and the expected joint intensity distribution obtained from the corresponding training image pair is minimized. The distance is measured using the divergence measure based on Tsallis entropy. Experimental results show that, compared with the widely-used Shannon mutual information as well as Tsallis mutual information, the proposed method is computationally more efficient without sacrificing registration accuracy.
Keywords: Multimodality images, image registration, Shannonentropy, Tsallis entropy, mutual information, Powell optimization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1635908 Health Risk Assessment for Sewer Workers using Bayesian Belief Networks
Authors: Kevin Fong-Rey Liu, Ken Yeh, Cheng-Wu Chen, Han-Hsi Liang
Abstract:
The sanitary sewerage connection rate becomes an important indicator of advanced cities. Following the construction of sanitary sewerages, the maintenance and management systems are required for keeping pipelines and facilities functioning well. These maintenance tasks often require sewer workers to enter the manholes and the pipelines, which are confined spaces short of natural ventilation and full of hazardous substances. Working in sewers could be easily exposed to a risk of adverse health effects. This paper proposes the use of Bayesian belief networks (BBN) as a higher level of noncarcinogenic health risk assessment of sewer workers. On the basis of the epidemiological studies, the actual hospital attendance records and expert experiences, the BBN is capable of capturing the probabilistic relationships between the hazardous substances in sewers and their adverse health effects, and accordingly inferring the morbidity and mortality of the adverse health effects. The provision of the morbidity and mortality rates of the related diseases is more informative and can alleviate the drawbacks of conventional methods.Keywords: Bayesian belief networks, sanitary sewerage, healthrisk assessment, hazard quotient, target organ-specific hazard index.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1706907 Protocol and Method for Preventing Attacks from the Web
Authors: Ryuya Uda
Abstract:
Nowadays, computer worms, viruses and Trojan horse become popular, and they are collectively called malware. Those malware just spoiled computers by deleting or rewriting important files a decade ago. However, recent malware seems to be born to earn money. Some of malware work for collecting personal information so that malicious people can find secret information such as password for online banking, evidence for a scandal or contact address which relates with the target. Moreover, relation between money and malware becomes more complex. Many kinds of malware bear bots to get springboards. Meanwhile, for ordinary internet users, countermeasures against malware come up against a blank wall. Pattern matching becomes too much waste of computer resources, since matching tools have to deal with a lot of patterns derived from subspecies. Virus making tools can automatically bear subspecies of malware. Moreover, metamorphic and polymorphic malware are no longer special. Recently there appears malware checking sites that check contents in place of users' PC. However, there appears a new type of malicious sites that avoids check by malware checking sites. In this paper, existing protocols and methods related with the web are reconsidered in terms of protection from current attacks, and new protocol and method are indicated for the purpose of security of the web.Keywords: Information Security, Malware, Network Security, World Wide Web
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2097906 Manual to Automated Testing: An Effort-Based Approach for Determining the Priority of Software Test Automation
Authors: Peter Sabev, Katalina Grigorova
Abstract:
Test automation allows performing difficult and time consuming manual software testing tasks efficiently, quickly and repeatedly. However, development and maintenance of automated tests is expensive, so it needs a proper prioritization what to automate first. This paper describes a simple yet efficient approach for such prioritization of test cases based on the effort needed for both manual execution and software test automation. The suggested approach is very flexible because it allows working with a variety of assessment methods, and adding or removing new candidates at any time. The theoretical ideas presented in this article have been successfully applied in real world situations in several software companies by the authors and their colleagues including testing of real estate websites, cryptographic and authentication solutions, OSGi-based middleware framework that has been applied in various systems for smart homes, connected cars, production plants, sensors, home appliances, car head units and engine control units (ECU), vending machines, medical devices, industry equipment and other devices that either contain or are connected to an embedded service gateway.Keywords: Automated Testing, Manual Testing, Test Automation, Software testing, Test Prioritization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3385905 A Dynamic Time-Lagged Correlation based Method to Learn Multi-Time Delay Gene Networks
Authors: Ankit Agrawal, Ankush Mittal
Abstract:
A gene network gives the knowledge of the regulatory relationships among the genes. Each gene has its activators and inhibitors that regulate its expression positively and negatively respectively. Genes themselves are believed to act as activators and inhibitors of other genes. They can even activate one set of genes and inhibit another set. Identifying gene networks is one of the most crucial and challenging problems in Bioinformatics. Most work done so far either assumes that there is no time delay in gene regulation or there is a constant time delay. We here propose a Dynamic Time- Lagged Correlation Based Method (DTCBM) to learn the gene networks, which uses time-lagged correlation to find the potential gene interactions, and then uses a post-processing stage to remove false gene interactions to common parents, and finally uses dynamic correlation thresholds for each gene to construct the gene network. DTCBM finds correlation between gene expression signals shifted in time, and therefore takes into consideration the multi time delay relationships among the genes. The implementation of our method is done in MATLAB and experimental results on Saccharomyces cerevisiae gene expression data and comparison with other methods indicate that it has a better performance.Keywords: Activators, correlation, dynamic time-lagged correlation based method, inhibitors, multi-time delay gene network.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1614904 An Improved K-Means Algorithm for Gene Expression Data Clustering
Authors: Billel Kenidra, Mohamed Benmohammed
Abstract:
Data mining technique used in the field of clustering is a subject of active research and assists in biological pattern recognition and extraction of new knowledge from raw data. Clustering means the act of partitioning an unlabeled dataset into groups of similar objects. Each group, called a cluster, consists of objects that are similar between themselves and dissimilar to objects of other groups. Several clustering methods are based on partitional clustering. This category attempts to directly decompose the dataset into a set of disjoint clusters leading to an integer number of clusters that optimizes a given criterion function. The criterion function may emphasize a local or a global structure of the data, and its optimization is an iterative relocation procedure. The K-Means algorithm is one of the most widely used partitional clustering techniques. Since K-Means is extremely sensitive to the initial choice of centers and a poor choice of centers may lead to a local optimum that is quite inferior to the global optimum, we propose a strategy to initiate K-Means centers. The improved K-Means algorithm is compared with the original K-Means, and the results prove how the efficiency has been significantly improved.
Keywords: Microarray data mining, biological pattern recognition, partitional clustering, k-means algorithm, centroid initialization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1284903 TiO2/Clay Minerals (Palygorskite/Halloysite) Nanocomposite Coatings for Water Disinfection
Authors: Dionisios Panagiotaras, Dimitrios Papoulis, Elias Stathatos
Abstract:
Microfibrous palygorskite and tubular halloysite clay mineral combined with nanocrystalline TiO2 are incorporating in the preparation of nanocomposite films on glass substrates via sol-gel route at 450oC. The synthesis is employing nonionic surfactant molecule as pore directing agent along with acetic acid-based sol-gel route without addition of water molecules. Drying and thermal treatment of composite films ensure elimination of organic material lead to the formation of TiO2 nanoparticles homogeneously distributed on the palygorskite or halloysite surfaces. Nanocomposite films without cracks of active anatase crystal phase on palygorskite and halloysite surfaces are characterized by microscopy techniques, UV-Vis spectroscopy, and porosimetry methods in order to examine their structural properties.
The composite palygorskite-TiO2 and halloysite-TiO2 films with variable quantities of palygorskite and halloysite were tested as photocatalysts in the photo-oxidation of Basic Blue 41 azo dye in water. These nanocomposite films proved to be most promising photocatalysts and highly effective to dye’s decoloration in spite of small amount of palygorskite-TiO2 or halloysite-TiO2 catalyst immobilized onto glass substrates mainly due to the high surface area and uniform distribution of TiO2 on clay minerals avoiding aggregation.
Keywords: Halloysite, Palygorskite, Photocatalysis, Titanium Dioxide.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3598902 A New Correlation between SPT and CPT for Various Soils
Authors: Fauzi Jarushi, S. AlKaabim, Paul Cosentino
Abstract:
The Standard Penetration Test (SPT) is the most common in situ test for soil investigations. On the other hand, the Cone Penetration Test (CPT) is considered one of the best investigation tools. Due to the fast and accurate results that can be obtained it complaints the SPT in many applications like field explorations, design parameters, and quality control assessments. Many soil index and engineering properties have been correlated to both of SPT and CPT. Various foundation design methods were developed based on the outcome of these tests. Therefore it is vital to correlate these tests to each other so that either one of the tests can be used in the absence of the other, especially for preliminary evaluation and design purposes. The primary purpose of this study was to investigate the relationships between the SPT and CPT for different type of sandy soils in Florida. Data for this research were collected from number of projects sponsored by the Florida Department of Transportation (FDOT), six sites served as the subject of SPT-CPT correlations. The correlations were established between the cone resistance (qc), sleeve friction (fs) and the uncorrected SPT blow counts (N) for various soils. A positive linear relationship was found between qc, fs and N for various sandy soils. In general, qc versus N showed higher correlation coefficients than fs versus N. qc/N ratios were developed for different soil types and compared to literature values, the results of this research revealed higher ratios than literature values.
Keywords: In situ tests, Correlation, SPT, CPT.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16576901 Money Laundering and Financing of Terrorism
Authors: C. Mallada Fernández
Abstract:
Economic development and globalization of international markets have created a favourable atmosphere for the emergence of new forms of crime such as money laundering or financing of terrorism, which may contribute to destabilized and damage economic systems. In particular, money laundering have acquired great importance since the 11S attacks, what has caused on the one hand, the establishment and development of preventive measures and, on the other hand, a progressive hardening of penal measures. Since then, the regulations imposed to fight against money laundering have been viewed as key components also in the fight against terrorist financing. Terrorism, at the beginning, was a “national” crime connected with internal problems of the State (for instance the RAF in Germany or ETA in Spain) but in the last 20 years has started to be an international problem that is connected with the defence and security of the States. Therefore, the new strategic concept for the defense and security of NATO has a comprehensive list of security threats to the Alliance, such as terrorism, international instability, money laundering or attacks on cyberspace, among others. With this new concept, money laundering and terrorism has become a priority in the national defense.
In this work we will analyze the methods to combat these new threats to the national security. We will study the preventive legislations to combat money laundering and financing of terrorism, the UIF that exchange information between States, and the hawala-Banking.
Keywords: Control of financial flows, money laundering, terrorism, financing of terrorism.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2882900 Content-Based Image Retrieval Using HSV Color Space Features
Authors: Hamed Qazanfari, Hamid Hassanpour, Kazem Qazanfari
Abstract:
In this paper, a method is provided for content-based image retrieval. Content-based image retrieval system searches query an image based on its visual content in an image database to retrieve similar images. In this paper, with the aim of simulating the human visual system sensitivity to image's edges and color features, the concept of color difference histogram (CDH) is used. CDH includes the perceptually color difference between two neighboring pixels with regard to colors and edge orientations. Since the HSV color space is close to the human visual system, the CDH is calculated in this color space. In addition, to improve the color features, the color histogram in HSV color space is also used as a feature. Among the extracted features, efficient features are selected using entropy and correlation criteria. The final features extract the content of images most efficiently. The proposed method has been evaluated on three standard databases Corel 5k, Corel 10k and UKBench. Experimental results show that the accuracy of the proposed image retrieval method is significantly improved compared to the recently developed methods.
Keywords: Content-based image retrieval, color difference histogram, efficient features selection, entropy, correlation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 660899 Knowledge Impact on Measurement: A Conceptual Metric for Evaluating Performance Improvement (PI) at the Kuwait Institute for Scientific Research (KISR)
Authors: AlMatrouk H. S., Juszczak M. D.
Abstract:
Research and development R&D work involves enormous amount of work that has to do with data measurement and collection. This process evolves as new information is fed, new technologies are utilized, and eventually new knowledge is created by the stakeholders i.e., researchers, clients, and end-users. When new knowledge is created, procedures of R&D work should evolve and produce better results within improved research skills and improved methods of data measurements and collection. This measurement improvement should then be benchmarked against a metric that should be developed at the organization. In this paper, we are suggesting a conceptual metric for R&D work performance improvement (PI) at the Kuwait Institute for Scientific Research (KISR). This PI is to be measured against a set of variables in the suggested metric, which are more closely correlated to organizational output, as opposed to organizational norms. The paper also mentions and discusses knowledge creation and management as an addedvalue to R&D work and measurement improvement. The research methodology followed in this work is qualitative in nature, based on a survey that was distributed to researchers and interviews held with senior researchers at KISR. Research and analyses in this paper also include looking at and analyzing KISR-s literature.Keywords: Knowledge Creation, Performance Improvement (PI), Conceptual Metric, Knowledge Management (KM) addedvalue.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1252898 Using Suffix Tree Document Representation in Hierarchical Agglomerative Clustering
Authors: Daniel I. Morariu, Radu G. Cretulescu, Lucian N. Vintan
Abstract:
In text categorization problem the most used method for documents representation is based on words frequency vectors called VSM (Vector Space Model). This representation is based only on words from documents and in this case loses any “word context" information found in the document. In this article we make a comparison between the classical method of document representation and a method called Suffix Tree Document Model (STDM) that is based on representing documents in the Suffix Tree format. For the STDM model we proposed a new approach for documents representation and a new formula for computing the similarity between two documents. Thus we propose to build the suffix tree only for any two documents at a time. This approach is faster, it has lower memory consumption and use entire document representation without using methods for disposing nodes. Also for this method is proposed a formula for computing the similarity between documents, which improves substantially the clustering quality. This representation method was validated using HAC - Hierarchical Agglomerative Clustering. In this context we experiment also the stemming influence in the document preprocessing step and highlight the difference between similarity or dissimilarity measures to find “closer" documents.Keywords: Text Clustering, Suffix tree documentrepresentation, Hierarchical Agglomerative Clustering
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1910897 A Method for Iris Recognition Based on 1D Coiflet Wavelet
Authors: Agus Harjoko, Sri Hartati, Henry Dwiyasa
Abstract:
There have been numerous implementations of security system using biometric, especially for identification and verification cases. An example of pattern used in biometric is the iris pattern in human eye. The iris pattern is considered unique for each person. The use of iris pattern poses problems in encoding the human iris. In this research, an efficient iris recognition method is proposed. In the proposed method the iris segmentation is based on the observation that the pupil has lower intensity than the iris, and the iris has lower intensity than the sclera. By detecting the boundary between the pupil and the iris and the boundary between the iris and the sclera, the iris area can be separated from pupil and sclera. A step is taken to reduce the effect of eyelashes and specular reflection of pupil. Then the four levels Coiflet wavelet transform is applied to the extracted iris image. The modified Hamming distance is employed to measure the similarity between two irises. This research yields the identification success rate of 84.25% for the CASIA version 1.0 database. The method gives an accuracy of 77.78% for the left eyes of MMU 1 database and 86.67% for the right eyes. The time required for the encoding process, from the segmentation until the iris code is generated, is 0.7096 seconds. These results show that the accuracy and speed of the method is better than many other methods.Keywords: Biometric, iris recognition, wavelet transform.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1906896 Effective Defect Prevention Approach in Software Process for Achieving Better Quality Levels
Authors: Suma. V., T. R. Gopalakrishnan Nair
Abstract:
Defect prevention is the most vital but habitually neglected facet of software quality assurance in any project. If functional at all stages of software development, it can condense the time, overheads and wherewithal entailed to engineer a high quality product. The key challenge of an IT industry is to engineer a software product with minimum post deployment defects. This effort is an analysis based on data obtained for five selected projects from leading software companies of varying software production competence. The main aim of this paper is to provide information on various methods and practices supporting defect detection and prevention leading to thriving software generation. The defect prevention technique unearths 99% of defects. Inspection is found to be an essential technique in generating ideal software generation in factories through enhanced methodologies of abetted and unaided inspection schedules. On an average 13 % to 15% of inspection and 25% - 30% of testing out of whole project effort time is required for 99% - 99.75% of defect elimination. A comparison of the end results for the five selected projects between the companies is also brought about throwing light on the possibility of a particular company to position itself with an appropriate complementary ratio of inspection testing.Keywords: Defect Detection and Prevention, Inspections, Software Engineering, Software Process, Testing.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1537895 Clustering Categorical Data Using the K-Means Algorithm and the Attribute’s Relative Frequency
Authors: Semeh Ben Salem, Sami Naouali, Moetez Sallami
Abstract:
Clustering is a well known data mining technique used in pattern recognition and information retrieval. The initial dataset to be clustered can either contain categorical or numeric data. Each type of data has its own specific clustering algorithm. In this context, two algorithms are proposed: the k-means for clustering numeric datasets and the k-modes for categorical datasets. The main encountered problem in data mining applications is clustering categorical dataset so relevant in the datasets. One main issue to achieve the clustering process on categorical values is to transform the categorical attributes into numeric measures and directly apply the k-means algorithm instead the k-modes. In this paper, it is proposed to experiment an approach based on the previous issue by transforming the categorical values into numeric ones using the relative frequency of each modality in the attributes. The proposed approach is compared with a previously method based on transforming the categorical datasets into binary values. The scalability and accuracy of the two methods are experimented. The obtained results show that our proposed method outperforms the binary method in all cases.
Keywords: Clustering, k-means, categorical datasets, pattern recognition, unsupervised learning, knowledge discovery.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3545894 Enhancing the Performance of H.264/AVC in Adaptive Group of Pictures Mode Using Octagon and Square Search Pattern
Authors: S. Sowmyayani, P. Arockia Jansi Rani
Abstract:
This paper integrates Octagon and Square Search pattern (OCTSS) motion estimation algorithm into H.264/AVC (Advanced Video Coding) video codec in Adaptive Group of Pictures (AGOP) mode. AGOP structure is computed based on scene change in the video sequence. Octagon and square search pattern block-based motion estimation method is implemented in inter-prediction process of H.264/AVC. Both these methods reduce bit rate and computational complexity while maintaining the quality of the video sequence respectively. Experiments are conducted for different types of video sequence. The results substantially proved that the bit rate, computation time and PSNR gain achieved by the proposed method is better than the existing H.264/AVC with fixed GOP and AGOP. With a marginal gain in quality of 0.28dB and average gain in bitrate of 132.87kbps, the proposed method reduces the average computation time by 27.31 minutes when compared to the existing state-of-art H.264/AVC video codec.Keywords: Block Distortion Measure, Block Matching Algorithms, H.264/AVC, Motion estimation, Search patterns, Shot cut detection.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1731893 Role of Facade in Sustainability Enhancement of Contemporary Iranian Buildings
Authors: H. Nejadriahi
Abstract:
A growing demand for sustainability makes sustainability as one of the significant debates of nowadays. Energy saving is one of the main criteria to be considered in the context of sustainability. Reducing energy use in buildings is one of the most important ways to reduce humans’ overall environmental impact. Taking this into consideration, study of different design strategies, which can assist in reducing energy use and subsequently improving the sustainability level of today's buildings would be an essential task. The sustainability level of a building is highly affected by the sustainability performance of its components. One of the main building components, which can have a great impact on energy saving and sustainability level of the building, is its facade. The aim of this study is to investigate on the role of facade in sustainability enhancement of the contemporary buildings of Iran. In this study, the concept of sustainability in architecture, the building facades, and their relationship to sustainability are explained briefly. Following that, a number of contemporary Iranian buildings are discussed and analyzed in terms of different design strategies used in their facades in accordance to the sustainability concepts. The methods used in this study are descriptive and analytic. The results of this paper would assist in generating a wider vision and a source of inspiration for the current designers to design and create environmental and sustainable buildings for the future.
Keywords: Building facade, contemporary buildings, Iran, sustainability.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 990892 A Complexity-Based Approach in Image Compression using Neural Networks
Authors: Hadi Veisi, Mansour Jamzad
Abstract:
In this paper we present an adaptive method for image compression that is based on complexity level of the image. The basic compressor/de-compressor structure of this method is a multilayer perceptron artificial neural network. In adaptive approach different Back-Propagation artificial neural networks are used as compressor and de-compressor and this is done by dividing the image into blocks, computing the complexity of each block and then selecting one network for each block according to its complexity value. Three complexity measure methods, called Entropy, Activity and Pattern-based are used to determine the level of complexity in image blocks and their ability in complexity estimation are evaluated and compared. In training and evaluation, each image block is assigned to a network based on its complexity value. Best-SNR is another alternative in selecting compressor network for image blocks in evolution phase which chooses one of the trained networks such that results best SNR in compressing the input image block. In our evaluations, best results are obtained when overlapping the blocks is allowed and choosing the networks in compressor is based on the Best-SNR. In this case, the results demonstrate superiority of this method comparing with previous similar works and JPEG standard coding.Keywords: Adaptive image compression, Image complexity, Multi-layer perceptron neural network, JPEG Standard, PSNR.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2222891 The Optimum Operating Conditions for the Synthesis of Zeolite from Waste Incineration Fly Ash by Alkali Fusion and Hydrothermal Methods
Authors: Yi-Jie Lin, Jyh-Cherng Chen
Abstract:
The fly ash of waste incineration processes is usually hazardous and the disposal or reuse of waste incineration fly ash is difficult. In this study, the waste incineration fly ash was converted to useful zeolites by the alkali fusion and hydrothermal synthesis method. The influence of different operating conditions (the ratio of Si/Al, the ratio of hydrolysis liquid to solid, and hydrothermal time) was investigated to seek the optimum operating conditions for the synthesis of zeolite from waste incineration fly ash. The results showed that concentrations of heavy metals in the leachate of Toxicity Characteristic Leaching Procedure (TCLP) were all lower than the regulatory limits except lead. The optimum operating conditions for the synthesis of zeolite from waste incineration fly ash by the alkali fusion and hydrothermal synthesis method were Si/Al=40, NaOH/ash=1.5, alkali fusion at 400 oC for 40 min, hydrolysis with Liquid to Solid ratio (L/S)= 200 at 105 oC for 24 h, and hydrothermal synthesis at 105 oC for 24 h. The specific surface area of fly ash could be significantly increased from 8.59 m2/g to 651.51 m2/g (synthesized zeolite). The influence of different operating conditions on the synthesis of zeolite from waste incineration fly ash followed the sequence of Si/Al ratio > hydrothermal time > hydrolysis L/S ratio. The synthesized zeolites can be reused as good adsorbents to control the air or wastewater pollutants. The purpose of fly ash detoxification, reduction and waste recycling/reuse is achieved successfully.
Keywords: Alkali fusion, hydrothermal, fly ash, zeolite.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 882890 Gender Based Variability Time Series Complexity Analysis
Authors: Ramesh K. Sunkaria, Puneeta Marwaha
Abstract:
Non linear methods of heart rate variability (HRV) analysis are becoming more popular. It has been observed that complexity measures quantify the regularity and uncertainty of cardiovascular RR-interval time series. In the present work, SampEn has been evaluated in healthy normal sinus rhythm (NSR) male and female subjects for different data lengths and tolerance level r. It is demonstrated that SampEn is small for higher values of tolerance r. Also SampEn value of healthy female group is higher than that of healthy male group for short data length and with increase in data length both groups overlap each other and it is difficult to distinguish them. The SampEn gives inaccurate results by assigning higher value to female group, because male subject have more complex HRV pattern than that of female subjects. Therefore, this traditional algorithm exhibits higher complexity for healthy female subjects than for healthy male subjects, which is misleading observation. This may be due to the fact that SampEn do not account for multiple time scales inherent in the physiologic time series and the hidden spatial and temporal fluctuations remains unexplored.
Keywords: Heart rate variability, normal sinus rhythm group, RR interval time series, sample entropy.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1766889 Family-size Biogas Plant Using Manure and Urine Mixture at Ambient Temperature in Semi-arid Regions of Northwestern China
Authors: Wenguang Ding, Yang Wu, Xia Wang, Yayu Gao
Abstract:
Biogas, a clean renewable energy, is attracting a growing concern of researchers and professionals in many fields. Based on the natural and climatic conditions in semi-arid regions of northwestern China, the present study introduces a specifically-designed family-size biogas plant (with a digester of 10m3) with manure and urine of animals and humanity as raw materials. The biogas plant is applicable to areas with altitudes of more than 2000 meters in northwestern China. In addition to the installation cost, a little operational expenditure, structure, characteristics, benefits of this small-scale biogas plant, this article introduces a wide range of specific popularization methods such as training, financial support, guided tour to the biogas plant, community-based group study and delivery of operational manuals. The feasibility of the biogas plant is explored on the basis of the availability of the raw materials. Simple operations contained in the current work increase the possibility of the wide use of this small-scale biogas plant in similar regions of the world.Keywords: biogas, family-size biogas plant, northwestern China, popularization
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2773888 Entrepreneurial Orientation and Customers Satisfaction: Evidences nearby Khao San Road
Authors: Vichada Chokesikarin
Abstract:
The study aims to determine which factors account for customer satisfaction and to investigate the relationship between entrepreneurial orientation and business success, in particular, context of the information understanding of hostel business in Pranakorn district, Bangkok and the significant element of entrepreneurship in tourism industry. This study covers 352 hostels customers and 61 hostel owners/managers nearby Khao San road. Data collection methods were used by survey questionnaire and a series of hypotheses were developed from services marketing literature. The findings suggest the customer satisfaction most influenced by image, service quality, room quality and price accordingly. Furthermore the findings revealed that significant relationships exist between entrepreneurial orientation and business success; while competitive aggressiveness was found unrelated. The ECSI model’s generic measuring customer satisfaction was found partially mediate the business success. A reconsideration of other variables applicable should be supported with the model of hostel business. The study provides context and overall view of hostel business while discussing from the entrepreneurial orientation to customer satisfaction, thereby reducing decision risk on hostel investment.
Keywords: Customer satisfaction, ECSI Model, entrepreneurial orientation, small hotels, hostel, business performance.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3452887 RUPSec: An Extension on RUP for Developing Secure Systems - Requirements Discipline
Authors: Mohammad Reza Ayatollahzadeh Shirazi, Pooya Jaferian, Golnaz Elahi, Hamid Baghi, Babak Sadeghian
Abstract:
The world is moving rapidly toward the deployment of information and communication systems. Nowadays, computing systems with their fast growth are found everywhere and one of the main challenges for these systems is increasing attacks and security threats against them. Thus, capturing, analyzing and verifying security requirements becomes a very important activity in development process of computing systems, specially in developing systems such as banking, military and e-business systems. For developing every system, a process model which includes a process, methods and tools is chosen. The Rational Unified Process (RUP) is one of the most popular and complete process models which is used by developers in recent years. This process model should be extended to be used in developing secure software systems. In this paper, the Requirement Discipline of RUP is extended to improve RUP for developing secure software systems. These proposed extensions are adding and integrating a number of Activities, Roles, and Artifacts to RUP in order to capture, document and model threats and security requirements of system. These extensions introduce a group of clear and stepwise activities to developers. By following these activities, developers assure that security requirements are captured and modeled. These models are used in design, implementation and test activitie Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2811886 Comparative Canadian Online News Coverage Analysis of Sex Trafficking Reported Cases in Ontario and Nova Scotia
Authors: Alisha Fisher
Abstract:
Sex trafficking is a worldwide crisis that requires trauma-informed and survivor-centered media attention to accurate disseminate information. Much of the previous literature of sex trafficking tends to focus on frequency of incidents, intervention, and support strategies for survivors, with few of them looking to how the media is conducting their reporting on sex trafficking cases to the public. Utilizing data of reports from the media of cases of sex trafficking in the two Canadian provinces with the highest cases of sex trafficking, Ontario and Nova Scotia, we sought to analyze the similarities and differences of how sex trafficking cases were being reported. A total of 20 articles were examined, with 10 based within the province of Ontario and the remaining 10 from the province of Nova Scotia. We coded in two processes, first, who the article was about, and second, the framing and content inclusion. The results suggest that there is high usage, and reliance of voices and images of authority, with male people of color being shown as the perpetrators, and white women being shown as the survivors. These findings can aid in the expansion of trauma-informed, survivor-centered media literacy of reports of sex trafficking to provide accurate insights, and further developing robust methods to intersectional approaches to reporting cases of sex trafficking.
Keywords: Sex Trafficking, media coverage, canada sex trafficking, content analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 772