Search results for: correlation features image fusion
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 3855

Search results for: correlation features image fusion

465 Concept to Enhance the Project Success and Promote the Implementation of Success Factors in Infrastructure Projects

Authors: A. Elbaz, K. Spang

Abstract:

Infrastructure projects are often subjected to delays and cost overruns and mistakenly described as unsuccessful projects. These projects have many peculiarities such as public attention, impact on the environment, subjected to special regulations, etc. They also deal with several stakeholders with different motivations and face unique risks. With this in mind we need to reconsider our approach to manage them, define their success factors and implement these success factors. Infrastructure projects are not only lacking a unified meaning of project success or a definition of success factors, but also a clear method to implement these factors. This paper investigates this gap and introduces a concept to implement success factors in an efficient way, taking into consideration the specific characteristics of infrastructure projects. This concept consists of six enablers such as project organization, project team, project management workflow, contract management, communication and knowledge transfer and project documentations. These enablers allow other success factors to be efficiently implemented in projects. In conclusion, this paper provides project managers as well as company managers with a tool to define and implement success factors efficiently in their projects, along with upgrading their assets for the coming projects. This tool consists of processes and validated checklists to ensure the best use of company resources and knowledge. Due to the special features of infrastructure projects this tool will be tested in the German infrastructure market. However, it is meant to be adaptable to other markets and industries.

Keywords: Infrastructure projects, enablers, project success, success factors, transportation projects.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 940
464 Performance Evaluation of Distributed Deep Learning Frameworks in Cloud Environment

Authors: Shuen-Tai Wang, Fang-An Kuo, Chau-Yi Chou, Yu-Bin Fang

Abstract:

2016 has become the year of the Artificial Intelligence explosion. AI technologies are getting more and more matured that most world well-known tech giants are making large investment to increase the capabilities in AI. Machine learning is the science of getting computers to act without being explicitly programmed, and deep learning is a subset of machine learning that uses deep neural network to train a machine to learn  features directly from data. Deep learning realizes many machine learning applications which expand the field of AI. At the present time, deep learning frameworks have been widely deployed on servers for deep learning applications in both academia and industry. In training deep neural networks, there are many standard processes or algorithms, but the performance of different frameworks might be different. In this paper we evaluate the running performance of two state-of-the-art distributed deep learning frameworks that are running training calculation in parallel over multi GPU and multi nodes in our cloud environment. We evaluate the training performance of the frameworks with ResNet-50 convolutional neural network, and we analyze what factors that result in the performance among both distributed frameworks as well. Through the experimental analysis, we identify the overheads which could be further optimized. The main contribution is that the evaluation results provide further optimization directions in both performance tuning and algorithmic design.

Keywords: Artificial Intelligence, machine learning, deep learning, convolutional neural networks.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1218
463 A Quantitative Approach to Strategic Design of Component-Based Business Process Models

Authors: Eakong Atiptamvaree, Twittie Senivongse

Abstract:

A new paradigm for software design and development models software by its business process, translates the model into a process execution language, and has it run by a supporting execution engine. This process-oriented paradigm promotes modeling of software by less technical users or business analysts as well as rapid development. Since business process models may be shared by different organizations and sometimes even by different business domains, it is interesting to apply a technique used in traditional software component technology to design reusable business processes. This paper discusses an approach to apply a technique for software component fabrication to the design of process-oriented software units, called process components. These process components result from decomposing a business process of a particular application domain into subprocesses with an aim that the process components can be reusable in different process-based software models. The approach is quantitative because the quality of process component design is measured from technical features of the process components. The approach is also strategic because the measured quality is determined against business-oriented component management goals. A software tool has been developed to measure how good a process component design is, according to the required managerial goals and comparing to other designs. We also discuss how we benefit from reusable process components.

Keywords: Business process model, process component, component management goals, measurement

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1646
462 Using Field Indices of Rill and Gully in order to Erosion Estimating and Sediment Analysis (Case Study: Menderjan Watershed in Isfahan Province, Iran)

Authors: Masoud Nasri, Sadat Feiznia, Mohammad Jafari, Hasan Ahmadi

Abstract:

Today, incorrect use of lands and land use changes, excessive grazing, no suitable using of agricultural farms, plowing on steep slopes, road construct, building construct, mine excavation etc have been caused increasing of soil erosion and sediment yield. For erosion and sediment estimation one can use statistical and empirical methods. This needs to identify land unit map and the map of effective factors. However, these empirical methods are usually time consuming and do not give accurate estimation of erosion. In this study, we applied GIS techniques to estimate erosion and sediment of Menderjan watershed at upstream Zayandehrud river in center of Iran. Erosion faces at each land unit were defined on the basis of land use, geology and land unit map using GIS. The UTM coordinates of each erosion type that showed more erosion amounts such as rills and gullies were inserted in GIS using GPS data. The frequency of erosion indicators at each land unit, land use and their sediment yield of these indices were calculated. Also using tendency analysis of sediment yield changes in watershed outlet (Menderjan hydrometric gauge station), was calculated related parameters and estimation errors. The results of this study according to implemented watershed management projects can be used for more rapid and more accurate estimation of erosion than traditional methods. These results can also be used for regional erosion assessment and can be used for remote sensing image processing.

Keywords: Erosion and sedimentation, Gully, Rill, GIS, GPS, Menderjan Watershed

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1886
461 Unsupervised Feature Learning by Pre-Route Simulation of Auto-Encoder Behavior Model

Authors: Youngjae Jin, Daeshik Kim

Abstract:

This paper describes a cycle accurate simulation results of weight values learned by an auto-encoder behavior model in terms of pre-route simulation. Given the results we visualized the first layer representations with natural images. Many common deep learning threads have focused on learning high-level abstraction of unlabeled raw data by unsupervised feature learning. However, in the process of handling such a huge amount of data, the learning method’s computation complexity and time limited advanced research. These limitations came from the fact these algorithms were computed by using only single core CPUs. For this reason, parallel-based hardware, FPGAs, was seen as a possible solution to overcome these limitations. We adopted and simulated the ready-made auto-encoder to design a behavior model in VerilogHDL before designing hardware. With the auto-encoder behavior model pre-route simulation, we obtained the cycle accurate results of the parameter of each hidden layer by using MODELSIM. The cycle accurate results are very important factor in designing a parallel-based digital hardware. Finally this paper shows an appropriate operation of behavior model based pre-route simulation. Moreover, we visualized learning latent representations of the first hidden layer with Kyoto natural image dataset.

Keywords: Auto-encoder, Behavior model simulation, Digital hardware design, Pre-route simulation, Unsupervised feature learning.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2670
460 Assessing Innovation Activity in Mexico and South Korea: An Econometric Approach

Authors: Mario Gómez, Won Ho Kim, Ángel Licona, José Carlos Rodríguez

Abstract:

This article analyzes innovation activity in Mexico and South Korea. It develops an econometric model to test for structural breaks in the number of patent applications filed by residents and nonresidents in these countries during the period of 1965 to 2012. These changes may suggest that firms’ innovative capabilities have changed because of implementing different science, technology and innovation (STI) policies in Mexico and South Korea. Two important features characterize this research from others already developed by these authors. First, the theoretical research framework in this research is the debate between the assimilation view of growth and the accumulation view of growth. This characteristic suggests that trade liberalization should be accompanied by an adequate STI policy to boost competitiveness among indigenous firms. Second, the analysis in this research stresses the importance of key actors (e.g. governments) to successfully develop innovation capabilities among indigenous firms. Therefore, the question conducting this research is how STI policies in Mexico and South Korea contributed to develop firms’ innovation capabilities in these countries during last decades? The results from this research suggests that STI policy in South Korea was more suitable to boost innovation firms to compete in markets. Data to develop this research was released by the World Intellectual Property Organization (WIPO).

Keywords: Econometric methods, innovation, Mexico, South Korea, STI Policy.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 932
459 Ensemble Approach for Predicting Student's Academic Performance

Authors: L. A. Muhammad, M. S. Argungu

Abstract:

Educational data mining (EDM) has recorded substantial considerations. Techniques of data mining in one way or the other have been proposed to dig out out-of-sight knowledge in educational data. The result of the study got assists academic institutions in further enhancing their process of learning and methods of passing knowledge to students. Consequently, the performance of students boasts and the educational products are by no doubt enhanced. This study adopted a student performance prediction model premised on techniques of data mining with Students' Essential Features (SEF). SEF are linked to the learner's interactivity with the e-learning management system. The performance of the student's predictive model is assessed by a set of classifiers, viz. Bayes Network, Logistic Regression, and Reduce Error Pruning Tree (REP). Consequently, ensemble methods of Bagging, Boosting, and Random Forest (RF) are applied to improve the performance of these single classifiers. The study reveals that the result shows a robust affinity between learners' behaviors and their academic attainment. Result from the study shows that the REP Tree and its ensemble record the highest accuracy of 83.33% using SEF. Hence, in terms of the Receiver Operating Curve (ROC), boosting method of REP Tree records 0.903, which is the best. This result further demonstrates the dependability of the proposed model.

Keywords: Ensemble, bagging, Random Forest, boosting, data mining, classifiers, machine learning.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 702
458 Robot Navigation and Localization Based on the Rat’s Brain Signals

Authors: Endri Rama, Genci Capi, Shigenori Kawahara

Abstract:

The mobile robot ability to navigate autonomously in its environment is very important. Even though the advances in technology, robot self-localization and goal directed navigation in complex environments are still challenging tasks. In this article, we propose a novel method for robot navigation based on rat’s brain signals (Local Field Potentials). It has been well known that rats accurately and rapidly navigate in a complex space by localizing themselves in reference to the surrounding environmental cues. As the first step to incorporate the rat’s navigation strategy into the robot control, we analyzed the rats’ strategies while it navigates in a multiple Y-maze, and recorded Local Field Potentials (LFPs) simultaneously from three brain regions. Next, we processed the LFPs, and the extracted features were used as an input in the artificial neural network to predict the rat’s next location, especially in the decision-making moment, in Y-junctions. We developed an algorithm by which the robot learned to imitate the rat’s decision-making by mapping the rat’s brain signals into its own actions. Finally, the robot learned to integrate the internal states as well as external sensors in order to localize and navigate in the complex environment.

Keywords: Brain machine interface, decision-making, local field potentials, mobile robot, navigation, neural network, rat, signal processing.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1450
457 A Study on Human Musculoskeletal Model for Cycle Fitting: Comparison with EMG

Authors: Yoon- Ho Shin, Jin-Seung Choi, Dong-Won Kang, Jeong-Woo Seo, Joo-Hack Lee, Ju-Young Kim, Dae-Hyeok Kim, Seung-Tae Yang, Gye-Rae Tack

Abstract:

It is difficult to study the effect of various variables on cycle fitting through actual experiment. To overcome such difficulty, the forward dynamics of a musculoskeletal model was applied to cycle fitting in this study. The measured EMG data weres compared with the muscle activities of the musculoskeletal model through forward dynamics. EMG data were measured from five cyclists who do not have musculoskeletal diseases during three minutes pedaling with a constant load (150 W) and cadence (90 RPM). The muscles used for the analysis were the Vastus Lateralis (VL), Tibialis Anterior (TA), Bicep Femoris (BF), and Gastrocnemius Medial (GM). Person’s correlation coefficients of the muscle activity patterns, the peak timing of the maximum muscle activities, and the total muscle activities were calculated and compared. BIKE3D model of AnyBody (Anybodytech, Denmark) was used for the musculoskeletal model simulation. The comparisons of the actual experiments with the simulation results showed significant correlations in the muscle activity patterns (VL: 0.789, TA: 0.503, BF: 0.468, GM: 0.670). The peak timings of the maximum muscle activities were distributed at particular phases. The total muscle activities were compared with the normalized muscle activities, and the comparison showed about 10% difference in the VL (+10%), TA (+9.7%), and BF (+10%), excluding the GM (+29.4%). Thus, it can be concluded that muscle activities of model & experiment showed similar results. The results of this study indicated that it was possible to apply the simulation of further improved musculoskeletal model to cycle fitting.

Keywords: Cycle fitting, EMG, Musculoskeletal modeling, Simulation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3150
456 Oral Examination: An Important Adjunct to the Diagnosis of Dermatological Disorders

Authors: Sanjay Saraf

Abstract:

The oral cavity can be the site for early manifestations of mucocutaneous disorders (MD) or the only site for occurrence of these disorders. It can also exhibit oral lesions with simultaneous associated skin lesions. The MD involving the oral mucosa commonly presents with signs such as ulcers, vesicles and bullae. The unique environment of the oral cavity may modify these signs of the disease, thereby making the clinical diagnosis an arduous task. In addition to the unique environment of oral cavity, the overlapping of the signs of various mucocutaneous disorders, also makes the clinical diagnosis more intricate. The aim of this review is to present the oral signs of dermatological disorders having common oral involvement and emphasize their   importance in   early detection of the systemic disorders. The aim is also to highlight the necessity of oral examination by a dermatologist while examining the skin lesions. Prior to the oral examination, it must be imperative for the dermatologists and the dental clinicians to have the knowledge of oral anatomy. It is also important to know the impact of various diseases on oral mucosa, and the characteristic features of various oral mucocutaneous lesions. An initial clinical oral examination is may help in the early diagnosis of the MD. Failure to identify the oral manifestations may reduce the likelihood of early treatment and lead to more serious problems. This paper reviews the oral manifestations of immune mediated dermatological disorders with common oral manifestations.

Keywords: Vesiculobullous lesions, Desquamative gingivitis, Nikolsky’s sign, Erythema.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1592
455 Non-Methane Hydrocarbons Emission during the Photocopying Process

Authors: Kiurski S. Jelena, Aksentijević M. Snežana, Kecić S. Vesna, Oros B. Ivana

Abstract:

Prosperity of electronic equipment in photocopying environment not only has improved work efficiency, but also has changed indoor air quality. Considering the number of photocopying employed, indoor air quality might be worse than in general office environments. Determining the contribution from any type of equipment to indoor air pollution is a complex matter. Non-methane hydrocarbons are known to have an important role on air quality due to their high reactivity. The presence of hazardous pollutants in indoor air has been detected in one photocopying shop in Novi Sad, Serbia. Air samples were collected and analyzed for five days, during 8-hr working time in three time intervals, whereas three different sampling points were determined. Using multiple linear regression model and software package STATISTICA 10 the concentrations of occupational hazards and microclimates parameters were mutually correlated. Based on the obtained multiple coefficients of determination (0.3751, 0.2389 and 0.1975), a weak positive correlation between the observed variables was determined. Small values of parameter F indicated that there was no statistically significant difference between the concentration levels of nonmethane hydrocarbons and microclimates parameters. The results showed that variable could be presented by the general regression model: y = b0 + b1xi1+ b2xi2. Obtained regression equations allow to measure the quantitative agreement between the variables and thus obtain more accurate knowledge of their mutual relations.

Keywords: Indoor air quality, multiple regression analysis, nonmethane hydrocarbons, photocopying process.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1948
454 Innovative Activity and Development: Analyzing Firm Data from Eurozone Country-Members

Authors: Ilias A. Makris

Abstract:

In this work, we attempt to associate firm characteristics with innovative activity. We collect microdata from listed firms of selected Eurozone Country-members, after the beginning of 2007 financial crisis. The following literature, several indicators of growth and performance were selected and tested for their ability to interpret innovative activity. The main scope is to examine the possible differences in performance and growth between innovative and non-innovative firms, during a severe recession. Additionally to that, a special focus will be held on whether macroeconomic performance and national innovation system, determines the extent of innovators' performance. Preliminary findings, through correlation matrices and non-parametric tests, strongly indicate the positive relation between innovative activity and most of the measures used (profitability, size, employment), confirming that even during a recessionary period, innovative firms not only survive but also seem to succeed better economic results in almost all indexes relative to non-innovative. However, even though innovators seem to perform better in all economies examined, the extent of that performance seems to be strongly affected by the supportive mechanisms (financial and structural) that their country provides. Thus, it is clear, that the technologically intensive 'gap' between European South and North, during the economic crisis, became chaotic, due to the harsh austerity measures and reduced budgets in those countries, even in sectors with high potentials in economic activity and employment, impairing the effects of crisis and enhancing the vicious circle of recession.

Keywords: Eurozone, innovative activity, development, firm performance, non-parametric tests.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1429
453 Bone Mineral Density and Quality, Body Composition of Women in the Postmenopausal Period

Authors: Vladyslav Povoroznyuk, Oksana Ivanyk, Nataliia Dzerovych

Abstract:

In the diagnostics of osteoporosis, the gold standard is considered to be bone mineral density; however, X-ray densitometry is not an accurate indicator of osteoporotic fracture risk under all circumstances. In this regard, the search for new methods that could determine the indicators not only of the mineral density, but of the bone tissue quality, is a logical step for diagnostic optimization. One of these methods is the evaluation of trabecular bone quality. The aim of this study was to examine the quality and mineral density of spine bone tissue, femoral neck, and body composition of women depending on the duration of the postmenopausal period, to determine the correlation of body fat with indicators of bone mineral density and quality. The study examined 179 women in premenopausal and postmenopausal periods. The patients were divided into the following groups: Women in the premenopausal period and women in the postmenopausal period at various stages (early, middle, late postmenopause). A general examination and study of the above parameters were conducted with General Electric X-ray densitometer. The results show that bone quality and mineral density probably deteriorate with advancing of postmenopausal period. Total fat and lean mass ratio is not likely to change with age. In the middle and late postmenopausal periods, the bone tissue mineral density of the spine and femoral neck increases along with total fat mass.

Keywords: Osteoporosis, bone tissue mineral density, bone quality, fat mass, lean mass, postmenopausal osteoporosis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 911
452 Numerical Simulation of Wall Treatment Effects on the Micro-Scale Combustion

Authors: R. Kamali, A. R. Binesh, S. Hossainpour

Abstract:

To understand working features of a micro combustor, a computer code has been developed to study combustion of hydrogen–air mixture in a series of chambers with same shape aspect ratio but various dimensions from millimeter to micrometer level. The prepared algorithm and the computer code are capable of modeling mixture effects in different fluid flows including chemical reactions, viscous and mass diffusion effects. The effect of various heat transfer conditions at chamber wall, e.g. adiabatic wall, with heat loss and heat conduction within the wall, on the combustion is analyzed. These thermal conditions have strong effects on the combustion especially when the chamber dimension goes smaller and the ratio of surface area to volume becomes larger. Both factors, such as larger heat loss through the chamber wall and smaller chamber dimension size, may lead to the thermal quenching of micro-scale combustion. Through such systematic numerical analysis, a proper operation space for the micro-combustor is suggested, which may be used as the guideline for microcombustor design. In addition, the results reported in this paper illustrate that the numerical simulation can be one of the most powerful and beneficial tools for the micro-combustor design, optimization and performance analysis.

Keywords: Numerical simulation, Micro-combustion, MEMS, CFD, Chemical reaction.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1779
451 Urban Growth Analysis Using Multi-Temporal Satellite Images, Non-stationary Decomposition Methods and Stochastic Modeling

Authors: Ali Ben Abbes, ImedRiadh Farah, Vincent Barra

Abstract:

Remotely sensed data are a significant source for monitoring and updating databases for land use/cover. Nowadays, changes detection of urban area has been a subject of intensive researches. Timely and accurate data on spatio-temporal changes of urban areas are therefore required. The data extracted from multi-temporal satellite images are usually non-stationary. In fact, the changes evolve in time and space. This paper is an attempt to propose a methodology for changes detection in urban area by combining a non-stationary decomposition method and stochastic modeling. We consider as input of our methodology a sequence of satellite images I1, I2, … In at different periods (t = 1, 2, ..., n). Firstly, a preprocessing of multi-temporal satellite images is applied. (e.g. radiometric, atmospheric and geometric). The systematic study of global urban expansion in our methodology can be approached in two ways: The first considers the urban area as one same object as opposed to non-urban areas (e.g. vegetation, bare soil and water). The objective is to extract the urban mask. The second one aims to obtain a more knowledge of urban area, distinguishing different types of tissue within the urban area. In order to validate our approach, we used a database of Tres Cantos-Madrid in Spain, which is derived from Landsat for a period (from January 2004 to July 2013) by collecting two frames per year at a spatial resolution of 25 meters. The obtained results show the effectiveness of our method.

Keywords: Multi-temporal satellite image, urban growth, Non-stationarity, stochastic modeling.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1485
450 3D Star Skeleton for Fast Human Posture Representation

Authors: Sungkuk Chun, Kwangjin Hong, Keechul Jung

Abstract:

In this paper, we propose an improved 3D star skeleton technique, which is a suitable skeletonization for human posture representation and reflects the 3D information of human posture. Moreover, the proposed technique is simple and then can be performed in real-time. The existing skeleton construction techniques, such as distance transformation, Voronoi diagram, and thinning, focus on the precision of skeleton information. Therefore, those techniques are not applicable to real-time posture recognition since they are computationally expensive and highly susceptible to noise of boundary. Although a 2D star skeleton was proposed to complement these problems, it also has some limitations to describe the 3D information of the posture. To represent human posture effectively, the constructed skeleton should consider the 3D information of posture. The proposed 3D star skeleton contains 3D data of human, and focuses on human action and posture recognition. Our 3D star skeleton uses the 8 projection maps which have 2D silhouette information and depth data of human surface. And the extremal points can be extracted as the features of 3D star skeleton, without searching whole boundary of object. Therefore, on execution time, our 3D star skeleton is faster than the “greedy" 3D star skeleton using the whole boundary points on the surface. Moreover, our method can offer more accurate skeleton of posture than the existing star skeleton since the 3D data for the object is concerned. Additionally, we make a codebook, a collection of representative 3D star skeletons about 7 postures, to recognize what posture of constructed skeleton is.

Keywords: computer vision, gesture recognition, skeletonization, human posture representation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2091
449 Procedure for Impact Testing of Fused Recycled Glass

Authors: David Halley, Tyra Oseng-Rees, Luca Pagano, Juan A Ferriz-Papi

Abstract:

Recycled glass material is made from 100% recycled bottle glass and consumes less energy than re-melt technology. It also uses no additives in the manufacturing process allowing the recycled glass material, in principal, to go back to the recycling stream after end-of-use, contributing to the circular economy with a low ecological impact. The aim of this paper is to investigate the procedure for testing the recycled glass material for impact resistance, so it can be applied to pavements and other surfaces which are at risk of impact during service. A review of different impact test procedures for construction materials was undertaken, comparing methodologies and international standards applied to other materials such as natural stone, ceramics and glass. A drop weight impact testing machine was designed and manufactured in-house to perform these tests. As a case study, samples of the recycled glass material were manufactured with two different thicknesses and tested. The impact energy was calculated theoretically, obtaining results with 5 and 10 J. The results on the material were subsequently discussed. Improvements on the procedure can be made using high speed video technology to calculate velocity just before and immediately after the impact to know the absorbed energy. The initial results obtained in this procedure were positive although repeatability needs to be developed to obtain a correlation of results and finally be able to validate the procedure. The experiment with samples showed the practicality of this procedure and application to the recycled glass material impact testing although further research needs to be developed.

Keywords: Construction materials, drop weight impact, impact testing, recycled glass.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1503
448 A Software Framework for Predicting Oil-Palm Yield from Climate Data

Authors: Mohd. Noor Md. Sap, A. Majid Awan

Abstract:

Intelligent systems based on machine learning techniques, such as classification, clustering, are gaining wide spread popularity in real world applications. This paper presents work on developing a software system for predicting crop yield, for example oil-palm yield, from climate and plantation data. At the core of our system is a method for unsupervised partitioning of data for finding spatio-temporal patterns in climate data using kernel methods which offer strength to deal with complex data. This work gets inspiration from the notion that a non-linear data transformation into some high dimensional feature space increases the possibility of linear separability of the patterns in the transformed space. Therefore, it simplifies exploration of the associated structure in the data. Kernel methods implicitly perform a non-linear mapping of the input data into a high dimensional feature space by replacing the inner products with an appropriate positive definite function. In this paper we present a robust weighted kernel k-means algorithm incorporating spatial constraints for clustering the data. The proposed algorithm can effectively handle noise, outliers and auto-correlation in the spatial data, for effective and efficient data analysis by exploring patterns and structures in the data, and thus can be used for predicting oil-palm yield by analyzing various factors affecting the yield.

Keywords: Pattern analysis, clustering, kernel methods, spatial data, crop yield

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1951
447 Methane Production from Biomedical Waste (Blood)

Authors: Fatima M. Kabbashi, Abdalla M. Abdalla, Hussam K. Hamad, Elias S. Hassan

Abstract:

This study investigates the production of renewable energy (biogas) from biomedical hazard waste (blood) and eco-friendly disposal. Biogas is produced by the bacterial anaerobic digestion of biomaterial (blood). During digestion process bacterial feeding result in breaking down chemical bonds of the biomaterial and changing its features, by the end of the digestion (biogas production) the remains become manure as known. That has led to the economic and eco-friendly disposal of hazard biomedical waste (blood). The samples (Whole blood, Red blood cells 'RBCs', Blood platelet and Fresh Frozen Plasma ‘FFP’) are collected and measured in terms of carbon to nitrogen C/N ratio and total solid, then filled in connected flasks (three flasks) using water displacement method. The results of trails showed that the platelet and FFP failed to produce flammable gas, but via a gas analyzer, it showed the presence of the following gases: CO, HC, CO₂, and NOX. Otherwise, the blood and RBCs produced flammable gases: Methane-nitrous CH₃NO (99.45%), which has a blue color flame and carbon dioxide CO₂ (0.55%), which has red/yellow color flame. Methane-nitrous is sometimes used as fuel for rockets, some aircraft and racing cars.

Keywords: Renewable energy, biogas, biomedical waste, blood, anaerobic digestion, eco-friendly disposal.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1183
446 Artificial Intelligence-Based Chest X-Ray Test of COVID-19 Patients

Authors: Dhurgham Al-Karawi, Nisreen Polus, Shakir Al-Zaidi, Sabah Jassim

Abstract:

The management of COVID-19 patients based on chest imaging is emerging as an essential tool for evaluating the spread of the pandemic which has gripped the global community. It has already been used to monitor the situation of COVID-19 patients who have issues in respiratory status. There has been increase to use chest imaging for medical triage of patients who are showing moderate-severe clinical COVID-19 features, this is due to the fast dispersal of the pandemic to all continents and communities. This article demonstrates the development of machine learning techniques for the test of COVID-19 patients using Chest X-Ray (CXR) images in nearly real-time, to distinguish the COVID-19 infection with a significantly high level of accuracy. The testing performance has covered a combination of different datasets of CXR images of positive COVID-19 patients, patients with viral and bacterial infections, also, people with a clear chest. The proposed AI scheme successfully distinguishes CXR scans of COVID-19 infected patients from CXR scans of viral and bacterial based pneumonia as well as normal cases with an average accuracy of 94.43%, sensitivity 95%, and specificity 93.86%. Predicted decisions would be supported by visual evidence to help clinicians speed up the initial assessment process of new suspected cases, especially in a resource-constrained environment.

Keywords: COVID-19, chest x-ray scan, artificial intelligence, texture analysis, local binary pattern transform, Gabor filter.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 641
445 Damage to Strawberries Caused by Simulated Transport

Authors: G. La Scalia, M. Enea, R. Micale, O. Corona, L. Settanni

Abstract:

The quality and condition of perishable products delivered to the market and their subsequent selling prices are directly affected by the care taken during harvesting and handling. Mechanical injury, in fact, occurs at all stages, from pre-harvest operations through post-harvest handling, packing and transport to the market. The main implications of this damage are the reduction of the product’s quality and economical losses related to the shelf life diminution. For most perishable products, the shelf life is relatively short and it is typically dictated by microbial growth related to the application of dynamic and static loads during transportation. This paper presents the correlation between vibration levels and microbiological growth on strawberries and woodland strawberries and detects the presence of volatile organic compounds (VOC) in order to develop an intelligent logistic unit capable of monitoring VOCs using a specific sensor system. Fresh fruits were exposed to vibrations by means of a vibrating table in a temperature-controlled environment. Microbiological analyses were conducted on samples, taken at different positions along the column of the crates. The values obtained were compared with control samples not exposed to vibrations and the results show that different positions along the column influence the development of bacteria, yeasts and filamentous fungi.

Keywords: Microbiological analysis, shelf life, transport damage, volatile organic compounds.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3099
444 Implementation of an Improved Secure System Detection for E-passport by using EPC RFID Tags

Authors: A. Baith Mohamed, Ayman Abdel-Hamid, Kareem Youssri Mohamed

Abstract:

Current proposals for E-passport or ID-Card is similar to a regular passport with the addition of tiny contactless integrated circuit (computer chip) inserted in the back cover, which will act as a secure storage device of the same data visually displayed on the photo page of the passport. In addition, it will include a digital photograph that will enable biometric comparison, through the use of facial recognition technology at international borders. Moreover, the e-passport will have a new interface, incorporating additional antifraud and security features. However, its problems are reliability, security and privacy. Privacy is a serious issue since there is no encryption between the readers and the E-passport. However, security issues such as authentication, data protection and control techniques cannot be embedded in one process. In this paper, design and prototype implementation of an improved E-passport reader is presented. The passport holder is authenticated online by using GSM network. The GSM network is the main interface between identification center and the e-passport reader. The communication data is protected between server and e-passport reader by using AES to encrypt data for protection will transferring through GSM network. Performance measurements indicate a 19% improvement in encryption cycles versus previously reported results.

Keywords: RFID "Radio Frequency Identification", EPC"Electronic Product Code", ICAO "International Civil Aviation Organization", IFF "Identify Friend or Foe"

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2581
443 Urban Renewal from the Perspective of Industrial Heritage Protection: Taking the Qiaokou District of Wuhan as an Example

Authors: Yue Sun, Yuan Wang

Abstract:

Most of the earliest national industries in Wuhan are located along the Hanjiang River, and Qiaokou is considered to be a gathering place for Dahankou old industrial base. Zongguan Waterworks, Pacific Soap Factory, Fuxin Flour Factory, Nanyang Tobacco Factory and other hundred-year-old factories are located along Hanjiang River in Qiaokou District, especially the Gutian Industrial Zone, which was listed as one of 156 national restoration projects at the beginning of the founding of the People’s Republic of China. After decades of development, Qiaokou has become the gathering place of the chemical industry and secondary industry, causing damage to the city and serious pollution, becoming a marginalized area forgotten by the central city. In recent years, with the accelerated pace of urban renewal, Qiaokou has been constantly reforming and innovating, and has begun drastic changes in the transformation of old cities and the development of new districts. These factories have been listed as key reconstruction projects, and a large number of industrial heritage with historical value and full urban memory have been relocated, demolished and reformed, with only a few factory buildings preserved. Through the methods of industrial archaeology, image analysis, typology and field investigation, this paper analyzes and summarizes the spatial characteristics of industrial heritage in Qiaokou District, explores urban renewal from the perspective of industrial heritage protection, and provides design strategies for the regeneration of urban industrial sites and industrial heritage.

Keywords: Industrial heritage, urban renewal, protection, urban memory.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 961
442 Multi-Layer Multi-Feature Background Subtraction Using Codebook Model Framework

Authors: Yun-Tao Zhang, Jong-Yeop Bae, Whoi-Yul Kim

Abstract:

Background modeling and subtraction in video analysis has been widely used as an effective method for moving objects detection in many computer vision applications. Recently, a large number of approaches have been developed to tackle different types of challenges in this field. However, the dynamic background and illumination variations are the most frequently occurred problems in the practical situation. This paper presents a favorable two-layer model based on codebook algorithm incorporated with local binary pattern (LBP) texture measure, targeted for handling dynamic background and illumination variation problems. More specifically, the first layer is designed by block-based codebook combining with LBP histogram and mean value of each RGB color channel. Because of the invariance of the LBP features with respect to monotonic gray-scale changes, this layer can produce block wise detection results with considerable tolerance of illumination variations. The pixel-based codebook is employed to reinforce the precision from the output of the first layer which is to eliminate false positives further. As a result, the proposed approach can greatly promote the accuracy under the circumstances of dynamic background and illumination changes. Experimental results on several popular background subtraction datasets demonstrate very competitive performance compared to previous models.

Keywords: Background subtraction, codebook model, local binary pattern, dynamic background, illumination changes.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1943
441 Empirical Roughness Progression Models of Heavy Duty Rural Pavements

Authors: Nahla H. Alaswadko, Rayya A. Hassan, Bayar N. Mohammed

Abstract:

Empirical deterministic models have been developed to predict roughness progression of heavy duty spray sealed pavements for a dataset representing rural arterial roads. The dataset provides a good representation of the relevant network and covers a wide range of operating and environmental conditions. A sample with a large size of historical time series data for many pavement sections has been collected and prepared for use in multilevel regression analysis. The modelling parameters include road roughness as performance parameter and traffic loading, time, initial pavement strength, reactivity level of subgrade soil, climate condition, and condition of drainage system as predictor parameters. The purpose of this paper is to report the approaches adopted for models development and validation. The study presents multilevel models that can account for the correlation among time series data of the same section and to capture the effect of unobserved variables. Study results show that the models fit the data very well. The contribution and significance of relevant influencing factors in predicting roughness progression are presented and explained. The paper concludes that the analysis approach used for developing the models confirmed their accuracy and reliability by well-fitting to the validation data.

Keywords: Roughness progression, empirical model, pavement performance, heavy duty pavement.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 776
440 Conversion in Chemical Reactors using Hollow Cylindrical Catalyst Pellet

Authors: Mohammad Asif

Abstract:

Heterogeneous catalysis is vital for a number of chemical, refinery and pollution control processes. The use of catalyst pellets of hollow cylindrical shape provide several distinct advantages over other common shapes, and can therefore help to enhance conversion levels in reactors. A better utilization of the catalytic material is probably most notable of these features due to the absence of the pellet core, which helps to significantly lower the effect of the internal transport resistance. This is reflected in the enhancement of the effectiveness factor. For the case of the first order irreversible kinetics, a substantial increase in the effectiveness factor can be obtained by varying shape parameters. Important shape parameters of a finite hollow cylinder are the ratio of the inside to the outside radii (κ) and the height to the diameter ratio (γ). A high value of κ the generally helps to enhance the effectiveness factor. On the other hand, lower values of the effectiveness factors are obtained when the dimension of the height and the diameter are comparable. Thus, the departure of parameter γ from the unity favors higher effectiveness factor. Since a higher effectiveness factor is a measure of a greater utilization of the catalytic material, higher conversion levels can be achieved using the hollow cylindrical pellets possessing optimized shape parameters.

Keywords: Finite hollow cylinder, Catalyst pellet, Effectiveness factor, Thiele Modulus, Conversion

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3687
439 Analysis of Precipitation Time Series of Urban Centers of Northeastern Brazil using Wavelet Transform

Authors: Celso A. G. Santos, Paula K. M. M. Freire

Abstract:

The urban centers within northeastern Brazil are mainly influenced by the intense rainfalls, which can occur after long periods of drought, when flood events can be observed during such events. Thus, this paper aims to study the rainfall frequencies in such region through the wavelet transform. An application of wavelet analysis is done with long time series of the total monthly rainfall amount at the capital cities of northeastern Brazil. The main frequency components in the time series are studied by the global wavelet spectrum and the modulation in separated periodicity bands were done in order to extract additional information, e.g., the 8 and 16 months band was examined by an average of all scales, giving a measure of the average annual variance versus time, where the periods with low or high variance could be identified. The important increases were identified in the average variance for some periods, e.g. 1947 to 1952 at Teresina city, which can be considered as high wet periods. Although, the precipitation in those sites showed similar global wavelet spectra, the wavelet spectra revealed particular features. This study can be considered an important tool for time series analysis, which can help the studies concerning flood control, mainly when they are applied together with rainfall-runoff simulations.

Keywords: rainfall data, urban center, wavelet transform.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2427
438 Authenticity Issues of Social Media: Credibility, Quality and Reality

Authors: Shahrinaz Ismail, Roslina Abdul Latif

Abstract:

Social media has led to paradigm shifts in ways people work and do business, interact and socialize, learn and obtain knowledge. So much so that social media has established itself as an important spatial extension of this nation-s historicity and challenges. Regardless of the enabling reputation and recommendation features through social networks embedded in the social media system, the overflow of broadcasted and publicized media contents turns the table around from engendering trust to doubting the trust system. When the trust is at doubt, the effects include deactivation of accounts and creation of multiple profiles, which lead to the overflow of 'ghost' contents (i.e. “the abundance of abandoned ships"). In most literature, the study of trust can be related to culture; hence the difference between Western-s “openness" and Eastern-s “blue-chip" concepts in networking and relationships. From a survey on issues and challenges among Malaysian social media users, 'authenticity' emerges as one of the main factors that causes and is caused by other factors. The other issue that has surfaced is credibility either in terms of message/content and source. Another is the quality of the knowledge that is shared. This paper explores the terrains of this critical space which in recent years has been dominated increasingly by, arguably, social networks embedded in the social media system, the overflow of broadcasted and publicized media content.

Keywords: Authenticity, credibility, knowledge quality and social media.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4511
437 Weighted-Distance Sliding Windows and Cooccurrence Graphs for Supporting Entity-Relationship Discovery in Unstructured Text

Authors: Paolo Fantozzi, Luigi Laura, Umberto Nanni

Abstract:

The problem of Entity relation discovery in structured data, a well covered topic in literature, consists in searching within unstructured sources (typically, text) in order to find connections among entities. These can be a whole dictionary, or a specific collection of named items. In many cases machine learning and/or text mining techniques are used for this goal. These approaches might be unfeasible in computationally challenging problems, such as processing massive data streams. A faster approach consists in collecting the cooccurrences of any two words (entities) in order to create a graph of relations - a cooccurrence graph. Indeed each cooccurrence highlights some grade of semantic correlation between the words because it is more common to have related words close each other than having them in the opposite sides of the text. Some authors have used sliding windows for such problem: they count all the occurrences within a sliding windows running over the whole text. In this paper we generalise such technique, coming up to a Weighted-Distance Sliding Window, where each occurrence of two named items within the window is accounted with a weight depending on the distance between items: a closer distance implies a stronger evidence of a relationship. We develop an experiment in order to support this intuition, by applying this technique to a data set consisting in the text of the Bible, split into verses.

Keywords: Cooccurrence graph, entity relation graph, unstructured text, weighted distance.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 659
436 Satellite Data Classification Accuracy Assessment Based from Reference Dataset

Authors: Mohd Hasmadi Ismail, Kamaruzaman Jusoff

Abstract:

In order to develop forest management strategies in tropical forest in Malaysia, surveying the forest resources and monitoring the forest area affected by logging activities is essential. There are tremendous effort has been done in classification of land cover related to forest resource management in this country as it is a priority in all aspects of forest mapping using remote sensing and related technology such as GIS. In fact classification process is a compulsory step in any remote sensing research. Therefore, the main objective of this paper is to assess classification accuracy of classified forest map on Landsat TM data from difference number of reference data (200 and 388 reference data). This comparison was made through observation (200 reference data), and interpretation and observation approaches (388 reference data). Five land cover classes namely primary forest, logged over forest, water bodies, bare land and agricultural crop/mixed horticultural can be identified by the differences in spectral wavelength. Result showed that an overall accuracy from 200 reference data was 83.5 % (kappa value 0.7502459; kappa variance 0.002871), which was considered acceptable or good for optical data. However, when 200 reference data was increased to 388 in the confusion matrix, the accuracy slightly improved from 83.5% to 89.17%, with Kappa statistic increased from 0.7502459 to 0.8026135, respectively. The accuracy in this classification suggested that this strategy for the selection of training area, interpretation approaches and number of reference data used were importance to perform better classification result.

Keywords: Image Classification, Reference Data, Accuracy Assessment, Kappa Statistic, Forest Land Cover

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3113