Search results for: large database of image
2559 Association Rules Mining and NOSQL Oriented Document in Big Data
Authors: Sarra Senhadji, Imene Benzeguimi, Zohra Yagoub
Abstract:
Big Data represents the recent technology of manipulating voluminous and unstructured data sets over multiple sources. Therefore, NOSQL appears to handle the problem of unstructured data. Association rules mining is one of the popular techniques of data mining to extract hidden relationship from transactional databases. The algorithm for finding association dependencies is well-solved with Map Reduce. The goal of our work is to reduce the time of generating of frequent itemsets by using Map Reduce and NOSQL database oriented document. A comparative study is given to evaluate the performances of our algorithm with the classical algorithm Apriori.
Keywords: Apriori, Association rules mining, Big Data, data mining, Hadoop, Map Reduce, MongoDB, NoSQL.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6942558 Correlation between Capacitance and Dissipation Factor used for Assessment of Stator Insulation
Authors: José Luis Oslinger, Luis Carlos Castro
Abstract:
Measurements of capacitance C and dissipation factor tand of the stator insulation system provide useful information about internal defects within the insulation. The index k is defined as the proportionality constant between the changes at high voltage of capacitance DC and of the dissipation factor Dtand . DC and Dtand values were highly correlated when small flat defects were within the insulation and that correlation was lost in the presence of large narrow defects like electrical treeing. The discrimination between small and large defects is made resorting to partial discharge PD phase angle analysis. For the validation of the results, C and tand measurements were carried out in a 15MVA 4160V steam turbine turbogenerator placed in a sugar mill. In addition, laboratory test results obtained by other authors were analyzed jointly. In such laboratory tests, model coil bars subjected to thermal cycling resulted highly degraded and DC and Dtand values were not correlated. Thus, the index k could not be calculated.Keywords: Aging, capacitance, dissipation factor, electrical treeing, insulation condition, partial discharge.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 29422557 Face Recognition using a Kernelization of Graph Embedding
Authors: Pang Ying Han, Hiew Fu San, Ooi Shih Yin
Abstract:
Linearization of graph embedding has been emerged as an effective dimensionality reduction technique in pattern recognition. However, it may not be optimal for nonlinearly distributed real world data, such as face, due to its linear nature. So, a kernelization of graph embedding is proposed as a dimensionality reduction technique in face recognition. In order to further boost the recognition capability of the proposed technique, the Fisher-s criterion is opted in the objective function for better data discrimination. The proposed technique is able to characterize the underlying intra-class structure as well as the inter-class separability. Experimental results on FRGC database validate the effectiveness of the proposed technique as a feature descriptor.Keywords: Face recognition, Fisher discriminant, graph embedding, kernelization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17012556 Automatic Segmentation of the Clean Speech Signal
Authors: M. A. Ben Messaoud, A. Bouzid, N. Ellouze
Abstract:
Speech Segmentation is the measure of the change point detection for partitioning an input speech signal into regions each of which accords to only one speaker. In this paper, we apply two features based on multi-scale product (MP) of the clean speech, namely the spectral centroid of MP, and the zero crossings rate of MP. We focus on multi-scale product analysis as an important tool for segmentation extraction. The MP is based on making the product of the speech wavelet transform coefficients (WTC). We have estimated our method on the Keele database. The results show the effectiveness of our method. It indicates that the two features can find word boundaries, and extracted the segments of the clean speech.
Keywords: Speech segmentation, Multi-scale product, Spectral centroid, Zero crossings rate.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 25082555 Mineralogical Characterization and Petrographic Classification of the Soil of Casablanca City
Authors: I. Fahi, T. Remmal, F. El Kamel, B. Ayoub
Abstract:
The treatment of the geotechnical database of the region of Casablanca was difficult to achieve due to the heterogeneity of the nomenclature of the lithological formations composing its soil. It appears necessary to harmonize the nomenclature of the facies and to produce cartographic documents useful for construction projects and studies before any investment program. To achieve this, more than 600 surveys made by the Public Laboratory for Testing and Studies (LPEE) in the agglomeration of Casablanca, were studied. Moreover, some local observations were made in different places of the metropolis. Each survey was the subject of a sheet containing lithological succession, macro and microscopic description of petrographic facies with photographic illustration, as well as measurements of geomechanical tests. In addition, an X-ray diffraction analysis was made in order to characterize the surficial formations of the region.
Keywords: Casablanca, guidebook, petrography, soil.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9192554 Methods for Business Process Simulation Based on Petri Nets
Authors: K. Shoylekova, K. Grigorova
Abstract:
The Petri nets are the first standard for business process modeling. Most probably, it is one of the core reasons why all new standards created afterwards have to be so reformed as to reach the stage of mapping the new standard onto Petri nets. The paper presents a business process repository based on a universal database. The repository provides the possibility the data about a given process to be stored in three different ways. Business process repository is developed with regard to the reformation of a given model to a Petri net in order to be easily simulated. Two different techniques for business process simulation based on Petri nets - Yasper and Woflan are discussed. Their advantages and drawbacks are outlined. The way of simulating business process models, stored in the Business process repository is shown.
Keywords: Business process repository, Petri nets, Simulation, Woflan, Yasper.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20662553 The Influence of Zeolitic Spent Refinery Admixture on the Rheological and Technological Properties of Steel Fiber Reinforced Self-Compacting Concrete
Authors: Ž. Rudžionis, P. Grigaliūnas, D. Vaičiukynienė
Abstract:
By planning this experimental work to investigate the effect of zeolitic waste on rheological and technological properties of self-compacting fiber reinforced concrete, we had an intention to draw attention to the environmental factor. Large amount of zeolitic waste, as secondary raw materials are not in use properly and large amount of it is collected without a clear view of its usage in future. The principal aim of this work is to assure, that zeolitic waste admixture takes positive effect to the self-compacting fiber reinforced concrete mixes stability, flowability and other properties by using the experimental research methods. In addition to that a research on cement and zeolitic waste mortars were implemented to clarify the effect of zeolitic waste on properties of cement paste and stone. Primary studies indicates that zeolitic waste characterizes clear pozzolanic behavior, do not deteriorate and in some cases ensure positive rheological and mechanical characteristics of self-compacting concrete mixes.
Keywords: Self compacting concrete, steel fiber reinforced concrete, zeolitic waste, rheological properties of concrete, slump flow.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17802552 Component-based Segmentation of Words from Handwritten Arabic Text
Authors: Jawad H AlKhateeb, Jianmin Jiang, Jinchang Ren, Stan S Ipson
Abstract:
Efficient preprocessing is very essential for automatic recognition of handwritten documents. In this paper, techniques on segmenting words in handwritten Arabic text are presented. Firstly, connected components (ccs) are extracted, and distances among different components are analyzed. The statistical distribution of this distance is then obtained to determine an optimal threshold for words segmentation. Meanwhile, an improved projection based method is also employed for baseline detection. The proposed method has been successfully tested on IFN/ENIT database consisting of 26459 Arabic words handwritten by 411 different writers, and the results were promising and very encouraging in more accurate detection of the baseline and segmentation of words for further recognition.Keywords: Arabic OCR, off-line recognition, Baseline estimation, Word segmentation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22062551 Web Search Engine Based Naming Procedure for Independent Topic
Authors: Takahiro Nishigaki, Takashi Onoda
Abstract:
In recent years, the number of document data has been increasing since the spread of the Internet. Many methods have been studied for extracting topics from large document data. We proposed Independent Topic Analysis (ITA) to extract topics independent of each other from large document data such as newspaper data. ITA is a method for extracting the independent topics from the document data by using the Independent Component Analysis. The topic represented by ITA is represented by a set of words. However, the set of words is quite different from the topics the user imagines. For example, the top five words with high independence of a topic are as follows. Topic1 = {"scor", "game", "lead", "quarter", "rebound"}. This Topic 1 is considered to represent the topic of "SPORTS". This topic name "SPORTS" has to be attached by the user. ITA cannot name topics. Therefore, in this research, we propose a method to obtain topics easy for people to understand by using the web search engine, topics given by the set of words given by independent topic analysis. In particular, we search a set of topical words, and the title of the homepage of the search result is taken as the topic name. And we also use the proposed method for some data and verify its effectiveness.Keywords: Independent topic analysis, topic extraction, topic naming, web search engine.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5002550 Automated Segmentation of ECG Signals using Piecewise Derivative Dynamic Time Warping
Authors: Ali Zifan, Mohammad Hassan Moradi, Sohrab Saberi, Farzad Towhidkhah
Abstract:
Electrocardiogram (ECG) segmentation is necessary to help reduce the time consuming task of manually annotating ECG-s. Several algorithms have been developed to segment the ECG automatically. We first review several of such methods, and then present a new single lead segmentation method based on Adaptive piecewise constant approximation (APCA) and Piecewise derivative dynamic time warping (PDDTW). The results are tested on the QT database. We compared our results to Laguna-s two lead method. Our proposed approach has a comparable mean error, but yields a slightly higher standard deviation than Laguna-s method.Keywords: Adaptive Piecewise Constant Approximation, Dynamic programming, ECG segmentation, Piecewise DerivativeDynamic Time Warping.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20672549 Using Perspective Schemata to Model the ETL Process
Authors: Valeria M. Pequeno, Joao Carlos G. M. Pires
Abstract:
Data Warehouses (DWs) are repositories which contain the unified history of an enterprise for decision support. The data must be Extracted from information sources, Transformed and integrated to be Loaded (ETL) into the DW, using ETL tools. These tools focus on data movement, where the models are only used as a means to this aim. Under a conceptual viewpoint, the authors want to innovate the ETL process in two ways: 1) to make clear compatibility between models in a declarative fashion, using correspondence assertions and 2) to identify the instances of different sources that represent the same entity in the real-world. This paper presents the overview of the proposed framework to model the ETL process, which is based on the use of a reference model and perspective schemata. This approach provides the designer with a better understanding of the semantic associated with the ETL process.
Keywords: conceptual data model, correspondence assertions, data warehouse, data integration, ETL process, object relational database.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15112548 The Contemporary Visual Spectacle — Critical Visual Literacy
Authors: Lai-Fen Yang
Abstract:
In this increasingly visual world, how can we best decipher and understand the many ways that our everyday lives are organized around looking practices and the many images we encounter each day? Indeed, how we interact with and interpret visual images is a basic component of human life. Today, however, we are living in one of the most artificial visual and image-saturated cultures in human history, which makes understanding the complex construction and multiple social functions of visual imagery more important than ever before. Themes regarding our experience of a visually pervasive mediated culture, here, termed visual spectacle.
Keywords: Visual culture, contemporary, visual spectacle.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19662547 Industrial Effects and Firm's Survival (Case Study: Iran- East Azarbaijan Province)
Authors: Ghaffar Tari
Abstract:
The aim of this paper is to investigate the effect of mean size of industry on survival of new firms in East-Azarbaijan province through 1981-2006 using hazard function. So the effect of two variables including mean employment of industry and mean capital of industry are investigated on firm's survival. The Industry & Mine Ministry database has used for data gathering and the data are analyzed using the semi-parametric cox regression model. The results of this study shows that there is a meaningful negative relationship between mean capital of industry and firm's survival, but the mean employment of industry has no meaningful effect on survival of new firms.Keywords: Firm's Survival, Hazard Function, Mean Capital of Industry, Mean Employment of Industry.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 12992546 JREM: An Approach for Formalising Models in the Requirements Phase with JSON and NoSQL Databases
Authors: Aitana Alonso-Nogueira, Helia Estévez-Fernández, Isaías García
Abstract:
This paper presents an approach to reduce some of its current flaws in the requirements phase inside the software development process. It takes the software requirements of an application, makes a conceptual modeling about it and formalizes it within JSON documents. This formal model is lodged in a NoSQL database which is document-oriented, that is, MongoDB, because of its advantages in flexibility and efficiency. In addition, this paper underlines the contributions of the detailed approach and shows some applications and benefits for the future work in the field of automatic code generation using model-driven engineering tools.
Keywords: Conceptual modeling, JSON, NoSQL databases, requirements engineering, software development.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10822545 Remote Vital Signs Monitoring in Neonatal Intensive Care Unit Using a Digital Camera
Authors: Fatema-Tuz-Zohra Khanam, Ali Al-Naji, Asanka G. Perera, Kim Gibson, Javaan Chahl
Abstract:
Conventional contact-based vital signs monitoring sensors such as pulse oximeters or electrocardiogram (ECG) may cause discomfort, skin damage, and infections, particularly in neonates with fragile, sensitive skin. Therefore, remote monitoring of the vital sign is desired in both clinical and non-clinical settings to overcome these issues. Camera-based vital signs monitoring is a recent technology for these applications with many positive attributes. However, there are still limited camera-based studies on neonates in a clinical setting. In this study, the heart rate (HR) and respiratory rate (RR) of eight infants at the Neonatal Intensive Care Unit (NICU) in Flinders Medical Centre were remotely monitored using a digital camera applying color and motion-based computational methods. The region-of-interest (ROI) was efficiently selected by incorporating an image decomposition method. Furthermore, spatial averaging, spectral analysis, band-pass filtering, and peak detection were also used to extract both HR and RR. The experimental results were validated with the ground truth data obtained from an ECG monitor and showed a strong correlation using the Pearson correlation coefficient (PCC) 0.9794 and 0.9412 for HR and RR, respectively. The root mean square errors (RMSE) between camera-based data and ECG data for HR and RR were 2.84 beats/min and 2.91 breaths/min, respectively. A Bland Altman analysis of the data also showed a close correlation between both data sets with a mean bias of 0.60 beats/min and 1 breath/min, and the lower and upper limit of agreement -4.9 to + 6.1 beats/min and -4.4 to +6.4 breaths/min for both HR and RR, respectively. Therefore, video camera imaging may replace conventional contact-based monitoring in NICU and has potential applications in other contexts such as home health monitoring.
Keywords: Neonates, NICU, digital camera, heart rate, respiratory rate, image decomposition.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5762544 Scaling up Detection Rates and Reducing False Positives in Intrusion Detection using NBTree
Authors: Dewan Md. Farid, Nguyen Huu Hoa, Jerome Darmont, Nouria Harbi, Mohammad Zahidur Rahman
Abstract:
In this paper, we present a new learning algorithm for anomaly based network intrusion detection using improved self adaptive naïve Bayesian tree (NBTree), which induces a hybrid of decision tree and naïve Bayesian classifier. The proposed approach scales up the balance detections for different attack types and keeps the false positives at acceptable level in intrusion detection. In complex and dynamic large intrusion detection dataset, the detection accuracy of naïve Bayesian classifier does not scale up as well as decision tree. It has been successfully tested in other problem domains that naïve Bayesian tree improves the classification rates in large dataset. In naïve Bayesian tree nodes contain and split as regular decision-trees, but the leaves contain naïve Bayesian classifiers. The experimental results on KDD99 benchmark network intrusion detection dataset demonstrate that this new approach scales up the detection rates for different attack types and reduces false positives in network intrusion detection.Keywords: Detection rates, false positives, network intrusiondetection, naïve Bayesian tree.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22812543 Automated ECG Segmentation Using Piecewise Derivative Dynamic Time Warping
Authors: Ali Zifan, Sohrab Saberi, Mohammad Hassan Moradi, Farzad Towhidkhah
Abstract:
Electrocardiogram (ECG) segmentation is necessary to help reduce the time consuming task of manually annotating ECG's. Several algorithms have been developed to segment the ECG automatically. We first review several of such methods, and then present a new single lead segmentation method based on Adaptive piecewise constant approximation (APCA) and Piecewise derivative dynamic time warping (PDDTW). The results are tested on the QT database. We compared our results to Laguna's two lead method. Our proposed approach has a comparable mean error, but yields a slightly higher standard deviation than Laguna's method.
Keywords: Adaptive Piecewise Constant Approximation, Dynamic programming, ECG segmentation, Piecewise Derivative Dynamic Time Warping.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 23922542 Development of a Vegetation Searching System
Authors: Rattanathip Rattanachai, Kunyanuth Kularbphettong
Abstract:
This paper describes the development of a Vegetation Searching System based on Web Application in case of Suan Sunandha Rajabhat University. The model was developed by PHP, JavaScript and MySQL database system and it was designed to support searching for endemic and rare species of trees on Web site. We describe the design methods and functional components of this prototype. To evaluate the system performance, questionnaires for the system usability and Black Box Testing were used to measure expert and user satisfaction. The results were satisfactory as followed: Means for experts and users were 4.30 and 4.50, and standard deviation for experts and users were 0.61and 0.73 respectively. Further analysis showed that the quality of the plant searching Website was also at a good level as well.
Keywords: Endemic species, Vegetation, Web based System, and Black Box Testing.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17922541 A Comparison of Some Thresholding Selection Methods for Wavelet Regression
Authors: Alsaidi M. Altaher, Mohd T. Ismail
Abstract:
In wavelet regression, choosing threshold value is a crucial issue. A too large value cuts too many coefficients resulting in over smoothing. Conversely, a too small threshold value allows many coefficients to be included in reconstruction, giving a wiggly estimate which result in under smoothing. However, the proper choice of threshold can be considered as a careful balance of these principles. This paper gives a very brief introduction to some thresholding selection methods. These methods include: Universal, Sure, Ebays, Two fold cross validation and level dependent cross validation. A simulation study on a variety of sample sizes, test functions, signal-to-noise ratios is conducted to compare their numerical performances using three different noise structures. For Gaussian noise, EBayes outperforms in all cases for all used functions while Two fold cross validation provides the best results in the case of long tail noise. For large values of signal-to-noise ratios, level dependent cross validation works well under correlated noises case. As expected, increasing both sample size and level of signal to noise ratio, increases estimation efficiency.
Keywords: wavelet regression, simulation, Threshold.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17672540 Learning Block Memories with Metric Networks
Authors: Mario Gonzalez, David Dominguez, Francisco B. Rodriguez
Abstract:
An attractor neural network on the small-world topology is studied. A learning pattern is presented to the network, then a stimulus carrying local information is applied to the neurons and the retrieval of block-like structure is investigated. A synaptic noise decreases the memory capability. The change of stability from local to global attractors is shown to depend on the long-range character of the network connectivity.Keywords: Hebbian learning, image recognition, small world, spatial information.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18662539 Development of Map of Gridded Basin Flash Flood Potential Index: GBFFPI Map of QuangNam, QuangNgai, DaNang, Hue Provinces
Authors: Le Xuan Cau
Abstract:
Flash flood is occurred in short time rainfall interval: from 1 hour to 12 hours in small and medium basins. Flash floods typically have two characteristics: large water flow and big flow velocity. Flash flood is occurred at hill valley site (strip of lowland of terrain) in a catchment with large enough distribution area, steep basin slope, and heavy rainfall. The risk of flash floods is determined through Gridded Basin Flash Flood Potential Index (GBFFPI). Flash Flood Potential Index (FFPI) is determined through terrain slope flash flood index, soil erosion flash flood index, land cover flash floods index, land use flash flood index, rainfall flash flood index. Determining GBFFPI, each cell in a map can be considered as outlet of a water accumulation basin. GBFFPI of the cell is determined as basin average value of FFPI of the corresponding water accumulation basin. Based on GIS, a tool is developed to compute GBFFPI using ArcObjects SDK for .NET. The maps of GBFFPI are built in two types: GBFFPI including rainfall flash flood index (real time flash flood warning) or GBFFPI excluding rainfall flash flood index. GBFFPI Tool can be used to determine a high flash flood potential site in a large region as quick as possible. The GBFFPI is improved from conventional FFPI. The advantage of GBFFPI is that GBFFPI is taking into account the basin response (interaction of cells) and determines more true flash flood site (strip of lowland of terrain) while conventional FFPI is taking into account single cell and does not consider the interaction between cells. The GBFFPI Map of QuangNam, QuangNgai, DaNang, Hue is built and exported to Google Earth. The obtained map proves scientific basis of GBFFPI.Keywords: ArcObjects SDK for .NET, Basin average value of FFPI, Gridded basin flash flood potential index, GBFFPI map.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19232538 CLASS, A New Tool for Nuclear Scenarios: Description and First Application
Authors: B. Mouginot, J.B. Clavel, N Thiolliere
Abstract:
The presented work is motivated by a french law regarding nuclear waste management. In order to avoid the limitation coming with the usage of the existing scenario codes, as COSI, VISION or FAMILY, the Core Library for Advance Scenario Simulation (CLASS) is being develop. CLASS is an open source tool, which allows any user to simulate an electronuclear scenario. The main CLASS asset, is the possibility to include any type of reactor, even a complitely new concept, through the generation of its ACSII evolution database. In the present article, the CLASS working basis will be presented as well as a simple exemple in order to show his potentiel. In the considered exemple, the effect of the transmutation will be assessed on Minor Actinide Inventory produced by PWR reactors.
Keywords: Electronuclear scenario, reactor, simulation, nuclear waste.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16832537 On the Impact of Reference Node Placement in Wireless Indoor Positioning Systems
Authors: Supattra Aomumpai, Chutima Prommak
Abstract:
This paper presents a studyof the impact of reference node locations on the accuracy of the indoor positioning systems. In particular, we analyze the localization accuracy of the RSSI database mapping techniques, deploying on the IEEE 802.15.4 wireless networks. The results show that the locations of the reference nodes used in the positioning systems affect the signal propagation characteristics in the service area. Thisin turn affects the accuracy of the wireless indoor positioning system. We found that suitable location of reference nodes could reduce the positioning error upto 35 %.Keywords: Indoor positioning systems, IEEE 802.15.4 wireless networks, Signal propagation characteristics
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14832536 Microstructure and Aging Behavior of Nonflammable AZ91D Mg Alloy
Authors: Seok Hong Min, Tae Kwon Ha
Abstract:
Phase equilibria of AZ91D Mg alloys for nonflammable use, containing Ca and Y, were carried out by using FactSage® and FTLite database, which revealed that solid solution treatment could be performed at temperatures from 400 to 450oC. Solid solution treatment of AZ91D Mg alloy without Ca and Y was successfully conducted at 420oC and supersaturated microstructure with all beta phase resolved into matrix was obtained. In the case of AZ91D Mg alloy with some Ca and Y; however, a little amount of intermetallic particles were observed after solid solution treatment. After solid solution treatment, each alloy was annealed at temperatures of 180 and 200oC for time intervals from 1 min to 48 hrs and hardness of each condition was measured by micro-Vickers method. Peak aging conditions were deduced as at the temperature of 200oC for 10 hrs.
Keywords: Mg alloy, AZ91D, nonflammable alloy, phase equilibrium, peak aging.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22362535 A Hybrid Nature Inspired Algorithm for Generating Optimal Query Plan
Authors: R. Gomathi, D. Sharmila
Abstract:
The emergence of the Semantic Web technology increases day by day due to the rapid growth of multiple web pages. Many standard formats are available to store the semantic web data. The most popular format is the Resource Description Framework (RDF). Querying large RDF graphs becomes a tedious procedure with a vast increase in the amount of data. The problem of query optimization becomes an issue in querying large RDF graphs. Choosing the best query plan reduces the amount of query execution time. To address this problem, nature inspired algorithms can be used as an alternative to the traditional query optimization techniques. In this research, the optimal query plan is generated by the proposed SAPSO algorithm which is a hybrid of Simulated Annealing (SA) and Particle Swarm Optimization (PSO) algorithms. The proposed SAPSO algorithm has the ability to find the local optimistic result and it avoids the problem of local minimum. Experiments were performed on different datasets by changing the number of predicates and the amount of data. The proposed algorithm gives improved results compared to existing algorithms in terms of query execution time.
Keywords: Semantic web, RDF, Query optimization, Nature inspired algorithms, PSO, SA.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22392534 Kinematic Hardening Parameters Identification with Respect to Objective Function
Authors: Marina Franulovic, Robert Basan, Bozidar Krizan
Abstract:
Constitutive modeling of material behavior is becoming increasingly important in prediction of possible failures in highly loaded engineering components, and consequently, optimization of their design. In order to account for large number of phenomena that occur in the material during operation, such as kinematic hardening effect in low cycle fatigue behavior of steels, complex nonlinear material models are used ever more frequently, despite of the complexity of determination of their parameters. As a method for the determination of these parameters, genetic algorithm is good choice because of its capability to provide very good approximation of the solution in systems with large number of unknown variables. For the application of genetic algorithm to parameter identification, inverse analysis must be primarily defined. It is used as a tool to fine-tune calculated stress-strain values with experimental ones. In order to choose proper objective function for inverse analysis among already existent and newly developed functions, the research is performed to investigate its influence on material behavior modeling.
Keywords: Genetic algorithm, kinematic hardening, material model, objective function
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 38012533 Impact of Extended Enterprise Resource Planning in the Context of Cloud Computing on Industries and Organizations
Authors: Gholamreza Momenzadeh, Forough Nematolahi
Abstract:
The Extended Enterprise Resource Planning (ERPII) system usually requires massive amounts of storage space, powerful servers, and large upfront and ongoing investments to purchase and manage the software and the related hardware which are not affordable for organizations. In recent decades, organizations prefer to adapt their business structures with new technologies for remaining competitive in the world economy. Therefore, cloud computing (which is one of the tools of information technology (IT)) is a modern system that reveals the next-generation application architecture. Also, cloud computing has had some advantages that reduce costs in many ways such as: lower upfront costs for all computing infrastructure and lower cost of maintaining and supporting. On the other hand, traditional ERPII is not responding for huge amounts of data and relations between the organizations. In this study, based on a literature study, ERPII is investigated in the context of cloud computing where the organizations operate more efficiently. Also, ERPII conditions have a response to needs of organizations in large amounts of data and relations between the organizations.Keywords: Extended enterprise resource planning, cloud computing, business process, enterprise information integration.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9792532 Teaching Approach and Self-Confidence Effect Model Consistency between Taiwan and Singapore Multi-Group HLM
Authors: PeiWen Liao, Tsung Hau Jen
Abstract:
This study was conducted to explore the effects of two countries model comparison program in Taiwan and Singapore in TIMSS database. The researchers used Multi-Group Hierarchical Linear Modeling techniques to compare the effects of two different country models and we tested our hypotheses on 4,046 Taiwan students and 4,599 Singapore students in 2007 at two levels: the class level and student (individual) level. Design quality is a class level variable. Student level variables are achievement and self-confidence. The results challenge the widely held view that retention has a positive impact on self-confidence. Suggestions for future research are discussed.
Keywords: Teaching approach, self-confidence, achievement, multi-group HLM.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18432531 Evaluation of a Multi-Resolution Dyadic Wavelet Transform Method for usable Speech Detection
Authors: Wajdi Ghezaiel, Amel Ben Slimane Rahmouni, Ezzedine Ben Braiek
Abstract:
Many applications of speech communication and speaker identification suffer from the problem of co-channel speech. This paper deals with a multi-resolution dyadic wavelet transform method for usable segments of co-channel speech detection that could be processed by a speaker identification system. Evaluation of this method is performed on TIMIT database referring to the Target to Interferer Ratio measure. Co-channel speech is constructed by mixing all possible gender speakers. Results do not show much difference for different mixtures. For the overall mixtures 95.76% of usable speech is correctly detected with false alarms of 29.65%.Keywords: Co-channel speech, usable speech, multi-resolutionanalysis, speaker identification
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13662530 Ethanol Fuelled HCCI Engine: A Review
Authors: B. Bahri, A. A. Aziz, M. Shahbakhti, M. F. Muhamad Said
Abstract:
The greenhouse effect and limitations on carbon dioxide emissions concern engine maker and the future of the internal combustion engines should go toward substantially and improved thermal efficiency engine. Homogeneous charge compression ignition (HCCI) is an alternative high-efficiency technology for combustion engines to reduce exhaust emissions and fuel consumption. However, there are still tough challenges in the successful operation of HCCI engines, such as controlling the combustion phasing, extending the operating range, and high unburned hydrocarbon and CO emissions. HCCI and the exploitation of ethanol as an alternative fuel is one way to explore new frontiers of internal combustion engines with an eye towards maintaining its sustainability. This study was done to extend database knowledge about HCCI with ethanol a fuel.
Keywords: Ethanol combustion, Ethanol fuel, HCCI.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2825