Search results for: labeling complexity.
562 Effective Context Lossless Image Coding Approach Based on Adaptive Prediction
Authors: Grzegorz Ulacha, Ryszard Stasiński
Abstract:
In the paper an effective context based lossless coding technique is presented. Three principal and few auxiliary contexts are defined. The predictor adaptation technique is an improved CoBALP algorithm, denoted CoBALP+. Cumulated predictor error combining 8 bias estimators is calculated. It is shown experimentally that indeed, the new technique is time-effective while it outperforms the well known methods having reasonable time complexity, and is inferior only to extremely computationally complex ones.Keywords: Adaptive prediction, context coding, image losslesscoding, prediction error bias correction.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1350561 Evolutionary Query Optimization for Heterogeneous Distributed Database Systems
Authors: Reza Ghaemi, Amin Milani Fard, Hamid Tabatabaee, Mahdi Sadeghizadeh
Abstract:
Due to new distributed database applications such as huge deductive database systems, the search complexity is constantly increasing and we need better algorithms to speedup traditional relational database queries. An optimal dynamic programming method for such high dimensional queries has the big disadvantage of its exponential order and thus we are interested in semi-optimal but faster approaches. In this work we present a multi-agent based mechanism to meet this demand and also compare the result with some commonly used query optimization algorithms.Keywords: Information retrieval systems, list fusion methods, document score, multi-agent systems.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3424560 Recognition of Isolated Speech Signals using Simplified Statistical Parameters
Authors: Abhijit Mitra, Bhargav Kumar Mitra, Biswajoy Chatterjee
Abstract:
We present a novel scheme to recognize isolated speech signals using certain statistical parameters derived from those signals. The determination of the statistical estimates is based on extracted signal information rather than the original signal information in order to reduce the computational complexity. Subtle details of these estimates, after extracting the speech signal from ambience noise, are first exploited to segregate the polysyllabic words from the monosyllabic ones. Precise recognition of each distinct word is then carried out by analyzing the histogram, obtained from these information.Keywords: Isolated speech signals, Block overlapping technique, Positive peaks, Histogram analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1426559 Evaluating Performance of Quality-of-Service Routing in Large Networks
Authors: V. Narasimha Raghavan, M. Venkatesh, T. Peer Meera Labbai, Praveen Dwarakanath Prabhu
Abstract:
The performance and complexity of QoS routing depends on the complex interaction between a large set of parameters. This paper investigated the scaling properties of source-directed link-state routing in large core networks. The simulation results show that the routing algorithm, network topology, and link cost function each have a significant impact on the probability of successfully routing new connections. The experiments confirm and extend the findings of other studies, and also lend new insight designing efficient quality-of-service routing policies in large networks.
Keywords: QoS, Link-State Routing, Dijkstra, Path Selection, Path Computation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1581558 Efficient Realization of an ADFE with a New Adaptive Algorithm
Authors: N. Praveen Kumar, Abhijit Mitra, C. Ardil
Abstract:
Decision feedback equalizers are commonly employed to reduce the error caused by intersymbol interference. Here, an adaptive decision feedback equalizer is presented with a new adaptation algorithm. The algorithm follows a block-based approach of normalized least mean square (NLMS) algorithm with set-membership filtering and achieves a significantly less computational complexity over its conventional NLMS counterpart with set-membership filtering. It is shown in the results that the proposed algorithm yields similar type of bit error rate performance over a reasonable signal to noise ratio in comparison with the latter one.Keywords: Decision feedback equalizer, Adaptive algorithm, Block based computation, Set membership filtering.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1677557 Reverse Logistics Information Management Using Ontological Approach
Authors: F. Lhafiane, A. Elbyed, M. Bouchoum
Abstract:
Reverse Logistics (RL) Network is considered as complex and dynamic network that involves many stakeholders such as: suppliers, manufactures, warehouse, retails and costumers, this complexity is inherent in such process due to lack of perfect knowledge or conflicting information. Ontologies on the other hand can be considered as an approach to overcome the problem of sharing knowledge and communication among the various reverse logistics partners. In this paper we propose a semantic representation based on hybrid architecture for building the Ontologies in ascendant way, this method facilitates the semantic reconciliation between the heterogeneous information systems that support reverse logistics processes and product data.
Keywords: Reverse Logistics, information management, heterogeneity, Ontologies, semantic web.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2966556 A study on a Generic Development Process for the BPM+SOA Design and Implementation
Authors: Toshimi Munehira
Abstract:
In order to optimize annual IT spending and to reduce the complexity of an entire system architecture, SOA trials have been started. It is common knowledge that to design an SOA system we have to adopt the top-down approach, but in reality silo systems are being made, so these companies cannot reuse newly designed services, and cannot enjoy SOA-s economic benefits. To prevent this situation, we designed a generic SOA development process referred to as the architecture of “mass customization." To define the generic detail development processes, we did a case study on an imaginary company. Through the case study, we could define the practical development processes and found this could vastly reduce updating development costs.Keywords: SOA, BPM, Generic Model, MassCustomization
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1456555 A Generic, Functionally Comprehensive Approach to Maintaining an Ontology as a Relational Database
Authors: Jennifer Leopold, Alton Coalter, Leong Lee
Abstract:
An ontology is a data model that represents a set of concepts in a given field and the relationships among those concepts. As the emphasis on achieving a semantic web continues to escalate, ontologies for all types of domains increasingly will be developed. These ontologies may become large and complex, and as their size and complexity grows, so will the need for multi-user interfaces for ontology curation. Herein a functionally comprehensive, generic approach to maintaining an ontology as a relational database is presented. Unlike many other ontology editors that utilize a database, this approach is entirely domain-generic and fully supports Webbased, collaborative editing including the designation of different levels of authorization for users.Keywords: Ontology Editor, Relational Database, CollaborativeCuration.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1446554 Turing Pattern in the Oregonator Revisited
Authors: Elragig Aiman, Dreiwi Hanan, Townley Stuart, Elmabrook Idriss
Abstract:
In this paper, we reconsider the analysis of the Oregonator model. We highlight an error in this analysis which leads to an incorrect depiction of the parameter region in which diffusion driven instability is possible. We believe that the cause of the oversight is the complexity of stability analyses based on eigenvalues and the dependence on parameters of matrix minors appearing in stability calculations. We regenerate the parameter space where Turing patterns can be seen, and we use the common Lyapunov function (CLF) approach, which is numerically reliable, to further confirm the dependence of the results on diffusion coefficients intensities.Keywords: Diffusion driven instability, common Lyapunov function (CLF), turing pattern, positive-definite matrix.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1048553 Development of a RAM Simulation Model for Acid Gas Removal System
Authors: Ainul Akmar Mokhtar, Masdi Muhammad, Hilmi Hussin, Mohd Amin Abdul Majid
Abstract:
A reliability, availability and maintainability (RAM) model has been built for acid gas removal plant for system analysis that will play an important role in any process modifications, if required, for achieving its optimum performance. Due to the complexity of the plant, the model was based on a Reliability Block Diagram (RBD) with a Monte Carlo simulation engine. The model has been validated against actual plant data as well as local expert opinions, resulting in an acceptable simulation model. The results from the model showed that the operation and maintenance can be further improved, resulting in reduction of the annual production loss.
Keywords: Acid gas removal plant, RAM model, Reliabilityblock diagram
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2343552 On the Symbol Based Decision Feedback Equalizer
Authors: Mohammed Nafie
Abstract:
Decision Feedback equalizers (DFEs) usually outperform linear equalizers for channels with intersymbol interference. However, the DFE performance is highly dependent on the availability of reliable past decisions. Hence, in coded systems, where reliable decisions are only available after decoding the full block, the performance of the DFE will be affected. A symbol based DFE is a DFE that only uses the decision after the block is decoded. In this paper we derive the optimal settings of both the feedforward and feedback taps of the symbol based equalizer. We present a novel symbol based DFE filterbank, and derive its taps optimal settings. We also show that it outperforms the classic DFE in terms of complexity and/or performance.Keywords: Coding, DFE, Equalization, Exponential Channelmodels.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2304551 Low-complexity Integer Frequency Offset Synchronization for OFDMA System
Authors: Young-Jae Kim, Young-Hwan You
Abstract:
This paper presents a integer frequency offset (IFO) estimation scheme for the 3GPP long term evolution (LTE) downlink system. Firstly, the conventional joint detection method for IFO and sector cell index (CID) information is introduced. Secondly, an IFO estimation without explicit sector CID information is proposed, which can operate jointly with the proposed IFO estimation and reduce the time delay in comparison with the conventional joint method. Also, the proposed method is computationally efficient and has almost similar performance in comparison with the conventional method over the Pedestrian and Vehicular channel models.Keywords: LTE, OFDMA, primary synchronization signal (PSS), IFO, CID
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2313550 Solving SPDEs by a Least Squares Method
Authors: Hassan Manouzi
Abstract:
We present in this paper a useful strategy to solve stochastic partial differential equations (SPDEs) involving stochastic coefficients. Using the Wick-product of higher order and the Wiener-Itˆo chaos expansion, the SPDEs is reformulated as a large system of deterministic partial differential equations. To reduce the computational complexity of this system, we shall use a decomposition-coordination method. To obtain the chaos coefficients in the corresponding deterministic equations, we use a least square formulation. Once this approximation is performed, the statistics of the numerical solution can be easily evaluated.
Keywords: Least squares, Wick product, SPDEs, finite element, Wiener chaos expansion, gradient method.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1801549 Design of Variable Fractional-Delay FIR Differentiators
Authors: Jong-Jy Shyu, Soo-Chang Pei, Min-Han Chang
Abstract:
In this paper, the least-squares design of variable fractional-delay (VFD) finite impulse response (FIR) digital differentiators is proposed. The used transfer function is formulated so that Farrow structure can be applied to realize the designed system. Also, the symmetric characteristics of filter coefficients are derived, which leads to the complexity reduction by saving almost a half of the number of coefficients. Moreover, all the elements of related vectors or matrices for the optimal process can be represented in closed forms, which make the design easier. Design example is also presented to illustrate the effectiveness of the proposed method.
Keywords: Differentiator, variable fractional-delay filter, FIR filter, least-squares method, Farrow structure.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1415548 A Fast Object Detection Method with Rotation Invariant Features
Authors: Zilong He, Yuesheng Zhu
Abstract:
Based on the combined shape feature and texture feature, a fast object detection method with rotation invariant features is proposed in this paper. A quick template matching scheme based online learning designed for online applications is also introduced in this paper. The experimental results have shown that the proposed approach has the features of lower computation complexity and higher detection rate, while keeping almost the same performance compared to the HOG-based method, and can be more suitable for run time applications.Keywords: gradient feature, online learning, rotationinvariance, template feature
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2477547 Performance Enhancement of Motion Estimation Using SSE2 Technology
Authors: Trung Hieu Tran, Hyo-Moon Cho, Sang-Bock Cho
Abstract:
Motion estimation is the most computationally intensive part in video processing. Many fast motion estimation algorithms have been proposed to decrease the computational complexity by reducing the number of candidate motion vectors. However, these studies are for fast search algorithms themselves while almost image and video compressions are operated with software based. Therefore, the timing constraints for running these motion estimation algorithms not only challenge for the video codec but also overwhelm for some of processors. In this paper, the performance of motion estimation is enhanced by using Intel's Streaming SIMD Extension 2 (SSE2) technology with Intel Pentium 4 processor.Keywords: Motion Estimation, Full Search, Three StepSearch, MMX/SSE/SSE2 Technologies, SIMD.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2100546 Thermal Carpet Cloaking Achieved by Layered Metamaterial
Authors: Bang-Shiuh Chen, Lien-Wen Chen
Abstract:
We have devised a thermal carpet cloak theoretically and implemented in silicon using layered metamaterial. The layered metamaterial is composed of single crystalline silicon and its phononic crystal. The design is based on a coordinate transformation. We demonstrate the result with numerical simulation. Great cloaking performance is achieved as a thermal insulator is well hidden under the thermal carpet cloak. We also show that the thermal carpet cloak can even the temperature on irregular surface. Using thermal carpet cloak to manipulate the heat conduction is effective because of its low complexity.
Keywords: Metamaterial, heat conduction, cloaking, phononic crystal.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2264545 Risk Allocation in Public-Private Partnership (PPP) Projects for Wastewater Treatment Plants
Authors: Samuel Capintero, Ole H. Petersen
Abstract:
This paper examines the utilization of public-private partnerships for the building and operation of wastewater treatment plants. Our research focuses on risk allocation in this kind of projects. Our analysis builds on more than hundred wastewater treatment plants built and operated through PPP projects in Aragon (Spain). The paper illustrates the consequences of an inadequate management of construction risk and an unsuitable transfer of demand risk in wastewater treatment plants. It also shows that the involvement of many public bodies at local, regional and national level further increases the complexity of this kind of projects and make time delays more likely.Keywords: Wastewater, treatment plants, PPP, construction.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2286544 Intelligent Aid-Analysis Based on the Use of Digital Twin: Application to Electronic Warfare System
Authors: L. Chaussy, M. Nouvel
Abstract:
Workload of the system engineers during Integration Validation Verification process of Electronic Warfare Systems (EWS) is growing with complexity of the systems and with the diversity of tested cases (diversity of operational scenario in front of EWS). Even if the use of Digital Twin makes easier conception and development phases in term of planning and test equipment availability, time to analyze tests results is still too long and too complex. The idea to reduce the system engineer’s workload and improve test coverage is to introduce some intelligent and aid-analysis algorithms to improve this step.
Keywords: Analysis tools, automatic testing, digital twin, electronic warfare system.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 385543 An Optimized Design of Non-uniform Filterbank
Authors: Ram Kumar Soni, Alok Jain, Rajiv Saxena
Abstract:
The tree structured approach of non-uniform filterbank (NUFB) is normally used in perfect reconstruction (PR). The PR is not always feasible due to certain limitations, i.e, constraints in selecting design parameters, design complexity and some times output is severely affected by aliasing error if necessary and sufficient conditions of PR is not satisfied perfectly. Therefore, there has been generalized interest of researchers to go for near perfect reconstruction (NPR). In this proposed work, an optimized tree structure technique is used for the design of NPR non-uniform filterbank. Window functions of Blackman family are used to design the prototype FIR filter. A single variable linear optimization is used to minimize the amplitude distortion. The main feature of the proposed design is its simplicity with linear phase property.Keywords: Tree structure, NUFB, QMF, NPR.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1738542 Optimal ECG Sampling Frequency for Multiscale Entropy-Based HRV
Authors: Manjit Singh
Abstract:
Multiscale entropy (MSE) is an extensively used index to provide a general understanding of multiple complexity of physiologic mechanism of heart rate variability (HRV) that operates on a wide range of time scales. Accurate selection of electrocardiogram (ECG) sampling frequency is an essential concern for clinically significant HRV quantification; high ECG sampling rate increase memory requirements and processing time, whereas low sampling rate degrade signal quality and results in clinically misinterpreted HRV. In this work, the impact of ECG sampling frequency on MSE based HRV have been quantified. MSE measures are found to be sensitive to ECG sampling frequency and effect of sampling frequency will be a function of time scale.Keywords: ECG, heart rate variability, HRV, multiscale entropy, sampling frequency.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1352541 A Simulation Software for DNA Computing Algorithms Implementation
Authors: M. S. Muhammad, S. M. W. Masra, K. Kipli, N. Zamhari
Abstract:
The capturing of gel electrophoresis image represents the output of a DNA computing algorithm. Before this image is being captured, DNA computing involves parallel overlap assembly (POA) and polymerase chain reaction (PCR) that is the main of this computing algorithm. However, the design of the DNA oligonucleotides to represent a problem is quite complicated and is prone to errors. In order to reduce these errors during the design stage before the actual in-vitro experiment is carried out; a simulation software capable of simulating the POA and PCR processes is developed. This simulation software capability is unlimited where problem of any size and complexity can be simulated, thus saving cost due to possible errors during the design process. Information regarding the DNA sequence during the computing process as well as the computing output can be extracted at the same time using the simulation software.Keywords: DNA computing, PCR, POA, simulation software
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1814540 A Hybrid Feature Subset Selection Approach based on SVM and Binary ACO. Application to Industrial Diagnosis
Authors: O. Kadri, M. D. Mouss, L.H. Mouss, F. Merah
Abstract:
This paper proposes a novel hybrid algorithm for feature selection based on a binary ant colony and SVM. The final subset selection is attained through the elimination of the features that produce noise or, are strictly correlated with other already selected features. Our algorithm can improve classification accuracy with a small and appropriate feature subset. Proposed algorithm is easily implemented and because of use of a simple filter in that, its computational complexity is very low. The performance of the proposed algorithm is evaluated through a real Rotary Cement kiln dataset. The results show that our algorithm outperforms existing algorithms.
Keywords: Binary Ant Colony algorithm, Support VectorMachine, feature selection, classification.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1608539 Capacity Flexibility within Production
Authors: Johannes Nywlt, Julian Becker, Sebastian Bertsch
Abstract:
Due to high dynamics in current markets the expectations regarding logistics increase steadily. However, the complexity and variety of products and production make it difficult to understand the interdependencies between logistical objectives and their determining factors. Therefore specific models are needed to meet this challenge. The Logistic Operating Curves Theory is such a model. With its aid the basic correlations between the logistic objectives can be described. Within this model the capacity flexibility represents an important parameter. However, a proper mathematical description for this parameter is still missing. Within this paper such a description will be developed in order to make the Logistic Operating Curves Theory more accurate.
Keywords: Capacity flexibility, Production controlling, Production logistics, Production management.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2086538 Combining Diverse Neural Classifiers for Complex Problem Solving: An ECOC Approach
Authors: R. Ebrahimpour, M. Abbasnezhad Arabi, H. Babamiri Moghaddam
Abstract:
Combining classifiers is a useful method for solving complex problems in machine learning. The ECOC (Error Correcting Output Codes) method has been widely used for designing combining classifiers with an emphasis on the diversity of classifiers. In this paper, in contrast to the standard ECOC approach in which individual classifiers are chosen homogeneously, classifiers are selected according to the complexity of the corresponding binary problem. We use SATIMAGE database (containing 6 classes) for our experiments. The recognition error rate in our proposed method is %10.37 which indicates a considerable improvement in comparison with the conventional ECOC and stack generalization methods.Keywords: Error correcting output code, combining classifiers, neural networks.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1401537 AI Applications to Metal Stamping Die Design– A Review
Authors: Vishal Naranje, Shailendra Kumar
Abstract:
Metal stamping die design is a complex, experiencebased and time-consuming task. Various artificial intelligence (AI) techniques are being used by worldwide researchers for stamping die design to reduce complexity, dependence on human expertise and time taken in design process as well as to improve design efficiency. In this paper a comprehensive review of applications of AI techniques in manufacturability evaluation of sheet metal parts, die design and process planning of metal stamping die is presented. Further the salient features of major research work published in the area of metal stamping are presented in tabular form and scope of future research work is identified.Keywords: Artificial Intelligence, Die design, ManufacturabilityEvaluation, Metal Stamping Die.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3869536 Modeling and Simulation of Robotic Arm Movement using Soft Computing
Authors: V. K. Banga, Jasjit Kaur, R. Kumar, Y. Singh
Abstract:
In this research paper we have presented control architecture for robotic arm movement and trajectory planning using Fuzzy Logic (FL) and Genetic Algorithms (GAs). This architecture is used to compensate the uncertainties like; movement, friction and settling time in robotic arm movement. The genetic algorithms and fuzzy logic is used to meet the objective of optimal control movement of robotic arm. This proposed technique represents a general model for redundant structures and may extend to other structures. Results show optimal angular movement of joints as result of evolutionary process. This technique has edge over the other techniques as minimum mathematics complexity used.Keywords: Kinematics, Genetic algorithms (GAs), Fuzzy logic(FL), Optimal control.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3009535 A Fast HRRP Synthesis Algorithm with Sensing Dictionary in GTD Model
Authors: R. Fan, Q. Wan, H. Chen, Y.L. Liu, Y.P. Liu
Abstract:
In the paper, a fast high-resolution range profile synthetic algorithm called orthogonal matching pursuit with sensing dictionary (OMP-SD) is proposed. It formulates the traditional HRRP synthetic to be a sparse approximation problem over redundant dictionary. As it employs a priori that the synthetic range profile (SRP) of targets are sparse, SRP can be accomplished even in presence of data lost. Besides, the computation complexity decreases from O(MNDK) flops for OMP to O(M(N + D)K) flops for OMP-SD by introducing sensing dictionary (SD). Simulation experiments illustrate its advantages both in additive white Gaussian noise (AWGN) and noiseless situation, respectively.
Keywords: GTD-based model, HRRP, orthogonal matching pursuit, sensing dictionary.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1923534 Strategic Management via System Dynamics Simulation Models
Authors: G. Papageorgiou, A. Hadjis
Abstract:
This paper examines the problem of strategic management in highly turbulent dynamic business environmental conditions. As shown the high complexity of the problem can be managed with the use of System Dynamics Models and Computer Simulation in obtaining insights, and thorough understanding of the interdependencies between the organizational structure and the business environmental elements, so that effective product –market strategies can be designed. Simulation reveals the underlying forces that hold together the structure of an organizational system in relation to its environment. Such knowledge will contribute to the avoidance of fundamental planning errors and enable appropriate proactive well focused action.Keywords: Strategic Management, System Dynamics, Modelingand Simulation, Strategic Planning, Organizational Dynamics
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2610533 Optimal External Merge Sorting Algorithm with Smart Block Merging
Authors: Mir Hadi Seyedafsari, Iraj Hasanzadeh
Abstract:
Like other external sorting algorithms, the presented algorithm is a two step algorithm including internal and external steps. The first part of the algorithm is like the other similar algorithms but second part of that is including a new easy implementing method which has reduced the vast number of inputoutput operations saliently. As decreasing processor operating time does not have any effect on main algorithm speed, any improvement in it should be done through decreasing the number of input-output operations. This paper propose an easy algorithm for choose the correct record location of the final list. This decreases the time complexity and makes the algorithm faster.Keywords: External sorting algorithm, internal sortingalgorithm, fast sorting, robust algorithm.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2189