Search results for: complexity theory
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 2303

Search results for: complexity theory

2213 Functional Decomposition Based Effort Estimation Model for Software-Intensive Systems

Authors: Nermin Sökmen

Abstract:

An effort estimation model is needed for softwareintensive projects that consist of hardware, embedded software or some combination of the two, as well as high level software solutions. This paper first focuses on functional decomposition techniques to measure functional complexity of a computer system and investigates its impact on system development effort. Later, it examines effects of technical difficulty and design team capability factors in order to construct the best effort estimation model. With using traditional regression analysis technique, the study develops a system development effort estimation model which takes functional complexity, technical difficulty and design team capability factors as input parameters. Finally, the assumptions of the model are tested.

Keywords: Functional complexity, functional decomposition, development effort, technical difficulty, design team capability, regression analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2238
2212 Learning through Shared Procedures -A Case of Using Technology to Bridge the Gap between Theory and Practice in Officer Education

Authors: O. Boe, S-T. Kristiansen, R. Wold

Abstract:

In this article we explore how computer assisted exercises may allow for bridging the traditional gap between theory and practice in professional education. To educate officers able to master the complexity of the battlefield the Norwegian Military Academy needs to develop a learning environment that allows for creating viable connections between the educational environment and the field of practice. In response to this challenge we explore the conditions necessary to make computer assisted training systems (CATS) a useful tool to create structural similarities between an educational context and the field of military practice. Although, CATS may facilitate work procedures close to real life situations, this case do demonstrate how professional competence also must build on viable learning theories and environments. This paper explores the conditions that allow for using simulators to facilitate professional competence from within an educational setting. We develop a generic didactic model that ascribes learning to participation in iterative cycles of action and reflection. The development of this model is motivated by the need to develop an interdisciplinary professional education rooted in the pattern of military practice.

Keywords: Development in higher education, experiential learning, professional education, simulation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1152
2211 The "Project" Approach in Urban: A Response to Uncertainty

Authors: Nedjima Mouhoubi, Souad Sassi Boudemagh

Abstract:

In this paper, we will try to demonstrate the importance of the project approach in the urban to deal with uncertainty, the importance of the involvement of all stakeholders in the urban project process and that the absence of an actor can lead to project failure but also the importance of the urban project management. These points are handled through the following questions: Does the urban adhere to the theory of complexity? Does the project approach bring hope and solution to make urban planning "sustainable"? How converging visions of actors for the same project? Is the management of urban project the solution to support the urban project approach?

Keywords: Strategic planning, project, urban project stakeholders, management.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1244
2210 Research Analysis in Eclectic Theory (Kaboudan and Sfandiar)

Authors: Farideh Alizadeh, Mohd Nasir Hashim

Abstract:

Present research investigates eclecticism in Iranian theatre on the basis of eclectic theory. Eclectic theatre is a new theory in postmodernism. The theory appeared during 60th – 70th century in some theatres such as “Conference of the Birds”. Special theatrical forms have been developed in many geographical- cultural areas of the world and are indigenous to that area. These forms, as compared with original forms, are considered to be traditional while being comprehensive, the form is considered to be national. Kaboudan and Sfandiar theatre has been influenced by elements of traditional form of Iran.

Keywords: Eclectic theatre, theatrical forms, tradition, play.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2072
2209 Discovery of Quantified Hierarchical Production Rules from Large Set of Discovered Rules

Authors: Tamanna Siddiqui, M. Afshar Alam

Abstract:

Automated discovery of Rule is, due to its applicability, one of the most fundamental and important method in KDD. It has been an active research area in the recent past. Hierarchical representation allows us to easily manage the complexity of knowledge, to view the knowledge at different levels of details, and to focus our attention on the interesting aspects only. One of such efficient and easy to understand systems is Hierarchical Production rule (HPRs) system. A HPR, a standard production rule augmented with generality and specificity information, is of the following form: Decision If < condition> Generality Specificity . HPRs systems are capable of handling taxonomical structures inherent in the knowledge about the real world. This paper focuses on the issue of mining Quantified rules with crisp hierarchical structure using Genetic Programming (GP) approach to knowledge discovery. The post-processing scheme presented in this work uses Quantified production rules as initial individuals of GP and discovers hierarchical structure. In proposed approach rules are quantified by using Dempster Shafer theory. Suitable genetic operators are proposed for the suggested encoding. Based on the Subsumption Matrix(SM), an appropriate fitness function is suggested. Finally, Quantified Hierarchical Production Rules (HPRs) are generated from the discovered hierarchy, using Dempster Shafer theory. Experimental results are presented to demonstrate the performance of the proposed algorithm.

Keywords: Knowledge discovery in database, quantification, dempster shafer theory, genetic programming, hierarchy, subsumption matrix.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1493
2208 Performance Analysis of HSDPA Systems using Low-Density Parity-Check (LDPC)Coding as Compared to Turbo Coding

Authors: K. Anitha Sheela, J. Tarun Kumar

Abstract:

HSDPA is a new feature which is introduced in Release-5 specifications of the 3GPP WCDMA/UTRA standard to realize higher speed data rate together with lower round-trip times. Moreover, the HSDPA concept offers outstanding improvement of packet throughput and also significantly reduces the packet call transfer delay as compared to Release -99 DSCH. Till now the HSDPA system uses turbo coding which is the best coding technique to achieve the Shannon limit. However, the main drawbacks of turbo coding are high decoding complexity and high latency which makes it unsuitable for some applications like satellite communications, since the transmission distance itself introduces latency due to limited speed of light. Hence in this paper it is proposed to use LDPC coding in place of Turbo coding for HSDPA system which decreases the latency and decoding complexity. But LDPC coding increases the Encoding complexity. Though the complexity of transmitter increases at NodeB, the End user is at an advantage in terms of receiver complexity and Bit- error rate. In this paper LDPC Encoder is implemented using “sparse parity check matrix" H to generate a codeword at Encoder and “Belief Propagation algorithm "for LDPC decoding .Simulation results shows that in LDPC coding the BER suddenly drops as the number of iterations increase with a small increase in Eb/No. Which is not possible in Turbo coding. Also same BER was achieved using less number of iterations and hence the latency and receiver complexity has decreased for LDPC coding. HSDPA increases the downlink data rate within a cell to a theoretical maximum of 14Mbps, with 2Mbps on the uplink. The changes that HSDPA enables includes better quality, more reliable and more robust data services. In other words, while realistic data rates are only a few Mbps, the actual quality and number of users achieved will improve significantly.

Keywords: AMC, HSDPA, LDPC, WCDMA, 3GPP.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2015
2207 Application the Queuing Theory in the Warehouse Optimization

Authors: Jaroslav Masek, Juraj Camaj, Eva Nedeliakova

Abstract:

The aim of optimization of store management is not only designing the situation of store management itself including its equipment, technology and operation. In optimization of store management we need to consider also synchronizing of technological, transport, store and service operations throughout the whole process of logistic chain in such a way that a natural flow of material from provider to consumer will be achieved the shortest possible way, in the shortest possible time in requested quality and quantity and with minimum costs. The paper deals with the application of the queuing theory for optimization of warehouse processes. The first part refers to common information about the problematic of warehousing and using mathematical methods for logistics chains optimization. The second part refers to preparing a model of a warehouse within queuing theory. The conclusion of the paper includes two examples of using queuing theory in praxis.

Keywords: Queuing theory, logistics system, mathematical methods, warehouse optimization.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6499
2206 Predicting Dietary Practice Behavior among Type 2 Diabetics Using the Theory of Planned Behavior and Mixed Methods Design

Authors: D.O. Omondi, M.K. Walingo, G.M. Mbagaya, L.O.A. Othuon

Abstract:

This study applied the Theory of Planned Behavior model in predicting dietary behavior among Type 2 diabetics in a Kenyan environment. The study was conducted for three months within the diabetic clinic at Kisii Hospital in Nyanza Province in Kenya and adopted sequential mixed methods design combing both qualitative and quantitative phases. Qualitative data was analyzed using grounded theory analysis method. Structural equation modeling using maximum likelihood was used to analyze quantitative data. The results based on the common fit indices revealed that the theory of planned behavior fitted the data acceptably well among the Type 2 diabetes and within dietary behavior {χ2 = 223.3, df = 77, p = .02, χ2/df = 2.9, n=237; TLI = .93; CFI =.91; RMSEA (90CI) = .090(.039, .146)}. This implies that the Theory of Planned Behavior holds and forms a framework for promoting dietary practice among Type 2 diabetics.

Keywords: Dietary practice, Kenya, Theory of PlannedBehavior, Type 2 diabetes, Mixed Methods Design.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2052
2205 Theoretical Considerations for Software Component Metrics

Authors: V. Lakshmi Narasimhan, Bayu Hendradjaya

Abstract:

We have defined two suites of metrics, which cover static and dynamic aspects of component assembly. The static metrics measure complexity and criticality of component assembly, wherein complexity is measured using Component Packing Density and Component Interaction Density metrics. Further, four criticality conditions namely, Link, Bridge, Inheritance and Size criticalities have been identified and quantified. The complexity and criticality metrics are combined to form a Triangular Metric, which can be used to classify the type and nature of applications. Dynamic metrics are collected during the runtime of a complete application. Dynamic metrics are useful to identify super-component and to evaluate the degree of utilisation of various components. In this paper both static and dynamic metrics are evaluated using Weyuker-s set of properties. The result shows that the metrics provide a valid means to measure issues in component assembly. We relate our metrics suite with McCall-s Quality Model and illustrate their impact on product quality and to the management of component-based product development.

Keywords: Component Assembly, Component Based SoftwareEngineering, CORBA Component Model, Software ComponentMetrics.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2241
2204 Climate Change and Environmental Education: The Application of Concept Map for Representing the Knowledge Complexity of Climate Change

Authors: Hsueh-Chih, Chen, Yau-Ting, Sung, Tsai-Wen, Lin, Hung-Teng, Chou

Abstract:

It has formed an essential issue that Climate Change, composed of highly knowledge complexity, reveals its significant impact on human existence. Therefore, specific national policies, some of which present the educational aspects, have been published for overcoming the imperative problem. Accordingly, the study aims to analyze as well as integrate the relationship between Climate Change and environmental education and apply the perspective of concept map to represent the knowledge contents and structures of Climate Change; by doing so, knowledge contents of Climate Change could be represented in an even more comprehensive way and manipulated as the tool for environmental education. The method adapted for this study is knowledge conversion model compounded of the platform for experts and teachers, who were the participants for this study, to cooperate and combine each participant-s standpoints into a complete knowledge framework that is the foundation for structuring the concept map. The result of this research contains the important concepts, the precise propositions and the entire concept map for representing the robust concepts of Climate Change.

Keywords: Climate Change, knowledge complexity, concept map.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1680
2203 Flexure of Cantilever Thick Beams Using Trigonometric Shear Deformation Theory

Authors: Yuwaraj M. Ghugal, Ajay G. Dahake

Abstract:

A trigonometric shear deformation theory for flexure of thick beams, taking into account transverse shear deformation effects, is developed. The number of variables in the present theory is same as that in the first order shear deformation theory. The sinusoidal function is used in displacement field in terms of thickness coordinate to represent the shear deformation effects. The noteworthy feature of this theory is that the transverse shear stresses can be obtained directly from the use of constitutive relations with excellent accuracy, satisfying the shear stress free conditions on the top and bottom surfaces of the beam. Hence, the theory obviates the need of shear correction factor. Governing differential equations and boundary conditions are obtained by using the principle of virtual work. The thick cantilever isotropic beams are considered for the numerical studies to demonstrate the efficiency of the. Results obtained are discussed critically with those of other theories.

Keywords: Trigonometric shear deformation, thick beam, flexure, principle of virtual work, equilibrium equations, stress.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3061
2202 A New Developed Formula to Determine the Shear Buckling Stress in Welded Aluminum Plate Girders

Authors: Badr Alsulami, Ahmed S. Elamary

Abstract:

This paper summarizes and presents main results of an in-depth numerical analysis dealing with the shear buckling resistance of aluminum plate girders. The studies conducted have permitted the development of a simple design expression to determine the critical shear buckling stress in aluminum web panels. This expression takes into account the effects of reduction of strength in aluminum alloys due to welding process. Ultimate shear resistance (USR) of plate girders can be obtained theoretically using Cardiff theory or Hӧglunds theory. USR of aluminum alloy plate girders predicted theoretically using BS8118 appear inconsistent when compared with test data. Theoretical predictions based on Hӧglunds theory, are more realistic. Cardiff theory proposed to predict the USR of steel plate girders only. Welded aluminum alloy plate girders studied experimentally by others; the USR resulted from tests are reviewed. Comparison between the test results with the values obtained from Hӧglunds theory, BS8118 design method and Cardiff theory performed theoretically. Finally, a new equation based on Cardiff tension-field theory, proposed to predict theoretically the USR of aluminum plate girders.

Keywords: Shear resistance, Aluminum, Cardiff theory, Hӧglund's theory, Plate girder.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2835
2201 Using Information Theory to Observe Natural Intelligence and Artificial Intelligence

Authors: Lipeng Zhang, Limei Li, Yanming Pearl Zhang

Abstract:

This paper takes a philosophical view as axiom, and reveals the relationship between information theory and Natural Intelligence and Artificial Intelligence under real world conditions. This paper also derives the relationship between natural intelligence and nature. According to communication principle of information theory, Natural Intelligence can be divided into real part and virtual part. Based on information theory principle that Information does not increase, the restriction mechanism of Natural Intelligence creativity is conducted. The restriction mechanism of creativity reveals the limit of natural intelligence and artificial intelligence. The paper provides a new angle to observe natural intelligence and artificial intelligence.

Keywords: Natural intelligence, artificial intelligence, creativity, information theory.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1910
2200 Flexure of Simply Supported Thick Beams Using Refined Shear Deformation Theory

Authors: Yuwaraj M. Ghugal, Ajay G. Dahake

Abstract:

A trigonometric shear deformation theory for flexure of thick beams, taking into account transverse shear deformation effects, is developed. The number of variables in the present theory is same as that in the first order shear deformation theory. The sinusoidal function is used in displacement field in terms of thickness coordinate to represent the shear deformation effects. The noteworthy feature of this theory is that the transverse shear stresses can be obtained directly from the use of constitutive relations with excellent accuracy, satisfying the shear stress free conditions on the top and bottom surfaces of the beam. Hence, the theory obviates the need of shear correction factor. Governing differential equations and boundary conditions are obtained by using the principle of virtual work. The thick simply supported isotropic beams are considered for the numerical studies to demonstrate the efficiency of the results obtained is discussed critically with those of other theories.

Keywords: Trigonometric shear deformation, thick beam, flexure, principle of virtual work, equilibrium equations, stress.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1452
2199 Fast and Efficient Algorithms for Evaluating Uniform and Nonuniform Lagrange and Newton Curves

Authors: Taweechai Nuntawisuttiwong, Natasha Dejdumrong

Abstract:

Newton-Lagrange Interpolations are widely used in numerical analysis. However, it requires a quadratic computational time for their constructions. In computer aided geometric design (CAGD), there are some polynomial curves: Wang-Ball, DP and Dejdumrong curves, which have linear time complexity algorithms. Thus, the computational time for Newton-Lagrange Interpolations can be reduced by applying the algorithms of Wang-Ball, DP and Dejdumrong curves. In order to use Wang-Ball, DP and Dejdumrong algorithms, first, it is necessary to convert Newton-Lagrange polynomials into Wang-Ball, DP or Dejdumrong polynomials. In this work, the algorithms for converting from both uniform and non-uniform Newton-Lagrange polynomials into Wang-Ball, DP and Dejdumrong polynomials are investigated. Thus, the computational time for representing Newton-Lagrange polynomials can be reduced into linear complexity. In addition, the other utilizations of using CAGD curves to modify the Newton-Lagrange curves can be taken.

Keywords: Newton interpolation, Lagrange interpolation, linear complexity.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 570
2198 A Physical Theory of Information vs. a Mathematical Theory of Communication

Authors: Manouchehr Amiri

Abstract:

This article presents a general notion of physical bit information that is compatible with the basics of quantum mechanics and incorporates the Shannon entropy as a special case. This notion of physical information leads to the Binary Data Matrix model (BDM), which predicts the basic results of quantum mechanics, general relativity, and black hole thermodynamics. The compatibility of the model with holographic, information conservation, and Landauer’s principle is investigated. After deriving the “Bit Information principle” as a consequence of BDM, the fundamental equations of Planck, De Broglie, Bekenstein, and mass-energy equivalence are derived.

Keywords: Physical theory of information, binary data matrix model, Shannon information theory, bit information principle.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 60
2197 Personalized Learning: An Analysis Using Item Response Theory

Authors: A. Yacob, N. Hj. Ali, M. H. Yusoff, M. Y. MohdSaman, W. M. A. F. W. Hamzah

Abstract:

Personalized learning becomes increasingly popular which not be restricted by time, place or any other barriers. This study proposes an analysis of Personalized Learning using Item Response Theory which considers course material difficulty and learner ability.The study investigates twenty undergraduate students at TATI University College, who are taking programming subject. By using the IRT,it was found that, finding the most appropriate problem levels to each student include high and low level test items together is not a problem. Thus, the student abilities can be asses more accurately and fairly. Learners who experience more anxiety will affect a heavier cognitive load and receive lower test scores.Instructors are encouraged to provide a supportive learning environment to enhance learning effectiveness because Cognitive Load Theory concerns the limited capacity of the brain to absorb new information.

Keywords: Analysis, Cognitive Load Theory, Item Response Theory, Learning, Motivation, Performance.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3038
2196 Understanding Physical Activity Behavior of Type 2 Diabetics Using the Theory of Planned Behavior and Structural Equation Modeling

Authors: D. O. Omondi, M. K. Walingo, G. M. Mbagaya, L. O. A. Othuon

Abstract:

Understanding patient factors related to physical activity behavior is important in the management of Type 2 Diabetes. This study applied the Theory of Planned Behavior model to understand physical activity behavior among sampled Type 2 diabetics in Kenya. The study was conducted within the diabetic clinic at Kisii Level 5 Hospital and adopted sequential mixed methods design beginning with qualitative phase and ending with quantitative phase. Qualitative data was analyzed using grounded theory analysis method. Structural equation modeling using maximum likelihood was used to analyze quantitative data. The common fit indices revealed that the theory of planned behavior fitted the data acceptably well among the Type 2 diabetes and within physical activity behavior {¤ç2 = 213, df = 84, n=230, p = .061, ¤ç2/df = 2.53; TLI = .97; CFI =.96; RMSEA (90CI) = .073(.029, .08)}. This theory proved to be useful in understanding physical activity behavior among Type 2 diabetics.

Keywords: Physical activity, Theory of Planned Behavior, Type2 diabetes, Kenya.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1945
2195 A Time-Reducible Approach to Compute Determinant |I-X|

Authors: Wang Xingbo

Abstract:

Computation of determinant in the form |I-X| is primary and fundamental because it can help to compute many other determinants. This article puts forward a time-reducible approach to compute determinant |I-X|. The approach is derived from the Newton’s identity and its time complexity is no more than that to compute the eigenvalues of the square matrix X. Mathematical deductions and numerical example are presented in detail for the approach. By comparison with classical approaches the new approach is proved to be superior to the classical ones and it can naturally reduce the computational time with the improvement of efficiency to compute eigenvalues of the square matrix.

Keywords: Algorithm, determinant, computation, eigenvalue, time complexity.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1114
2194 Towards a Simulation Model to Ensure the Availability of Machines in Maintenance Activities

Authors: Maryam Gallab, Hafida Bouloiz, Youness Chater, Mohamed Tkiouat

Abstract:

The aim of this paper is to present a model based on multi-agent systems in order to manage the maintenance activities and to ensure the reliability and availability of machines just with the required resources (operators, tools). The interest of the simulation is to solve the complexity of the system and to find results without cost or wasting time. An implementation of the model is carried out on the AnyLogic platform to display the defined performance indicators.

Keywords: Maintenance, complexity, simulation, multi-agent systems, AnyLogic platform.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1472
2193 Promoting Complex Systems Learning through the use of Computer Modeling

Authors: Kamel Hashem, David Mioduser

Abstract:

This paper describes part of a project about Learningby- Modeling (LbM). Studying complex systems is increasingly important in teaching and learning many science domains. Many features of complex systems make it difficult for students to develop deep understanding. Previous research indicates that involvement with modeling scientific phenomena and complex systems can play a powerful role in science learning. Some researchers argue with this view indicating that models and modeling do not contribute to understanding complexity concepts, since these increases the cognitive load on students. This study will investigate the effect of different modes of involvement in exploring scientific phenomena using computer simulation tools, on students- mental model from the perspective of structure, behavior and function. Quantitative and qualitative methods are used to report about 121 freshmen students that engaged in participatory simulations about complex phenomena, showing emergent, self-organized and decentralized patterns. Results show that LbM plays a major role in students' concept formation about complexity concepts.

Keywords: Complexity, Educational technology, Learning by modeling, Mental models

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1539
2192 A Multi-Signature Scheme based on Coding Theory

Authors: Mohammed Meziani, Pierre-Louis Cayrel

Abstract:

In this paper we propose two first non-generic constructions of multisignature scheme based on coding theory. The first system make use of the CFS signature scheme and is secure in random oracle while the second scheme is based on the KKS construction and is a few times. The security of our construction relies on a difficult problems in coding theory: The Syndrome Decoding problem which has been proved NP-complete [4].

Keywords: Post-quantum cryptography, Coding-based cryptography, Digital signature, Multisignature scheme.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1831
2191 Low Complexity Multi Mode Interleaver Core for WiMAX with Support for Convolutional Interleaving

Authors: Rizwan Asghar, Dake Liu

Abstract:

A hardware efficient, multi mode, re-configurable architecture of interleaver/de-interleaver for multiple standards, like DVB, WiMAX and WLAN is presented. The interleavers consume a large part of silicon area when implemented by using conventional methods as they use memories to store permutation patterns. In addition, different types of interleavers in different standards cannot share the hardware due to different construction methodologies. The novelty of the work presented in this paper is threefold: 1) Mapping of vital types of interleavers including convolutional interleaver onto a single architecture with flexibility to change interleaver size; 2) Hardware complexity for channel interleaving in WiMAX is reduced by using 2-D realization of the interleaver functions; and 3) Silicon cost overheads reduced by avoiding the use of small memories. The proposed architecture consumes 0.18mm2 silicon area for 0.12μm process and can operate at a frequency of 140 MHz. The reduced complexity helps in minimizing the memory utilization, and at the same time provides strong support to on-the-fly computation of permutation patterns.

Keywords: Hardware interleaver implementation, WiMAX, DVB, block interleaver, convolutional interleaver, hardwaremultiplexing.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1995
2190 Conformal Invariance in F (R, T) Gravity

Authors: Pyotr Tsyba, Olga Razina, Ertan Güdekli, Ratbay Myrzakulov

Abstract:

In this paper we consider the equation of motion for the F (R, T) gravity on their property of conformal invariance. It is shown that in the general case, such a theory is not conformal invariant. Studied special cases for the functions v and u in which can appear properties of the theory. Also we consider cosmological aspects F (R, T) theory of gravity, having considered particular case F (R, T) = μR+νT^2. Showed that in this case there is a nonlinear dependence of the parameter equation of state from time to time, which affects its evolution.

Keywords: Conformally invariance, F (R, T) gravity, metric FRW, equation of motion, dark energy.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2598
2189 Optimal Solution of Constraint Satisfaction Problems

Authors: Jeffrey L. Duffany

Abstract:

An optimal solution for a large number of constraint satisfaction problems can be found using the technique of substitution and elimination of variables analogous to the technique that is used to solve systems of equations. A decision function f(A)=max(A2) is used to determine which variables to eliminate. The algorithm can be expressed in six lines and is remarkable in both its simplicity and its ability to find an optimal solution. However it is inefficient in that it needs to square the updated A matrix after each variable elimination. To overcome this inefficiency the algorithm is analyzed and it is shown that the A matrix only needs to be squared once at the first step of the algorithm and then incrementally updated for subsequent steps, resulting in significant improvement and an algorithm complexity of O(n3).

Keywords: Algorithm, complexity, constraint, np-complete.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1393
2188 An Efficient Multi Join Algorithm Utilizing a Lattice of Double Indices

Authors: Hanan A. M. Abd Alla, Lilac A. E. Al-Safadi

Abstract:

In this paper, a novel multi join algorithm to join multiple relations will be introduced. The novel algorithm is based on a hashed-based join algorithm of two relations to produce a double index. This is done by scanning the two relations once. But instead of moving the records into buckets, a double index will be built. This will eliminate the collision that can happen from a complete hash algorithm. The double index will be divided into join buckets of similar categories from the two relations. The algorithm then joins buckets with similar keys to produce joined buckets. This will lead at the end to a complete join index of the two relations. without actually joining the actual relations. The time complexity required to build the join index of two categories is Om log m where m is the size of each category. Totaling time complexity to O n log m for all buckets. The join index will be used to materialize the joined relation if required. Otherwise, it will be used along with other join indices of other relations to build a lattice to be used in multi-join operations with minimal I/O requirements. The lattice of the join indices can be fitted into the main memory to reduce time complexity of the multi join algorithm.

Keywords: Multi join, Relation, Lattice, Join indices.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1261
2187 Estimation of Component Reusability through Reusability Metrics

Authors: Aditya Pratap Singh, Pradeep Tomar

Abstract:

Software reusability is an essential characteristic of Component-Based Software (CBS). The component reusability is an important assess for the effective reuse of components in CBS. The attributes of reusability proposed by various researchers are studied and four of them are identified as potential factors affecting reusability. This paper proposes metric for reusability estimation of black-box software component along with metrics for Interface Complexity, Understandability, Customizability and Reliability. An experiment is performed for estimation of reusability through a case study on a sample web application using a real world component.

Keywords: Component-based software, component reusability, customizability, interface complexity, reliability, understandability.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3022
2186 The Comparative Analysis of Micro-reading and Traditional Reading Based On Schema Theory

Authors: Haiyan Wang

Abstract:

Micro-reading is a new way of reading depended on short messages of mobile phones, network articles and short literary forms, which impacts greatly on traditional way of reading. The effect of "micro-reading" is deeper especially for those growing middle school students and college students. Aiming at the problem with the development of college students' micro-reading and based on the influence of schema theory on the research of cognition of reading, this paper is to analyze the comparison between micro-reading and traditional reading and explore reading strategies in micro-era based on the negative and positive effect which schema theory has on micro-reading.

Keywords: Schema theory, comparative analysis, micro-reading, traditional reading

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3664
2185 A Vortex Plate Theory of Hovering Animal Flight

Authors: Khaled. M. Faqih

Abstract:

A model of vortex wake is suggested to determine the induced power during animal hovering flight. The wake is modeled by a series of equi-spaced rigid rectangular vortex plates, positioned horizontally and moving vertically downwards with identical speeds; each plate is generated during powering of the functionally wing stroke. The vortex representation of the wake considered in the current theory allows a considerable loss of momentum to occur. The current approach accords well with the nature of the wingbeat since it considers the unsteadiness in the wake as an important fluid dynamical characteristic. Induced power in hovering is calculated as the aerodynamic power required to generate the vortex wake system. Specific mean induced power to mean wing tip velocity ratio is determined by solely the normal spacing parameter (f) for a given wing stroke amplitude. The current theory gives much higher specific induced power estimate than anticipated by classical methods.

Keywords: vortex theory, hovering flight, induced power, Prandlt's tip theory.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1701
2184 Peridynamic Modeling of an Isotropic Plate under Tensile and Flexural Loading

Authors: Eda Gök

Abstract:

Peridynamics is a new modeling concept of non-local interactions for solid structures. The formulations of Peridynamic (PD) theory are based on integral equations rather than differential equations. Through, undefined equations of associated problems are avoided. PD theory might be defined as continuum version of molecular dynamics. The medium is usually modeled with mass particles bonded together. Particles interact with each other directly across finite distances through central forces named as bonds. The main assumption of this theory is that the body is composed of material points which interact with other material points within a finite distance. Although, PD theory developed for discontinuities, it gives good results for structures which have no discontinuities. In this paper, displacement control of the isotropic plate under the effect of tensile and bending loading has been investigated by means of PD theory. A MATLAB code is generated to create PD bonds and corresponding surface correction factors. Using generated MATLAB code the geometry of the specimen is generated, and the code is implemented in Finite Element Software. The results obtained from non-local continuum theory are compared with the Finite Element Analysis results and analytical solution. The results show good agreement.

Keywords: Flexural loading, non-local continuum mechanics, Peridynamic theory, solid structures, tensile loading.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1136