Search results for: modified cyclomatic complexity
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 1728

Search results for: modified cyclomatic complexity

1608 Theoretical Considerations for Software Component Metrics

Authors: V. Lakshmi Narasimhan, Bayu Hendradjaya

Abstract:

We have defined two suites of metrics, which cover static and dynamic aspects of component assembly. The static metrics measure complexity and criticality of component assembly, wherein complexity is measured using Component Packing Density and Component Interaction Density metrics. Further, four criticality conditions namely, Link, Bridge, Inheritance and Size criticalities have been identified and quantified. The complexity and criticality metrics are combined to form a Triangular Metric, which can be used to classify the type and nature of applications. Dynamic metrics are collected during the runtime of a complete application. Dynamic metrics are useful to identify super-component and to evaluate the degree of utilisation of various components. In this paper both static and dynamic metrics are evaluated using Weyuker-s set of properties. The result shows that the metrics provide a valid means to measure issues in component assembly. We relate our metrics suite with McCall-s Quality Model and illustrate their impact on product quality and to the management of component-based product development.

Keywords: Component Assembly, Component Based SoftwareEngineering, CORBA Component Model, Software ComponentMetrics.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2238
1607 Climate Change and Environmental Education: The Application of Concept Map for Representing the Knowledge Complexity of Climate Change

Authors: Hsueh-Chih, Chen, Yau-Ting, Sung, Tsai-Wen, Lin, Hung-Teng, Chou

Abstract:

It has formed an essential issue that Climate Change, composed of highly knowledge complexity, reveals its significant impact on human existence. Therefore, specific national policies, some of which present the educational aspects, have been published for overcoming the imperative problem. Accordingly, the study aims to analyze as well as integrate the relationship between Climate Change and environmental education and apply the perspective of concept map to represent the knowledge contents and structures of Climate Change; by doing so, knowledge contents of Climate Change could be represented in an even more comprehensive way and manipulated as the tool for environmental education. The method adapted for this study is knowledge conversion model compounded of the platform for experts and teachers, who were the participants for this study, to cooperate and combine each participant-s standpoints into a complete knowledge framework that is the foundation for structuring the concept map. The result of this research contains the important concepts, the precise propositions and the entire concept map for representing the robust concepts of Climate Change.

Keywords: Climate Change, knowledge complexity, concept map.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1679
1606 A Computational Cost-Effective Clustering Algorithm in Multidimensional Space Using the Manhattan Metric: Application to the Global Terrorism Database

Authors: Semeh Ben Salem, Sami Naouali, Moetez Sallami

Abstract:

The increasing amount of collected data has limited the performance of the current analyzing algorithms. Thus, developing new cost-effective algorithms in terms of complexity, scalability, and accuracy raised significant interests. In this paper, a modified effective k-means based algorithm is developed and experimented. The new algorithm aims to reduce the computational load without significantly affecting the quality of the clusterings. The algorithm uses the City Block distance and a new stop criterion to guarantee the convergence. Conducted experiments on a real data set show its high performance when compared with the original k-means version.

Keywords: Pattern recognition, partitional clustering, K-means clustering, Manhattan distance, terrorism data analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1317
1605 Efficient Secured Lossless Coding of Medical Images– Using Modified Runlength Coding for Character Representation

Authors: S. Annadurai, P. Geetha

Abstract:

Lossless compression schemes with secure transmission play a key role in telemedicine applications that helps in accurate diagnosis and research. Traditional cryptographic algorithms for data security are not fast enough to process vast amount of data. Hence a novel Secured lossless compression approach proposed in this paper is based on reversible integer wavelet transform, EZW algorithm, new modified runlength coding for character representation and selective bit scrambling. The use of the lifting scheme allows generating truly lossless integer-to-integer wavelet transforms. Images are compressed/decompressed by well-known EZW algorithm. The proposed modified runlength coding greatly improves the compression performance and also increases the security level. This work employs scrambling method which is fast, simple to implement and it provides security. Lossless compression ratios and distortion performance of this proposed method are found to be better than other lossless techniques.

Keywords: EZW algorithm, lifting scheme, losslesscompression, reversible integer wavelet transform, securetransmission, selective bit scrambling, modified runlength coding .

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1332
1604 Modified Diffie-Hellman Protocol By Extend The Theory of The Congruence

Authors: Rand Alfaris, Mohamed Rushdan MD Said, Mohamed Othman, Fudziah Ismail

Abstract:

This paper is introduced a modification to Diffie- Hellman protocol to be applicable on the decimal numbers, which they are the numbers between zero and one. For this purpose we extend the theory of the congruence. The new congruence is over the set of the real numbers and it is called the “real congruence" or the “real modulus". We will refer to the existing congruence by the “integer congruence" or the “integer modulus". This extension will define new terms and redefine the existing terms. As the properties and the theorems of the integer modulus are extended as well. Modified Diffie-Hellman key exchange protocol is produced a sharing, secure and decimal secret key for the the cryptosystems that depend on decimal numbers.

Keywords: Extended theory of the congruence, modified Diffie- Hellman protocol.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1528
1603 Titanium Dioxide Modified with Glutathione as Potential Drug Carrier with Reduced Toxic Properties

Authors: Olga Długosz, Jolanta Pulit-Prociak, Marcin Banach

Abstract:

The paper presents a process to obtain glutathione-modified titanium oxide nanoparticles. The processes were carried out in a microwave radiation field. The influence of the molar ratio of glutathione to titanium oxide and the effect of the fold of NaOH vs. stoichiometric amount on the size of the formed TiO2 nanoparticles was determined. The physicochemical properties of the obtained products were evaluated using dynamic light scattering (DLS), transmission electron microscope- energy-dispersive X-ray spectroscopy (TEM-EDS), low-temperature nitrogen adsorption method (BET), X-Ray Diffraction (XRD) and Fourier-transform infrared spectroscopy (FTIR) microscopy methods. The size of TiO2 nanoparticles was characterized from 30 nm to 336 nm. The release of titanium ions from the prepared products was evaluated. These studies were carried out using different media in which the powders were incubated for a specific time. These were: water, SBF and Ringer's solution. The release of titanium ions from modified products is weaker compared to unmodified titanium oxide nanoparticles. The reduced release of titanium ions may allow the use of such modified materials as substances in drug delivery systems.

Keywords: titanium dioxide, nanoparticles, drug carrier, glutathione

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 487
1602 The Two Layers of Food Safety and GMOs in the Hungarian Agricultural Law

Authors: Gergely Horváth

Abstract:

The study presents the complexity of food safety dividing it into two layers. Beyond the basic layer of requirements, there is a more demanding higher level linked with quality and purity aspects. It would be important to give special prominence to both layers, given that massive illnesses are caused by foods even though officially licensed. Then the study discusses an exciting safety challenge stemming from the risks of genetically modified organisms (GMOs). Furthermore, it features legal case examples that illustrate how certain liability questions are solved or not yet decided in connection with the production of genetically modified crops. In addition, a special kind of land grabbing, more precisely land grabbing from non-GMO farming systems can also be noticed as well as a new phenomenon eroding food sovereignty. Coexistence, the state where organic, conventional, and GM farming systems are standing alongside each other is an unsuitable experiment that cannot be successful, because of biophysical reasons (such as cross-pollination). Agricultural and environmental lawyers both try to find the optimal solution. Agri-environmental measures are introduced as a special subfield of law maintaining also food safety. The important steps of agri-environmental legislation are aiming at the protection of natural values, the environmental media and strengthening food safety as well, practically the quality of agricultural products intended for human consumption. The major findings of the study focus on searching for the appropriate approach capable of solving the security and safety problems of food production. The most interesting concepts of the Hungarian national and EU food law legislation are analyzed in more detail with descriptive, analytic and comparative methods.

Keywords: Food law, food safety, food security, GMO, agri-environmental measures.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1164
1601 Optimization of Soybean Oil by Modified Supercritical Carbon Dioxide

Authors: N. R. Putra, A. H. Abdul Aziz, A. S. Zaini, Z. Idham, F. Idrus, M. Z. Bin Zullyadini, M. A. Che Yunus

Abstract:

The content of omega-3 in soybean oil is important in the development of infants and is an alternative for the omega-3 in fish oils. The investigation of extraction of soybean oil is needed to obtain the bioactive compound in the extract. Supercritical carbon dioxide extraction is modern and green technology to extract herbs and plants to obtain high quality extract due to high diffusivity and solubility of the solvent. The aim of this study was to obtain the optimum condition of soybean oil extraction by modified supercritical carbon dioxide. The soybean oil was extracted by using modified supercritical carbon dioxide (SC-CO2) under the temperatures of 40, 60, 80 °C, pressures of 150, 250, 350 Bar, and constant flow-rate of 10 g/min as the parameters of extraction processes. An experimental design was performed in order to optimize three important parameters of SC-CO2 extraction which are pressure (X1), temperature (X2) to achieve optimum yields of soybean oil. Box Behnken Design was applied for experimental design. From the optimization process, the optimum condition of extraction of soybean oil was obtained at pressure 338 Bar and temperature 80 °C with oil yield of 2.713 g. Effect of pressure is significant on the extraction of soybean oil by modified supercritical carbon dioxide. Increasing of pressure will increase the oil yield of soybean oil.

Keywords: Soybean oil, SC-CO2 extraction, yield, optimization.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 888
1600 Environmental Competency Framework: Development of a Modified 2-Tuple Delphi Approach

Authors: M. Bouri, L. Chraïbi, N. Sefiani

Abstract:

Currently, industries endeavor to align their environmental management system with the ISO 14001:2015 international standard, while preserving competitiveness and sustainability. Then, a key driver for these industries is to develop a skilled workforce that is able to implement, continuously improve and audit the environmental management system. The purpose of this paper is to provide an environmental competency framework that aims to identify, rank and categorize the competencies required by both the environmental managers and auditors. This competency framework is expected to be useful during competency assessment, recruitment, and training processes. To achieve this end, a modified 2-tuple Delphi approach is here proposed based on a combination of the modified Delphi approach and the 2-tuple linguistic representation model. The adopted approach is presented as numerous questionnaires that are spread over multiple rounds in order to obtain a consensus among the different Moroccan experts participating to this study.

Keywords: Competency framework, Delphi, environmental competency, 2-tuple.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 337
1599 Efficient Moment Frame Structure

Authors: Mircea I. Pastrav, Cornelia Baera, Florea Dinu

Abstract:

A different concept for designing and detailing of reinforced concrete precast frame structures is analyzed in this paper. The new detailing of the joints derives from the special hybrid moment frame joints. The special reinforcements of this alternative detailing, named modified special hybrid joint, are bondless with respect to both column and beams. Full scale tests were performed on a plan model, which represents a part of 5 story structure, cropped in the middle of the beams and columns spans. Theoretical approach was developed, based on testing results on twice repaired model, subjected to lateral seismic type loading. Discussion regarding the modified special hybrid joint behavior and further on widening research needed concludes the presentation.

Keywords: Acceptance criteria, modified hybrid joint, repair, seismic loading type.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1970
1598 Second-order Time Evolution Scheme for Time-dependent Neutron Transport Equation

Authors: Zhenying Hong, Guangwei Yuan, Xuedong Fu, Shulin Yang

Abstract:

In this paper, the typical exponential method, diamond difference and modified time discrete scheme is researched for self adaptive time step. The second-order time evolution scheme is applied to time-dependent spherical neutron transport equation by discrete ordinates method. The numerical results show that second-order time evolution scheme associated exponential method has some good properties. The time differential curve about neutron current is more smooth than that of exponential method and diamond difference and modified time discrete scheme.

Keywords: Exponential method, diamond difference, modified time discrete scheme, second-order time evolution scheme.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1543
1597 Staling and Quality of Iranian Flat Bread Stored at Modified Atmosphere in Different Packaging

Authors: A. Hematian Sourki, F. Tabatabaei Yazdi, M. Ghiafeh Davoodi, S.A. Mortazavi, M. Karimi, S.H. Razavizadegan Jahromi, A. Pourfarzad

Abstract:

This study investigated the use of modified atmosphere packaging (MAP) and different packaging to extend the shelf life of Barbari flat bread. Three atmospheres including 70%CO2 and 30%N2, 50% CO2 and 50%N2 and a normal air as control were used. The bread samples were packaged in three type pouches. The shelf life was determined by appearance of mold and yeast (M +Y) in Barbari bread samples stored at 25 ± 1°C and 38 ± 2% relative humidity. The results showed that it is possible to prolong the shelf life of Barbari bread from four days to about 21 days by using modified atmosphere packaging with high carbon dioxide concentration and high-barrier laminated and vacuum bags packages. However, the hardness of samples kept in MAP increase significantly by increase of carbon dioxide concentration. The correlation coefficient (r) between headspace CO2 concentration and hardness was 0.997, 0.997 and 0.599 for A, B and C packaging respectively. High negative correlation coefficients were found between the crumb moisture and the hardness values in various packaging. There were significant negative correlation coefficients between sensory parameters and hardness of texture.

Keywords: modified atmosphere packaging, flat bread, Iranian bread, staling, correlation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2680
1596 Analyzing the Performance Properties of Stress Absorbing Membrane Interlayer Modified with Recycled Crumb Rubber

Authors: Seyed Mohammad Asgharzadeh, Moein Biglari

Abstract:

Asphalt overlay is the most commonly used technique of pavement rehabilitation. However, the reflective cracks which occur on the overlay surface after a short period of time are the most important distresses threatening the durability of new overlays. Stress Absorbing Membrane Interlayers (SAMIs) are used to postpone the reflective cracking in the overlays. Sand asphalt mixtures, in unmodified or crumb rubber modified (CRM) conditions, can be used as an SAMI material. In this research, the performance properties of different SAMI applications were evaluated in the laboratory using an Indirect Tensile (IDT) fracture energy. The IDT fracture energy of sand asphalt samples was also evaluated and then compared to that of the regular dense graded asphalt used as an overlay. Texas boiling water and modified Lottman tests were also conducted to evaluate the moisture susceptibility of sand asphalt mixtures. The test results showed that sand asphalt mixtures can stand higher levels of energy before cracking, and this is even more pronounced for the CRM sand mix. Sand asphalt mixture using CRM binder was also shown to be more resistance to moisture induced distresses.

Keywords: SAMI, sand asphalt, crumb rubber, Lottman Modified Test.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1235
1595 Fast and Efficient Algorithms for Evaluating Uniform and Nonuniform Lagrange and Newton Curves

Authors: Taweechai Nuntawisuttiwong, Natasha Dejdumrong

Abstract:

Newton-Lagrange Interpolations are widely used in numerical analysis. However, it requires a quadratic computational time for their constructions. In computer aided geometric design (CAGD), there are some polynomial curves: Wang-Ball, DP and Dejdumrong curves, which have linear time complexity algorithms. Thus, the computational time for Newton-Lagrange Interpolations can be reduced by applying the algorithms of Wang-Ball, DP and Dejdumrong curves. In order to use Wang-Ball, DP and Dejdumrong algorithms, first, it is necessary to convert Newton-Lagrange polynomials into Wang-Ball, DP or Dejdumrong polynomials. In this work, the algorithms for converting from both uniform and non-uniform Newton-Lagrange polynomials into Wang-Ball, DP and Dejdumrong polynomials are investigated. Thus, the computational time for representing Newton-Lagrange polynomials can be reduced into linear complexity. In addition, the other utilizations of using CAGD curves to modify the Newton-Lagrange curves can be taken.

Keywords: Newton interpolation, Lagrange interpolation, linear complexity.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 570
1594 A Novel Hopfield Neural Network for Perfect Calculation of Magnetic Resonance Spectroscopy

Authors: Hazem M. El-Bakry

Abstract:

In this paper, an automatic determination algorithm for nuclear magnetic resonance (NMR) spectra of the metabolites in the living body by magnetic resonance spectroscopy (MRS) without human intervention or complicated calculations is presented. In such method, the problem of NMR spectrum determination is transformed into the determination of the parameters of a mathematical model of the NMR signal. To calculate these parameters efficiently, a new model called modified Hopfield neural network is designed. The main achievement of this paper over the work in literature [30] is that the speed of the modified Hopfield neural network is accelerated. This is done by applying cross correlation in the frequency domain between the input values and the input weights. The modified Hopfield neural network can accomplish complex dignals perfectly with out any additinal computation steps. This is a valuable advantage as NMR signals are complex-valued. In addition, a technique called “modified sequential extension of section (MSES)" that takes into account the damping rate of the NMR signal is developed to be faster than that presented in [30]. Simulation results show that the calculation precision of the spectrum improves when MSES is used along with the neural network. Furthermore, MSES is found to reduce the local minimum problem in Hopfield neural networks. Moreover, the performance of the proposed method is evaluated and there is no effect on the performance of calculations when using the modified Hopfield neural networks.

Keywords: Hopfield Neural Networks, Cross Correlation, Nuclear Magnetic Resonance, Magnetic Resonance Spectroscopy, Fast Fourier Transform.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1802
1593 Computer Proven Correctness of the Rabin Public-Key Scheme

Authors: Johannes Buchmann, Markus Kaiser

Abstract:

We decribe a formal specification and verification of the Rabin public-key scheme in the formal proof system Is-abelle/HOL. The idea is to use the two views of cryptographic verification: the computational approach relying on the vocabulary of probability theory and complexity theory and the formal approach based on ideas and techniques from logic and programming languages. The analysis presented uses a given database to prove formal properties of our implemented functions with computer support. Thema in task in designing a practical formalization of correctness as well as security properties is to cope with the complexity of cryptographic proving. We reduce this complexity by exploring a light-weight formalization that enables both appropriate formal definitions as well as eficient formal proofs. This yields the first computer-proved implementation of the Rabin public-key scheme in Isabelle/HOL. Consequently, we get reliable proofs with a minimal error rate augmenting the used database. This provides a formal basis for more computer proof constructions in this area.

Keywords: public-key encryption, Rabin public-key scheme, formalproof system, higher-order logic, formal verification.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1557
1592 Radiation Stability of Pigment ZnO Modified by Nanopowder

Authors: Chundong Li, V. V. Neshchimenko, M. M. Mikhailov

Abstract:

The effect of the modification of ZnO powders by ZrO2, Al2O3, TiO2, SiO2, CeO2 and Y2O3 nanoparticles with a concentration of 1-30 wt % is investigated by diffuse reflectance spectra within the wavelength range 200 to 2500 nm before and after 100 keV proton and electron irradiation. It has been established that the introduction of nanoparticles ZrO2, Al2O3 enhances the optical stability of the pigments under proton irradiation, but reduces it under electron irradiation. Modifying with TiO2, SiO2, CeO2, Y2O3 nanopowders leads to decrease radiation stability in both types of irradiation. Samples modified by 5 wt. % of ZrO2 nanoparticles have the highest stability of optical properties after proton exposure. The degradation of optical properties under electron irradiation is not high for this concentration of nanoparticles. A decrease in the absorption of pigments modified with nanoparticles proton exposure is determined by a decrease in the intensity of bands located in the UV and visible regions. After electron exposure the absorption bands have in the whole spectrum range.

Keywords: Irradiation, nanopowders, radiation stability, zinc oxide.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2173
1591 Visual Analytics in K 12 Education - Emerging Dimensions of Complexity

Authors: Linnea Stenliden

Abstract:

The aim of this paper is to understand emerging learning conditions, when a visual analytics is implemented and used in K 12 (education). To date, little attention has been paid to the role visual analytics (digital media and technology that highlight visual data communication in order to support analytical tasks) can play in education, and to the extent to which these tools can process actionable data for young students. This study was conducted in three public K 12 schools, in four social science classes with students aged 10 to 13 years, over a period of two to four weeks at each school. Empirical data were generated using video observations and analyzed with help of metaphors within Actor-network theory (ANT). The learning conditions are found to be distinguished by broad complexity, characterized by four dimensions. These emerge from the actors’ deeply intertwined relations in the activities. The paper argues in relation to the found dimensions that novel approaches to teaching and learning could benefit students’ knowledge building as they work with visual analytics, analyzing visualized data.

Keywords: Analytical reasoning, complexity, data use, problem space, visual analytics, visual storytelling, translation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1660
1590 A Novel Compression Algorithm for Electrocardiogram Signals based on Wavelet Transform and SPIHT

Authors: Sana Ktata, Kaïs Ouni, Noureddine Ellouze

Abstract:

Electrocardiogram (ECG) data compression algorithm is needed that will reduce the amount of data to be transmitted, stored and analyzed, but without losing the clinical information content. A wavelet ECG data codec based on the Set Partitioning In Hierarchical Trees (SPIHT) compression algorithm is proposed in this paper. The SPIHT algorithm has achieved notable success in still image coding. We modified the algorithm for the one-dimensional (1-D) case and applied it to compression of ECG data. By this compression method, small percent root mean square difference (PRD) and high compression ratio with low implementation complexity are achieved. Experiments on selected records from the MIT-BIH arrhythmia database revealed that the proposed codec is significantly more efficient in compression and in computation than previously proposed ECG compression schemes. Compression ratios of up to 48:1 for ECG signals lead to acceptable results for visual inspection.

Keywords: Discrete Wavelet Transform, ECG compression, SPIHT.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2086
1589 A Time-Reducible Approach to Compute Determinant |I-X|

Authors: Wang Xingbo

Abstract:

Computation of determinant in the form |I-X| is primary and fundamental because it can help to compute many other determinants. This article puts forward a time-reducible approach to compute determinant |I-X|. The approach is derived from the Newton’s identity and its time complexity is no more than that to compute the eigenvalues of the square matrix X. Mathematical deductions and numerical example are presented in detail for the approach. By comparison with classical approaches the new approach is proved to be superior to the classical ones and it can naturally reduce the computational time with the improvement of efficiency to compute eigenvalues of the square matrix.

Keywords: Algorithm, determinant, computation, eigenvalue, time complexity.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1113
1588 Towards a Simulation Model to Ensure the Availability of Machines in Maintenance Activities

Authors: Maryam Gallab, Hafida Bouloiz, Youness Chater, Mohamed Tkiouat

Abstract:

The aim of this paper is to present a model based on multi-agent systems in order to manage the maintenance activities and to ensure the reliability and availability of machines just with the required resources (operators, tools). The interest of the simulation is to solve the complexity of the system and to find results without cost or wasting time. An implementation of the model is carried out on the AnyLogic platform to display the defined performance indicators.

Keywords: Maintenance, complexity, simulation, multi-agent systems, AnyLogic platform.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1471
1587 Promoting Complex Systems Learning through the use of Computer Modeling

Authors: Kamel Hashem, David Mioduser

Abstract:

This paper describes part of a project about Learningby- Modeling (LbM). Studying complex systems is increasingly important in teaching and learning many science domains. Many features of complex systems make it difficult for students to develop deep understanding. Previous research indicates that involvement with modeling scientific phenomena and complex systems can play a powerful role in science learning. Some researchers argue with this view indicating that models and modeling do not contribute to understanding complexity concepts, since these increases the cognitive load on students. This study will investigate the effect of different modes of involvement in exploring scientific phenomena using computer simulation tools, on students- mental model from the perspective of structure, behavior and function. Quantitative and qualitative methods are used to report about 121 freshmen students that engaged in participatory simulations about complex phenomena, showing emergent, self-organized and decentralized patterns. Results show that LbM plays a major role in students' concept formation about complexity concepts.

Keywords: Complexity, Educational technology, Learning by modeling, Mental models

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1539
1586 Evaluating Complexity – Ethical Challenges in Computational Design Processes

Authors: J.Partanen

Abstract:

Complexity, as a theoretical background has made it easier to understand and explain the features and dynamic behavior of various complex systems. As the common theoretical background has confirmed, borrowing the terminology for design from the natural sciences has helped to control and understand urban complexity. Phenomena like self-organization, evolution and adaptation are appropriate to describe the formerly inaccessible characteristics of the complex environment in unpredictable bottomup systems. Increased computing capacity has been a key element in capturing the chaotic nature of these systems. A paradigm shift in urban planning and architectural design has forced us to give up the illusion of total control in urban environment, and consequently to seek for novel methods for steering the development. New methods using dynamic modeling have offered a real option for more thorough understanding of complexity and urban processes. At best new approaches may renew the design processes so that we get a better grip on the complex world via more flexible processes, support urban environmental diversity and respond to our needs beyond basic welfare by liberating ourselves from the standardized minimalism. A complex system and its features are as such beyond human ethics. Self-organization or evolution is either good or bad. Their mechanisms are by nature devoid of reason. They are common in urban dynamics in both natural processes and gas. They are features of a complex system, and they cannot be prevented. Yet their dynamics can be studied and supported. The paradigm of complexity and new design approaches has been criticized for a lack of humanity and morality, but the ethical implications of scientific or computational design processes have not been much discussed. It is important to distinguish the (unexciting) ethics of the theory and tools from the ethics of computer aided processes based on ethical decisions. Urban planning and architecture cannot be based on the survival of the fittest; however, the natural dynamics of the system cannot be impeded on grounds of being “non-human". In this paper the ethical challenges of using the dynamic models are contemplated in light of a few examples of new architecture and dynamic urban models and literature. It is suggested that ethical challenges in computational design processes could be reframed under the concepts of responsibility and transparency.

Keywords: urban planning, architecture, dynamic modeling, ethics, complexity theory.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1853
1585 Use of Zeolite and Surfactant Modified Zeolite as Ion Exchangers to Control Nitrate Leaching

Authors: R. Malekian, J. Abedi-Koupai, S. S. Eslamian

Abstract:

Nitrogen loss from irrigated cropland, particularly sandy soils, significantly contributes to nitrate (NO3 -) levels in surface and groundwaters. Thus, it is of great interest to use inexpensive natural products that can increase the fertilizer efficiency and decrease nitrate leaching. In this study, the ability of natural Iranian zeolite clinoptilolite (Cp) and surfactant modified zeolite clinoptilolite (SMZ) to remove NH4 + and NO3 -, respectively, from aqueous solutions was determined. The feasibility of using Cp and SMZ as soil amendment to reduce nitrate leaching from soil using lysimeters was also investigated. Zeolite showed 10.23% to 88.42% NH4 + removal efficiency over a wide range of initial NH4 + concentrations. Nitrate removal efficiency by SMZ was 32.26% to 82.26%. Field study results showed that Cp and SMZ significantly (p < 0.05) reduced leachate NO3-N concentration compared to control. There was no significant difference between maximum and mean leachate NO3-N concentration of SMZ lysimeters and those of Cp lysimeters.

Keywords: Ammonium removal, Leaching, Nitrate removal, Surfactant modified zeolite

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2709
1584 Grid–SVC: An Improvement in SVC Algorithm, Based On Grid Based Clustering

Authors: Farhad Hadinejad, Hasan Saberi, Saeed Kazem

Abstract:

Support vector clustering (SVC) is an important kernelbased clustering algorithm in multi applications. It has got two main bottle necks, the high computation price and labeling piece. In this paper, we presented a modified SVC method, named Grid–SVC, to improve the original algorithm computationally. First we normalized and then we parted the interval, where the SVC is processing, using a novel Grid–based clustering algorithm. The algorithm parts the intervals, based on the density function of the data set and then applying the cartesian multiply makes multi-dimensional grids. Eliminating many outliers and noise in the preprocess, we apply an improved SVC method to each parted grid in a parallel way. The experimental results show both improvement in time complexity order and the accuracy.

Keywords: Grid–based clustering, SVC, Density function, Radial basis function.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1700
1583 Modified Poly(pyrrole) Film Based Biosensors for Phenol Detection

Authors: S. Korkut, M. S. Kilic, E. Erhan

Abstract:

In order to detect and quantify the phenolic contents of a wastewater with biosensors, two working electrodes based on modified Poly(Pyrrole) films were fabricated. Enzyme horseradish peroxidase was used as biomolecule of the prepared electrodes. Various phenolics were tested at the biosensor. Phenol detection was realized by electrochemical reduction of quinones produced by enzymatic activity. Analytical parameters were calculated and the results were compared with each other.

Keywords: Carbon nanotube, Phenol biosensor, Polypyrrole, Poly(glutaraldehyde).

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2145
1582 A Modified Run Length Coding Technique for Test Data Compression Based on Multi-Level Selective Huffman Coding

Authors: C. Kalamani, K. Paramasivam

Abstract:

Test data compression is an efficient method for reducing the test application cost. The problem of reducing test data has been addressed by researchers in three different aspects: Test Data Compression, Built-in-Self-Test (BIST) and Test set compaction. The latter two methods are capable of enhancing fault coverage with cost of hardware overhead. The drawback of the conventional methods is that they are capable of reducing the test storage and test power but when test data have redundant length of runs, no additional compression method is followed. This paper presents a modified Run Length Coding (RLC) technique with Multilevel Selective Huffman Coding (MLSHC) technique to reduce test data volume, test pattern delivery time and power dissipation in scan test applications where redundant length of runs is encountered then the preceding run symbol is replaced with tiny codeword. Experimental results show that the presented method not only improves the test data compression but also reduces the overall test data volume compared to recent schemes. Experiments for the six largest ISCAS-98 benchmarks show that our method outperforms most known techniques.

Keywords: Modified run length coding, multilevel selective Huffman coding, built-in-self-test modified selective Huffman coding, automatic test equipment.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1223
1581 Modified Fuzzy PID Control for Networked Control Systems with Random Delays

Authors: Yong-can Cao, Wei-dong Zhang

Abstract:

To deal with random delays in Networked Control System (NCS), Modified Fuzzy PID Controller is introduced in this paper to implement real-time control adaptively. Via adjusting the control signal dynamically, the system performance is improved. In this paper, the design process and the ultimate simulation results are represented. Finally, examples and corresponding comparisons prove the significance of this method.

Keywords: Fuzzy Control, Networked Control System, PID, Random Delays

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1510
1580 Low Complexity Multi Mode Interleaver Core for WiMAX with Support for Convolutional Interleaving

Authors: Rizwan Asghar, Dake Liu

Abstract:

A hardware efficient, multi mode, re-configurable architecture of interleaver/de-interleaver for multiple standards, like DVB, WiMAX and WLAN is presented. The interleavers consume a large part of silicon area when implemented by using conventional methods as they use memories to store permutation patterns. In addition, different types of interleavers in different standards cannot share the hardware due to different construction methodologies. The novelty of the work presented in this paper is threefold: 1) Mapping of vital types of interleavers including convolutional interleaver onto a single architecture with flexibility to change interleaver size; 2) Hardware complexity for channel interleaving in WiMAX is reduced by using 2-D realization of the interleaver functions; and 3) Silicon cost overheads reduced by avoiding the use of small memories. The proposed architecture consumes 0.18mm2 silicon area for 0.12μm process and can operate at a frequency of 140 MHz. The reduced complexity helps in minimizing the memory utilization, and at the same time provides strong support to on-the-fly computation of permutation patterns.

Keywords: Hardware interleaver implementation, WiMAX, DVB, block interleaver, convolutional interleaver, hardwaremultiplexing.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1993
1579 Robust Numerical Scheme for Pricing American Options under Jump Diffusion Models

Authors: Salah Alrabeei, Mohammad Yousuf

Abstract:

The goal of option pricing theory is to help the investors to manage their money, enhance returns and control their financial future by theoretically valuing their options. However, most of the option pricing models have no analytical solution. Furthermore, not all the numerical methods are efficient to solve these models because they have nonsmoothing payoffs or discontinuous derivatives at the exercise price. In this paper, we solve the American option under jump diffusion models by using efficient time-dependent numerical methods. several techniques are integrated to reduced the overcome the computational complexity. Fast Fourier Transform (FFT) algorithm is used as a matrix-vector multiplication solver, which reduces the complexity from O(M2) into O(M logM). Partial fraction decomposition technique is applied to rational approximation schemes to overcome the complexity of inverting polynomial of matrices. The proposed method is easy to implement on serial or parallel versions. Numerical results are presented to prove the accuracy and efficiency of the proposed method.

Keywords: Integral differential equations, American options, jump–diffusion model, rational approximation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 504