Search results for: normalized systems theory
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 13178

Search results for: normalized systems theory

13178 Normalized Enterprises Architectures: Portugal's Public Procurement System Application

Authors: Tiago Sampaio, André Vasconcelos, Bruno Fragoso

Abstract:

The Normalized Systems Theory, which is designed to be applied to software architectures, provides a set of theorems, elements and rules, with the purpose of enabling evolution in Information Systems, as well as ensuring that they are ready for change. In order to make that possible, this work’s solution is to apply the Normalized Systems Theory to the domain of enterprise architectures, using Archimate. This application is achieved through the adaptation of the elements of this theory, making them artifacts of the modeling language. The theorems are applied through the identification of the viewpoints to be used in the architectures, as well as the transformation of the theory’s encapsulation rules into architectural rules. This way, it is possible to create normalized enterprise architectures, thus fulfilling the needs and requirements of the business. This solution was demonstrated using the Portuguese Public Procurement System. The Portuguese government aims to make this system as fair as possible, allowing every organization to have the same business opportunities. The aim is for every economic operator to have access to all public tenders, which are published in any of the 6 existing platforms, independently of where they are registered. In order to make this possible, we applied our solution to the construction of two different architectures, which are able of fulfilling the requirements of the Portuguese government. One of those architectures, TO-BE A, has a Message Broker that performs the communication between the platforms. The other, TO-BE B, represents the scenario in which the platforms communicate with each other directly. Apart from these 2 architectures, we also represent the AS-IS architecture that demonstrates the current behavior of the Public Procurement Systems. Our evaluation is based on a comparison between the AS-IS and the TO-BE architectures, regarding the fulfillment of the rules and theorems of the Normalized Systems Theory and some quality metrics.

Keywords: archimate, architecture, broker, enterprise, evolvable systems, interoperability, normalized architectures, normalized systems, normalized systems theory, platforms

Procedia PDF Downloads 326
13177 Normalized Laplacian Eigenvalues of Graphs

Authors: Shaowei Sun

Abstract:

Let G be a graph with vertex set V(G)={v_1,v_2,...,v_n} and edge set E(G). For any vertex v belong to V(G), let d_v denote the degree of v. The normalized Laplacian matrix of the graph G is the matrix where the non-diagonal (i,j)-th entry is -1/(d_id_j) when vertex i is adjacent to vertex j and 0 when they are not adjacent, and the diagonal (i,i)-th entry is the di. In this paper, we discuss some bounds on the largest and the second smallest normalized Laplacian eigenvalue of trees and graphs. As following, we found some new bounds on the second smallest normalized Laplacian eigenvalue of tree T in terms of graph parameters. Moreover, we use Sage to give some conjectures on the second largest and the third smallest normalized eigenvalues of graph.

Keywords: graph, normalized Laplacian eigenvalues, normalized Laplacian matrix, tree

Procedia PDF Downloads 301
13176 Blueprinting of a Normalized Supply Chain Processes: Results in Implementing Normalized Software Systems

Authors: Bassam Istanbouli

Abstract:

With the technology evolving every day and with the increase in global competition, industries are always under the pressure to be the best. They need to provide good quality products at competitive prices, when and how the customer wants them.  In order to achieve this level of service, products and their respective supply chain processes need to be flexible and evolvable; otherwise changes will be extremely expensive, slow and with many combinatorial effects. Those combinatorial effects impact the whole organizational structure, from a management, financial, documentation, logistics and specially the information system Enterprise Requirement Planning (ERP) perspective. By applying the normalized system concept/theory to segments of the supply chain, we believe minimal effects, especially at the time of launching an organization global software project. The purpose of this paper is to point out that if an organization wants to develop a software from scratch or implement an existing ERP software for their business needs and if their business processes are normalized and modular then most probably this will yield to a normalized and modular software system that can be easily modified when the business evolves. Another important goal of this paper is to increase the awareness regarding the design of the business processes in a software implementation project. If the blueprints created are normalized then the software developers and configurators will use those modular blueprints to map them into modular software. This paper only prepares the ground for further studies;  the above concept will be supported by going through the steps of developing, configuring and/or implementing a software system for an organization by using two methods: The Software Development Lifecycle method (SDLC) and the Accelerated SAP implementation method (ASAP). Both methods start with the customer requirements, then blue printing of its business processes and finally mapping those processes into a software system.  Since those requirements and processes are the starting point of the implementation process, then normalizing those processes will end up in a normalizing software.

Keywords: blueprint, ERP, modular, normalized

Procedia PDF Downloads 104
13175 Normalized Compression Distance Based Scene Alteration Analysis of a Video

Authors: Lakshay Kharbanda, Aabhas Chauhan

Abstract:

In this paper, an application of Normalized Compression Distance (NCD) to detect notable scene alterations occurring in videos is presented. Several research groups have been developing methods to perform image classification using NCD, a computable approximation to Normalized Information Distance (NID) by studying the degree of similarity in images. The timeframes where significant aberrations between the frames of a video have occurred have been identified by obtaining a threshold NCD value, using two compressors: LZMA and BZIP2 and defining scene alterations using Pixel Difference Percentage metrics.

Keywords: image compression, Kolmogorov complexity, normalized compression distance, root mean square error

Procedia PDF Downloads 303
13174 Socio-Technical Systems: Transforming Theory into Practice

Authors: L. Ngowi, N. H. Mvungi

Abstract:

This paper critically examines the evolution of socio-technical systems theory, its practices, and challenges in system design and development. It examines concepts put forward by researchers focusing on the application of the theory in software engineering. There are various methods developed that use socio-technical concepts based on systems engineering without remarkable success. The main constraint is the large amount of data and inefficient techniques used in the application of the concepts in system engineering for developing time-bound systems and within a limited/controlled budget. This paper critically examines each of the methods, highlight bottlenecks and suggest the way forward. Since socio-technical systems theory only explains what to do, but not how doing it, hence engineers are not using the concept to save time, costs and reduce risks associated with new frameworks. Hence, a new framework, which can be considered as a practical approach is proposed that borrows concepts from soft systems method, agile systems development and object-oriented analysis and design to bridge the gap between theory and practice. The approach will enable the development of systems using socio-technical systems theory to attract/enable the system engineers/software developers to use socio-technical systems theory in building worthwhile information systems to avoid fragilities and hostilities in the work environment.

Keywords: socio-technical systems, human centered design, software engineering, cognitive engineering, soft systems, systems engineering

Procedia PDF Downloads 258
13173 Some New Bounds for a Real Power of the Normalized Laplacian Eigenvalues

Authors: Ayşe Dilek Maden

Abstract:

For a given a simple connected graph, we present some new bounds via a new approach for a special topological index given by the sum of the real number power of the non-zero normalized Laplacian eigenvalues. To use this approach presents an advantage not only to derive old and new bounds on this topic but also gives an idea how some previous results in similar area can be developed.

Keywords: degree Kirchhoff index, normalized Laplacian eigenvalue, spanning tree, simple connected graph

Procedia PDF Downloads 341
13172 Classification on Statistical Distributions of a Complex N-Body System

Authors: David C. Ni

Abstract:

Contemporary models for N-body systems are based on temporal, two-body, and mass point representation of Newtonian mechanics. Other mainstream models include 2D and 3D Ising models based on local neighborhood the lattice structures. In Quantum mechanics, the theories of collective modes are for superconductivity and for the long-range quantum entanglement. However, these models are still mainly for the specific phenomena with a set of designated parameters. We are therefore motivated to develop a new construction directly from the complex-variable N-body systems based on the extended Blaschke functions (EBF), which represent a non-temporal and nonlinear extension of Lorentz transformation on the complex plane – the normalized momentum spaces. A point on the complex plane represents a normalized state of particle momentums observed from a reference frame in the theory of special relativity. There are only two key parameters, normalized momentum and nonlinearity for modelling. An algorithm similar to Jenkins-Traub method is adopted for solving EBF iteratively. Through iteration, the solution sets show a form of σ + i [-t, t], where σ and t are the real numbers, and the [-t, t] shows various distributions, such as 1-peak, 2-peak, and 3-peak etc. distributions and some of them are analog to the canonical distributions. The results of the numerical analysis demonstrate continuum-to-discreteness transitions, evolutional invariance of distributions, phase transitions with conjugate symmetry, etc., which manifest the construction as a potential candidate for the unification of statistics. We hereby classify the observed distributions on the finite convergent domains. Continuous and discrete distributions both exist and are predictable for given partitions in different regions of parameter-pair. We further compare these distributions with canonical distributions and address the impacts on the existing applications.

Keywords: blaschke, lorentz transformation, complex variables, continuous, discrete, canonical, classification

Procedia PDF Downloads 273
13171 Implementation in Python of a Method to Transform One-Dimensional Signals in Graphs

Authors: Luis Andrey Fajardo Fajardo

Abstract:

We are immersed in complex systems. The human brain, the galaxies, the snowflakes are examples of complex systems. An area of interest in Complex systems is the chaos theory. This revolutionary field of science presents different ways of study than determinism and reductionism. Here is where in junction with the Nonlinear DSP, chaos theory offer valuable techniques that establish a link between time series and complex theory in terms of complex networks, so that, the study of signals can be explored from the graph theory. Recently, some people had purposed a method to transform time series in graphs, but no one had developed a suitable implementation in Python with signals extracted from Chaotic Systems or Complex systems. That’s why the implementation in Python of an existing method to transform one dimensional chaotic signals from time domain to graph domain and some measures that may reveal information not extracted in the time domain is proposed.

Keywords: Python, complex systems, graph theory, dynamical systems

Procedia PDF Downloads 478
13170 Ecological Systems Theory, the SCERTS Model, and the Autism Spectrum, Node and Nexus

Authors: C. Surmei

Abstract:

Autism Spectrum Disorder (ASD) is a complex developmental disorder that can affect an individual’s (but is not limited to) cognitive development, emotional development, language acquisition and the capability to relate to others. Ecological Systems Theory is a sociocultural theory that focuses on environmental systems with which an individual interacts. The SCERTS Model is an educational approach and multidisciplinary framework that addresses the challenges confronted by individuals on the autism spectrum and other developmental disabilities. To aid the understanding of ASD and educational philosophies for families, educators, and the global community alike, a Comparative Analysis was undertaken to examine key variables (the child, society, education, nurture/care, relationships, communication). The results indicated that the Ecological Systems Theory and the SCERTS Model were comparable in focus, motivation, and application, attaining to a viable and notable relationship between both theories. This paper unpacks two child development philosophies and their relationship to each other.

Keywords: autism spectrum disorder, ecological systems theory, education, SCERTS model

Procedia PDF Downloads 535
13169 Robust Image Registration Based on an Adaptive Normalized Mutual Information Metric

Authors: Huda Algharib, Amal Algharib, Hanan Algharib, Ali Mohammad Alqudah

Abstract:

Image registration is an important topic for many imaging systems and computer vision applications. The standard image registration techniques such as Mutual information/ Normalized mutual information -based methods have a limited performance because they do not consider the spatial information or the relationships between the neighbouring pixels or voxels. In addition, the amount of image noise may significantly affect the registration accuracy. Therefore, this paper proposes an efficient method that explicitly considers the relationships between the adjacent pixels, where the gradient information of the reference and scene images is extracted first, and then the cosine similarity of the extracted gradient information is computed and used to improve the accuracy of the standard normalized mutual information measure. Our experimental results on different data types (i.e. CT, MRI and thermal images) show that the proposed method outperforms a number of image registration techniques in terms of the accuracy.

Keywords: image registration, mutual information, image gradients, image transformations

Procedia PDF Downloads 213
13168 Understanding Workplace Behavior through Organizational Culture and Complex Adaptive Systems Theory

Authors: Péter Restás, Andrea Czibor, Zsolt Péter Szabó

Abstract:

Purpose: This article aims to rethink the phenomena of employee behavior as a product of a system. Both organizational culture and Complex Adaptive Systems (CAS) theory emphasize that individual behavior depends on the specific system and the unique organizational culture. These two major theories are both represented in the field of organizational studies; however, they are rarely used together for the comprehensive understanding of workplace behavior. Methodology: By reviewing the literature we use key concepts stemming from organizational culture and CAS theory in order to show the similarities between these theories and create an enriched understanding of employee behavior. Findings: a) Workplace behavior is defined here as social cognition issue. b) Organizations are discussed here as complex systems, and cultures which drive and dictate the cognitive processes of agents in the system. c) Culture gives CAS theory a context which lets us see organizations not just as ever-changing and unpredictable, but as such systems that aim to create and maintain stability by recurring behavior. Conclusion: Applying the knowledge from culture and CAS theory sheds light on our present understanding of employee behavior, also emphasizes the importance of novel ways in organizational research and management.

Keywords: complex adaptive systems theory, employee behavior, organizational culture, stability

Procedia PDF Downloads 377
13167 Whether Chaos Theory Could Reconstruct the Ancient Societies

Authors: Zahra Kouzehgari

Abstract:

Since the early emergence of chaos theory in the 1970s in mathematics and physical science, it has increasingly been developed and adapted in social sciences as well. The non-linear and dynamic characteristics of the theory make it a useful conceptual framework to interpret the complex social systems behavior. Regarding chaotic approach principals, sensitivity to initial conditions, dynamic adoption, strange attractors and unpredictability this paper aims to examine whether chaos approach could interpret the ancient social changes. To do this, at first, a brief history of the chaos theory, its development and application in social science as well as the principals making the theory, then its application in archaeological since has been reviewed. The study demonstrates that although based on existing archaeological records reconstruct the whole social system of the human past, the non-linear approaches in studying social complex systems would be of a great help in finding general order of the ancient societies and would enable us to shed light on some of the social phenomena in the human history or to make sense of them.

Keywords: archaeology, non-linear approach, chaos theory, ancient social systems

Procedia PDF Downloads 241
13166 Normalized P-Laplacian: From Stochastic Game to Image Processing

Authors: Abderrahim Elmoataz

Abstract:

More and more contemporary applications involve data in the form of functions defined on irregular and topologically complicated domains (images, meshs, points clouds, networks, etc). Such data are not organized as familiar digital signals and images sampled on regular lattices. However, they can be conveniently represented as graphs where each vertex represents measured data and each edge represents a relationship (connectivity or certain affinities or interaction) between two vertices. Processing and analyzing these types of data is a major challenge for both image and machine learning communities. Hence, it is very important to transfer to graphs and networks many of the mathematical tools which were initially developed on usual Euclidean spaces and proven to be efficient for many inverse problems and applications dealing with usual image and signal domains. Historically, the main tools for the study of graphs or networks come from combinatorial and graph theory. In recent years there has been an increasing interest in the investigation of one of the major mathematical tools for signal and image analysis, which are Partial Differential Equations (PDEs) variational methods on graphs. The normalized p-laplacian operator has been recently introduced to model a stochastic game called tug-of-war-game with noise. Part interest of this class of operators arises from the fact that it includes, as particular case, the infinity Laplacian, the mean curvature operator and the traditionnal Laplacian operators which was extensiveley used to models and to solve problems in image processing. The purpose of this paper is to introduce and to study a new class of normalized p-Laplacian on graphs. The introduction is based on the extension of p-harmonious function introduced in as discrete approximation for both infinity Laplacian and p-Laplacian equations. Finally, we propose to use these operators as a framework for solving many inverse problems in image processing.

Keywords: normalized p-laplacian, image processing, stochastic game, inverse problems

Procedia PDF Downloads 482
13165 Application of EEG Wavelet Power to Prediction of Antidepressant Treatment Response

Authors: Dorota Witkowska, Paweł Gosek, Lukasz Swiecicki, Wojciech Jernajczyk, Bruce J. West, Miroslaw Latka

Abstract:

In clinical practice, the selection of an antidepressant often degrades to lengthy trial-and-error. In this work we employ a normalized wavelet power of alpha waves as a biomarker of antidepressant treatment response. This novel EEG metric takes into account both non-stationarity and intersubject variability of alpha waves. We recorded resting, 19-channel EEG (closed eyes) in 22 inpatients suffering from unipolar (UD, n=10) or bipolar (BD, n=12) depression. The EEG measurement was done at the end of the short washout period which followed previously unsuccessful pharmacotherapy. The normalized alpha wavelet power of 11 responders was markedly different than that of 11 nonresponders at several, mostly temporoparietal sites. Using the prediction of treatment response based on the normalized alpha wavelet power, we achieved 81.8% sensitivity and 81.8% specificity for channel T4.

Keywords: alpha waves, antidepressant, treatment outcome, wavelet

Procedia PDF Downloads 282
13164 Inverse Matrix in the Theory of Dynamical Systems

Authors: Renata Masarova, Bohuslava Juhasova, Martin Juhas, Zuzana Sutova

Abstract:

In dynamic system theory a mathematical model is often used to describe their properties. In order to find a transfer matrix of a dynamic system we need to calculate an inverse matrix. The paper contains the fusion of the classical theory and the procedures used in the theory of automated control for calculating the inverse matrix. The final part of the paper models the given problem by the Matlab.

Keywords: dynamic system, transfer matrix, inverse matrix, modeling

Procedia PDF Downloads 480
13163 Adaptive Filtering in Subbands for Supervised Source Separation

Authors: Bruna Luisa Ramos Prado Vasques, Mariane Rembold Petraglia, Antonio Petraglia

Abstract:

This paper investigates MIMO (Multiple-Input Multiple-Output) adaptive filtering techniques for the application of supervised source separation in the context of convolutive mixtures. From the observation that there is correlation among the signals of the different mixtures, an improvement in the NSAF (Normalized Subband Adaptive Filter) algorithm is proposed in order to accelerate its convergence rate. Simulation results with mixtures of speech signals in reverberant environments show the superior performance of the proposed algorithm with respect to the performances of the NLMS (Normalized Least-Mean-Square) and conventional NSAF, considering both the convergence speed and SIR (Signal-to-Interference Ratio) after convergence.

Keywords: adaptive filtering, multi-rate processing, normalized subband adaptive filter, source separation

Procedia PDF Downloads 404
13162 Identifying Chaotic Architecture: Origins of Nonlinear Design Theory

Authors: Mohammadsadegh Zanganehfar

Abstract:

Since the modernism, movement, and appearance of modern architecture, an aggressive desire for a general design theory in the theoretical works of architects in the form of books and essays emerges. Since Robert Venturi and Denise Scott Brown’s published complexity and contradiction in architecture in 1966, the discourse of complexity and volumetric composition has been an important and controversial issue in the discipline. Ever since various theories and essays were involved in this discourse, this paper attempt to identify chaos theory as a scientific model of complexity and its relation to architecture design theory by conducting a qualitative analysis and multidisciplinary critical approach through architecture and basic sciences resources. As a result, we identify chaotic architecture as the correlation of chaos theory and architecture as an independent nonlinear design theory with specific characteristics and properties.

Keywords: architecture complexity, chaos theory, fractals, nonlinear dynamic systems, nonlinear ontology

Procedia PDF Downloads 338
13161 Acoustic Echo Cancellation Using Different Adaptive Algorithms

Authors: Hamid Sharif, Nazish Saleem Abbas, Muhammad Haris Jamil

Abstract:

An adaptive filter is a filter that self-adjusts its transfer function according to an optimization algorithm driven by an error signal. Because of the complexity of the optimization algorithms, most adaptive filters are digital filters. Adaptive filtering constitutes one of the core technologies in digital signal processing and finds numerous application areas in science as well as in industry. Adaptive filtering techniques are used in a wide range of applications, including adaptive noise cancellation and echo cancellation. Acoustic echo cancellation is a common occurrence in today’s telecommunication systems. The signal interference caused by acoustic echo is distracting to both users and causes a reduction in the quality of the communication. In this paper, we review different techniques of adaptive filtering to reduce this unwanted echo. In this paper, we see the behavior of techniques and algorithms of adaptive filtering like Least Mean Square (LMS), Normalized Least Mean Square (NLMS), Variable Step-Size Least Mean Square (VSLMS), Variable Step-Size Normalized Least Mean Square (VSNLMS), New Varying Step Size LMS Algorithm (NVSSLMS) and Recursive Least Square (RLS) algorithms to reduce this unwanted echo, to increase communication quality.

Keywords: adaptive acoustic, echo cancellation, LMS algorithm, adaptive filter, normalized least mean square (NLMS), variable step-size least mean square (VSLMS)

Procedia PDF Downloads 46
13160 Analysis of Cardiac Health Using Chaotic Theory

Authors: Chandra Mukherjee

Abstract:

The prevalent knowledge of the biological systems is based on the standard scientific perception of natural equilibrium, determination and predictability. Recently, a rethinking of concepts was presented and a new scientific perspective emerged that involves complexity theory with deterministic chaos theory, nonlinear dynamics and theory of fractals. The unpredictability of the chaotic processes probably would change our understanding of diseases and their management. The mathematical definition of chaos is defined by deterministic behavior with irregular patterns that obey mathematical equations which are critically dependent on initial conditions. The chaos theory is the branch of sciences with an interest in nonlinear dynamics, fractals, bifurcations, periodic oscillations and complexity. Recently, the biomedical interest for this scientific field made these mathematical concepts available to medical researchers and practitioners. Any biological network system is considered to have a nominal state, which is recognized as a homeostatic state. In reality, the different physiological systems are not under normal conditions in a stable state of homeostatic balance, but they are in a dynamically stable state with a chaotic behavior and complexity. Biological systems like heart rhythm and brain electrical activity are dynamical systems that can be classified as chaotic systems with sensitive dependence on initial conditions. In biological systems, the state of a disease is characterized by a loss of the complexity and chaotic behavior, and by the presence of pathological periodicity and regulatory behavior. The failure or the collapse of nonlinear dynamics is an indication of disease rather than a characteristic of health.

Keywords: HRV, HRVI, LF, HF, DII

Procedia PDF Downloads 390
13159 Trajectory Tracking Controller Based on Normalized Right Coprime Factorization Technique for the Ball and Plate System

Authors: Martins Olatunbosun Babatunde, Muhammed Bashir Muazu, Emmanuel Adewale Adedokun

Abstract:

This paper presents the development of a double-loop trajectory-tracking controller for the ball and plate system (BPS) using the Normalized Right Coprime Factorization (NRCF) scheme.The Linear Algebraic (LA) method is used to design the inner loop required to stabilize the ball, while H-infinity NRCF method, that involved the lead-lag compensator design approach, is used to develop the outer loop that controls the plate. Simulation results show that the plate was stabilized at 0.2989 seconds and the ball was able to settle after 0.9646 seconds, with a trajectory tracking error of 0.0036. This shows that the controller has good adaptability and robustness.

Keywords: ball and plate system, normalized right coprime factorization, linear algebraic method, compensator, controller, tracking.

Procedia PDF Downloads 92
13158 Applying Bowen’s Theory to Intern Supervision

Authors: Jeff A. Tysinger, Dawn P. Tysinger

Abstract:

The aim of this paper is to theoretically apply Bowen’s understanding of triangulation and triads to school psychology intern supervision so that it can assist in the conceptualization of the dynamics of intern supervision and provide some key methods to address common issues. The school psychology internship is the capstone experience for the school psychologist in training. It involves three key participants whose relationships will determine the success of the internship.  To understand the potential effect, Bowen’s family systems theory can be applied to the supervision relationship. He describes a way to resolve stress between two people by triangulating or binging in a third person. He applies this to a nuclear family, but school psychology intern supervision requires the marriage of an intern, field supervisor, and university supervisor; thus, setting all up for possible triangulation. The consequences of triangulation can apply to standards and requirements, direct supervision, and intern evaluation. Strategies from family systems theory to decrease the negative impact of supervision triangulation.

Keywords: family systems theory, intern supervision, school psychology training, triangulation

Procedia PDF Downloads 98
13157 Normalized Difference Vegetation Index and Hyperspectral: Plant Health Assessment

Authors: Srushti R. Joshi, Ujjwal Rakesh, Spoorthi Sripad

Abstract:

The rapid advancement of remote sensing technologies has revolutionized plant health monitoring, offering valuable insights for precision agriculture and environmental management. This paper presents a comprehensive comparative analysis between the widely employed normalized difference vegetation index (NDVI) and state-of-the-art hyperspectral sensors in the context of plant health assessment. The study aims to elucidate the weigh ups of spectral resolution. Employing a diverse range of vegetative environments, the research utilizes simulated datasets to evaluate the performance of NDVI and hyperspectral sensors in detecting subtle variations indicative of plant stress, disease, and overall vitality. Through meticulous data analysis and statistical validation, this study highlights the superior performance of hyperspectral sensors across the parameters used.

Keywords: normalized difference vegetation index, hyperspectral sensor, spectral resolution, infrared

Procedia PDF Downloads 32
13156 The Magic Bullet in Africa: Exploring an Alternative Theoretical Model

Authors: Daniel Nkrumah

Abstract:

The Magic Bullet theory was a popular media effect theory that defined the power of the mass media in altering beliefs and perceptions of its audiences. However, following the People's Choice study, the theory was said to have been disproved and was supplanted by the Two-Step Flow Theory. This paper examines the relevance of the Magic Bullet theory in Africa and establishes whether it is still relevant in Africa's media spaces and societies. Using selected cases on the continent, it adopts a grounded theory approach and explores a new theoretical model that attempts to enforce an argument that the Two-Step Flow theory though important and valid, was ill-conceived as a direct replacement to the Magic Bullet theory.

Keywords: magic bullet theory, two-step flow theory, media effects, african media

Procedia PDF Downloads 86
13155 Use of Landsat OLI Images in the Mapping of Landslides: Case of the Taounate Province in Northern Morocco

Authors: S. Benchelha, H. Chennaoui, M. Hakdaoui, L. Baidder, H. Mansouri, H. Ejjaaouani, T. Benchelha

Abstract:

Northern Morocco is characterized by relatively young mountains experiencing a very important dynamic compared to other areas of Morocco. The dynamics associated with the formation of the Rif chain (Alpine tectonics), is accompanied by instabilities essentially related to tectonic movements. The realization of important infrastructures (Roads, Highways,...) represents a triggering factor and favoring landslides. This paper is part of the establishment of landslides susceptibility map and concerns the mapping of unstable areas in the province of Taounate. The landslide was identified using the components of the false color (FCC) of images Landsat OLI: i) the first independent component (IC1), ii) The main component (PC), iii) Normalized difference index (NDI). This mapping for landslides class is validated by in-situ surveys.

Keywords: landslides, False Color Composite (FCC), Independent Component Analysis (ICA), Principal Component Analysis (PCA), Normalized Difference Index (NDI), Normalized Difference Mid Red Index (NDMIDR)

Procedia PDF Downloads 261
13154 Optimized Weight Selection of Control Data Based on Quotient Space of Multi-Geometric Features

Authors: Bo Wang

Abstract:

The geometric processing of multi-source remote sensing data using control data of different scale and different accuracy is an important research direction of multi-platform system for earth observation. In the existing block bundle adjustment methods, as the controlling information in the adjustment system, the approach using single observation scale and precision is unable to screen out the control information and to give reasonable and effective corresponding weights, which reduces the convergence and adjustment reliability of the results. Referring to the relevant theory and technology of quotient space, in this project, several subjects are researched. Multi-layer quotient space of multi-geometric features is constructed to describe and filter control data. Normalized granularity merging mechanism of multi-layer control information is studied and based on the normalized scale factor, the strategy to optimize the weight selection of control data which is less relevant to the adjustment system can be realized. At the same time, geometric positioning experiment is conducted using multi-source remote sensing data, aerial images, and multiclass control data to verify the theoretical research results. This research is expected to break through the cliché of the single scale and single accuracy control data in the adjustment process and expand the theory and technology of photogrammetry. Thus the problem to process multi-source remote sensing data will be solved both theoretically and practically.

Keywords: multi-source image geometric process, high precision geometric positioning, quotient space of multi-geometric features, optimized weight selection

Procedia PDF Downloads 253
13153 The Effect on Lead Times When Normalizing a Supply Chain Process

Authors: Bassam Istanbouli

Abstract:

Organizations are living in a very competitive and dynamic environment which is constantly changing. In order to achieve a high level of service, the products and processes of these organizations need to be flexible and evolvable. If the supply chains are not modular and well designed, changes can bring combinatorial effects to most areas of a company from its management, financial, documentation, logistics and its information structure. Applying the normalized system’s concept to segments of the supply chain may help in reducing those ripple effects, but it may also increase lead times. Lead times are important and can become a decisive element in gaining customers. Industries are always under the pressure in providing good quality products, at competitive prices, when and how the customer wants them. Most of the time, the customers want their orders now, if not yesterday. The above concept will be proven by examining lead times in a manufacturing example before and after applying normalized systems concept to that segment of the chain. We will then show that although we can minimize the combinatorial effects when changes occur, the lead times will be increased.

Keywords: supply chain, lead time, normalization, modular

Procedia PDF Downloads 94
13152 [Keynote Talk]: Evidence Fusion in Decision Making

Authors: Mohammad Abdullah-Al-Wadud

Abstract:

In the current era of automation and artificial intelligence, different systems have been increasingly keeping on depending on decision-making capabilities of machines. Such systems/applications may range from simple classifiers to sophisticated surveillance systems based on traditional sensors and related equipment which are becoming more common in the internet of things (IoT) paradigm. However, the available data for such problems are usually imprecise and incomplete, which leads to uncertainty in decisions made based on traditional probability-based classifiers. This requires a robust fusion framework to combine the available information sources with some degree of certainty. The theory of evidence can provide with such a method for combining evidence from different (may be unreliable) sources/observers. This talk will address the employment of the Dempster-Shafer Theory of evidence in some practical applications.

Keywords: decision making, dempster-shafer theory, evidence fusion, incomplete data, uncertainty

Procedia PDF Downloads 390
13151 Value from Environmental and Cultural Perspectives or Two Sides of the Same Coin

Authors: Vilem Paril, Dominika Tothova

Abstract:

This paper discusses the value theory in cultural heritage and the value theory in environmental economics. Two economic views of the value theory are compared within the field of cultural heritage maintenance and within the field of the environment. The main aims are to find common features in these two differently structured theories under the layer of differently defined terms as well as really differing features of these two approaches, to clear the confusion which stems from different terminology as in fact these terms capture the same aspects of reality and to show possible inspiration these two perspectives can offer one another. Another aim is to present these two value systems in one value framework. First, important moments of the value theory from the economic perspective are presented, leading to the marginal revolution of (not only) the Austrian School. Then the theory of value within cultural heritage and environmental economics are explored. Finally, individual approaches are compared and their potential mutual inspiration searched for.

Keywords: cultural heritage, environmental economics, existence value, value theory

Procedia PDF Downloads 289
13150 Disintegration of Deuterons by Photons Reaction Model for GEANT4 with Dibaryon Formalism

Authors: Jae Won Shin, Chang Ho Hyun

Abstract:

A disintegration of deuterons by photons (dγ → np) reaction model for GEANT4 is developed in this work. An effective field theory with dibaryon fields Introducing a dibaryon field, we can take into account the effective range contribution to the propagator up to infinite order, and it consequently makes the convergence of the theory better than the pionless effective field theory without dibaryon fields. We develop a hadronic model for GEANT4 which is specialized for the disintegration of the deuteron by photons, dγ → np. For the description of two-nucleon interactions, we employ an effective field theory so called pionless theory with dibaryon fields (dEFT). In spite of its simplicity, the theory has proven very effective and useful in the applications to various two-nucleon systems and processes at low energies. We apply the new model of GEANT4 (G4dEFT) to the calculation of total and differential cross sections in dγ → np, and obtain good agreements to experimental data for a wide range of incoming photon energies.

Keywords: dγ → np, dibaryon fields, effective field theory, GEANT4

Procedia PDF Downloads 345
13149 Cotton Crops Vegetative Indices Based Assessment Using Multispectral Images

Authors: Muhammad Shahzad Shifa, Amna Shifa, Muhammad Omar, Aamir Shahzad, Rahmat Ali Khan

Abstract:

Many applications of remote sensing to vegetation and crop response depend on spectral properties of individual leaves and plants. Vegetation indices are usually determined to estimate crop biophysical parameters like crop canopies and crop leaf area indices with the help of remote sensing. Cotton crops assessment is performed with the help of vegetative indices. Remotely sensed images from an optical multispectral radiometer MSR5 are used in this study. The interpretation is based on the fact that different materials reflect and absorb light differently at different wavelengths. Non-normalized and normalized forms of these datasets are analyzed using two complementary data mining algorithms; K-means and K-nearest neighbor (KNN). Our analysis shows that the use of normalized reflectance data and vegetative indices are suitable for an automated assessment and decision making.

Keywords: cotton, condition assessment, KNN algorithm, clustering, MSR5, vegetation indices

Procedia PDF Downloads 290