Search results for: probability approach
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 14331

Search results for: probability approach

13881 A Bundled Approach to Explaining Technological Change: The Case of E-Estonia

Authors: Andrew Adjah Sai, Portia Opoku Boadi

Abstract:

Explaining change is an abstract endeavor. Many management scholars have adopted metaphors to explain change. In this paper, we deal with the drivers of technological change. We use a historical and theoretical approach to review and elaborate on the concepts and context about a specific case. We discuss the limitations of each approach proffered and the implications as a consequence on technological change. We present plurality and multiplicity of perspectives using a socio-technical approach to explain technological change contextually on an organizational level. We show by using our model how technology absorption and diffusion can be accelerated through artefactual institutions to enable social change. The multiplicity of perspectives and plurality of our arguments creates a fine explanation of the e-Estonia case as an example.

Keywords: artefactual institutions, e-Estonia, social change, technological trajectories

Procedia PDF Downloads 429
13880 Non-Destructive Visual-Statistical Approach to Detect Leaks in Water Mains

Authors: Alaa Al Hawari, Mohammad Khader, Tarek Zayed, Osama Moselhi

Abstract:

In this paper, an effective non-destructive, non-invasive approach for leak detection was proposed. The process relies on analyzing thermal images collected by an IR viewer device that captures thermo-grams. In this study a statistical analysis of the collected thermal images of the ground surface along the expected leak location followed by a visual inspection of the thermo-grams was performed in order to locate the leak. In order to verify the applicability of the proposed approach the predicted leak location from the developed approach was compared with the real leak location. The results showed that the expected leak location was successfully identified with an accuracy of more than 95%.

Keywords: thermography, leakage, water pipelines, thermograms

Procedia PDF Downloads 331
13879 Optimization of Switched Reluctance Motor for Drive System in Automotive Applications

Authors: A. Peniak, J. Makarovič, P. Rafajdus, P. Dúbravka

Abstract:

The purpose of this work is to optimize a Switched Reluctance Motor (SRM) for an automotive application, specifically for a fully electric car. A new optimization approach is proposed. This unique approach transforms automotive customer requirements into an optimization problem, based on sound knowledge of a SRM theory. The approach combines an analytical and a finite element analysis of the motor to quantify static nonlinear and dynamic performance parameters, as phase currents and motor torque maps, an output power and power losses in order to find the optimal motor as close to the reality as possible, within reasonable time. The new approach yields the optimal motor which is competitive with other types of already proposed motors for automotive applications. This distinctive approach can also be used to optimize other types of electrical motors, when parts specifically related to the SRM are adjusted accordingly.

Keywords: automotive, drive system, electric car, finite element method, hybrid car, optimization, switched reluctance motor

Procedia PDF Downloads 498
13878 Strengthening Deradicalizing Islamist Extremism in Indonesia: A Victim-Centred Approach

Authors: Milda Istiqomah

Abstract:

Deradicalization program has long been the subject of investigation. There is a steadily growing interest in examining the results on how Islamist terrorists agree to abandon violence and leave radicalism; however, it is argued that de-radicalization program on terrorism in many countries is still questionable for its effectiveness. This article aims to provide an overview of the deradicalization program specifically related to the victim-centred approach conducted by the Indonesian government and investigates critical issues surrounding the analysis of their effectiveness and outcomes. This research employs several case studies of a victim-centred approach conducted by the Indonesian Witness and Victim Protection Agency as well as the Indonesian Counter-terrorism Agency. This paper argues that the victim-centred approach to de-radicalize former terrorist prisoners faces several implemental challenges; however, the initiative may offer promise for future successful de-radicalization program. Furthermore, until more data surrounding the efficacy of this initiative available, the victim-centred approach may also constitute a significant and essential component of disengagement, de-radicalisation, and reintegration of terrorist prisoners. In conclusion, this paper suggests that further empirical research concerning prevention policies and disengagement interventions related to victim-centred approach need to be explored to give more inputs to the Indonesian government to achieve the effectiveness of de-radicalization program.

Keywords: terrorism, victim-centred approach, de-radicalization, Islamist extremism

Procedia PDF Downloads 299
13877 The “Ecological Approach” to GIS Implementation in Low Income Countries’ and the Role of Universities: Union of Municipalities of Joumeh Case Study

Authors: A. Iaaly, O. Jadayel, R. Jadayel

Abstract:

This paper explores the effectiveness of approaches used for the implementation of technology within central governments specifically Geographic Information Systems (GIS). It examines the extent to which various strategies to GIS implementation and its roll out to users within an organization is crucial for its long term assimilation. Depending on the contextual requirements, various implementation strategies exist spanning from the most revolutionary to the most evolutionary, which have an influence on the success of GIS projects and the realization of resulting business benefits within the central governments. This research compares between two strategies of GIS implementation within the Lebanese Municipalities. The first strategy is the “Technological Approach” which is focused on technology acquisition, overlaid on existing governmental frameworks. This approach gives minimal attention to capability building and the long term sustainability of the implemented program. The second strategy, referred to as the “Ecological Approach”, is naturally oriented to the function of the organization. This approach stresses on fostering the evolution of the program and on building the human capabilities. The Union of the Joumeh Municipalities will be presented as a case study under the “Ecological Approach” and the role of the GIS Center at the University of Balamand will be highlighted. Thus, this research contributes to the development of knowledge on technology implementation and the vital role of academia in the specific context of the Lebanese public sector so that this experience may pave the way for further applications.

Keywords: ecological approach GIS, low income countries, technological approach

Procedia PDF Downloads 288
13876 Graph Similarity: Algebraic Model and Its Application to Nonuniform Signal Processing

Authors: Nileshkumar Vishnav, Aditya Tatu

Abstract:

A recent approach of representing graph signals and graph filters as polynomials is useful for graph signal processing. In this approach, the adjacency matrix plays pivotal role; instead of the more common approach involving graph-Laplacian. In this work, we follow the adjacency matrix based approach and corresponding algebraic signal model. We further expand the theory and introduce the concept of similarity of two graphs. The similarity of graphs is useful in that key properties (such as filter-response, algebra related to graph) get transferred from one graph to another. We demonstrate potential applications of the relation between two similar graphs, such as nonuniform filter design, DTMF detection and signal reconstruction.

Keywords: graph signal processing, algebraic signal processing, graph similarity, isospectral graphs, nonuniform signal processing

Procedia PDF Downloads 330
13875 Towards a Competence Management Approach Based on Continuous Improvement

Authors: N. Sefiani, C. Fikri Benbrahim, A. Boumane, K. Reklaoui

Abstract:

Nowadays, the reflection on competence management is the basic for new competitive strategies. It is considered as the core of the problems of the global supply chain. It interacts a variety of actors: information, physical and activities flows, etc. Even though competence management is seen as the key factor for any business success, the existing approaches demonstrate the deficiencies and limitations of the competence concept. This research has two objectives: The first is to make a contribution by focusing on the development of a competence approach, based on continuous improvement. It allows the enterprise to spot key competencies, mobilize them in order to serve its strategic objectives and to develop future competencies. The second is to propose a method to evaluate the level of Collective Competence. The approach was confirmed through an application carried out at an automotive company.

Keywords: competence, competencies’ approach, competence management, continuous improvement, collective competence level, performance indicator

Procedia PDF Downloads 472
13874 Neuro-Fuzzy Based Model for Phrase Level Emotion Understanding

Authors: Vadivel Ayyasamy

Abstract:

The present approach deals with the identification of Emotions and classification of Emotional patterns at Phrase-level with respect to Positive and Negative Orientation. The proposed approach considers emotion triggered terms, its co-occurrence terms and also associated sentences for recognizing emotions. The proposed approach uses Part of Speech Tagging and Emotion Actifiers for classification. Here sentence patterns are broken into phrases and Neuro-Fuzzy model is used to classify which results in 16 patterns of emotional phrases. Suitable intensities are assigned for capturing the degree of emotion contents that exist in semantics of patterns. These emotional phrases are assigned weights which supports in deciding the Positive and Negative Orientation of emotions. The approach uses web documents for experimental purpose and the proposed classification approach performs well and achieves good F-Scores.

Keywords: emotions, sentences, phrases, classification, patterns, fuzzy, positive orientation, negative orientation

Procedia PDF Downloads 359
13873 Survey of the Role of Contextualism in the Designing of Cultural Constructions Based on Rapoport Views

Authors: E. Zarei, M. Bazaei, A. Seifi, A. Keshavarzi

Abstract:

Amos Rapoport, based on his anthropology approach, believed that the space origins from the human body and influences on human body mutually. As a holistic approach in architecture, Contextualism describes a collection of views in philosophy which emphasize the context in which an action, utterance, or expression occurs, and argues that, in some important respect, the action, utterance, or expression can only be understood relative to that context. In this approach, the main goal – studying the role of cultural component in the Contextualism construction shaping up, based on Amos Rapoport’s anthropology approach- has being done by descriptive- analytic method. The results of the research indicate that in the field of Contextualism designing, referring to the cultural aspects are as necessary as the physical dimensions of a construction. Rapoport believes that the shape of a construction is influenced by cultural aspects and he suggests a kind of mutual interaction between human and environment that should be considered in housing. The mail goal of contextual architecture is to establish an interaction between environment, human and culture. According to this approach, a desirable design should be in harmony with this approach.

Keywords: Amos Rapoport, anthropology, contextual architecture, culture

Procedia PDF Downloads 383
13872 From User's Requirements to UML Class Diagram

Authors: Zeineb Ben Azzouz, Wahiba Ben Abdessalem Karaa

Abstract:

The automated extraction of UML class diagram from natural language requirements is a highly challenging task. Many approaches, frameworks and tools have been presented in this field. Nonetheless, the experiments of these tools have shown that there is no approach that can work best all the time. In this context, we propose a new accurate approach to facilitate the automatic mapping from textual requirements to UML class diagram. Our new approach integrates the best properties of statistical Natural Language Processing (NLP) techniques to reduce ambiguity when analysing natural language requirements text. In addition, our approach follows the best practices defined by conceptual modelling experts to determine some patterns indispensable for the extraction of basic elements and concepts of the class diagram. Once the relevant information of class diagram is captured, a XMI document is generated and imported with a CASE tool to build the corresponding UML class diagram.

Keywords: class diagram, user’s requirements, XMI, software engineering

Procedia PDF Downloads 452
13871 Using Gene Expression Programming in Learning Process of Rough Neural Networks

Authors: Sanaa Rashed Abdallah, Yasser F. Hassan

Abstract:

The paper will introduce an approach where a rough sets, gene expression programming and rough neural networks are used cooperatively for learning and classification support. The Objective of gene expression programming rough neural networks (GEP-RNN) approach is to obtain new classified data with minimum error in training and testing process. Starting point of gene expression programming rough neural networks (GEP-RNN) approach is an information system and the output from this approach is a structure of rough neural networks which is including the weights and thresholds with minimum classification error.

Keywords: rough sets, gene expression programming, rough neural networks, classification

Procedia PDF Downloads 357
13870 Measuring the Economic Impact of Cultural Heritage: Comparative Analysis of the Multiplier Approach and the Value Chain Approach

Authors: Nina Ponikvar, Katja Zajc Kejžar

Abstract:

While the positive impacts of heritage on a broad societal spectrum have long been recognized and measured, the economic effects of the heritage sector are often less visible and frequently underestimated. At macro level, economic effects are usually studied based on one of the two mainstream approach, i.e. either the multiplier approach or the value chain approach. Consequently, there is limited comparability of the empirical results due to the use of different methodological approach in the literature. Furthermore, it is also not clear on which criteria the used approach was selected. Our aim is to bring the attention to the difference in the scope of effects that are encompassed by the two most frequent methodological approaches to valuation of economic effects of cultural heritage on macroeconomic level, i.e. the multiplier approach and the value chain approach. We show that while the multiplier approach provides a systematic, theory-based view of economic impacts but requires more data and analysis, the value chain approach has less solid theoretical foundations and depends on the availability of appropriate data to identify the contribution of cultural heritage to other sectors. We conclude that the multiplier approach underestimates the economic impact of cultural heritage, mainly due to the narrow definition of cultural heritage in the statistical classification and the inability to identify part of the contribution of cultural heritage that is hidden in other sectors. Yet it is not possible to clearly determine whether the value chain method overestimates or underestimates the actual economic impact of cultural heritage since there is a risk that the direct effects are overestimated and double counted, but not all indirect and induced effects are considered. Accordingly, these two approaches are not substitutes but rather complementary. Consequently, a direct comparison of the estimated impacts is not possible and should not be done due to the different scope. To illustrate the difference of the impact assessment of the cultural heritage, we apply both approaches to the case of Slovenia in the 2015-2022 period and measure the economic impact of cultural heritage sector in terms of turnover, gross value added and employment. The empirical results clearly show that the estimation of the economic impact of a sector using the multiplier approach is more conservative, while the estimates based on value added capture a much broader range of impacts. According to the multiplier approach, each euro in cultural heritage sector generates an additional 0.14 euros in indirect effects and an additional 0.44 euros in induced effects. Based on the value-added approach, the indirect economic effect of the “narrow” heritage sectors is amplified by the impact of cultural heritage activities on other sectors. Accordingly, every euro of sales and every euro of gross value added in the cultural heritage sector generates approximately 6 euros of sales and 4 to 5 euros of value added in other sectors. In addition, each employee in the cultural heritage sector is linked to 4 to 5 jobs in other sectors.

Keywords: economic value of cultural heritage, multiplier approach, value chain approach, indirect effects, slovenia

Procedia PDF Downloads 59
13869 3D-Mesh Robust Watermarking Technique for Ownership Protection and Authentication

Authors: Farhan A. Alenizi

Abstract:

Digital watermarking has evolved in the past years as an important means for data authentication and ownership protection. The images and video watermarking was well known in the field of multimedia processing; however, 3D objects' watermarking techniques have emerged as an important means for the same purposes, as 3D mesh models are in increasing use in different areas of scientific, industrial, and medical applications. Like the image watermarking techniques, 3D watermarking can take place in either space or transform domains. Unlike images and video watermarking, where the frames have regular structures in both space and temporal domains, 3D objects are represented in different ways as meshes that are basically irregular samplings of surfaces; moreover, meshes can undergo a large variety of alterations which may be hard to tackle. This makes the watermarking process more challenging. While the transform domain watermarking is preferable in images and videos, they are still difficult to implement in 3d meshes due to the huge number of vertices involved and the complicated topology and geometry, and hence the difficulty to perform the spectral decomposition, even though significant work was done in the field. Spatial domain watermarking has attracted significant attention in the past years; they can either act on the topology or on the geometry of the model. Exploiting the statistical characteristics in the 3D mesh models from both geometrical and topological aspects was useful in hiding data. However, doing that with minimal surface distortions to the mesh attracted significant research in the field. A 3D mesh blind watermarking technique is proposed in this research. The watermarking method depends on modifying the vertices' positions with respect to the center of the object. An optimal method will be developed to reduce the errors, minimizing the distortions that the 3d object may experience due to the watermarking process, and reducing the computational complexity due to the iterations and other factors. The technique relies on the displacement process of the vertices' locations depending on the modification of the variances of the vertices’ norms. Statistical analyses were performed to establish the proper distributions that best fit each mesh, and hence establishing the bins sizes. Several optimizing approaches were introduced in the realms of mesh local roughness, the statistical distributions of the norms, and the displacements in the mesh centers. To evaluate the algorithm's robustness against other common geometry and connectivity attacks, the watermarked objects were subjected to uniform noise, Laplacian smoothing, vertices quantization, simplification, and cropping. Experimental results showed that the approach is robust in terms of both perceptual and quantitative qualities. It was also robust against both geometry and connectivity attacks. Moreover, the probability of true positive detection versus the probability of false-positive detection was evaluated. To validate the accuracy of the test cases, the receiver operating characteristics (ROC) curves were drawn, and they’ve shown robustness from this aspect. 3D watermarking is still a new field but still a promising one.

Keywords: watermarking, mesh objects, local roughness, Laplacian Smoothing

Procedia PDF Downloads 144
13868 Decision Making in Medicine and Treatment Strategies

Authors: Kamran Yazdanbakhsh, Somayeh Mahmoudi

Abstract:

Three reasons make good use of the decision theory in medicine: 1. Increased medical knowledge and their complexity makes it difficult treatment information effectively without resorting to sophisticated analytical methods, especially when it comes to detecting errors and identify opportunities for treatment from databases of large size. 2. There is a wide geographic variability of medical practice. In a context where medical costs are, at least in part, by the patient, these changes raise doubts about the relevance of the choices made by physicians. These differences are generally attributed to differences in estimates of probabilities of success of treatment involved, and differing assessments of the results on success or failure. Without explicit criteria for decision, it is difficult to identify precisely the sources of these variations in treatment. 3. Beyond the principle of informed consent, patients need to be involved in decision-making. For this, the decision process should be explained and broken down. A decision problem is to select the best option among a set of choices. The problem is what is meant by "best option ", or know what criteria guide the choice. The purpose of decision theory is to answer this question. The systematic use of decision models allows us to better understand the differences in medical practices, and facilitates the search for consensus. About this, there are three types of situations: situations certain, risky situations, and uncertain situations: 1. In certain situations, the consequence of each decision are certain. 2. In risky situations, every decision can have several consequences, the probability of each of these consequences is known. 3. In uncertain situations, each decision can have several consequences, the probability is not known. Our aim in this article is to show how decision theory can usefully be mobilized to meet the needs of physicians. The decision theory can make decisions more transparent: first, by clarifying the data systematically considered the problem and secondly by asking a few basic principles should guide the choice. Once the problem and clarified the decision theory provides operational tools to represent the available information and determine patient preferences, and thus assist the patient and doctor in their choices.

Keywords: decision making, medicine, treatment strategies, patient

Procedia PDF Downloads 567
13867 Self-Organization-Based Approach for Embedded Real-Time System Design

Authors: S. S. Bendib, L. W. Mouss, S. Kalla

Abstract:

This paper proposes a self-organization-based approach for real-time systems design. The addressed issue is the mapping of an application onto an architecture of heterogeneous processors while optimizing both makespan and reliability. Since this problem is NP-hard, a heuristic algorithm is used to obtain efficiently approximate solutions. The proposed approach takes into consideration the quality as well as the diversity of solutions. Indeed, an alternate treatment of the two objectives allows to produce solutions of good quality while a self-organization approach based on the neighborhood structure is used to reorganize solutions and consequently to enhance their diversity. Produced solutions make different compromises between the makespan and the reliability giving the user the possibility to select the solution suited to his (her) needs.

Keywords: embedded real-time systems design, makespan, reliability, self-organization, compromises

Procedia PDF Downloads 116
13866 Biodiesel Production from Edible Oil Wastewater Sludge with Bioethanol Using Nano-Magnetic Catalysis

Authors: Wighens Ngoie Ilunga, Pamela J. Welz, Olewaseun O. Oyekola, Daniel Ikhu-Omoregbe

Abstract:

Currently, most sludge from the wastewater treatment plants of edible oil factories is disposed to landfills, but landfill sites are finite and potential sources of environmental pollution. Production of biodiesel from wastewater sludge can contribute to energy production and waste minimization. However, conventional biodiesel production is energy and waste intensive. Generally, biodiesel is produced from the transesterification reaction of oils with alcohol (i.e., Methanol, ethanol) in the presence of a catalyst. Homogeneously catalysed transesterification is the conventional approach for large-scale production of biodiesel as reaction times are relatively short. Nevertheless, homogenous catalysis presents several challenges such as high probability of soap. The current study aimed to reuse wastewater sludge from the edible oil industry as a novel feedstock for both monounsaturated fats and bioethanol for the production of biodiesel. Preliminary results have shown that the fatty acid profile of the oilseed wastewater sludge is favourable for biodiesel production with 48% (w/w) monounsaturated fats and that the residue left after the extraction of fats from the sludge contains sufficient fermentable sugars after steam explosion followed by an enzymatic hydrolysis for the successful production of bioethanol [29% (w/w)] using a commercial strain of Saccharomyces cerevisiae. A novel nano-magnetic catalyst was synthesised from mineral processing alkaline tailings, mainly containing dolomite originating from cupriferous ores using a modified sol-gel. The catalyst elemental chemical compositions and structural properties were characterised by X-ray diffraction (XRD), scanning electron microscopy (SEM), Fourier transform infra-red (FTIR) and the BET for the surface area with 14.3 m²/g and 34.1 nm average pore diameter. The mass magnetization of the nano-magnetic catalyst was 170 emu/g. Both the catalytic properties and reusability of the catalyst were investigated. A maximum biodiesel yield of 78% was obtained, which dropped to 52% after the fourth transesterification reaction cycle. The proposed approach has the potential to reduce material costs, energy consumption and water usage associated with conventional biodiesel production technologies. It may also mitigate the impact of conventional biodiesel production on food and land security, while simultaneously reducing waste.

Keywords: biodiesel, bioethanol, edible oil wastewater sludge, nano-magnetism

Procedia PDF Downloads 129
13865 Image Retrieval Based on Multi-Feature Fusion for Heterogeneous Image Databases

Authors: N. W. U. D. Chathurani, Shlomo Geva, Vinod Chandran, Proboda Rajapaksha

Abstract:

Selecting an appropriate image representation is the most important factor in implementing an effective Content-Based Image Retrieval (CBIR) system. This paper presents a multi-feature fusion approach for efficient CBIR, based on the distance distribution of features and relative feature weights at the time of query processing. It is a simple yet effective approach, which is free from the effect of features' dimensions, ranges, internal feature normalization and the distance measure. This approach can easily be adopted in any feature combination to improve retrieval quality. The proposed approach is empirically evaluated using two benchmark datasets for image classification (a subset of the Corel dataset and Oliva and Torralba) and compared with existing approaches. The performance of the proposed approach is confirmed with the significantly improved performance in comparison with the independently evaluated baseline of the previously proposed feature fusion approaches.

Keywords: feature fusion, image retrieval, membership function, normalization

Procedia PDF Downloads 330
13864 An Efficient Hybrid Approach Based on Multi-Agent System and Emergence Method for the Integration of Systematic Preventive Maintenance Policies

Authors: Abdelhadi Adel, Kadri Ouahab

Abstract:

This paper proposes a hybrid algorithm for the integration of systematic preventive maintenance policies in hybrid flow shop scheduling to minimize makespan. We have implemented a problem-solving approach for optimizing the processing time, methods based on metaheuristics. The proposed approach is inspired by the behavior of the human body. This hybridization is between a multi-agent system and inspirations of the human body, especially genetics. The effectiveness of our approach has been demonstrated repeatedly in this paper. To solve such a complex problem, we proposed an approach which we have used advanced operators such as uniform crossover set and single point mutation. The proposed approach is applied to three preventive maintenance policies. These policies are intended to maximize the availability or to maintain a minimum level of reliability during the production chain. The results show that our algorithm outperforms existing algorithms. We assumed that the machines might be unavailable periodically during the production scheduling.

Keywords: multi-agent systems, emergence, genetic algorithm, makespan, systematic maintenance, scheduling, hybrid flow shop scheduling

Procedia PDF Downloads 317
13863 Reliability Analysis of Construction Schedule Plan Based on Building Information Modelling

Authors: Lu Ren, You-Liang Fang, Yan-Gang Zhao

Abstract:

In recent years, the application of BIM (Building Information Modelling) to construction schedule plan has been the focus of more and more researchers. In order to assess the reasonable level of the BIM-based construction schedule plan, that is whether the schedule can be completed on time, some researchers have introduced reliability theory to evaluate. In the process of evaluation, the uncertain factors affecting the construction schedule plan are regarded as random variables, and probability distributions of the random variables are assumed to be normal distribution, which is determined using two parameters evaluated from the mean and standard deviation of statistical data. However, in practical engineering, most of the uncertain influence factors are not normal random variables. So the evaluation results of the construction schedule plan will be unreasonable under the assumption that probability distributions of random variables submitted to the normal distribution. Therefore, in order to get a more reasonable evaluation result, it is necessary to describe the distribution of random variables more comprehensively. For this purpose, cubic normal distribution is introduced in this paper to describe the distribution of arbitrary random variables, which is determined by the first four moments (mean, standard deviation, skewness and kurtosis). In this paper, building the BIM model firstly according to the design messages of the structure and making the construction schedule plan based on BIM, then the cubic normal distribution is used to describe the distribution of the random variables due to the collecting statistical data of the random factors influencing construction schedule plan. Next the reliability analysis of the construction schedule plan based on BIM can be carried out more reasonably. Finally, the more accurate evaluation results can be given providing reference for the implementation of the actual construction schedule plan. In the last part of this paper, the more efficiency and accuracy of the proposed methodology for the reliability analysis of the construction schedule plan based on BIM are conducted through practical engineering case.

Keywords: BIM, construction schedule plan, cubic normal distribution, reliability analysis

Procedia PDF Downloads 124
13862 The Misuse of Free Cash and Earnings Management: An Analysis of the Extent to Which Board Tenure Mitigates Earnings Management

Authors: Michael McCann

Abstract:

Managerial theories propose that, in joint stock companies, executives may be tempted to waste excess free cash on unprofitable projects to keep control of resources. In order to conceal their projects' poor performance, they may seek to engage in earnings management. On the one hand, managers may manipulate earnings upwards in order to post ‘good’ performances and safeguard their position. On the other, since managers pursuit of unrewarding investments are likely to lead to low long-term profitability, managers will use negative accruals to reduce current year’s earnings, smoothing earnings over time in order to conceal the negative effects. Agency models argue that boards of directors are delegated by shareholders to ensure that companies are governed properly. Part of that responsibility is ensuring the reliability of financial information. Analyses of the impact of board characteristics, particularly board independence on the misuse of free cash flow and earnings management finds conflicting evidence. However, existing characterizations of board independence do not account for such directors gaining firm-specific knowledge over time, influencing their monitoring ability. Further, there is little analysis of the influence of the relative experience of independent directors and executives on decisions surrounding the use of free cash. This paper contributes to this literature regarding the heterogeneous characteristics of boards by investigating the influence of independent director tenure on earnings management and the relative tenures of independent directors and Chief Executives. A balanced panel dataset comprising 51 companies across 11 annual periods from 2005 to 2015 is used for the analysis. In each annual period, firms were classified as conducting earnings management if they had discretionary accruals in the bottom quartile (downwards) and top quartile (upwards) of the distributed values for the sample. Logistical regressions were conducted to determine the marginal impact of independent board tenure and a number of control variables on the probability of conducting earnings management. The findings indicate that both absolute and relative measures of board independence and experience do not have a significant impact on the likelihood of earnings management. It is the level of free cash flow which is the major influence on the probability of earnings management. Higher free cash flow increases the probability of earnings management significantly. The research also investigates whether board monitoring of earnings management is contingent on the level of free cash flow. However, the results suggest that board monitoring is not amplified when free cash flow is higher. This suggests that the extent of earnings management in companies is determined by a range of company, industry and situation-specific factors.

Keywords: corporate governance, boards of directors, agency theory, earnings management

Procedia PDF Downloads 209
13861 Multiscale Connected Component Labelling and Applications to Scientific Microscopy Image Processing

Authors: Yayun Hsu, Henry Horng-Shing Lu

Abstract:

In this paper, a new method is proposed to extending the method of connected component labeling from processing binary images to multi-scale modeling of images. By using the adaptive threshold of multi-scale attributes, this approach minimizes the possibility of missing those important components with weak intensities. In addition, the computational cost of this approach remains similar to that of the typical approach of component labeling. Then, this methodology is applied to grain boundary detection and Drosophila Brain-bow neuron segmentation. These demonstrate the feasibility of the proposed approach in the analysis of challenging microscopy images for scientific discovery.

Keywords: microscopic image processing, scientific data mining, multi-scale modeling, data mining

Procedia PDF Downloads 419
13860 Mapping Feature Models to Code Using a Reference Architecture: A Case Study

Authors: Karam Ignaim, Joao M. Fernandes, Andre L. Ferreira

Abstract:

Mapping the artifacts coming from a set of similar products family developed in an ad-hoc manner to make up the resulting software product line (SPL) plays a key role to maintain the consistency between requirements and code. This paper presents a feature mapping approach that focuses on tracing the artifact coming from the migration process, the current feature model (FM), to the other artifacts of the resulting SPL, the reference architecture, and code. Thus, our approach relates each feature of the current FM to its locations in the implementation code, using the reference architecture as an intermediate artifact (as a centric point) to preserve consistency among them during an SPL evolution. The approach uses a particular artifact (i.e., traceability tree) as a solution for managing the mapping process. Tool support is provided using friendlyMapper. We have evaluated the feature mapping approach and tool support by putting the approach into practice (i.e., conducting a case study) of the automotive domain for Classical Sensor Variants Family at Bosch Car Multimedia S.A. The evaluation reveals that the mapping approach presented by this paper fits the automotive domain.

Keywords: feature location, feature models, mapping, software product lines, traceability

Procedia PDF Downloads 105
13859 A Novel Approach to Design and Implement Context Aware Mobile Phone

Authors: G. S. Thyagaraju, U. P. Kulkarni

Abstract:

Context-aware computing refers to a general class of computing systems that can sense their physical environment, and adapt their behaviour accordingly. Context aware computing makes systems aware of situations of interest, enhances services to users, automates systems and personalizes applications. Context-aware services have been introduced into mobile devices, such as PDA and mobile phones. In this paper we are presenting a novel approaches used to realize the context aware mobile. The context aware mobile phone (CAMP) proposed in this paper senses the users situation automatically and provides user context required services. The proposed system is developed by using artificial intelligence techniques like Bayesian Network, fuzzy logic and rough sets theory based decision table. Bayesian Network to classify the incoming call (high priority call, low priority call and unknown calls), fuzzy linguistic variables and membership degrees to define the context situations, the decision table based rules for service recommendation. To exemplify and demonstrate the effectiveness of the proposed methods, the context aware mobile phone is tested for college campus scenario including different locations like library, class room, meeting room, administrative building and college canteen.

Keywords: context aware mobile, fuzzy logic, decision table, Bayesian probability

Procedia PDF Downloads 347
13858 Mechanical-Reliability Coupling for a Bearing Capacity Assessment of Shallow Foundations

Authors: Amal Hentati, Mbarka Selmi, Tarek Kormi, Julien Baroth, Barthelemy Harthong

Abstract:

The impact of uncertainties on the performance assessment of shallow foundations is often significant. The need of the geotechnical engineers to a more objective and rigorous description of soil variations permitting to quantify these uncertainties and to incorporate them into calculation methods led to the development of reliability approaches. In this context, a mechanical-reliability coupling was developed in this paper, using a program coded in Matlab and the finite element software Abaqus, for the bearing capacity assessment of shallow foundations. The reliability analysis, based on the finite element method, assumed both soil cohesion and friction angle as uncertain parameters characterized by normal or lognormal probability distributions. The inherent spatial variability of both soil properties was, then, taken into account using 1D stationary random fields. The application of the proposed methodology to a shallow foundation subjected to a centered vertical loading permitted to highlight the proposed process interest. Findings proved the insufficiency of the conventional approach to predict the foundation failure and a high sensitivity of the ultimate loads to the soil properties uncertainties, mainly those related to the friction angle, was noted. Moreover, an asymmetry of both displacement and velocity fields was obtained.

Keywords: mechanical-reliability coupling, finite element method, shallow foundation, random fields, spatial variability

Procedia PDF Downloads 644
13857 Techno-Economic Analysis Framework for Wave Energy Conversion Schemes under South African Conditions: Modeling and Simulations

Authors: Siyanda S. Biyela, Willie A. Cronje

Abstract:

This paper presents a desktop study of comparing two different wave energy to electricity technologies (WECs) using a techno-economic approach. This techno-economic approach forms basis of a framework for rapid comparison of current and future technologies. The approach also seeks to assist in investment and strategic decision making expediting future deployment of wave energy harvesting in South Africa.

Keywords: cost of energy (COE) tool, sea state, wave energy converter (WEC), WEC-Sim

Procedia PDF Downloads 273
13856 Multidimensional Poverty and Child Cognitive Development

Authors: Bidyadhar Dehury, Sanjay Kumar Mohanty

Abstract:

According to the Right to Education Act of India, education is the fundamental right of all children of age group 6-14 year irrespective of their status. Using the unit level data from India Human Development Survey (IHDS), we tried to understand the inter-relationship between the level of poverty and the academic performance of the children aged 8-11 years. The level of multidimensional poverty is measured using five dimensions and 10 indicators using Alkire-Foster approach. The weighted deprivation score was obtained by giving equal weight to each dimension and indicators within the dimension. The weighted deprivation score varies from 0 to 1 and grouped into four categories as non-poor, vulnerable, multidimensional poor and sever multidimensional poor. The academic performance index was measured using three variables reading skills, math skills and writing skills using PCA. The bivariate and multivariate analysis was used in the analysis. The outcome variable was ordinal. So the predicted probabilities were calculated using the ordinal logistic regression. The predicted probabilities of good academic performance index was 0.202 if the child was sever multidimensional poor, 0.235 if the child was multidimensional poor, 0.264 if the child was vulnerable, and 0.316 if the child was non-poor. Hence, if the level of poverty among the children decreases from sever multidimensional poor to non-poor, the probability of good academic performance increases.

Keywords: multidimensional poverty, academic performance index, reading skills, math skills, writing skills, India

Procedia PDF Downloads 571
13855 A Problem-Based Learning Approach in a Writing Classroom: Tutors’ Experiences and Perceptions

Authors: Muhammad Mukhtar Aliyu

Abstract:

This study investigated tutors’ experiences and perceptions of a problem-based learning approach (PBL) in a writing classroom. The study involved two Nigerian lecturers who facilitated an intact class of second-year students in an English composition course for the period of 12 weeks. Semi-structured interviews were employed to collect data of the study. The lecturers were interviewed before and after the implementation of the PBL process. The overall findings of the study show that the lecturers had positive perceptions of the use of PBL in a writing classroom. Specifically, the findings reveal the lecturers’ positive experiences and perception of the group activities. Finally, the paper gives some pedagogical implications which would give insight for better implementation of the PBL approach.

Keywords: experiences and perception, Nigeria, problem-based learning approach, writing classroom

Procedia PDF Downloads 148
13854 Applying Systems Thinking and a System of Systems Approach to Facilitate Sustainable Grid Integration of Variable Renewable Energy

Authors: Edward B. Ssekulima, Amir Etemadi

Abstract:

This paper presents a Systems Thinking and System of Systems (SoS) viewpoint for managing requirements complexity in the grid integration of Variable Renewable Energy (VRE). To achieve a SoS approach, it is often necessary to inculcate a Systems Thinking (ST) perspective in the planning and design of the attendant system. We show how this approach can support the enhanced integration of VRE (wind, solar small hydro) for which intermittency is a key inhibiting factor to their sustainable grid integration. The results indicate that a ST and SoS approach are a critical tool for decision makers in the planning, design and deployment of VRE Sources for their sustainable grid-integration in accordance with relevant techno-economic, social and environmental requirements.

Keywords: sustainable grid-integration, system of systems, systems thinking, variable energy resources

Procedia PDF Downloads 100
13853 Robust Diagnosis Efficiency by Bond-Graph Approach

Authors: Benazzouz Djamel, Termeche Adel, Touati Youcef, Alem Said, Ouziala Mahdi

Abstract:

This paper presents an approach which detect and isolate efficiently a fault in a system. This approach avoids false alarms, non-detections and delays in detecting faults. A study case have been proposed to show the importance of taking into consideration the uncertainties in the decision-making procedure and their effect on the degradation diagnostic performance and advantage of using Bond Graph (BG) for such degradation. The use of BG in the Linear Fractional Transformation (LFT) form allows generating robust Analytical Redundancy Relations (ARR’s), where the uncertain part of ARR’s is used to generate the residuals adaptive thresholds. The study case concerns an electromechanical system composed of a motor, a reducer and an external load. The aim of this application is to show the effectiveness of the BG-LFT approach to robust fault detection.

Keywords: bond graph, LFT, uncertainties, detection and faults isolation, ARR

Procedia PDF Downloads 288
13852 Flame Volume Prediction and Validation for Lean Blowout of Gas Turbine Combustor

Authors: Ejaz Ahmed, Huang Yong

Abstract:

The operation of aero engines has a critical importance in the vicinity of lean blowout (LBO) limits. Lefebvre’s model of LBO based on empirical correlation has been extended to flame volume concept by the authors. The flame volume takes into account the effects of geometric configuration, the complex spatial interaction of mixing, turbulence, heat transfer and combustion processes inside the gas turbine combustion chamber. For these reasons, flame volume based LBO predictions are more accurate. Although LBO prediction accuracy has improved, it poses a challenge associated with Vf estimation in real gas turbine combustors. This work extends the approach of flame volume prediction previously based on fuel iterative approximation with cold flow simulations to reactive flow simulations. Flame volume for 11 combustor configurations has been simulated and validated against experimental data. To make prediction methodology robust as required in the preliminary design stage, reactive flow simulations were carried out with the combination of probability density function (PDF) and discrete phase model (DPM) in FLUENT 15.0. The criterion for flame identification was defined. Two important parameters i.e. critical injection diameter (Dp,crit) and critical temperature (Tcrit) were identified, and their influence on reactive flow simulation was studied for Vf estimation. Obtained results exhibit ±15% error in Vf estimation with experimental data.

Keywords: CFD, combustion, gas turbine combustor, lean blowout

Procedia PDF Downloads 250