Search results for: mixing approaches
758 Local Steerable Pyramid Binary Pattern Sequence LSPBPS for Face Recognition Method
Authors: Mohamed El Aroussi, Mohammed El Hassouni, Sanaa Ghouzali, Mohammed Rziza, Driss Aboutajdine
Abstract:
In this paper the problem of face recognition under variable illumination conditions is considered. Most of the works in the literature exhibit good performance under strictly controlled acquisition conditions, but the performance drastically drop when changes in pose and illumination occur, so that recently number of approaches have been proposed to deal with such variability. The aim of this work is to introduce an efficient local appearance feature extraction method based steerable pyramid (SP) for face recognition. Local information is extracted from SP sub-bands using LBP(Local binary Pattern). The underlying statistics allow us to reduce the required amount of data to be stored. The experiments carried out on different face databases confirm the effectiveness of the proposed approach.
Keywords: Face recognition (FR), Steerable pyramid (SP), localBinary Pattern (LBP).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2184757 Optimal Maintenance and Improvement Policies in Water Distribution System: Markov Decision Process Approach
Authors: Jong Woo Kim, Go Bong Choi, Sang Hwan Son, Dae Shik Kim, Jung Chul Suh, Jong Min Lee
Abstract:
The Markov decision process (MDP) based methodology is implemented in order to establish the optimal schedule which minimizes the cost. Formulation of MDP problem is presented using the information about the current state of pipe, improvement cost, failure cost and pipe deterioration model. The objective function and detailed algorithm of dynamic programming (DP) are modified due to the difficulty of implementing the conventional DP approaches. The optimal schedule derived from suggested model is compared to several policies via Monte Carlo simulation. Validity of the solution and improvement in computational time are proved.
Keywords: Markov decision processes, Dynamic Programming, Monte Carlo simulation, Periodic replacement, Weibull distribution.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2817756 EMD-Based Signal Noise Reduction
Authors: A.O. Boudraa, J.C. Cexus, Z. Saidi
Abstract:
This paper introduces a new signal denoising based on the Empirical mode decomposition (EMD) framework. The method is a fully data driven approach. Noisy signal is decomposed adaptively into oscillatory components called Intrinsic mode functions (IMFs) by means of a process called sifting. The EMD denoising involves filtering or thresholding each IMF and reconstructs the estimated signal using the processed IMFs. The EMD can be combined with a filtering approach or with nonlinear transformation. In this work the Savitzky-Golay filter and shoftthresholding are investigated. For thresholding, IMF samples are shrinked or scaled below a threshold value. The standard deviation of the noise is estimated for every IMF. The threshold is derived for the Gaussian white noise. The method is tested on simulated and real data and compared with averaging, median and wavelet approaches.
Keywords: Empirical mode decomposition, Signal denoisingnonstationary process.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4003755 Advanced Information Extraction with n-gram based LSI
Authors: Ahmet Güven, Ö. Özgür Bozkurt, Oya Kalıpsız
Abstract:
Number of documents being created increases at an increasing pace while most of them being in already known topics and little of them introducing new concepts. This fact has started a new era in information retrieval discipline where the requirements have their own specialties. That is digging into topics and concepts and finding out subtopics or relations between topics. Up to now IR researches were interested in retrieving documents about a general topic or clustering documents under generic subjects. However these conventional approaches can-t go deep into content of documents which makes it difficult for people to reach to right documents they were searching. So we need new ways of mining document sets where the critic point is to know much about the contents of the documents. As a solution we are proposing to enhance LSI, one of the proven IR techniques by supporting its vector space with n-gram forms of words. Positive results we have obtained are shown in two different application area of IR domain; querying a document database, clustering documents in the document database.Keywords: Document clustering, Information Extraction, Information Retrieval, LSI, n-gram.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1803754 Non-negative Principal Component Analysis for Face Recognition
Abstract:
Principle component analysis is often combined with the state-of-art classification algorithms to recognize human faces. However, principle component analysis can only capture these features contributing to the global characteristics of data because it is a global feature selection algorithm. It misses those features contributing to the local characteristics of data because each principal component only contains some levels of global characteristics of data. In this study, we present a novel face recognition approach using non-negative principal component analysis which is added with the constraint of non-negative to improve data locality and contribute to elucidating latent data structures. Experiments are performed on the Cambridge ORL face database. We demonstrate the strong performances of the algorithm in recognizing human faces in comparison with PCA and NREMF approaches.Keywords: classification, face recognition, non-negativeprinciple component analysis (NPCA)
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1695753 Proposed Alternative System to Existing Traffic Signal System
Authors: Alluri Swaroopa, Lakkakula Venkata Narasimha Prasad
Abstract:
Alone with fast urbanization in world, traffic control became a big issue in urban construction. Having an efficient and reliable traffic control system is crucial to macro-traffic control. Traffic signal is used to manage conflicting requirement by allocating different sets of mutually compatible traffic movement during distinct time interval. Many approaches have been made proposed to solve this discrete stochastic problem. Recognizing the need to minimize right-of-way impacts while efficiently handling the anticipated high traffic volumes, the proposed alternative system gives effective design. This model allows for increased traffic capacity and reduces delays by eliminating a step in maneuvering through the freeway interchange. The concept proposed in this paper involves construction of bridges and ramps at intersection of four roads to control the vehicular congestion and to prevent traffic breakdown.
Keywords: Bridges, junctions, ramps, urban traffic control.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3182752 Fastest Growing Crime with Invisible Chains: A Review of Escaping Sex Trafficking Frameworks in Canada
Authors: Alisha Fisher
Abstract:
Survivors of sex trafficking often report extensive harm not just from the violence itself, but multiple levels such as internalized shame, societal misunderstandings, and the process of reporting, exiting, and healing. The aim of this article is to examine the multi-layered approach to supporting survivors who are exiting sex trafficking through immediate, short-term, and long-term care approaches. We present a systematic review of the current barriers structurally, psychosocially, and psychologically through a Canadian perspective, and apply them to the interventions within the service continuum, basic needs, and further needs and supports to consider. This article suggests that ongoing and additional funding to survivor’s support services, specialized police and heath care training, and increased prevention and public education on the realities of sex trafficking in Canada is a necessity for survivor healing, and the prevention of further harm.
Keywords: Canada Sex Trafficking, exiting sex trafficking, sex trafficking survivors, sex trafficking supports.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 498751 A Concept of Successful Collaborative Design towards Sustainability of Project Development
Authors: Yani Rahmawati, Nadjadji Anwar, Christiono Utomo
Abstract:
Development in construction industry is leading to involve complexities in engineering systems; whereas it also required to its sustainability towards social, environmental, and economical aspects. Experts with requisite background and expertise are involved in order to integrate knowledge in achieving whole criteria through design process. Collaborative design is needed in order to attain optimum design through shared solution and goal from experts. This study is conducted to explore issues and approaches development of collaborative design research in construction and its influence to sustainability of the development. Literature review method is used in order to conceptually figure future research direction of collaborative design research. This research is a part of beginning process in doctoral research program, and will be used to support dissertation’s conceptual definition.
Keywords: Collaborative design, multi-discipline design, design process, sustainability of project development.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2227750 Large-Scale Production of High-Performance Fiber-Metal-Laminates by Prepreg-Press-Technology
Authors: Christian Lauter, Corin Reuter, Shuang Wu, Thomas Troester
Abstract:
Lightweight construction became more and more important over the last decades in several applications, e.g. in the automotive or aircraft sector. This is the result of economic and ecological constraints on the one hand and increasing safety and comfort requirements on the other hand. In the field of lightweight design, different approaches are used due to specific requirements towards the technical systems. The use of endless carbon fiber reinforced plastics (CFRP) offers the largest weight saving potential of sometimes more than 50% compared to conventional metal-constructions. However, there are very limited industrial applications because of the cost-intensive manufacturing of the fibers and production technologies. Other disadvantages of pure CFRP-structures affect the quality control or the damage resistance. One approach to meet these challenges is hybrid materials. This means CFRP and sheet metal are combined on a material level. Therefore, new opportunities for innovative process routes are realizable. Hybrid lightweight design results in lower costs due to an optimized material utilization and the possibility to integrate the structures in already existing production processes of automobile manufacturers. In recent and current research, the advantages of two-layered hybrid materials have been pointed out, i.e. the possibility to realize structures with tailored mechanical properties or to divide the curing cycle of the epoxy resin into two steps. Current research work at the Chair for Automotive Lightweight Design (LiA) at the Paderborn University focusses on production processes for fiber-metal-laminates. The aim of this work is the development and qualification of a large-scale production process for high-performance fiber-metal-laminates (FML) for industrial applications in the automotive or aircraft sector. Therefore, the prepreg-press-technology is used, in which pre-impregnated carbon fibers and sheet metals are formed and cured in a closed, heated mold. The investigations focus e.g. on the realization of short process chains and cycle times, on the reduction of time-consuming manual process steps, and the reduction of material costs. This paper gives an overview over the considerable steps of the production process in the beginning. Afterwards experimental results are discussed. This part concentrates on the influence of different process parameters on the mechanical properties, the laminate quality and the identification of process limits. Concluding the advantages of this technology compared to conventional FML-production-processes and other lightweight design approaches are carried out.
Keywords: Composite material, Fiber metal laminate, Lightweight construction, Prepreg press technology, Large-series production.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1892749 Research on Weakly Hard Real-Time Constraints and Their Boolean Combination to Support Adaptive QoS
Authors: Xiangbin Zhu
Abstract:
Advances in computing applications in recent years have prompted the demand for more flexible scheduling models for QoS demand. Moreover, in practical applications, partly violated temporal constraints can be tolerated if the violation meets certain distribution. So we need extend the traditional Liu and Lanland model to adapt to these circumstances. There are two extensions, which are the (m, k)-firm model and Window-Constrained model. This paper researches on weakly hard real-time constraints and their combination to support QoS. The fact that a practical application can tolerate some violations of temporal constraint under certain distribution is employed to support adaptive QoS on the open real-time system. The experiment results show these approaches are effective compared to traditional scheduling algorithms.Keywords: Weakly Hard Real-Time, Real-Time, Scheduling, Quality of Service.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1579748 Extended Constraint Mask Based One-Bit Transform for Low-Complexity Fast Motion Estimation
Authors: Oğuzhan Urhan
Abstract:
In this paper, an improved motion estimation (ME) approach based on weighted constrained one-bit transform is proposed for block-based ME employed in video encoders. Binary ME approaches utilize low bit-depth representation of the original image frames with a Boolean exclusive-OR based hardware efficient matching criterion to decrease computational burden of the ME stage. Weighted constrained one-bit transform (WC‑1BT) based approach improves the performance of conventional C-1BT based ME employing 2-bit depth constraint mask instead of a 1-bit depth mask. In this work, the range of constraint mask is further extended to increase ME performance of WC-1BT approach. Experiments reveal that the proposed method provides better ME accuracy compared existing similar ME methods in the literature.
Keywords: Fast motion estimation, low-complexity motion estimation, video coding.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 854747 Shadow Detection for Increased Accuracy of Privacy Enhancing Methods in Video Surveillance Edge Devices
Authors: F. Matusek, G. Pujolle, R. Reda
Abstract:
Shadow detection is still considered as one of the potential challenges for intelligent automated video surveillance systems. A pre requisite for reliable and accurate detection and tracking is the correct shadow detection and classification. In such a landscape of conditions, privacy issues add more and more complexity and require reliable shadow detection. In this work the intertwining between security, accuracy, reliability and privacy is analyzed and, accordingly, a novel architecture for Privacy Enhancing Video Surveillance (PEVS) is introduced. Shadow detection and masking are dealt with through the combination of two different approaches simultaneously. This results in a unique privacy enhancement, without affecting security. Subsequently, the methodology was employed successfully in a large-scale wireless video surveillance system; privacy relevant information was stored and encrypted on the unit, without transferring it over an un-trusted network.Keywords: Video Surveillance, Intelligent Video Surveillance, Physical Security, WSSU, Privacy, Shadow Detection.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1342746 Operations Research Applications in Audit Planning and Scheduling
Authors: Abdel-Aziz M. Mohamed
Abstract:
This paper presents a state-of-the-art survey of the operations research models developed for internal audit planning. Two alternative approaches have been followed in the literature for audit planning: (1) identifying the optimal audit frequency; and (2) determining the optimal audit resource allocation. The first approach identifies the elapsed time between two successive audits, which can be presented as the optimal number of audits in a given planning horizon, or the optimal number of transactions after which an audit should be performed. It also includes the optimal audit schedule. The second approach determines the optimal allocation of audit frequency among all auditable units in the firm. In our review, we discuss both the deterministic and probabilistic models developed for audit planning. In addition, game theory models are reviewed to find the optimal auditing strategy based on the interactions between the auditors and the clients.Keywords: Operations research applications, audit frequency, audit planning, audit-staff scheduling.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2934745 Universities Strategic Evaluation Using Balanced Scorecard
Authors: M. D. Nayeri, M. M. Mashhadi, K. Mohajeri
Abstract:
Defining strategic position of the organizations within the industry environment is one of the basic and most important phases of strategic planning to which extent that one of the fundamental schools of strategic planning is the strategic positioning school. In today-s knowledge-based economy and dynamic environment, it is essential for universities as the centers of education, knowledge creation and knowledge worker evolvement. Till now, variant models with different approaches to strategic positioning are deployed in defining the strategic position within the various industries. Balanced Scorecard as one of the powerful models for strategic positioning, analyzes all aspects of the organization evenly. In this paper with the consideration of BSC strength in strategic evaluation, it is used for analyzing the environmental position of the best-s Iranian Business Schools. The results could be used in developing strategic plans for these schools as well as other Iranian Management and Business Schools.Keywords: Strategic planning, Strategic position, Balancedscorecard, Higher education institutions.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4442744 An Intelligent Fuzzy-Neural Diagnostic System for Osteoporosis Risk Assessment
Authors: Chin-Ming Hong, Chin-Teng Lin, Chao-Yen Huang, Yi-Ming Lin
Abstract:
In this article, we propose an Intelligent Medical Diagnostic System (IMDS) accessible through common web-based interface, to on-line perform initial screening for osteoporosis. The fundamental approaches which construct the proposed system are mainly based on the fuzzy-neural theory, which can exhibit superiority over other conventional technologies in many fields. In diagnosis process, users simply answer a series of directed questions to the system, and then they will immediately receive a list of results which represents the risk degrees of osteoporosis. According to clinical testing results, it is shown that the proposed system can provide the general public or even health care providers with a convenient, reliable, inexpensive approach to osteoporosis risk assessment.Keywords: BMD, osteoporosis, IMDS, fuzzy-neural theory, web interface.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1949743 The Effect of Nonnormality on CB-SEM and PLS-SEM Path Estimates
Authors: Z. Jannoo, B. W. Yap, N. Auchoybur, M. A. Lazim
Abstract:
The two common approaches to Structural Equation Modeling (SEM) are the Covariance-Based SEM (CB-SEM) and Partial Least Squares SEM (PLS-SEM). There is much debate on the performance of CB-SEM and PLS-SEM for small sample size and when distributions are nonnormal. This study evaluates the performance of CB-SEM and PLS-SEM under normality and nonnormality conditions via a simulation. Monte Carlo Simulation in R programming language was employed to generate data based on the theoretical model with one endogenous and four exogenous variables. Each latent variable has three indicators. For normal distributions, CB-SEM estimates were found to be inaccurate for small sample size while PLS-SEM could produce the path estimates. Meanwhile, for a larger sample size, CB-SEM estimates have lower variability compared to PLS-SEM. Under nonnormality, CB-SEM path estimates were inaccurate for small sample size. However, CB-SEM estimates are more accurate than those of PLS-SEM for sample size of 50 and above. The PLS-SEM estimates are not accurate unless sample size is very large.
Keywords: CB-SEM, Monte Carlo simulation, Normality conditions, Nonnormality, PLS-SEM.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5139742 Origami Theory and Its Applications: A Literature Review
Abstract:
This paper presents the fundamentals of Origami engineering and its application in nowadays as well as future industry. Several main cores of mathematical approaches such as Huzita- Hatori axioms, Maekawa and Kawasaki-s theorems are introduced briefly. Meanwhile flaps and circle packing by Robert Lang is explained to make understood the underlying principles in designing crease pattern. Rigid origami and its corrugation patterns which are potentially applicable for creating transformable or temporary spaces is discussed to show the transition of origami from paper to thick material. Moreover, some innovative applications of origami such as eyeglass, origami stent and high tech origami based on mentioned theories and principles are showcased in section III; while some updated origami technology such as Vacuumatics, self-folding of polymer sheets and programmable matter folding which could greatlyenhance origami structureare demonstrated in Section IV to offer more insight in future origami.
Keywords: Origami, origami application, origami engineering, origami technology, rigid origami.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6810741 On the Noise Distance in Robust Fuzzy C-Means
Authors: M. G. C. A. Cimino, G. Frosini, B. Lazzerini, F. Marcelloni
Abstract:
In the last decades, a number of robust fuzzy clustering algorithms have been proposed to partition data sets affected by noise and outliers. Robust fuzzy C-means (robust-FCM) is certainly one of the most known among these algorithms. In robust-FCM, noise is modeled as a separate cluster and is characterized by a prototype that has a constant distance δ from all data points. Distance δ determines the boundary of the noise cluster and therefore is a critical parameter of the algorithm. Though some approaches have been proposed to automatically determine the most suitable δ for the specific application, up to today an efficient and fully satisfactory solution does not exist. The aim of this paper is to propose a novel method to compute the optimal δ based on the analysis of the distribution of the percentage of objects assigned to the noise cluster in repeated executions of the robust-FCM with decreasing values of δ . The extremely encouraging results obtained on some data sets found in the literature are shown and discussed.Keywords: noise prototype, robust fuzzy clustering, robustfuzzy C-means
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1822740 A Game Design Framework for Vocational Education
Authors: Heide Lukosch, Roy Van Bussel, Sebastiaan Meijer
Abstract:
Serious games have proven to be a useful instrument to engage learners and increase motivation. Nevertheless, a broadly accepted, practical instructional design approach to serious games does not exist. In this paper, we introduce the use of an instructional design model that has not been applied to serious games yet, and has some advantages compared to other design approaches. We present the case of mechanics mechatronics education to illustrate the close match with timing and role of knowledge and information that the instructional design model prescribes and how this has been translated to a rigidly structured game design. The structured approach answers the learning needs of applicable knowledge within the target group. It combines advantages of simulations with strengths of entertainment games to foster learner-s motivation in the best possible way. A prototype of the game will be evaluated along a well-respected evaluation method within an advanced test setting including test and control group.Keywords: Serious Gaming, Simulation, Complex Learning.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1766739 A Simplified Analytical Approach for Coupled Injection Method of Colloidal Silica with Time Dependent Properties
Authors: M. A. Nozari, R. Ziaie Moayed
Abstract:
Electro-osmosis in clayey soils and sediments, for purposes of clay consolidation, dewatering, or cleanup, and electro injection in porous media is widespread recent decades. It is experimentally found that the chemical properties of porous media especially PH change the characteristics of media. Electro-osmotic conductivity is a function of soil and grout material chemistry, altering with time. Many numerical approaches exist to simulate the of electro kinetic flow rate considering chemical changes. This paper presents a simplified analytical solution for constant flow rate based on varying electro osmotic conductivity and time dependent viscosity for injection of colloidal silica.
Keywords: Colloidal silica, electro-osmosis, pH, viscosity, zeta potential.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1339738 Dry Binder Mixing of Field Trial Investigation Using Soil Mix Technology: A Case Study on Contaminated Site Soil
Authors: M. Allagoa, A. Al-Tabbaa
Abstract:
The study explores the use of binders and additives, such as Portland cement, pulverized fuel ash, ground granulated blast furnace slag, and MgO, to reduce the concentration and leachability of pollutants in contaminated site soils. The research investigates their effectiveness and associated risks of binders, with a focus on Total Heavy Metals (THM) and Total Petroleum Hydrocarbon (TPH). The goal of this research is to evaluate the performance and effectiveness of binders and additives in remediating soil pollutants. The study aims to assess the suitability of the mixtures for ground improvement purposes, determine the optimal dosage, and investigate the associated risks. The research utilizes physical (unconfined compressive strength) and chemical tests (batch leachability test) to assess the efficacy of the binders and additives. A completely randomized design one-way ANOVA is used to determine the significance within mix binders of THM. The study also employs incremental lifetime cancer risk (ILCR) assessments and other indices to evaluate the associated risks. The study finds that Ground Granulated Blast Furnace Slag (GGBS): MgO is the most effective binder for remediation, particularly when using low dosages of MgO combined with higher dosages of GGBS binders on TPH. The results indicate that binders and additives can encapsulate and immobilize pollutants, thereby reducing their leachability and toxicity. The mean unconfined compressive strength of the soil ranges from 285.0-320.5 kPa, while THM levels with a combination of Ground granulated blast furnace slag and Magnesium oxide, Portland cement and Pulverised fuel ash were less than 10 µg/l. Portland cement was below 1 µg/l. The ILCR ranged from 6.77E-02 - 2.65E-01 and 5.444E-01 - 3.20 E+00, with the highest values observed under extreme conditions. The hazard index (HI), risk allowable daily dose intake (ADI), and risk chronic daily intake (CDI) were all less than 1 for the THM. The study identifies MgO as the best additive for use in soil remediation.
Keywords: Risk daily dose intake, risk chronic daily intake, incremental lifetime cancer risk, ILCR, novel binders, additives binders, hazard index.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 250737 Mean Velocity Modeling of Open-Channel Flow with Submerged Rigid Vegetation
Authors: M. Morri, A. Soualmia, P. Belleudy
Abstract:
Vegetation affects the mean and turbulent flow structure. It may increase flood risks and sediment transport. Therefore, it is important to develop analytical approaches for the bed shear stress on vegetated bed, to predict resistance caused by vegetation. In the recent years, experimental and numerical models have both been developed to model the effects of submerged vegetation on open-channel flow. In this paper, different analytic models are compared and tested using the criteria of deviation, to explore their capacity for predicting the mean velocity and select the suitable one that will be applied in real case of rivers. The comparison between the measured data in vegetated flume and simulated mean velocities indicated, a good performance, in the case of rigid vegetation, whereas, Huthoff model shows the best agreement with a high coefficient of determination (R2=80%) and the smallest error in the prediction of the average velocities.
Keywords: Analytic Models, Comparison, Mean Velocity, Vegetation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2538736 Flagging Critical Components to Prevent Transient Faults in Real-Time Systems
Authors: Muhammad Sheikh Sadi, D. G. Myers, Cesar Ortega Sanchez
Abstract:
This paper proposes the use of metrics in design space exploration that highlight where in the structure of the model and at what point in the behaviour, prevention is needed against transient faults. Previous approaches to tackle transient faults focused on recovery after detection. Almost no research has been directed towards preventive measures. But in real-time systems, hard deadlines are performance requirements that absolutely must be met and a missed deadline constitutes an erroneous action and a possible system failure. This paper proposes the use of metrics to assess the system design to flag where transient faults may have significant impact. These tools then allow the design to be changed to minimize that impact, and they also flag where particular design techniques – such as coding of communications or memories – need to be applied in later stages of design.
Keywords: Criticality, Metrics, Real-Time Systems, Transient Faults.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1341735 Introduction of Self-Healing Concrete and Different Methods of Its Scientific Implementation
Authors: Davoud Beheshtizadeh, Davood Jafari
Abstract:
Concrete, with its unique properties and advantages, has gained widespread and increasing use in the construction industry, particularly in a country's infrastructure. However, concrete exhibits certain defects, most notably the presence of micro-cracks that occur after the setting process, leading to increased costs for infrastructure repair and maintenance. As a result, self-healing concretes have garnered attention in various countries in recent years. These concretes employ different mechanisms for repair, including physical, chemical, biological, and combined approaches, each with its own subsets and implementation methods. Certain mechanisms hold significant importance, leading to specialized production methods. Given the novelty of this subject in Iran, there is limited knowledge or, in some cases, a complete lack of understanding. This paper presents various self-healing concrete mechanisms and the advantages, disadvantages, and application scope of each method.
Keywords: Micro-cracks, self-healing concrete, microcapsules, concrete, cement, self-sensitive.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 241734 Using Collaborative Pictures to Understand Student Experience
Authors: Tessa Berg, Emma Guion Akdag
Abstract:
Summative feedback forms are used in academia for gathering data on course quality and student understanding. Students answer a series of questions based on the course they are soon to finish in these forms. Feedback forms are notorious for being homogenised and limiting and thus the data captured is often neutral and lacking in tacit emotional responses. This paper contrasts student feedback forms with collaborative drawing. We analyse 19 pictures drawn by international students on a pre-sessional course. Through visuals we present an approach to enable a holistic level of student understanding. Visuals communicate irrespective of possible language, cultural and educational barriers. This paper sought to discover if the pictures mirrored the feedback given on a typical feedback form. Findings indicate a considerable difference in the two approaches and thus we highlight the value of collaborative drawing as a complimentary resource to aid the understanding of student experience.
Keywords: Feedback forms, visualisation, student experience, collaborative drawing.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1257733 Knowledge Audit Model for Requirement Elicitation Process
Authors: Laleh Taheri, Noraini C. Pa., Rusli Abdullah, Salfarina Abdullah
Abstract:
Knowledge plays an important role to the success of any organization. Software development organizations are highly knowledge-intensive organizations especially in their requirement elicitation process (REP). There are several problems regarding communicating and using the knowledge in REP such as misunderstanding, being out of scope, conflicting information and changes of requirements. All of these problems occurred in transmitting the requirements knowledge during REP. Several researches have been done in REP in order to solve the problem towards requirements. Knowledge Audit (KA) approaches were proposed in order to solve managing knowledge in human resources, financial and manufacturing. There is lack of study applying the KA in requirements elicitation process. Therefore, this paper proposes a KA model for REP in supporting to acquire good requirements.
Keywords: Knowledge Audit, Requirement Elicitation Process, KA Model, Knowledge in Requirement Elicitation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2901732 Evaluating and Selecting Optimization Software Packages: A Framework for Business Applications
Authors: Waleed Abohamad, Amr Arisha
Abstract:
Owing the fact that optimization of business process is a crucial requirement to navigate, survive and even thrive in today-s volatile business environment, this paper presents a framework for selecting a best-fit optimization package for solving complex business problems. Complexity level of the problem and/or using incorrect optimization software can lead to biased solutions of the optimization problem. Accordingly, the proposed framework identifies a number of relevant factors (e.g. decision variables, objective functions, and modeling approach) to be considered during the evaluation and selection process. Application domain, problem specifications, and available accredited optimization approaches are also to be regarded. A recommendation of one or two optimization software is the output of the framework which is believed to provide the best results of the underlying problem. In addition to a set of guidelines and recommendations on how managers can conduct an effective optimization exercise is discussed.Keywords: Complex Business Problems, Optimization, Selection Criteria, Software Evaluation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2910731 A Survey on Quasi-Likelihood Estimation Approaches for Longitudinal Set-ups
Authors: Naushad Mamode Khan
Abstract:
The Com-Poisson (CMP) model is one of the most popular discrete generalized linear models (GLMS) that handles both equi-, over- and under-dispersed data. In longitudinal context, an integer-valued autoregressive (INAR(1)) process that incorporates covariate specification has been developed to model longitudinal CMP counts. However, the joint likelihood CMP function is difficult to specify and thus restricts the likelihood-based estimating methodology. The joint generalized quasi-likelihood approach (GQL-I) was instead considered but is rather computationally intensive and may not even estimate the regression effects due to a complex and frequently ill-conditioned covariance structure. This paper proposes a new GQL approach for estimating the regression parameters (GQL-III) that is based on a single score vector representation. The performance of GQL-III is compared with GQL-I and separate marginal GQLs (GQL-II) through some simulation experiments and is proved to yield equally efficient estimates as GQL-I and is far more computationally stable.
Keywords: Longitudinal, Com-Poisson, Ill-conditioned, INAR(1), GLMS, GQL.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1776730 End To End Process to Automate Batch Application
Authors: Nagmani Lnu
Abstract:
Often, quality engineering refers to testing the applications that either have a User Interface (UI) or an Application Programming Interface (API). We often find mature test practices, standards, and automation regarding UI or API testing. However, another kind is present in almost all types of industries that deal with data in bulk and often get handled through something called a batch application. This is primarily an offline application companies develop to process large data sets that often deal with multiple business rules. The challenge gets more prominent when we try to automate batch testing. This paper describes the approaches taken to test a batch application from a financial industry to test the payment settlement process (a critical use case in all kinds of FinTech companies), resulting in 100% test automation in test creation and test execution. One can follow this approach for any other batch use cases to achieve a higher efficiency in their testing process.
Keywords: Batch testing, batch test automation, batch test strategy, payments testing, payments settlement testing.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 65729 The Impact of NICTBB in Facilitating the E-Services and M-Services in Tanzania
Authors: S. Pazi, C. Chatwin
Abstract:
ICT services are a key element of communications and important for socio-economic development. In recognition of the importance of this, the Tanzanian Government started to implement a National ICT Broadband Infrastructure Fibre Optic Backbone (NICTBB) in 2009; this development was planned to be implemented in four phases using an optical dense wavelength division multiplexing (DWDM) network technology in collaboration with the Chinese Government through the Chinese International Telecommunications Construction Corporation (CITCC) under a bilateral agreement. This paper briefly explores the NICTBB network technologies implementation, operations and Internet bandwidth costs. It also provides an in depth assessment of the delivery of ICT services such as e-services and m-services in both urban and rural areas following commissioning of the NICTBB system. Following quantitative and qualitative approaches, the study shows that there have been significant improvements in utilization efficiency, effectiveness and the reliability of the ICT service such as e-services and m-services the NICTCBB was commissioned.
Keywords: NICTBB, DWDM, Optic Fibre, Internet, ICT services, e-services, m-services.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3221