Search results for: Bayesian approach
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 14150

Search results for: Bayesian approach

12830 The Development Learning Module Physics based on Guided Inquiry Approach on Model Cooperative Learning Type STAD (Student Team Achievement Division) in the Main Subject of Temperature and Heat

Authors: Fani Firmahandari

Abstract:

The development learning module physics based on guided inquiry approach on model cooperative learning type STAD (Student Team Achievement Division) in the main subject of temperature and heat. The research development aimed to produce physics learning module based on guided cooperative learning type STAD (Student Team Achievement Division) in the main subject of temperature and heat to the student in X class. The research method used Research and Development approach. The development procedure of this module includes potential problems, data collection to meet the need, product design, and feasibility of this module. The impact of learning can be seen or observed clearly when the learning process takes place, the teachers or the students already implemented measures cooperative learning model type STAD, so that the learning process goes well, the interaction of teachers and students, students with student looks good, besides that students can interact and work together in group.

Keywords: cooperative learning type STAD (student team achievement division), development, inquiry, interaction students

Procedia PDF Downloads 362
12829 A Comparative Study between Displacement and Strain Based Formulated Finite Elements Applied to the Analysis of Thin Shell Structures

Authors: Djamal Hamadi, Oussama Temami, Abdallah Zatar, Sifeddine Abderrahmani

Abstract:

The analysis and design of thin shell structures is a topic of interest in a variety of engineering applications. In structural mechanics problems the analyst seeks to determine the distribution of stresses throughout the structure to be designed. It is also necessary to calculate the displacements of certain points of the structure to ensure that specified allowable values are not exceeded. In this paper a comparative study between displacement and strain based finite elements applied to the analysis of some thin shell structures is presented. The results obtained from some examples show the efficiency and the performance of the strain based approach compared to the well known displacement formulation.

Keywords: displacement formulation, finite elements, strain based approach, shell structures

Procedia PDF Downloads 420
12828 Adapted Intersection over Union: A Generalized Metric for Evaluating Unsupervised Classification Models

Authors: Prajwal Prakash Vasisht, Sharath Rajamurthy, Nishanth Dara

Abstract:

In a supervised machine learning approach, metrics such as precision, accuracy, and coverage can be calculated using ground truth labels to help in model tuning, evaluation, and selection. In an unsupervised setting, however, where the data has no ground truth, there are few interpretable metrics that can guide us to do the same. Our approach creates a framework to adapt the Intersection over Union metric, referred to as Adapted IoU, usually used to evaluate supervised learning models, into the unsupervised domain, which solves the problem by factoring in subject matter expertise and intuition about the ideal output from the model. This metric essentially provides a scale that allows us to compare the performance across numerous unsupervised models or tune hyper-parameters and compare different versions of the same model.

Keywords: general metric, unsupervised learning, classification, intersection over union

Procedia PDF Downloads 50
12827 Challenges in the Material and Action-Resistance Factor Design for Embedded Retaining Wall Limit State Analysis

Authors: Kreso Ivandic, Filip Dodigovic, Damir Stuhec

Abstract:

The paper deals with the proposed 'Material' and 'Action-resistance factor' design methods in designing the embedded retaining walls. The parametric analysis of evaluating the differences of the output values mutually and compared with classic approach computation was performed. There is a challenge with the criteria for choosing the proposed calculation design methods in Eurocode 7 with respect to current technical regulations and regular engineering practice. The basic criterion for applying a particular design method is to ensure minimum an equal degree of reliability in relation to the current practice. The procedure of combining the relevant partial coefficients according to design methods was carried out. The use of mentioned partial coefficients should result in the same level of safety, regardless of load combinations, material characteristics and problem geometry. This proposed approach of the partial coefficients related to the material and/or action-resistance should aimed at building a bridge between calculations used so far and pure probability analysis. The measure to compare the results was to determine an equivalent safety factor for each analysis. The results show a visible wide span of equivalent values of the classic safety factors.

Keywords: action-resistance factor design, classic approach, embedded retaining wall, Eurocode 7, limit states, material factor design

Procedia PDF Downloads 232
12826 Microscopic Simulation of Toll Plaza Safety and Operations

Authors: Bekir O. Bartin, Kaan Ozbay, Sandeep Mudigonda, Hong Yang

Abstract:

The use of microscopic traffic simulation in evaluating the operational and safety conditions at toll plazas is demonstrated. Two toll plazas in New Jersey are selected as case studies and were developed and validated in Paramics traffic simulation software. In order to simulate drivers’ lane selection behavior in Paramics, a utility-based lane selection approach is implemented in Paramics Application Programming Interface (API). For each vehicle approaching the toll plaza, a utility value is assigned to each toll lane by taking into account the factors that are likely to impact drivers’ lane selection behavior, such as approach lane, exit lane and queue lengths. The results demonstrate that similar operational conditions, such as lane-by-lane toll plaza traffic volume can be attained using this approach. In addition, assessment of safety at toll plazas is conducted via a surrogate safety measure. In particular, the crash index (CI), an improved surrogate measure of time-to-collision (TTC), which reflects the severity of a crash is used in the simulation analyses. The results indicate that the spatial and temporal frequency of observed crashes can be simulated using the proposed methodology. Further analyses can be conducted to evaluate and compare various different operational decisions and safety measures using microscopic simulation models.

Keywords: microscopic simulation, toll plaza, surrogate safety, application programming interface

Procedia PDF Downloads 184
12825 Adopting a Comparative Cultural Studies Approach to Teaching Writing in the Global Classroom

Authors: Madhura Bandyopadhyay

Abstract:

Teaching writing within multicultural and multiethnic communities poses many unique challenges not the least of which is that of intercultural communication. When the writing is in English, pedagogical imperatives often encounter the universalizing tendencies of standardization of both language use and structural parameters which are often at odds with maintaining local practices which preserve cultural pluralism. English often becomes the contact zone within which individual identities of students play out against the standardization imperatives of the larger world. Writing classes can serve as places which become instruments of assimilation of ethnic minorities to a larger globalizing or nationalistic agenda. Hence, for those outside of the standard practices of writing English, adaptability towards a mastery of those practices valued as standard become the focus of teaching taking away from diversity of local English use and other modes of critical thinking. In a very multicultural and multiethnic context such as the US or Singapore, these dynamics become very important. This paper will argue that multiethnic writing classrooms can greatly benefit from taking up a cultural studies approach whereby the students’ lived environments and experiences are analyzed as cultural texts to produce writing. Such an approach eliminates limitations of using both literary texts as foci of discussion as in traditional approaches to teaching writing and the current trend in teaching composition without using texts at all. By bringing in students’ lived experiences into the classroom and analyzing them as cultural compositions stressing the ability to communicate across cultures, cultural competency is valued rather than adaptability while privileging pluralistic experiences as valuable even as universal shared experience are found. Specifically, while teaching writing in English in a multicultural classroom, a cultural studies approach makes both teacher and student aware of the diversity of the English language as it exists in our global context in the students’ experience while making space for diversity in critical thinking, structure and organization of writing effective in an intercultural context.

Keywords: English, multicultural, teaching, writing

Procedia PDF Downloads 511
12824 To Include or Not to Include: Resolving Ethical Concerns over the 20% High Quality Cassava Flour Inclusion in Wheat Flour Policy in Nigeria

Authors: Popoola I. Olayinka, Alamu E. Oladeji, B. Maziya-Dixon

Abstract:

Cassava, an indigenous crop grown locally by subsistence farmers in Nigeria has potential to bring economic benefits to the country. Consumption of bread and other confectionaries has been on the rise due to lifestyle changes of Nigerian consumers. However, wheat, being the major ingredient for bread and confectionery production does not thrive well under Nigerian climate hence the huge spending on wheat importation. To reduce spending on wheat importation, the Federal Government of Nigeria intends passing into law mandatory inclusion of 20% high-quality cassava flour (HQCF) in wheat flour. While the proposed policy may reduce post harvest loss of cassava, and also increase food security and domestic agricultural productivity, there are downsides to the policy which include reduction in nutritional quality and low sensory appeal of cassava-wheat bread, reluctance of flour millers to use HQCF, technology and processing challenges among others. The policy thus presents an ethical dilemma which must be resolved for its successful implementation. While inclusion of HQCF to wheat flour in bread and confectionery is a topic that may have been well addressed, resolving the ethical dilemma resulting from the act has not received much attention. This paper attempts to resolve this dilemma using various approaches in food ethics (cost benefits, utilitarianism, deontological and deliberative). The Cost-benefit approach did not provide adequate resolution of the dilemma as all the costs and benefits of the policy could not be stated in the quantitative term. The utilitarianism approach suggests that the policy delivers greatest good to the greatest number while the deontological approach suggests that the act (inclusion of HQCF to wheat flour) is right hence the policy is not utterly wrong. The deliberative approach suggests a win-win situation through deliberation with the parties involved.

Keywords: HQCF, ethical dilemma, food security, composite flour, cassava bread

Procedia PDF Downloads 409
12823 Using Divergent Nozzle with Aerodynamic Lens to Focus Nanoparticles

Authors: Hasan Jumaah Mrayeh, Fue-Sang Lien

Abstract:

ANSYS Fluent will be used to simulate Computational Fluid Dynamics (CFD) for an efficient lens and nozzle design which will be explained in this paper. We have designed and characterized an aerodynamic lens and a divergent nozzle for focusing flow that transmits sub 25 nm particles through the aerodynamic lens. The design of the lens and nozzle has been improved using CFD for particle trajectories. We obtained a case for calculating nanoparticles (25 nm) flowing through the aerodynamic lens and divergent nozzle. Nanoparticles are transported by air, which is pumped into the aerodynamic lens through the nozzle at 1 atmospheric pressure. We have also developed a computational methodology that can determine the exact focus characteristics of aerodynamic lens systems. Particle trajectories were traced using the Lagrange approach. The simulation shows the ability of the aerodynamic lens to focus on 25 nm particles after using a divergent nozzle.

Keywords: aerodynamic lens, divergent nozzle, ANSYS Fluent, Lagrange approach

Procedia PDF Downloads 308
12822 Applying Critical Realism to Qualitative Social Work Research: A Critical Realist Approach for Social Work Thematic Analysis Method

Authors: Lynne Soon-Chean Park

Abstract:

Critical Realism (CR) has emerged as an alternative to both the positivist and constructivist perspectives that have long dominated social work research. By unpacking the epistemic weakness of two dogmatic perspectives, CR provides a useful philosophical approach that incorporates the ontological objectivist and subjectivist stance. The CR perspective suggests an alternative approach for social work researchers who have long been looking to engage in the complex interplay between perceived reality at the empirical level and the objective reality that lies behind the empirical event as a causal mechanism. However, despite the usefulness of CR in informing social work research, little practical guidance is available about how CR can inform methodological considerations in social work research studies. This presentation aims to provide a detailed description of CR-informed thematic analysis by drawing examples from a social work doctoral research of Korean migrants’ experiences and understanding of trust associated with their settlement experience in New Zealand. Because of its theoretical flexibility and accessibility as a qualitative analysis method, thematic analysis can be applied as a method that works both to search for the demi-regularities of the collected data and to identify the causal mechanisms that lay behind the empirical data. In so doing, this presentation seeks to provide a concrete and detailed exemplar for social work researchers wishing to employ CR in their qualitative thematic analysis process.

Keywords: critical Realism, data analysis, epistemology, research methodology, social work research, thematic analysis

Procedia PDF Downloads 213
12821 Key Factors for Stakeholder Engagement and Sustainable Development

Authors: Jo Rhodes, Bruce Bergstrom, Peter Lok, Vincent Cheng

Abstract:

The aim of this study is to determine key factors and processes for multinationals (MNCs) to develop an effective stakeholder engagement and sustainable development framework. A qualitative multiple-case approach was used. A triangulation method was adopted (interviews, archival documents and observations) to collect data on three global firms (MNCs). 9 senior executives were interviewed for this study (3 from each firm). An initial literature review was conducted to explore possible practices and factors (the deductive approach) to sustainable development. Interview data were analysed using Nvivo to obtain appropriate nodes and themes for the framework. A comparison of findings from interview data and themes, factors developed from the literature review and cross cases comparison were used to develop the final conceptual framework (the inductive approach). The results suggested that stakeholder engagement is a key mediator between ‘stakeholder network’ (internal and external factors) and outcomes (corporate social responsibility, social capital, shared value and sustainable development). Key internal factors such as human capital/talent, technology, culture, leadership and processes such as collaboration, knowledge sharing and co-creation of value with stakeholders were identified. These internal factors and processes must be integrated and aligned with external factors such as social, political, cultural, environment and NGOs to achieve effective stakeholder engagement.

Keywords: stakeholder, engagement, sustainable development, shared value, corporate social responsibility

Procedia PDF Downloads 515
12820 Housing First, Not Housing Only: The Life Skills Project

Authors: Sara Cumming, Julianne DiSanto, Leah Burton

Abstract:

Homelessness in Canada is a persistent problem. It has been widely argued that the best tactic for eradicating homelessness is to approach social issues from a Housing First perspective—an approach that centers on quickly moving people into permanent and independent housing and then providing them additional support and services as needed. It is recognized that life skills training is both necessary and an effective way to reduce cyclical homelessness; however, there is a scarcity of research on effective ways to teach life skills; this problem was exacerbated in a pandemic context, where in-person delivery was severely restricted or no longer possible. Very little attention has been paid to the diverse cultural needs of clients in a multicultural context and the need to foster cultural knowledge/awareness in individuals to successfully contribute to the cultural safety of communities. This research attempts to fill these gaps in the literature and in practice by employing a community-engaged research (CER) approach. Academic, government, funders, front-line staff, and clients at 15 not-for-profits from across the Greater Toronto Area in Ontario, Canada, collaborated to co-create a virtual, client-centric, equity, diversity, and inclusion (EDI) informed life skill learning management system. We employed a triangulation methodology for this research. An environmental scan was conducted for best practices. Two separate Creative Problem Solving Sessions were held with over 100 front-line workers, managers, and executive directors who work with homeless populations. Quantitative and open-ended surveys were completed by over 200 individuals with experience with homelessness. All sections of this research aimed to discover the areas of skills that individuals need to maintain housing and to ascertain what a more client-driven EDI approach to life skills training should include. This research will showcase which life skills are deemed essential for homeless and precariously housed individuals.

Keywords: homelessness, Housing First, life skills, community engaged research

Procedia PDF Downloads 69
12819 A Text Classification Approach Based on Natural Language Processing and Machine Learning Techniques

Authors: Rim Messaoudi, Nogaye-Gueye Gning, François Azelart

Abstract:

Automatic text classification applies mostly natural language processing (NLP) and other AI-guided techniques to automatically classify text in a faster and more accurate manner. This paper discusses the subject of using predictive maintenance to manage incident tickets inside the sociality. It focuses on proposing a tool that treats and analyses comments and notes written by administrators after resolving an incident ticket. The goal here is to increase the quality of these comments. Additionally, this tool is based on NLP and machine learning techniques to realize the textual analytics of the extracted data. This approach was tested using real data taken from the French National Railways (SNCF) company and was given a high-quality result.

Keywords: machine learning, text classification, NLP techniques, semantic representation

Procedia PDF Downloads 105
12818 Complexity in Managing Higher Education Institutions in Mexico: A System Dynamics Approach

Authors: José Carlos Rodríguez, Mario Gómez, Medardo Serna

Abstract:

This paper analyses managing higher education institutions in emerging economies. The paper investigates the case of postgraduate studies development at public universities. In so doing, it adopts the complex theory approach to evaluate how postgraduate studies have evolved in these countries. The investigation suggests that the postgraduate studies sector at public universities can be seen as a complex adaptive system (CAS). Therefore, the paper adopts system dynamics (SD) methods to develop this analysis. The case of postgraduate studies at Universidad Michoacana de San Nicolás de Hidalgo in Mexico is investigated in this paper.

Keywords: complex adaptive systems, higher education institutions, Mexico, system dynamics

Procedia PDF Downloads 320
12817 Determining the Causality Variables in Female Genital Mutilation: A Factor Screening Approach

Authors: Ekele Alih, Enejo Jalija

Abstract:

Female Genital Mutilation (FGM) is made up of three types namely: Clitoridectomy, Excision and Infibulation. In this study, we examine the factors responsible for FGM in order to identify the causality variables in a logistic regression approach. From the result of the survey conducted by the Public Health Division, Nigeria Institute of Medical Research, Yaba, Lagos State, the tau statistic, τ was used to screen 9 factors that causes FGM in order to select few of the predictors before multiple regression equation is obtained. The need for this may be that the sample size may not be able to sustain having a regression with all the predictors or to avoid multi-collinearity. A total of 300 respondents, comprising 150 adult males and 150 adult females were selected for the household survey based on the multi-stage sampling procedure. The tau statistic,

Keywords: female genital mutilation, logistic regression, tau statistic, African society

Procedia PDF Downloads 262
12816 Prefabricated Integral Design of Building Services

Authors: Mina Mortazavi

Abstract:

The common approach in the construction industry for restraint requirements in existing structures or new constructions is to have Non-Structural Components (NSCs) assembled and installed on-site by different MEP subcontractors. This leads to a lack of coordination and higher costs, construction time, and complications due to inaccurate building information modelling (BIM) systems. Introducing NSCs to a consistent BIM system from the beginning of the design process and considering their seismic loads in the analysis and design process can improve coordination and reduce costs and time. One solution is to use prefabricated mounts with attached MEPs delivered as an integral module. This eliminates the majority of coordination complications and reduces design and installation costs and time. An advanced approach is to have as many NSCs as possible installed in the same prefabricated module, which gives the structural engineer the opportunity to consider the involved component weights and locations in the analysis and design of the prefabricated support. This efficient approach eliminates coordination and access issues, leading to enhanced quality control. This research will focus on the existing literature on modular sub-assemblies that are integrated with architectural and structural components. Modular MEP systems take advantage of the precision provided by BIM tools to meet exact requirements and achieve a buildable design every time. Modular installations that include MEP systems provide efficient solutions for the installation of MEP services or components.

Keywords: building services, modularisation, prefabrication, integral building design

Procedia PDF Downloads 73
12815 A Qualitative Approach to Engineering Design Issues, Problems, and Solutions

Authors: M. U. Arshid, M. A. Kamal

Abstract:

The engineering design process is the activities formulation, to help an engineer raising a plan with a specified goal and performance. The engineering design process is a multi-stage course of action including the conceptualization, research, feasibility studies, establishment of design parameters, preliminary and finally the detailed design. It is a progression from the abstract to the concrete; starting with probably abstract ideas about need, and thereafter elaborating detailed specifications of the object that would satisfy the needs, identified. Engineering design issues, problems, and solutions are discussed in this paper using qualitative approach from an information structure perspective. The objective is to identify the problems, to analyze them and propose solutions by integrating; innovation, practical experience, time and resource management, communications skills, isolating the problem in coordination with all stakeholders. Consequently, this would be beneficial for the engineering community to improve the Engineering design practices.

Keywords: engineering design, engineering design issues, innovation, public sector projects

Procedia PDF Downloads 348
12814 Critical and Strategic Issues in Compensation, Staffing and Personnel Management in Nigeria

Authors: Shonuga Olajumoke Adedoyinsola

Abstract:

Staffing and Compensation are at the core of any employment exchange, and they serve as the defining characteristics of any employment relationship. Most organizations understand the benefits that a longer term approach to staff planning can bring and the answer to this problem lies not in trying to implement the traditional approach more effectively, but in implementing a completely different kind of process for strategic staffing. The study focuses on critical points of compensation, staffing and personnel management. The fundamentals of these programs include the elements of vision, potential, communication and motivation. The aim of the paper is to identify the most important attributes of compensation and incentives, staffing and personnel management. Research method is the analysis and synthesis of scientific literature, logical, comparative and graphic representation. On the basis of analysis, the author presents the models of these systems for positive employee attitudes and behaviors.

Keywords: compensation, employees, incentives, staffing, personnel management

Procedia PDF Downloads 301
12813 Defining a Reference Architecture for Predictive Maintenance Systems: A Case Study Using the Microsoft Azure IoT-Cloud Components

Authors: Walter Bernhofer, Peter Haber, Tobias Mayer, Manfred Mayr, Markus Ziegler

Abstract:

Current preventive maintenance measures are cost intensive and not efficient. With the available sensor data of state of the art internet of things devices new possibilities of automated data processing emerge. Current advances in data science and in machine learning enable new, so called predictive maintenance technologies, which empower data scientists to forecast possible system failures. The goal of this approach is to cut expenses in preventive maintenance by automating the detection of possible failures and to improve efficiency and quality of maintenance measures. Additionally, a centralization of the sensor data monitoring can be achieved by using this approach. This paper describes the approach of three students to define a reference architecture for a predictive maintenance solution in the internet of things domain with a connected smartphone app for service technicians. The reference architecture is validated by a case study. The case study is implemented with current Microsoft Azure cloud technologies. The results of the case study show that the reference architecture is valid and can be used to achieve a system for predictive maintenance execution with the cloud components of Microsoft Azure. The used concepts are technology platform agnostic and can be reused in many different cloud platforms. The reference architecture is valid and can be used in many use cases, like gas station maintenance, elevator maintenance and many more.

Keywords: case study, internet of things, predictive maintenance, reference architecture

Procedia PDF Downloads 253
12812 Modelling Agricultural Commodity Price Volatility with Markov-Switching Regression, Single Regime GARCH and Markov-Switching GARCH Models: Empirical Evidence from South Africa

Authors: Yegnanew A. Shiferaw

Abstract:

Background: commodity price volatility originating from excessive commodity price fluctuation has been a global problem especially after the recent financial crises. Volatility is a measure of risk or uncertainty in financial analysis. It plays a vital role in risk management, portfolio management, and pricing equity. Objectives: the core objective of this paper is to examine the relationship between the prices of agricultural commodities with oil price, gas price, coal price and exchange rate (USD/Rand). In addition, the paper tries to fit an appropriate model that best describes the log return price volatility and estimate Value-at-Risk and expected shortfall. Data and methods: the data used in this study are the daily returns of agricultural commodity prices from 02 January 2007 to 31st October 2016. The data sets consists of the daily returns of agricultural commodity prices namely: white maize, yellow maize, wheat, sunflower, soya, corn, and sorghum. The paper applies the three-state Markov-switching (MS) regression, the standard single-regime GARCH and the two regime Markov-switching GARCH (MS-GARCH) models. Results: to choose the best fit model, the log-likelihood function, Akaike information criterion (AIC), Bayesian information criterion (BIC) and deviance information criterion (DIC) are employed under three distributions for innovations. The results indicate that: (i) the price of agricultural commodities was found to be significantly associated with the price of coal, price of natural gas, price of oil and exchange rate, (ii) for all agricultural commodities except sunflower, k=3 had higher log-likelihood values and lower AIC and BIC values. Thus, the three-state MS regression model outperformed the two-state MS regression model (iii) MS-GARCH(1,1) with generalized error distribution (ged) innovation performs best for white maize and yellow maize; MS-GARCH(1,1) with student-t distribution (std) innovation performs better for sorghum; MS-gjrGARCH(1,1) with ged innovation performs better for wheat, sunflower and soya and MS-GARCH(1,1) with std innovation performs better for corn. In conclusion, this paper provided a practical guide for modelling agricultural commodity prices by MS regression and MS-GARCH processes. This paper can be good as a reference when facing modelling agricultural commodity price problems.

Keywords: commodity prices, MS-GARCH model, MS regression model, South Africa, volatility

Procedia PDF Downloads 204
12811 Inventory Control for Purchased Part under Long Lead Time and Uncertain Demand: MRP vs Demand-Driven MRP Approach

Authors: M. J. Shofa, A. Hidayatno, O. M. Armand

Abstract:

MRP as a production control system is appropriate for the deterministic environment. Unfortunately, most production systems such as customer demands are stochastic. Demand-Driven MRP (DDMRP) is a new approach for inventory control system, and it deals with demand uncertainty. The objective of this paper is to compare the MRP and DDMRP work for a long lead time and uncertain demand in terms of on-hand inventory levels. The evaluation is conducted through a discrete event simulation using purchased part data from an automotive company. The result is MRP gives 50,759 pcs / day while DDMRP gives 34,835 pcs / day (reduce 32%), it means DDMRP is more effective inventory control than MRP in terms of on-hand inventory levels.

Keywords: Demand-Driven MRP, long lead time, MRP, uncertain demand

Procedia PDF Downloads 302
12810 A New OvS Approach in Assembly Line Balancing Problem

Authors: P. Azimi, B. Behtoiy, A. A. Najafi, H. R. Charmchi

Abstract:

According to the previous studies, one of the most famous techniques which affect the efficiency of a production line is the assembly line balancing (ALB) technique. This paper examines the balancing effect of a whole production line of a real auto glass manufacturer in three steps. In the first step, processing time of each activity in the workstations is generated according to a practical approach. In the second step, the whole production process is simulated and the bottleneck stations have been identified, and finally in the third step, several improvement scenarios are generated to optimize the system throughput, and the best one is proposed. The main contribution of the current research is the proposed framework which combines two famous approaches including Assembly Line Balancing and Optimization via Simulation technique (OvS). The results show that the proposed framework could be applied in practical environments, easily.

Keywords: assembly line balancing problem, optimization via simulation, production planning

Procedia PDF Downloads 526
12809 Machine Learning Approach to Project Control Threshold Reliability Evaluation

Authors: Y. Kim, H. Lee, M. Park, B. Lee

Abstract:

Planning is understood as the determination of what has to be performed, how, in which sequence, when, what resources are needed, and their cost within the organization before execution. In most construction project, it is evident that the inherent nature of planning is dynamic, and initial planning is subject to be changed due to various uncertain conditions of construction project. Planners take a continuous revision process during the course of a project and until the very end of project. However, current practice lacks reliable, systematic tool for setting variance thresholds to determine when and what corrective actions to be taken. Rather it is heavily dependent on the level of experience and knowledge of the planner. Thus, this paper introduces a machine learning approach to evaluate project control threshold reliability incorporating project-specific data and presents a method to automate the process. The results have shown that the model improves the efficiency and accuracy of the monitoring process as an early warning.

Keywords: machine learning, project control, project progress monitoring, schedule

Procedia PDF Downloads 245
12808 Notice and Block?

Authors: Althaf Marsoof

Abstract:

The blocking injunction, giving rise to a ‘notice and block’ regime, has become the new approach to curtail the infringement of Intellectual Property rights on the Internet. As such, the blocking injunction is an addition to the arsenal of copyright owners, and more recently has also benefited trademark owners, in their battle against piracy and counterfeiting. Yet, the blocking injunction, notwithstanding the usefulness of its ‘notice and block’ outcome, is not without limitations. In the circumstances, it is argued that ‘notice and takedown’, the approach that has been adopted by right-holders for some years, is still an important remedy against the proliferation of online content that infringe the rights of copyright and trademark owners, which is both viable and effective. Thus, it is suggested that the battle against online piracy and counterfeiting could be won only if both the blocking injunction and the practice of ‘notice and takedown’ are utilised by right-holders as complementary and simultaneous remedies.

Keywords: blocking injunctions, internet intermediaries, notice and takedown, intellectual property

Procedia PDF Downloads 419
12807 Hybrid Approach for Software Defect Prediction Using Machine Learning with Optimization Technique

Authors: C. Manjula, Lilly Florence

Abstract:

Software technology is developing rapidly which leads to the growth of various industries. Now-a-days, software-based applications have been adopted widely for business purposes. For any software industry, development of reliable software is becoming a challenging task because a faulty software module may be harmful for the growth of industry and business. Hence there is a need to develop techniques which can be used for early prediction of software defects. Due to complexities in manual prediction, automated software defect prediction techniques have been introduced. These techniques are based on the pattern learning from the previous software versions and finding the defects in the current version. These techniques have attracted researchers due to their significant impact on industrial growth by identifying the bugs in software. Based on this, several researches have been carried out but achieving desirable defect prediction performance is still a challenging task. To address this issue, here we present a machine learning based hybrid technique for software defect prediction. First of all, Genetic Algorithm (GA) is presented where an improved fitness function is used for better optimization of features in data sets. Later, these features are processed through Decision Tree (DT) classification model. Finally, an experimental study is presented where results from the proposed GA-DT based hybrid approach is compared with those from the DT classification technique. The results show that the proposed hybrid approach achieves better classification accuracy.

Keywords: decision tree, genetic algorithm, machine learning, software defect prediction

Procedia PDF Downloads 332
12806 Integrative-Cyclical Approach to the Study of Quality Control of Resource Saving by the Use of Innovation Factors

Authors: Anatoliy A. Alabugin, Nikolay K. Topuzov, Sergei V. Aliukov

Abstract:

It is well known, that while we do a quantitative evaluation of the quality control of some economic processes (in particular, resource saving) with help innovation factors, there are three groups of problems: high uncertainty of indicators of the quality management, their considerable ambiguity, and high costs to provide a large-scale research. These problems are defined by the use of contradictory objectives of enhancing of the quality control in accordance with innovation factors and preservation of economic stability of the enterprise. The most acutely, such factors are felt in the countries lagging behind developed economies of the world according to criteria of innovativeness and effectiveness of management of the resource saving. In our opinion, the following two methods for reconciling of the above-mentioned objectives and reducing of conflictness of the problems are to solve this task most effectively: 1) the use of paradigms and concepts of evolutionary improvement of quality of resource-saving management in the cycle "from the project of an innovative product (technology) - to its commercialization and update parameters of customer value"; 2) the application of the so-called integrative-cyclical approach which consistent with complexity and type of the concept, to studies allowing to get quantitative assessment of the stages of achieving of the consistency of these objectives (from baseline of imbalance, their compromise to achievement of positive synergies). For implementation, the following mathematical tools are included in the integrative-cyclical approach: index-factor analysis (to identify the most relevant factors); regression analysis of relationship between the quality control and the factors; the use of results of the analysis in the model of fuzzy sets (to adjust the feature space); method of non-parametric statistics (for a decision on the completion or repetition of the cycle in the approach in depending on the focus and the closeness of the connection of indicator ranks of disbalance of purposes). The repetition is performed after partial substitution of technical and technological factors ("hard") by management factors ("soft") in accordance with our proposed methodology. Testing of the proposed approach has shown that in comparison with the world practice there are opportunities to improve the quality of resource-saving management using innovation factors. We believe that the implementation of this promising research, to provide consistent management decisions for reducing the severity of the above-mentioned contradictions and increasing the validity of the choice of resource-development strategies in terms of parameters of quality management and sustainability of enterprise, is perspective. Our existing experience in the field of quality resource-saving management and the achieved level of scientific competence of the authors allow us to hope that the use of the integrative-cyclical approach to the study and evaluation of the resulting and factor indicators will help raise the level of resource-saving characteristics up to the value existing in the developed economies of post-industrial type.

Keywords: integrative-cyclical approach, quality control, evaluation, innovation factors. economic sustainability, innovation cycle of management, disbalance of goals of development

Procedia PDF Downloads 247
12805 An Improved Two-dimensional Ordered Statistical Constant False Alarm Detection

Authors: Weihao Wang, Zhulin Zong

Abstract:

Two-dimensional ordered statistical constant false alarm detection is a widely used method for detecting weak target signals in radar signal processing applications. The method is based on analyzing the statistical characteristics of the noise and clutter present in the radar signal and then using this information to set an appropriate detection threshold. In this approach, the reference cell of the unit to be detected is divided into several reference subunits. These subunits are used to estimate the noise level and adjust the detection threshold, with the aim of minimizing the false alarm rate. By using an ordered statistical approach, the method is able to effectively suppress the influence of clutter and noise, resulting in a low false alarm rate. The detection process involves a number of steps, including filtering the input radar signal to remove any noise or clutter, estimating the noise level based on the statistical characteristics of the reference subunits, and finally, setting the detection threshold based on the estimated noise level. One of the main advantages of two-dimensional ordered statistical constant false alarm detection is its ability to detect weak target signals in the presence of strong clutter and noise. This is achieved by carefully analyzing the statistical properties of the signal and using an ordered statistical approach to estimate the noise level and adjust the detection threshold. In conclusion, two-dimensional ordered statistical constant false alarm detection is a powerful technique for detecting weak target signals in radar signal processing applications. By dividing the reference cell into several subunits and using an ordered statistical approach to estimate the noise level and adjust the detection threshold, this method is able to effectively suppress the influence of clutter and noise and maintain a low false alarm rate.

Keywords: two-dimensional, ordered statistical, constant false alarm, detection, weak target signals

Procedia PDF Downloads 80
12804 Development of Residual Power Series Methods for Efficient Solutions of Stiff Differential Equations

Authors: Gebreegziabher Hailu

Abstract:

This paper presents the development of residual power series methods aimed at efficiently solving stiff differential equations, which pose significant challenges in numerical analysis due to their rapid changes in solution behavior. The RPSM is a numerical approach that generates polynomial-based approximate solutions without the need for linearization, discretization, or perturbation techniques, making it straightforward to implement and less prone to computational errors. We introduce an approach that utilizes power series expansions combined with residual minimization techniques to enhance convergence and stability. By analyzing the theoretical foundations of stiffness, we delve into the formulation of the residual power series method, detailing how it effectively captures the dynamics of stiff systems while maintaining computational efficiency. Numerical experiments demonstrate the method's superiority in terms of accuracy and computational cost when compared to traditional methods like implicit Runge-Kutta or multistep techniques. We also explore adaptive strategies within our framework to automatically adjust parameters based on the stiffness characteristics of the problem at hand. Ultimately, our findings contribute to the broader toolkit for tackling stiff differential equations, offering a robust alternative that promises to streamline computational workflows in various applied mathematics and engineering contexts.

Keywords: residual power series methods, stiff differential equoations, numerical approach, Runge Kutta methods

Procedia PDF Downloads 27
12803 A Hybrid-Evolutionary Optimizer for Modeling the Process of Obtaining Bricks

Authors: Marius Gavrilescu, Sabina-Adriana Floria, Florin Leon, Silvia Curteanu, Costel Anton

Abstract:

Natural sciences provide a wide range of experimental data whose related problems require study and modeling beyond the capabilities of conventional methodologies. Such problems have solution spaces whose complexity and high dimensionality require correspondingly complex regression methods for proper characterization. In this context, we propose an optimization method which consists in a hybrid dual optimizer setup: a global optimizer based on a modified variant of the popular Imperialist Competitive Algorithm (ICA), and a local optimizer based on a gradient descent approach. The ICA is modified such that intermediate solution populations are more quickly and efficiently pruned of low-fitness individuals by appropriately altering the assimilation, revolution and competition phases, which, combined with an initialization strategy based on low-discrepancy sampling, allows for a more effective exploration of the corresponding solution space. Subsequently, gradient-based optimization is used locally to seek the optimal solution in the neighborhoods of the solutions found through the modified ICA. We use this combined approach to find the optimal configuration and weights of a fully-connected neural network, resulting in regression models used to characterize the process of obtained bricks using silicon-based materials. Installations in the raw ceramics industry, i.e., bricks, are characterized by significant energy consumption and large quantities of emissions. Thus, the purpose of our approach is to determine by simulation the working conditions, including the manufacturing mix recipe with the addition of different materials, to minimize the emissions represented by CO and CH4. Our approach determines regression models which perform significantly better than those found using the traditional ICA for the aforementioned problem, resulting in better convergence and a substantially lower error.

Keywords: optimization, biologically inspired algorithm, regression models, bricks, emissions

Procedia PDF Downloads 84
12802 Epidemiological and Clinical Characteristics of Five Rare Pathological Subtypes of Hepatocellular Carcinoma

Authors: Xiaoyuan Chen

Abstract:

Background: This study aimed to characterize the epidemiological and clinical features of five rare subtypes of hepatocellular carcinoma (HCC) and to create a competing risk nomogram for predicting cancer-specific survival. Methods: This study used the Surveillance, Epidemiology, and End Results database to analyze the clinicopathological data of 50,218 patients with classic HCC and five rare subtypes (ICD-O-3 Histology Code=8170/3-8175/3) between 2004 and 2018. The annual percent change (APC) was calculated using Joinpoint regression, and a nomogram was developed based on multivariable competing risk survival analyses. The prognostic performance of the nomogram was evaluated using the Akaike information criterion, Bayesian information criterion, C-index, calibration curve, and area under the receiver operating characteristic curve. Decision curve analysis was used to assess the clinical value of the models. Results: The incidence of scirrhous carcinoma showed a decreasing trend (APC=-6.8%, P=0.025), while the morbidity of other rare subtypes remained stable from 2004 to 2018. The incidence-based mortality plateau in all subtypes during the period. Clear cell carcinoma was the most common subtype (n=551, 1.1%), followed by fibrolamellar (n=241, 0.5%), scirrhous (n=82, 0.2%), spindle cell (n=61, 0.1%), and pleomorphic (n=17, ~0%) carcinomas. Patients with fibrolamellar carcinoma were younger and more likely to have non-cirrhotic liver and better prognoses. Scirrhous carcinoma shared almost the same macro clinical characteristics and outcomes as classic HCC. Clear cell carcinoma tended to occur in the Asia-Pacific elderly male population, and more than half of them were large HCC (Size>5cm). Sarcomatoid (including spindle cell and pleomorphic) carcinoma was associated with larger tumor size, poorer differentiation, and more dismal prognoses. The pathological subtype, T stage, M stage, surgery, alpha-fetoprotein, and cancer history were identified as independent predictors in patients with rare subtypes. The nomogram showed good calibration, discrimination, and net benefits in clinical practice. Conclusion: The rare subtypes of HCC had distinct clinicopathological features and biological behaviors compared with classic HCC. Our findings could provide a valuable reference for clinicians. The constructed nomogram could accurately predict prognoses, which is beneficial for individualized management.

Keywords: hepatocellular carcinoma, pathological subtype, fibrolamellar carcinoma, scirrhous carcinoma, clear cell carcinoma, spindle cell carcinoma, pleomorphic carcinoma

Procedia PDF Downloads 80
12801 Construction Project Planning Using Fuzzy Critical Path Approach

Authors: Omar M. Aldenali

Abstract:

Planning is one of the most important phases of the management science and network planning, which represents the project activities relationship. Critical path is one of the project management techniques used to plan and control the execution of a project activities. The objective of this paper is to implement a fuzzy logic approach to arrange network planning on construction projects. This method is used to finding out critical path in the fuzzy construction project network. The trapezoidal fuzzy numbers are used to represent the activity construction project times. A numerical example that represents a house construction project is introduced. The critical path method is implemented on the fuzzy construction network activities, and the results showed that this method significantly affects the completion time of the construction projects.

Keywords: construction project, critical path, fuzzy network project, planning

Procedia PDF Downloads 144