Search results for: software process improvement
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 21684

Search results for: software process improvement

20184 Process Simulation of 1-Butene Separation from C4 Mixture by Extractive Distillation

Authors: Muhammad Naeem, Abdulrahman A. Al-Rabiah, Wasif Mughees

Abstract:

Technical mixture of C4 containing 1-butene and n-butane are very close to each other with regard to their boiling points i.e. -6.3°C for 1-butene and -1°C for n-butane. Extractive distillation process is used for the separation of 1-butene from the existing mixture of C4. The solvent is the essential of extractive distillation, and an appropriate solvent plays an important role in the process economy of extractive distillation. Aspen Plus has been applied for the separation of these hydrocarbons as a simulator. Moreover, NRTL activity coefficient model was used in the simulation. This model indicated that the material balances in this separation process were accurate for several solvent flow rates. Mixture of acetonitrile and water used as a solvent and 99% pure 1-butene was separated. This simulation proposed the ratio of the feed to solvent as 1: 7.9 and 15 plates for the solvent recovery column. Previously feed to solvent ratio was more than this and the number of proposed plates were 30, which shows that the separation process can be economized.

Keywords: extractive distillation, 1-butene, aspen plus, ACN solvent

Procedia PDF Downloads 518
20183 Artificial Neural Network Modeling and Genetic Algorithm Based Optimization of Hydraulic Design Related to Seepage under Concrete Gravity Dams on Permeable Soils

Authors: Muqdad Al-Juboori, Bithin Datta

Abstract:

Hydraulic structures such as gravity dams are classified as essential structures, and have the vital role in providing strong and safe water resource management. Three major aspects must be considered to achieve an effective design of such a structure: 1) The building cost, 2) safety, and 3) accurate analysis of seepage characteristics. Due to the complexity and non-linearity relationships of the seepage process, many approximation theories have been developed; however, the application of these theories results in noticeable errors. The analytical solution, which includes the difficult conformal mapping procedure, could be applied for a simple and symmetrical problem only. Therefore, the objectives of this paper are to: 1) develop a surrogate model based on numerical simulated data using SEEPW software to approximately simulate seepage process related to a hydraulic structure, 2) develop and solve a linked simulation-optimization model based on the developed surrogate model to describe the seepage occurring under a concrete gravity dam, in order to obtain optimum and safe design at minimum cost. The result shows that the linked simulation-optimization model provides an efficient and optimum design of concrete gravity dams.

Keywords: artificial neural network, concrete gravity dam, genetic algorithm, seepage analysis

Procedia PDF Downloads 212
20182 Access Control System for Big Data Application

Authors: Winfred Okoe Addy, Jean Jacques Dominique Beraud

Abstract:

Access control systems (ACs) are some of the most important components in safety areas. Inaccuracies of regulatory frameworks make personal policies and remedies more appropriate than standard models or protocols. This problem is exacerbated by the increasing complexity of software, such as integrated Big Data (BD) software for controlling large volumes of encrypted data and resources embedded in a dedicated BD production system. This paper proposes a general access control strategy system for the diffusion of Big Data domains since it is crucial to secure the data provided to data consumers (DC). We presented a general access control circulation strategy for the Big Data domain by describing the benefit of using designated access control for BD units and performance and taking into consideration the need for BD and AC system. We then presented a generic of Big Data access control system to improve the dissemination of Big Data.

Keywords: access control, security, Big Data, domain

Procedia PDF Downloads 118
20181 Identification and Selection of a Supply Chain Target Process for Re-Design

Authors: Jaime A. Palma-Mendoza

Abstract:

A supply chain consists of different processes and when conducting supply chain re-design is necessary to identify the relevant processes and select a target for re-design. A solution was developed which consists to identify first the relevant processes using the Supply Chain Operations Reference (SCOR) model, then to use Analytical Hierarchy Process (AHP) for target process selection. An application was conducted in an Airline MRO supply chain re-design project which shows this combination can clearly aid the identification of relevant supply chain processes and the selection of a target process for re-design.

Keywords: decision support systems, multiple criteria analysis, supply chain management

Procedia PDF Downloads 475
20180 Hyperchaos-Based Video Encryption for Device-To-Device Communications

Authors: Samir Benzegane, Said Sadoudi, Mustapha Djeddou

Abstract:

In this paper, we present a software development of video streaming encryption for Device-to-Device (D2D) communications by using Hyperchaos-based Random Number Generator (HRNG) implemented in C#. The software implements and uses the proposed HRNG to generate key stream for encrypting and decrypting real-time video data. The used HRNG consists of Hyperchaos Lorenz system which produces four signal outputs taken as encryption keys. The generated keys are characterized by high quality randomness which is confirmed by passing standard NIST statistical tests. Security analysis of the proposed encryption scheme confirms its robustness against different attacks.

Keywords: hyperchaos Lorenz system, hyperchaos-based random number generator, D2D communications, C#

Procedia PDF Downloads 349
20179 Locating the Best Place for Earthquake Refugee Camps by OpenSource Software: A Case Study for Tehran, Iran

Authors: Reyhaneh Saeedi

Abstract:

Iran is one of the regions which are most prone for earthquakes annually having a large number of financial and mortality and financial losses. Every year around the world, a large number of people lose their home and life due to natural disasters such as earthquakes. It is necessary to provide and specify some suitable places for settling the homeless people before the occurrence of the earthquake, one of the most important factors in crisis planning and management. Some of the natural disasters can be Modeling and shown by Geospatial Information System (GIS). By using GIS, it would be possible to manage the spatial data and reach several goals by making use of the analyses existing in it. GIS has a determining role in disaster management because it can determine the best places for temporary resettling after such a disaster. In this research QuantumGIS software is used that It is an OpenSource software so that easy to access codes and It is also free. In this system, AHP method is used as decision model and to locate the best places for temporary resettling, is done based on the related organizations criteria with their weights and buffers. Also in this research are made the buffer layers of criteria and change them to the raster layers. Later on, the raster layers are multiplied on desired weights then, the results are added together. Eventually, there are suitable places for resettling of victims by desired criteria by different colors with their optimum rate in QuantumGIS platform.

Keywords: disaster management, temporary resettlement, earthquake, QuantumGIS

Procedia PDF Downloads 384
20178 Plantar Neuro-Receptor Activation in Total Knee Arthroplasty Patients: Impact on Clinical Function, Pain, and Stiffness - A Randomized Controlled Trial

Authors: Woolfrey K., Woolfrey M., Bolton C. L., Warchuk D.

Abstract:

Objectives: Osteoarthritis is the most common joint disease of adults worldwide. Despite total knee arthroplasty (TKA) demonstrating high levels of success, 20% of patients report dissatisfaction with their result. VOXX Wellness Stasis Socks are embedded with a proprietary pattern of neuro-receptor activation points that have been proven to activate a precise neuro-response, according to the pattern theory of haptic perception, which stimulates improvements in pain and function. The use of this technology in TKA patients may prove beneficial as an adjunct to recovery as many patients suffer from deficits to their proprioceptive system caused by ligamentous damage and alterations to mechanoreceptors during the procedure. We hypothesized that VOXX Wellness Stasis Socks are a safe, cost-effective, and easily scalable strategy to support TKA patients through their recovery. Design: Double-blinded, placebo-controlled randomized trial. Participants: Patients scheduled to receive TKA were considered eligible for inclusion in the trial. Interventions: Intervention group (I): VOXX Wellness Stasis socks containing receptor point-activation technology. Control group (C): VOXX Wellness Stasis socks without receptor point-activation technology. Sock use during the waking hours x 6 weeks. Main Outcome Measures: Western Ontario McMaster Universities Osteoarthritis Index (WOMAC) questionnaire completed at baseline, 2 weeks, and 6 weeks to assess pain, stiffness, and physical function. Results: Data analysis using SPSS software. P-values, effect sizes, and confidence intervals are reported to assess clinical relevance of the finding. Physical status classifications were compared using t-test. Within-subject and between-subject differences in the mean WOMAC were analyzed by ANOVA. Effect size was analyzed using Cramer’s V. Consistent improvement in WOMAC scores for pain and stiffness at 2 weeks post op in the I over the C group. The womac scores assessing physical function showed a consistent improvement at both 2 and 6 weeks post op in the I group compared to C group. Conclusions: VOXX proved to be a low cost, safe intervention in TKA to help patients improve with regard to pain, stiffness, and physical function. Disclosures: None

Keywords: osteoarthritis, RCT, pain management, total knee arthroplasty

Procedia PDF Downloads 516
20177 On the Exergy Analysis of the Aluminum Smelter

Authors: Ayoola T. Brimmo, Mohamed I. Hassan

Abstract:

The push to mitigate the aluminum smelting industry’s enormous energy consumption and high emission releases is now even more persistent with the recent climate change happenings. Common approaches to achieve this have been focused on improving energy efficiency in the pot line and cast house sections of the smelter. However, the conventional energy efficiency analyses are based on the first law of thermodynamics, which do not shed proper light on the smelter’s degradation of energy. This just gives a general idea of the furnace’s performance with no reference to locations where improvement is a possibility based on the second law of thermodynamics. In this study, we apply exergy analyses on the pot line and cast house sections of the smelter to identify the locality and causes of energy degradation. The exergy analyses, which are based on a real life smelter conditions, highlight the possible locations for technology improvement in a typical smelter. With this established, methods of minimizing the smelter’s exergy losses are assessed.

Keywords: exergy analysis, electrolytic cell, furnace, heat transfer

Procedia PDF Downloads 272
20176 Optimum Dispatching Rule in Solar Ingot-Wafer Manufacturing System

Authors: Wheyming Song, Hung-Hsiang Lin, Scott Lian

Abstract:

In this research, we investigate the optimal dispatching rule for machines and manpower allocation in the solar ingot-wafer systems. The performance of the method is measured by the sales profit for each dollar paid to the operators in a one week at steady-state. The decision variables are identification-number of machines and operators when each job is required to be served in each process. We propose a rule which is a function of operator’s ability, corresponding salary, and standing location while in the factory. The rule is named ‘Multi-nominal distribution dispatch rule’. The proposed rule performs better than many traditional rules including generic algorithm and particle swarm optimization. Simulation results show that the proposed Multi-nominal distribution dispatch rule improvement on the sales profit dramatically.

Keywords: dispatching, solar ingot, simulation, flexsim

Procedia PDF Downloads 287
20175 Contemporary Challenges in Public Relations in the Context of Globalization

Authors: Marine Kobalava, Eter Narimanishvili, Nino Grigolaia

Abstract:

The paper analyzes the contemporary problems of public relations in Georgia. The approaches to public attitudes towards the relationship with the population of the country are studied on a global scale, the importance of forming the concept of public relations in Georgia in terms of globalization is justified. The basic components of public relations are characterized by the RACE system, namely analyzing research, action, communication, evaluation. The main challenges of public relations are identified in the research process; taking into consideration the scope of globalization, the influence of social, economic, and political changes in Georgia on PR development are identified. The article discusses the public relations as the strategic management function that facilitates communication with the society, recognition of public interests, and their prediction. In addition, the feminization of the sector is considered to be the most important achievement of public relations in the modern world. The conclusion is that the feminization indicator of the field is an unconditional increase in the employment rates of women. In the paper, the problems of globalization and public relations in the industrial countries are studied, the directions of improvement of public relations with the background of peculiarities of different countries and globalization process are proposed. Public relations under globalization are assessed in accordance with the theory of benefits and requirements, and the requirements are classified according to informational, self-identification, integration, social interaction, and other types of signs. In the article, conclusions on the current challenges of public relations in Georgia are made, and the recommendations for their solution, taking into consideration globalization processes in the world, are proposed.

Keywords: public relations, globalization, RACE system, public relationship concept, feminization

Procedia PDF Downloads 150
20174 The Comparative Study of Binary Artifact Repository Managers

Authors: Evgeny Chugunnyy, Alena Gerasimova, Kirill Chernyavskiy, Alexander Krasnov

Abstract:

One of the primary component of Continuous deployment (CD) is a binary artifact repository — the place where artifacts are stored with metadata in a structured way. The binary artifact repository manager (BARM) is a software, which implements this repository logic and exposes a public application programming interface (API) for managing these artifacts. Almost every programming language ecosystem has its own artifact repository kind. During creating Artipie — BARM constructor and server, we analyzed and implemented a lot of different artifact repositories. In this paper we present criterias for comparing artifact repositories, and analyze the most popular repositories using these metrics. We also describe some of the notable features of different repositories. This paper aimed to help people who are creating, maintaining or optimizing software repository and CI tools.

Keywords: artifact, repository, continuous deployment, build automation, artifacts management

Procedia PDF Downloads 129
20173 Optimal Bayesian Chart for Controlling Expected Number of Defects in Production Processes

Authors: V. Makis, L. Jafari

Abstract:

In this paper, we develop an optimal Bayesian chart to control the expected number of defects per inspection unit in production processes with long production runs. We formulate this control problem in the optimal stopping framework. The objective is to determine the optimal stopping rule minimizing the long-run expected average cost per unit time considering partial information obtained from the process sampling at regular epochs. We prove the optimality of the control limit policy, i.e., the process is stopped and the search for assignable causes is initiated when the posterior probability that the process is out of control exceeds a control limit. An algorithm in the semi-Markov decision process framework is developed to calculate the optimal control limit and the corresponding average cost. Numerical examples are presented to illustrate the developed optimal control chart and to compare it with the traditional u-chart.

Keywords: Bayesian u-chart, economic design, optimal stopping, semi-Markov decision process, statistical process control

Procedia PDF Downloads 557
20172 Development of the Internal Educational Quality Assurance System of Suan Sunandha Rajabhat University

Authors: Nipawan Tharasak, Sajeewan Darbavasu

Abstract:

This research aims 1) to study the opinion, problems and obstacles to internal educational quality assurance system for individual and the university levels, 2) to propose an approach to the development of quality assurance system of Suan Sunandha Rajabhat University. A study of problems and obstacles to internal educational quality assurance system of the university conducted with sample group consisting of staff and quality assurance committee members of the year 2010. There were 152 respondents. 5 executives were interviewed. Tool used in the research was document analysis. The structure of the interview questions and questionnaires with 5-rate scale. Reliability was 0.981. Data analysis were percentage, mean and standard deviation with content analysis. Results can be divided into 3 main points: (1) The implementation of the internal quality assurance system of the university. It was found that in overall, input, process and output factors received high scores. Each item is considered, the preparation, planning, monitoring and evaluation. The results of evaluation to improve the reporting and improvement according to an evaluation received high scores. However, the process received an average score. (2) Problems and obstacles. It was found that the personnel responsible for the duty still lack understanding of indicators and criteria of the quality assurance. (3) Development approach: -Staff should be encouraged to develop a better understanding of the quality assurance system. -Database system for quality assurance should be developed. -The results and suggestions should be applied in the next year development planning.

Keywords: development system, internal quality assurance, education, educational quality assurance

Procedia PDF Downloads 279
20171 The Analysis of Different Classes of Weighted Fuzzy Petri Nets and Their Features

Authors: Yurii Bloshko, Oksana Olar

Abstract:

This paper presents the analysis of 6 different classes of Petri nets: fuzzy Petri nets (FPN), generalized fuzzy Petri nets (GFPN), parameterized fuzzy Petri nets (PFPN), T2GFPN, flexible generalized fuzzy Petri nets (FGFPN), binary Petri nets (BPN). These classes were simulated in the special software PNeS® for the analysis of its pros and cons on the example of models which are dedicated to the decision-making process of passenger transport logistics. The paper includes the analysis of two approaches: when input values are filled with the experts’ knowledge; when fuzzy expectations represented by output values are added to the point. These approaches fulfill the possibilities of triples of functions which are replaced with different combinations of t-/s-norms.

Keywords: fuzzy petri net, intelligent computational techniques, knowledge representation, triangular norms

Procedia PDF Downloads 129
20170 Wait-Optimized Scheduler Algorithm for Efficient Process Scheduling in Computer Systems

Authors: Md Habibur Rahman, Jaeho Kim

Abstract:

Efficient process scheduling is a crucial factor in ensuring optimal system performance and resource utilization in computer systems. While various algorithms have been proposed over the years, there are still limitations to their effectiveness. This paper introduces a new Wait-Optimized Scheduler (WOS) algorithm that aims to minimize process waiting time by dividing them into two layers and considering both process time and waiting time. The WOS algorithm is non-preemptive and prioritizes processes with the shortest WOS. In the first layer, each process runs for a predetermined duration, and any unfinished process is subsequently moved to the second layer, resulting in a decrease in response time. Whenever the first layer is free or the number of processes in the second layer is twice that of the first layer, the algorithm sorts all the processes in the second layer based on their remaining time minus waiting time and sends one process to the first layer to run. This ensures that all processes eventually run, optimizing waiting time. To evaluate the performance of the WOS algorithm, we conducted experiments comparing its performance with traditional scheduling algorithms such as First-Come-First-Serve (FCFS) and Shortest-Job-First (SJF). The results showed that the WOS algorithm outperformed the traditional algorithms in reducing the waiting time of processes, particularly in scenarios with a large number of short tasks with long wait times. Our study highlights the effectiveness of the WOS algorithm in improving process scheduling efficiency in computer systems. By reducing process waiting time, the WOS algorithm can improve system performance and resource utilization. The findings of this study provide valuable insights for researchers and practitioners in developing and implementing efficient process scheduling algorithms.

Keywords: process scheduling, wait-optimized scheduler, response time, non-preemptive, waiting time, traditional scheduling algorithms, first-come-first-serve, shortest-job-first, system performance, resource utilization

Procedia PDF Downloads 75
20169 The Impact of the Business Process Reengineering on the Practices of the Human Resources Management in the Franco Tunisian Company-Network

Authors: Nesrine Bougarech, Habib Affes

Abstract:

This research lays the emphasis on the business process reengineering (BPR) which consists in radically altering the organizational processes through the optimal use of information technology (IT) to attain major enhancements in terms of quality, performance and productivity. A survey of the business process reengineering (BPR) was carried out in three French groups and their subsidiaries in Tunisia. The data collected were qualitatively analyzed in an attempt to test the main indicators of the success of a business process reengineering project (BPR) and to compare the importance of these indicators in the context of France versus Tunisia. The study corroborates that the respect of the inherent principles of the business process reengineering (BPR) and the diversity of the human resources involved in the project can lead to better productivity, higher quality of the goods or services and lower cost. Additionally, our results mirror the extent to which the respect of the principles and the diversity of resources are more important in the French companies than in their Tunisian subsidiaries.

Keywords: business process reengineering (BPR), human resources management (HRM), information technology (IT), management

Procedia PDF Downloads 390
20168 Understanding Team Member Autonomy and Team Collaboration: A Qualitative Study

Authors: Ayşen Bakioğlu, Gökçen Seyra Çakır

Abstract:

This study aims to explore how research assistants who work in project teams experience team member autonomy and how they reconcile team member autonomy with team collaboration. The study utilizes snowball sampling. 20 research assistants who work the faculties of education in Marmara University and Yıldız Technical University have been interviewed. The analysis of data involves a content analysis MAXQDAPlus 11 which is a qualitative data analysis software is used as the data analysis tool. According to the findings of this study, emerging themes include team norm formation, team coordination management, the role of individual tasks in team collaboration, leadership distribution. According to the findings, interviewees experience team norm formation process in terms of processes, which pertain to task fulfillment, and processes, which pertain to the regulation of team dynamics. Team norm formation process instills a sense of responsibility amongst individual team members. Apart from that, the interviewees’ responses indicate that the realization of the obligation to work in a team contributes to the team norm formation process. The participants indicate that individual expectations are taken into consideration during the coordination of the team. The supervisor of the project team also has a crucial role in maintaining team collaboration. Coordination problems arise when an individual team member does not relate his/her academic field with the research topic of the project team. The findings indicate that the leadership distribution in the project teams involves two leadership processes: leadership distribution which is based on the processes that focus on individual team members and leadership distribution which is based on the processes that focus on team interaction. Apart from that, individual tasks serve as a facilitator of collaboration amongst team members. Interviewees also indicate that individual tasks also facilitate the expression of individuality.

Keywords: project teams in higher education, research assistant teams, team collaboration, team member autonomy

Procedia PDF Downloads 350
20167 A Tool to Represent People Approach to the Use of Pharmaceuticals and Related Criticality and Needs: A Territory Experience

Authors: Barbara Pittau, Piergiorgio Palla, Antonio Mastino

Abstract:

Communication is fundamental to health education. The proper use of medicinal products is a crucial aspect of the health of citizens that affects both safety and health care spending. Therefore, encouraging/promoting communication, concerning the importance of proper use of pharmaceuticals, has substantial implications in terms of individual health, health care, and health care system sustainability. In view of these considerations, in the context of two projects, one of which is still in progress, a relational database-backed web application named COLLABORAFARMACISOLA has been designed and developed as a tool to analyze and visualize how people approach the use of medicinal products, with the aim of improving and enhancing communication efficacy. The software application is being used to collect information (anonymously and voluntarily) from the citizens of Sardinia, an Italian region, regarding their knowledge, experiences, and opinions towards pharmaceuticals. This study that was conducted to date on thousand of interviewed people, has focused on different aspects such as: the treatment interruption and the "self-prescription” without medical consultation, the attention paid to reading the leaflets, the awareness of the economic value of the pharmaceuticals, the importance of avoiding the waste of medicinal products and the attitudes towards the use of generics. To this purpose, our software application provides a set of ad hoc parsing routines, to store information into the structure of a relational database and to process and visualize it through a set of interactive tools aimed to emphasize the findings and the insights obtained. The results of our preliminary analysis show the efficacy of the awareness plan and, at the same time, the criticality and the needs of the territory under examination. The ultimate goal of our study is to provide a contribution to the community by improving communication that can result in a benefit for public health in a context strictly connected to the reality of the territory.

Keywords: communication, pharmaceuticals, public health, relational database, tool, web application

Procedia PDF Downloads 122
20166 Automatic Algorithm for Processing and Analysis of Images from the Comet Assay

Authors: Yeimy L. Quintana, Juan G. Zuluaga, Sandra S. Arango

Abstract:

The comet assay is a method based on electrophoresis that is used to measure DNA damage in cells and has shown important results in the identification of substances with a potential risk to the human population as innumerable physical, chemical and biological agents. With this technique is possible to obtain images like a comet, in which the tail of these refers to damaged fragments of the DNA. One of the main problems is that the image has unequal luminosity caused by the fluorescence microscope and requires different processing to condition it as well as to know how many optimal comets there are per sample and finally to perform the measurements and determine the percentage of DNA damage. In this paper, we propose the design and implementation of software using Image Processing Toolbox-MATLAB that allows the automation of image processing. The software chooses the optimum comets and measuring the necessary parameters to detect the damage.

Keywords: artificial vision, comet assay, DNA damage, image processing

Procedia PDF Downloads 289
20165 Information Communication Technology Based Road Traffic Accidents’ Identification, and Related Smart Solution Utilizing Big Data

Authors: Ghulam Haider Haidaree, Nsenda Lukumwena

Abstract:

Today the world of research enjoys abundant data, available in virtually any field, technology, science, and business, politics, etc. This is commonly referred to as big data. This offers a great deal of precision and accuracy, supportive of an in-depth look at any decision-making process. When and if well used, Big Data affords its users with the opportunity to produce substantially well supported and good results. This paper leans extensively on big data to investigate possible smart solutions to urban mobility and related issues, namely road traffic accidents, its casualties, and fatalities based on multiple factors, including age, gender, location occurrences of accidents, etc. Multiple technologies were used in combination to produce an Information Communication Technology (ICT) based solution with embedded technology. Those technologies include principally Geographic Information System (GIS), Orange Data Mining Software, Bayesian Statistics, to name a few. The study uses the Leeds accident 2016 to illustrate the thinking process and extracts thereof a model that can be tested, evaluated, and replicated. The authors optimistically believe that the proposed model will significantly and smartly help to flatten the curve of road traffic accidents in the fast-growing population densities, which increases considerably motor-based mobility.

Keywords: accident factors, geographic information system, information communication technology, mobility

Procedia PDF Downloads 197
20164 Seismic Performance Evaluation of Existing Building Using Structural Information Modeling

Authors: Byungmin Cho, Dongchul Lee, Taejin Kim, Minhee Lee

Abstract:

The procedure for the seismic retrofit of existing buildings includes the seismic evaluation. In the evaluation step, it is assessed whether the buildings have satisfactory performance against seismic load. Based on the results of that, the buildings are upgraded. To evaluate seismic performance of the buildings, it usually goes through the model transformation from elastic analysis to inelastic analysis. However, when the data is not delivered through the interwork, engineers should manually input the data. In this process, since it leads to inaccuracy and loss of information, the results of the analysis become less accurate. Therefore, in this study, the process for the seismic evaluation of existing buildings using structural information modeling is suggested. This structural information modeling makes the work economic and accurate. To this end, it is determined which part of the process could be computerized through the investigation of the process for the seismic evaluation based on ASCE 41. The structural information modeling process is developed to apply to the seismic evaluation using Perform 3D program usually used for the nonlinear response history analysis. To validate this process, the seismic performance of an existing building is investigated.

Keywords: existing building, nonlinear analysis, seismic performance, structural information modeling

Procedia PDF Downloads 367
20163 Performance Comparison and Visualization of COMSOL Multiphysics, Matlab, and Fortran for Predicting the Reservoir Pressure on Oil Production in a Multiple Leases Reservoir with Boundary Element Method

Authors: N. Alias, W. Z. W. Muhammad, M. N. M. Ibrahim, M. Mohamed, H. F. S. Saipol, U. N. Z. Ariffin, N. A. Zakaria, M. S. Z. Suardi

Abstract:

This paper presents the performance comparison of some computation software for solving the boundary element method (BEM). BEM formulation is the numerical technique and high potential for solving the advance mathematical modeling to predict the production of oil well in arbitrarily shaped based on multiple leases reservoir. The limitation of data validation for ensuring that a program meets the accuracy of the mathematical modeling is considered as the research motivation of this paper. Thus, based on this limitation, there are three steps involved to validate the accuracy of the oil production simulation process. In the first step, identify the mathematical modeling based on partial differential equation (PDE) with Poisson-elliptic type to perform the BEM discretization. In the second step, implement the simulation of the 2D BEM discretization using COMSOL Multiphysic and MATLAB programming languages. In the last step, analyze the numerical performance indicators for both programming languages by using the validation of Fortran programming. The performance comparisons of numerical analysis are investigated in terms of percentage error, comparison graph and 2D visualization of pressure on oil production of multiple leases reservoir. According to the performance comparison, the structured programming in Fortran programming is the alternative software for implementing the accurate numerical simulation of BEM. As a conclusion, high-level language for numerical computation and numerical performance evaluation are satisfied to prove that Fortran is well suited for capturing the visualization of the production of oil well in arbitrarily shaped.

Keywords: performance comparison, 2D visualization, COMSOL multiphysic, MATLAB, Fortran, modelling and simulation, boundary element method, reservoir pressure

Procedia PDF Downloads 476
20162 A Model for Analyzing the Startup Dynamics of a Belt Transmission Driven by a DC Motor

Authors: Giovanni Incerti

Abstract:

In this paper the dynamic behavior of a synchronous belt drive during start-up is analyzed and discussed. Besides considering the belt elasticity, the mathematical model here proposed also takes into consideration the electrical behaviour of the DC motor. The solution of the motion equations is obtained by means of the modal analysis in state space, which allows to obtain the decoupling of all equations of the mathematical model without introducing the hypothesis of proportional damping. The mathematical model of the transmission and the solution algorithms have been implemented within a computing software that allows the user to simulate the dynamics of the system and to evaluate the effects due to the elasticity of the belt branches and to the electromagnetic behavior of the DC motor. In order to show the details of the calculation procedure, the paper presents a case study developed with the aid of the abovementioned software.

Keywords: belt drive, vibrations, startup, DC motor

Procedia PDF Downloads 557
20161 Understanding Tacit Knowledge and DIKW

Authors: Bahadir Aydin

Abstract:

Today it is difficult to reach accurate knowledge because of mass data. This huge data makes the environment more and more caotic. Data is a main piller of intelligence. There is a close tie between knowledge and intelligence. Information gathered from different sources can be modified, interpreted and classified by using knowledge development process. This process is applied in order to attain intelligence. Within this process the effect of knowledge is crucial. Knowledge is classified as explicit and tacit knowledge. Tacit knowledge can be seen as "only the tip of the iceberg”. This tacit knowledge accounts for much more than we guess in all intelligence cycle. If the concept of intelligence scrutinized, it can be seen that it contains risks, threats as well as success. The main purpose for all organization is to be succesful by eliminating risks and threats. Therefore, there is a need to connect or fuse existing information and the processes which can be used to develop it. By the help of process the decision-maker can be presented with a clear holistic understanding, as early as possible in the decision making process. Planning, execution and assessments are the key functions that connects to information to knowledge. Altering from the current traditional reactive approach to a proactive knowledge development approach would reduce extensive duplication of work in the organization. By new approach to this process, knowledge can be used more effectively.

Keywords: knowledge, intelligence cycle, tacit knowledge, KIDW

Procedia PDF Downloads 503
20160 The Extended Skew Gaussian Process for Regression

Authors: M. T. Alodat

Abstract:

In this paper, we propose a generalization to the Gaussian process regression(GPR) model called the extended skew Gaussian process for regression(ESGPr) model. The ESGPR model works better than the GPR model when the errors are skewed. We derive the predictive distribution for the ESGPR model at a new input. Also we apply the ESGPR model to FOREX data and we find that it fits the Forex data better than the GPR model.

Keywords: extended skew normal distribution, Gaussian process for regression, predictive distribution, ESGPr model

Procedia PDF Downloads 534
20159 Study on Adding Story and Seismic Strengthening of Old Masonry Buildings

Authors: Youlu Huang, Huanjun Jiang

Abstract:

A large number of old masonry buildings built in the last century still remain in the city. It generates the problems of unsafety, obsolescence, and non-habitability. In recent years, many old buildings have been reconstructed through renovating façade, strengthening, and adding floors. However, most projects only provide a solution for a single problem. It is difficult to comprehensively solve problems of poor safety and lack of building functions. Therefore, a comprehensive functional renovation program of adding reinforced concrete frame story at the bottom via integrally lifting the building and then strengthening the building was put forward. Based on field measurement and YJK calculation software, the seismic performance of an actual three-story masonry structure in Shanghai was identified. The results show that the material strength of masonry is low, and the bearing capacity of some masonry walls could not meet the code requirements. The elastoplastic time history analysis of the structure was carried out by using SAP2000 software. The results show that under the 7 degrees rare earthquake, the seismic performance of the structure reaches 'serious damage' performance level. Based on the code requirements of the stiffness ration of the bottom frame (lateral stiffness ration of the transition masonry story and frame story), the bottom frame story was designed. The integral lifting process of the masonry building was introduced based on many engineering examples. The reinforced methods for the bottom frame structure strengthened by the steel-reinforced mesh mortar surface layer (SRMM) and base isolators, respectively, were proposed. The time history analysis of the two kinds of structures, under the frequent earthquake, the fortification earthquake, and the rare earthquake, was conducted by SAP2000 software. For the bottom frame structure, the results show that the seismic response of the masonry floor is significantly reduced after reinforced by the two methods compared to the masonry structure. The previous earthquake disaster indicated that the bottom frame is vulnerable to serious damage under a strong earthquake. The analysis results showed that under the rare earthquake, the inter-story displacement angle of the bottom frame floor meets the 1/100 limit value of the seismic code. The inter-story drift of the masonry floor for the base isolated structure under different levels of earthquakes is similar to that of structure with SRMM, while the base-isolated program is better to protect the bottom frame. Both reinforced methods could significantly improve the seismic performance of the bottom frame structure.

Keywords: old buildings, adding story, seismic strengthening, seismic performance

Procedia PDF Downloads 112
20158 Simulation of the Reactive Rotational Molding Using Smoothed Particle Hydrodynamics

Authors: A. Hamidi, S. Khelladi, L. Illoul, A. Tcharkhtchi

Abstract:

Reactive rotational molding (RRM) is a process to manufacture hollow plastic parts with reactive material has several advantages compared to conventional roto molding of thermoplastic powders: process cycle time is shorter; raw material is less expensive because polymerization occurs during processing and high-performance polymers may be used such as thermosets, thermoplastics or blends. However, several phenomena occur during this process which makes the optimization of the process quite complex. In this study, we have used a mixture of isocyanate and polyol as a reactive system. The chemical transformation of this system to polyurethane has been studied by thermal analysis and rheology tests. Thanks to these results of the curing process and rheological measurements, the kinetic and rheokinetik of polyurethane was identified. Smoothed Particle Hydrodynamics, a Lagrangian meshless method, was chosen to simulate reactive fluid flow in 2 and 3D configurations of the polyurethane during the process taking into account the chemical, and chemiorehological results obtained experimentally in this study.

Keywords: reactive rotational molding, simulation, smoothed particle hydrodynamics, surface tension, rheology, free surface flows, viscoelastic, interpolation

Procedia PDF Downloads 274
20157 Buoyancy Effects in Pressure Retarded Osmosis with Extremely High Draw Solution Concentration

Authors: Ivonne Tshuma, Ralf Cord-Ruwisch, Wendell Ela

Abstract:

Water crisis is a world-wide problem because of population growth and climate change. Hence, desalination is a solution to water scarcity, which threatens the world. Reverse osmosis (RO) is the most used technique for desalination; unfortunately, this process, usually requires high-pressure requirement hence requires a lot of energy about 3 – 5.5 KWhr/m³ of electrical energy. The pressure requirements of RO can be alleviated by the use of PRO (pressure retarded osmosis) to drive the RO process. This paper proposes a process of utilizing the energy directly from PRO to drive an RO process. The paper mostly analyses the PRO process parameters such as cross-flow velocity, density, and buoyancy and how these have an effect on PRO hence ultimately the RO process. The experimental study of the PRO with various feed solution concentrations and cross-flow velocities at fixed applied pressure with different orientations of the PRO cell was performed. The study revealed that without cross-flow velocity, buoyancy effects were observed but not with cross-flow velocity.

Keywords: cross-flow velocity, pressure retarded osmosis, density, buoyancy

Procedia PDF Downloads 128
20156 A Regression Model for Predicting Sugar Crystal Size in a Fed-Batch Vacuum Evaporative Crystallizer

Authors: Sunday B. Alabi, Edikan P. Felix, Aniediong M. Umo

Abstract:

Crystal size distribution is of great importance in the sugar factories. It determines the market value of granulated sugar and also influences the cost of production of sugar crystals. Typically, sugar is produced using fed-batch vacuum evaporative crystallizer. The crystallization quality is examined by crystal size distribution at the end of the process which is quantified by two parameters: the average crystal size of the distribution in the mean aperture (MA) and the width of the distribution of the coefficient of variation (CV). Lack of real-time measurement of the sugar crystal size hinders its feedback control and eventual optimisation of the crystallization process. An attractive alternative is to use a soft sensor (model-based method) for online estimation of the sugar crystal size. Unfortunately, the available models for sugar crystallization process are not suitable as they do not contain variables that can be measured easily online. The main contribution of this paper is the development of a regression model for estimating the sugar crystal size as a function of input variables which are easy to measure online. This has the potential to provide real-time estimates of crystal size for its effective feedback control. Using 7 input variables namely: initial crystal size (Lo), temperature (T), vacuum pressure (P), feed flowrate (Ff), steam flowrate (Fs), initial super-saturation (S0) and crystallization time (t), preliminary studies were carried out using Minitab 14 statistical software. Based on the existing sugar crystallizer models, and the typical ranges of these 7 input variables, 128 datasets were obtained from a 2-level factorial experimental design. These datasets were used to obtain a simple but online-implementable 6-input crystal size model. It seems the initial crystal size (Lₒ) does not play a significant role. The goodness of the resulting regression model was evaluated. The coefficient of determination, R² was obtained as 0.994, and the maximum absolute relative error (MARE) was obtained as 4.6%. The high R² (~1.0) and the reasonably low MARE values are an indication that the model is able to predict sugar crystal size accurately as a function of the 6 easy-to-measure online variables. Thus, the model can be used as a soft sensor to provide real-time estimates of sugar crystal size during sugar crystallization process in a fed-batch vacuum evaporative crystallizer.

Keywords: crystal size, regression model, soft sensor, sugar, vacuum evaporative crystallizer

Procedia PDF Downloads 196
20155 Timely Screening for Palliative Needs in Ambulatory Oncology

Authors: Jaci Mastrandrea

Abstract:

Background: The National Comprehensive Cancer Network (NCCN) recommends that healthcare institutions have established processes for integrating palliative care (PC) into cancer treatment and that all cancer patients be screened for PC needs upon initial diagnosis as well as throughout the entire continuum of care (National Comprehensive Cancer Network, 2021). Early PC screening is directly correlated with improved patient outcomes. The Sky Lakes Cancer Treatment Center (SLCTC) is an institution that has access to PC services yet does not have protocols in place for identifying patients with palliative needs or a standardized referral process. The aim of this quality improvement project is to improve early access to PC services by establishing a standardized screening and referral process for outpatient oncology patients. Method: The sample population included all adult patients with an oncology diagnosis who presented to the SLCTC for treatment during the project timeline from March 15th, 2022, to April 29th, 2022. The “Palliative and Supportive Needs Assessment'' (PSNA) screening tool was developed from validated and evidence-based PC referral criteria. The tool was initially implemented using paper forms and later was integrated into the Epic-Beacon EHR system. Patients were screened by registered nurses on the SLCTC treatment team. Nurses responsible for screening patients received an educational inservice prior to implementation. Patients with a PSNA score of three or higher were considered to be a positive screen. Scores of five or higher triggered a PC referral order in the patient’s EHR for the oncologist to review and approve. All patients with a positive screen received an educational handout on the topic of PC, and the EHR was flagged for follow-up. Results: Prior to implementation of the PSCNA screening tool, the SLCTC had zero referrals to PC in the past year, excluding referrals to hospice. Data was collected from the first 100 patient screenings completed within the eight-week data collection period. Seventy-three percent of patients met criteria for PC referral with a score greater than or equal to three. Of those patients who met referral criteria, 53.4% (39 patients) were referred for a palliative and supportive care consultation. Patients that were not referred to PC upon meeting the criteria were flagged in the EHR for re-screening within one to three months. Patients with lung cancer, chronic hematologic malignancies, breast cancer, and gastrointestinal malignancy most frequently met criteria for PC referral and scored highest overall on the scale of 0-12. Conclusion: The implementation of a standardized PC screening tool at the SLCTC significantly increased awareness of PC needs among cancer patients in the outpatient setting. Additionally, data derived from this quality improvement project supports the national recommendation for PC to be an integral component of cancer treatment across the entire continuum of care.

Keywords: oncology, palliative care, symptom management, symptom screening, ambulatory oncology, cancer, supportive care

Procedia PDF Downloads 66