Search results for: Stochastic Branching Process
14877 Online Monitoring Rheological Property of Polymer Melt during Injection Molding
Authors: Chung-Chih Lin, Chien-Liang Wu
Abstract:
The detection of the polymer melt state during manufacture process is regarded as an efficient way to control the molded part quality in advance. Online monitoring rheological property of polymer melt during processing procedure provides an approach to understand the melt state immediately. Rheological property reflects the polymer melt state at different processing parameters and is very important in injection molding process especially. An approach that demonstrates how to calculate rheological property of polymer melt through in-process measurement, using injection molding as an example, is proposed in this study. The system consists of two sensors and a data acquisition module can process the measured data, which are used for the calculation of rheological properties of polymer melt. The rheological properties of polymer melt discussed in this study include shear rate and viscosity which are investigated with respect to injection speed and melt temperature. The results show that the effect of injection speed on the rheological properties is apparent, especially for high melt temperature and should be considered for precision molding process.Keywords: injection molding, melt viscosity, shear rate, monitoring
Procedia PDF Downloads 37914876 CMMI Key Process Areas and FDD Practices
Authors: Rituraj Deka, Nomi Baruah
Abstract:
The development of information technology during the past few years resulted in designing of more and more complex software. The outsourcing of software development makes a higher requirement for the management of software development project. Various software enterprises follow various paths in their pursuit of excellence, applying various principles, methods and techniques along the way. The new research is proving that CMMI and Agile methodologies can benefit from using both methods within organizations with the potential to dramatically improve business performance. The paper describes a mapping between CMMI key process areas (KPAs) and Feature-Driven Development (FDD) communication perspective, so as to increase the understanding of how improvements can be made in the software development process.Keywords: Agile, CMMI, FDD, KPAs
Procedia PDF Downloads 45614875 Interior Design Pedagogy in the 21st Century: Personalised Design Process
Authors: Roba Zakariah Shaheen
Abstract:
In the 21st-century Interior, design pedagogy has developed rapidly due to social and economical factors. Socially, this paper presents research findings that shows a significant relationship between educators and students in interior design education. It shows that students’ personal traits, design process, and thinking process are significantly interrelated. Constructively, this paper presented how personal traits can guide educators in the interior design education domain to develop students’ thinking process. In the same time, it demonstrated how students should use their own personal traits to create their own design process. Constructivism was the theory underneath this research, as it supports the grounded theory, which is the methodological approach of this research. Moreover, Mayer’s Briggs Type Indicator strategy was used to investigate the personality traits scientifically, as a psychological strategy that related to cognitive ability. Conclusions from this research strongly recommends that educators and students should utilize their personal traits to foster interior design education.Keywords: interior design, pedagogy, constructivism, grounded theory, personality traits, creativity
Procedia PDF Downloads 20514874 Transfer Knowledge From Multiple Source Problems to a Target Problem in Genetic Algorithm
Authors: Terence Soule, Tami Al Ghamdi
Abstract:
To study how to transfer knowledge from multiple source problems to the target problem, we modeled the Transfer Learning (TL) process using Genetic Algorithms as the model solver. TL is the process that aims to transfer learned data from one problem to another problem. The TL process aims to help Machine Learning (ML) algorithms find a solution to the problems. The Genetic Algorithms (GA) give researchers access to information that we have about how the old problem is solved. In this paper, we have five different source problems, and we transfer the knowledge to the target problem. We studied different scenarios of the target problem. The results showed combined knowledge from multiple source problems improves the GA performance. Also, the process of combining knowledge from several problems results in promoting diversity of the transferred population.Keywords: transfer learning, genetic algorithm, evolutionary computation, source and target
Procedia PDF Downloads 13814873 Residual Life Estimation Based on Multi-Phase Nonlinear Wiener Process
Authors: Hao Chen, Bo Guo, Ping Jiang
Abstract:
Residual life (RL) estimation based on multi-phase nonlinear Wiener process was studied in this paper, which is significant for complicated products with small samples. Firstly, nonlinear Wiener model with random parameter was introduced and multi-phase nonlinear Wiener model was proposed to model degradation process of products that were nonlinear and separated into different phases. Then the multi-phase RL probability density function based on the presented model was derived approximately in a closed form and parameters estimation was achieved with the method of maximum likelihood estimation (MLE). Finally, the method was applied to estimate the RL of high voltage plus capacitor. Compared with the other three different models by log-likelihood function (Log-LF) and Akaike information criterion (AIC), the results show that the proposed degradation model can capture degradation process of high voltage plus capacitors in a better way and provide a more reliable result.Keywords: multi-phase nonlinear wiener process, residual life estimation, maximum likelihood estimation, high voltage plus capacitor
Procedia PDF Downloads 45214872 Techno-Economic Assessment of Aluminum Waste Management
Authors: Hamad Almohamadi, Abdulrahman AlKassem, Majed Alamoudi
Abstract:
Dumping Aluminum (Al) waste into landfills causes several health and environmental problems. The pyrolysis process could treat Al waste to produce AlCl₃ and H₂. Using the Aspen Plus software, a techno-economic and feasibility assessment has been performed for Al waste pyrolysis. The Aspen Plus simulation was employed to estimate the plant's mass and energy balance, which was assumed to process 100 dry metric tons of Al waste per day. This study looked at two cases of Al waste treatment. The first case produces 355 tons of AlCl₃ per day and 9 tons of H₂ per day without recycling. The conversion rate must be greater than 50% in case 1 to make a profit. In this case, the MSP for AlCl₃ is $768/ton. The plant would generate $25 million annually if the AlCl₃ were sold at $1000 per ton. In case 2 with recycling, the conversion has less impact on the plant's profitability than in case 1. Moreover, compared to case 1, the MSP of AlCl₃ has no significant influence on process profitability. In this scenario, if AlCl₃ were sold at $1000/ton, the process profit would be $58 million annually. Case 2 is better than case 1 because recycling Al generates a higher yield than converting it to AlCl₃ and H₂.Keywords: aluminum waste, aspen plus, process modelling, fast pyrolysis, techno-economic assessment
Procedia PDF Downloads 9014871 The Effect of Tacit Knowledge for Intelligence Cycle
Authors: Bahadir Aydin
Abstract:
It is difficult to access accurate knowledge because of mass data. This huge data make environment more and more caotic. Data are main piller of intelligence. The affiliation between intelligence and knowledge is quite significant to understand underlying truths. The data gathered from different sources can be modified, interpreted and classified by using intelligence cycle process. This process is applied in order to progress to wisdom as well as intelligence. Within this process the effect of tacit knowledge is crucial. Knowledge which is classified as explicit and tacit knowledge is the key element for any purpose. Tacit knowledge can be seen as "the tip of the iceberg”. This tacit knowledge accounts for much more than we guess in all intelligence cycle. If the concept of intelligence cycle is scrutinized, it can be seen that it contains risks, threats as well as success. The main purpose of all organizations is to be successful by eliminating risks and threats. Therefore, there is a need to connect or fuse existing information and the processes which can be used to develop it. Thanks to this process the decision-makers can be presented with a clear holistic understanding, as early as possible in the decision making process. Altering from the current traditional reactive approach to a proactive intelligence cycle approach would reduce extensive duplication of work in the organization. Applying new result-oriented cycle and tacit knowledge intelligence can be procured and utilized more effectively and timely.Keywords: information, intelligence cycle, knowledge, tacit Knowledge
Procedia PDF Downloads 51114870 Taguchi-Based Surface Roughness Optimization for Slotted and Tapered Cylindrical Products in Milling and Turning Operations
Authors: Vineeth G. Kuriakose, Joseph C. Chen, Ye Li
Abstract:
The research follows a systematic approach to optimize the parameters for parts machined by turning and milling processes. The quality characteristic chosen is surface roughness since the surface finish plays an important role for parts that require surface contact. A tapered cylindrical surface is designed as a test specimen for the research. The material chosen for machining is aluminum alloy 6061 due to its wide variety of industrial and engineering applications. HAAS VF-2 TR computer numerical control (CNC) vertical machining center is used for milling and HAAS ST-20 CNC machine is used for turning in this research. Taguchi analysis is used to optimize the surface roughness of the machined parts. The L9 Orthogonal Array is designed for four controllable factors with three different levels each, resulting in 18 experimental runs. Signal to Noise (S/N) Ratio is calculated for achieving the specific target value of 75 ± 15 µin. The controllable parameters chosen for turning process are feed rate, depth of cut, coolant flow and finish cut and for milling process are feed rate, spindle speed, step over and coolant flow. The uncontrollable factors are tool geometry for turning process and tool material for milling process. Hypothesis testing is conducted to study the significance of different uncontrollable factors on the surface roughnesses. The optimal parameter settings were identified from the Taguchi analysis and the process capability Cp and the process capability index Cpk were improved from 1.76 and 0.02 to 3.70 and 2.10 respectively for turning process and from 0.87 and 0.19 to 3.85 and 2.70 respectively for the milling process. The surface roughnesses were improved from 60.17 µin to 68.50 µin, reducing the defect rate from 52.39% to 0% for the turning process and from 93.18 µin to 79.49 µin, reducing the defect rate from 71.23% to 0% for the milling process. The purpose of this study is to efficiently utilize the Taguchi design analysis to improve the surface roughness.Keywords: surface roughness, Taguchi parameter design, CNC turning, CNC milling
Procedia PDF Downloads 15514869 Exploring Open Process Innovation: Insights from a Systematic Review and Framework Development
Authors: Saeed Nayeri
Abstract:
This paper explores the feasibility of openness within firms' boundaries during process innovation and identifies the key determinants of open process innovation (OPI). Through a systematic review of 78 research studies published between 2001 and 2024, the author synthesized diverse findings into a comprehensive framework detailing OPI attributes and pillars. The identified OPI attributes encompass themes such as technology intensity, significance, magnitude, and locus of exploitation, while the OPI pillars include mechanisms, partners, achievements, and antecedents. Additionally, the author critically analysed gaps in the literature, proposing future research directions that advocate for a broader methodological approach, increased emphasis on theory development and testing, and more cross-national and cross-sectoral studies to advance understanding in this field.Keywords: open innovation, process innovation, OPI attributes, systematic literature review, organizational openness
Procedia PDF Downloads 6714868 D6tions: A Serious Game to Learn Software Engineering Process and Design
Authors: Hector G. Perez-Gonzalez, Miriam Vazquez-Escalante, Sandra E. Nava-Muñoz, Francisco E. Martinez-Perez, Alberto S. Nunez-Varela
Abstract:
The software engineering teaching process has been the subject of many studies. To improve this process, researchers have proposed merely illustrative techniques in the classroom, such as topic presentations and dynamics between students on one side or attempts to involve students in real projects with companies and institutions to bring them to a real software development problem on the other hand. Simulators and serious games have been used as auxiliary tools to introduce students to topics that are too abstract when these are presented in the traditional way. Most of these tools cover a limited area of the huge software engineering scope. To address this problem, we have developed D6tions, an educational serious game that simulates the software engineering process and is designed to experiment the different stages a software engineer (playing roles as project leader or as a developer or designer) goes through, while participating in a software project. We describe previous approaches to this problem, how D6tions was designed, its rules, directions, and the results we obtained of the use of this game involving undergraduate students playing the game.Keywords: serious games, software engineering, software engineering education, software engineering teaching process
Procedia PDF Downloads 49214867 Colonialism and Modernism in Architecture, the Case of a Blank Page Opportunity in Casablanka
Authors: Nezha Alaoui
Abstract:
The early 1950s French colonial context in Morocco provided an opportunity for architects to question the modernist established order by building dwellings for the local population. The dwellings were originally designed to encourage Muslims to adopt an urban lifestyle based on local customs. However, the inhabitants transformed their dwelling into a hybrid habitation. This paper aims to prove the relevance of the design process in accordance with the local colonial context by analyzing the dwellers' appropriation process and the modification of their habitat.Keywords: colonial heritage, appropriation process, islamic spatial habit, housing experiment, modernist mass housing
Procedia PDF Downloads 12614866 Science Process Skill and Interest Preschooler in Learning Early Science through Mobile Application
Authors: Seah Siok Peh, Hashimah Mohd Yunus, Nor Hashimah Hashim, Mariam Mohamad
Abstract:
A country needs a workforce that encompasses knowledge, skilled labourers to generate innovation, productivity and being able to solve problems creatively via technology. Science education experts believe that the mastery of science skills help preschoolers to generate such knowledge on scientific concepts by providing constructive experiences. Science process skills are skills used by scientists to study or investigate a problem, issue, problem or phenomenon of science. In line with the skills used by scientists. The purpose of this study is to investigate the basic science process skill and interest in learning early science through mobile application. This study aimed to explore six spesific basic science process skills by the use of a mobile application as a learning support tool. The descriptive design also discusses on the extent of the use of mobile application in improving basic science process skill in young children. This study consists of six preschoolers and two preschool teachers from two different classes located in Perak, Malaysia. Techniques of data collection are inclusive of observations, interviews and document analysis. This study will be useful to provide information and give real phenomena to policy makers especially Ministry of education in Malaysia.Keywords: science education, basic science process skill, interest, early science, mobile application
Procedia PDF Downloads 24414865 Using Analytical Hierarchy Process and TOPSIS Approaches in Designing a Finite Element Analysis Automation Program
Authors: Ming Wen, Nasim Nezamoddini
Abstract:
Sophisticated numerical simulations like finite element analysis (FEA) involve a complicated process from model setup to post-processing tasks that require replication of time-consuming steps. Utilizing FEA automation program simplifies the complexity of the involved steps while minimizing human errors in analysis set up, calculations, and results processing. One of the main challenges in designing FEA automation programs is to identify user requirements and link them to possible design alternatives. This paper presents a decision-making framework to design a Python based FEA automation program for modal analysis, frequency response analysis, and random vibration fatigue (RVF) analysis procedures. Analytical hierarchy process (AHP) and technique for order preference by similarity to ideal solution (TOPSIS) are applied to evaluate design alternatives considering the feedback received from experts and program users.Keywords: finite element analysis, FEA, random vibration fatigue, process automation, analytical hierarchy process, AHP, TOPSIS, multiple-criteria decision-making, MCDM
Procedia PDF Downloads 11014864 CFD Study on the Effect of Primary Air on Combustion of Simulated MSW Process in the Fixed Bed
Authors: Rui Sun, Tamer M. Ismail, Xiaohan Ren, M. Abd El-Salam
Abstract:
Incineration of municipal solid waste (MSW) is one of the key scopes in the global clean energy strategy. A computational fluid dynamics (CFD) model was established. In order to reveal these features of the combustion process in a fixed porous bed of MSW. Transporting equations and process rate equations of the waste bed were modeled and set up to describe the incineration process, according to the local thermal conditions and waste property characters. Gas phase turbulence was modeled using k-ε turbulent model and the particle phase was modeled using the kinetic theory of granular flow. The heterogeneous reaction rates were determined using Arrhenius eddy dissipation and the Arrhenius-diffusion reaction rates. The effects of primary air flow rate and temperature in the burning process of simulated MSW are investigated experimentally and numerically. The simulation results in bed are accordant with experimental data well. The model provides detailed information on burning processes in the fixed bed, which is otherwise very difficult to obtain by conventional experimental techniques.Keywords: computational fluid dynamics (CFD) model, waste incineration, municipal solid waste (MSW), fixed bed, primary air
Procedia PDF Downloads 40014863 Depyritization of US Coal Using Iron-Oxidizing Bacteria: Batch Stirred Reactor Study
Authors: Ashish Pathak, Dong-Jin Kim, Haragobinda Srichandan, Byoung-Gon Kim
Abstract:
Microbial depyritization of coal using chemoautotrophic bacteria is gaining acceptance as an efficient and eco-friendly technique. The process uses the metabolic activity of chemoautotrophic bacteria in removing sulfur and pyrite from the coal. The aim of the present study was to investigate the potential of Acidithiobacillus ferrooxidans in removing the pyritic sulfur and iron from high iron and sulfur containing US coal. The experiment was undertaken in 8 L bench scale stirred tank reactor having 1% (w/v) pulp density of coal. The reactor was operated at 35ºC and aerobic conditions were maintained by sparging the air into the reactor. It was found that at the end of bio-depyritization process, about 90% of pyrite and 67% of pyritic sulfur was removed from the coal. The results indicate that the bio-depyritization process is an efficient process in treating the high pyrite and sulfur containing coal.Keywords: At.ferrooxidans, batch reactor, coal desulfurization, pyrite
Procedia PDF Downloads 27414862 Quadrature Mirror Filter Bank Design Using Population Based Stochastic Optimization
Authors: Ju-Hong Lee, Ding-Chen Chung
Abstract:
The paper deals with the optimal design of two-channel linear-phase (LP) quadrature mirror filter (QMF) banks using a metaheuristic based optimization technique. Based on the theory of two-channel QMF banks using two recursive digital all-pass filters (DAFs), the design problem is appropriately formulated to result in an objective function which is a weighted sum of the group delay error of the designed QMF bank and the magnitude response error of the designed low-pass analysis filter. Through a frequency sampling and a weighted least squares approach, the optimization problem of the objective function can be solved by utilizing a particle swarm optimization algorithm. The resulting two-channel QMF banks can possess approximately LP response without magnitude distortion. Simulation results are presented for illustration and comparison.Keywords: quadrature mirror filter bank, digital all-pass filter, weighted least squares algorithm, particle swarm optimization
Procedia PDF Downloads 51814861 A Time-Varying and Non-Stationary Convolution Spectral Mixture Kernel for Gaussian Process
Authors: Kai Chen, Shuguang Cui, Feng Yin
Abstract:
Gaussian process (GP) with spectral mixture (SM) kernel demonstrates flexible non-parametric Bayesian learning ability in modeling unknown function. In this work a novel time-varying and non-stationary convolution spectral mixture (TN-CSM) kernel with a significant enhancing of interpretability by using process convolution is introduced. A way decomposing the SM component into an auto-convolution of base SM component and parameterizing it to be input dependent is outlined. Smoothly, performing a convolution between two base SM component yields a novel structure of non-stationary SM component with much better generalized expression and interpretation. The TN-CSM perfectly allows compatibility with the stationary SM kernel in terms of kernel form and spectral base ignored and confused by previous non-stationary kernels. On synthetic and real-world datatsets, experiments show the time-varying characteristics of hyper-parameters in TN-CSM and compare the learning performance of TN-CSM with popular and representative non-stationary GP.Keywords: Gaussian process, spectral mixture, non-stationary, convolution
Procedia PDF Downloads 19314860 Pakistan’s Taxation System: A Critical Appraisal
Authors: Khalid Javed, Rashid Mahmood
Abstract:
The constitution empowers the Federal Government to collect taxes on income other than agricultural income, taxes on capital value, customs, excise duties and sales taxes. The Central Board of Revenue (CBR) and its subordinate departments administer the tax system. Each of the three principal taxes has a different history and different set of issues. For a large number of income tax payers the core of the business process is pre-audit and assessment by a tax official. This process gives considerable discretion to tax officials, with potential for abuse. Moreover, this process is also not tenable as the number of taxpayers increase. The report is focused on a total overhaul of the process and organization of income tax. Sales tax is recent and its process and organization is adjusted to the needs of an expanding tax base. These are based on self-assessment and selective audit. Similarly, in customs the accent is on accelerating and broadening the changes begun in recent years. Before long, central excise will be subsumed in sales tax. During the nineties, despite many changes in the tax regime and introduction of withholding and presumptive taxes, Federal Government tax to GDP ratio has varied narrowly around eleven percent. The tax base has grown but still remains narrow and skewed. The number of income tax filers is around one million.Keywords: central board of revenue, GDP, sale tax, income tax
Procedia PDF Downloads 44014859 Designing Cultural-Creative Products with the Six Categories of Hanzi (Chinese Character Classification)
Authors: Pei-Jun Xue, Ming-Yu Hsiao
Abstract:
Chinese characters, or hanzi, represent a process of simplifying three-dimensional signs into plane signifiers. From pictograms at the beginning to logograms today, a Han linguist thus classified them into six categories known as the six categories of Chinese characters. Design is a process of signification, and cultural-creative design is a process translating ideas into design with creativity upon culture. Aiming to investigate the process of cultural-creative design transforming cultural text into cultural signs, this study analyzed existing cultural-creative products with the six categories of Chinese characters by treating such products as representations which accurately communicate the designer’s ideas to users through the categorization, simplification, and interpretation of sign features. This is a two-phase pilot study on designing cultural-creative products with the six categories of Chinese characters. Phase I reviews the related literature on the theory of the six categories of Chinese characters investigated and concludes with the process and principles of character evolution. Phase II analyzes the design of existing cultural-creative products with the six categories of Chinese characters and explores the conceptualization of product design.Keywords: six categories of Chinese characters, cultural-creative product design, cultural signs, cultural product
Procedia PDF Downloads 33714858 Bleaching Liquor Recovery of Batch-Wise and Continuous Method
Authors: Sidra Saleemi, Arsalan Khan, Urooj Baig, Tahir Jamil
Abstract:
In this research, it was examined that some residual amount of bleaching chemicals left in the liquor, this amount is more in Batch-wise process as compared to continuous process. These chemicals can be recovered and reused for bleaching by adding more quantity of fresh bleaching chemicals and water, this quantity will be required to balance the recipe for fabric. This liquor is recovered and samples were bleached with different modified recipe of liquor for both processes i.e. Batch-wise and continuous process. Every time good results were achieved with negligible variation in the quality parameter between the fabric bleached with fresh liquor and the fabric bleached with Recovered Liquor. Additionally, samples were dyed, and found that dyeing can be done easily on samples bleached with recover liquor.Keywords: bleaching process, hydrogen peroxide, sodium hydroxide, liquor recovery
Procedia PDF Downloads 36014857 Scoring System for the Prognosis of Sepsis Patients in Intensive Care Units
Authors: Javier E. García-Gallo, Nelson J. Fonseca-Ruiz, John F. Duitama-Munoz
Abstract:
Sepsis is a syndrome that occurs with physiological and biochemical abnormalities induced by severe infection and carries a high mortality and morbidity, therefore the severity of its condition must be interpreted quickly. After patient admission in an intensive care unit (ICU), it is necessary to synthesize the large volume of information that is collected from patients in a value that represents the severity of their condition. Traditional severity of illness scores seeks to be applicable to all patient populations, and usually assess in-hospital mortality. However, the use of machine learning techniques and the data of a population that shares a common characteristic could lead to the development of customized mortality prediction scores with better performance. This study presents the development of a score for the one-year mortality prediction of the patients that are admitted to an ICU with a sepsis diagnosis. 5650 ICU admissions extracted from the MIMICIII database were evaluated, divided into two groups: 70% to develop the score and 30% to validate it. Comorbidities, demographics and clinical information of the first 24 hours after the ICU admission were used to develop a mortality prediction score. LASSO (least absolute shrinkage and selection operator) and SGB (Stochastic Gradient Boosting) variable importance methodologies were used to select the set of variables that make up the developed score; each of this variables was dichotomized and a cut-off point that divides the population into two groups with different mean mortalities was found; if the patient is in the group that presents a higher mortality a one is assigned to the particular variable, otherwise a zero is assigned. These binary variables are used in a logistic regression (LR) model, and its coefficients were rounded to the nearest integer. The resulting integers are the point values that make up the score when multiplied with each binary variables and summed. The one-year mortality probability was estimated using the score as the only variable in a LR model. Predictive power of the score, was evaluated using the 1695 admissions of the validation subset obtaining an area under the receiver operating characteristic curve of 0.7528, which outperforms the results obtained with Sequential Organ Failure Assessment (SOFA), Oxford Acute Severity of Illness Score (OASIS) and Simplified Acute Physiology Score II (SAPSII) scores on the same validation subset. Observed and predicted mortality rates within estimated probabilities deciles were compared graphically and found to be similar, indicating that the risk estimate obtained with the score is close to the observed mortality, it is also observed that the number of events (deaths) is indeed increasing as the outcome go from the decile with the lowest probabilities to the decile with the highest probabilities. Sepsis is a syndrome that carries a high mortality, 43.3% for the patients included in this study; therefore, tools that help clinicians to quickly and accurately predict a worse prognosis are needed. This work demonstrates the importance of customization of mortality prediction scores since the developed score provides better performance than traditional scoring systems.Keywords: intensive care, logistic regression model, mortality prediction, sepsis, severity of illness, stochastic gradient boosting
Procedia PDF Downloads 22114856 Iranian Processed Cheese under Effect of Emulsifier Salts and Cooking Time in Process
Authors: M. Dezyani, R. Ezzati bbelvirdi, M. Shakerian, H. Mirzaei
Abstract:
Sodium Hexametaphosphate (SHMP) is commonly used as an Emulsifying Salt (ES) in process cheese, although rarely as the sole ES. It appears that no published studies exist on the effect of SHMP concentration on the properties of process cheese when pH is kept constant; pH is well known to affect process cheese functionality. The detailed interactions between the added phosphate, Casein (CN), and indigenous Ca phosphate are poorly understood. We studied the effect of the concentration of SHMP (0.25-2.75%) and holding time (0-20 min) on the textural and Rheological properties of pasteurized process Cheddar cheese using a central composite rotatable design. All cheeses were adjusted to pH 5.6. The meltability of process cheese (as indicated by the decrease in loss tangent parameter from small amplitude oscillatory rheology, degree of flow, and melt area from the Schreiber test) decreased with an increase in the concentration of SHMP. Holding time also led to a slight reduction in meltability. Hardness of process cheese increased as the concentration of SHMP increased. Acid-base titration curves indicated that the buffering peak at pH 4.8, which is attributable to residual colloidal Ca phosphate, was shifted to lower pH values with increasing concentration of SHMP. The insoluble Ca and total and insoluble P contents increased as concentration of SHMP increased. The proportion of insoluble P as a percentage of total (indigenous and added) P decreased with an increase in ES concentration because of some of the (added) SHMP formed soluble salts. The results of this study suggest that SHMP chelated the residual colloidal Ca phosphate content and dispersed CN; the newly formed Ca-phosphate complex remained trapped within the process cheese matrix, probably by cross-linking CN. Increasing the concentration of SHMP helped to improve fat emulsification and CN dispersion during cooking, both of which probably helped to reinforce the structure of process cheese.Keywords: Iranian processed cheese, emulsifying salt, rheology, texture
Procedia PDF Downloads 43014855 The Design of the Blended Learning System via E-Media and Online Learning for the Asynchronous Learning: Case Study of Process Management Subject
Authors: Pimploi Tirastittam, Suppara Charoenpoom
Abstract:
Nowadays the asynchronous learning has granted the permission to the anywhere and anything learning via the technology and E-media which give the learner more convenient. This research is about the design of the blended and online learning for the asynchronous learning of the process management subject in order to create the prototype of this subject asynchronous learning which will create the easiness and increase capability in the learning. The pattern of learning is the integration between the in-class learning and online learning via the internet. This research is mainly focused on the online learning and the online learning can be divided into 5 parts which are virtual classroom, online content, collaboration, assessment and reference material. After the system design was finished, it was evaluated and tested by 5 experts in blended learning design and 10 students which the user’s satisfaction level is good. The result is as good as the assumption so the system can be used in the process management subject for a real usage.Keywords: blended learning, asynchronous learning, design, process management
Procedia PDF Downloads 40414854 The Use of Haar Wavelet Mother Signal Tool for Performance Analysis Response of Distillation Column (Application to Moroccan Case Study)
Authors: Mahacine Amrani
Abstract:
This paper aims at reviewing some Moroccan industrial applications of wavelet especially in the dynamic identification of a process model using Haar wavelet mother response. Two recent Moroccan study cases are described using dynamic data originated by a distillation column and an industrial polyethylene process plant. The purpose of the wavelet scheme is to build on-line dynamic models. In both case studies, a comparison is carried out between the Haar wavelet mother response model and a linear difference equation model. Finally it concludes, on the base of the comparison of the process performances and the best responses, which may be useful to create an estimated on-line internal model control and its application towards model-predictive controllers (MPC). All calculations were implemented using AutoSignal Software.Keywords: process performance, model, wavelets, Haar, Moroccan
Procedia PDF Downloads 31614853 Enhancing Student Learning Outcomes Using Engineering Design Process: Case Study in Physics Course
Authors: Thien Van Ngo
Abstract:
The engineering design process is a systematic approach to solving problems. It involves identifying a problem, brainstorming solutions, prototyping and testing solutions, and evaluating the results. The engineering design process can be used to teach students how to solve problems in a creative and innovative way. The research aim of this study was to investigate the effectiveness of using the engineering design process to enhance student learning outcomes in a physics course. A mixed research method was used in this study. The quantitative data were collected using a pretest-posttest control group design. The qualitative data were collected using semi-structured interviews. The sample was 150 first-year students in the Department of Mechanical Engineering Technology at Cao Thang Technical College in Vietnam in the 2022-2023 school year. The quantitative data were collected using a pretest-posttest control group design. The pretest was administered to both groups at the beginning of the study. The posttest was administered to both groups at the end of the study. The qualitative data were collected using semi-structured interviews with a sample of eight students in the experimental group. The interviews were conducted after the posttest. The quantitative data were analyzed using independent sample T-tests. The qualitative data were analyzed using thematic analysis. The quantitative data showed that students in the experimental group, who were taught using the engineering design process, had significantly higher post-test scores on physics problem-solving than students in the control group, who were taught using the conventional method. The qualitative data showed that students in the experimental group were more motivated and engaged in the learning process than students in the control group. Students in the experimental group also reported that they found the engineering design process to be a more effective way of learning physics. The findings of this study suggest that the engineering design process can be an effective way of enhancing student learning outcomes in physics courses. The engineering design process engages students in the learning process and helps them to develop problem-solving skills.Keywords: engineering design process, problem-solving, learning outcome of physics, students’ physics competencies, deep learning
Procedia PDF Downloads 6414852 Quantification Of Uncertainties Related To The Implementation Of Reverse Logistics Process
Authors: Dnaya Soukaina
Abstract:
It’s over six decades that Reverse logistics had appeared as a research area, and it is emerging again and again in the scientific fields. As reverse logistics presents real potential for value recovery and environmental impacts decrease, it’s still necessary to extend this concept more in the industrial and commercial field especially in developing countries. The process of reverse logistics is a progression of steps beginning with the customer and finishing with the organization or even the customer, however the issue is that this cycle must be adjustable to the organization concerned, in addition of legislative, operational, financial and social obstacles. Literature had demonstrated that there are many other uncertainties while the implementation of this process that vary in function of the sector concerned and the kind of activity. Besides, even if literature is developing this topic over the last years, reseraches about uncertainties quantification in reverse logistics process still being few. the paper has the objective to fill this gap, and carry out a study to identify sustainable strategies that can be adapted to different industrial or commercial sectors to facilitate the implementation of reverse logistics.Keywords: reverse logistics, implementation, unceratinties quantification, mathematical model
Procedia PDF Downloads 1614851 A Simulated Scenario of WikiGIS to Support the Iteration and Traceability Management of the Geodesign Process
Authors: Wided Batita, Stéphane Roche, Claude Caron
Abstract:
Geodesign is an emergent term related to a new and complex process. Hence, it needs to rethink tools, technologies and platforms in order to efficiently achieve its goals. A few tools have emerged since 2010 such as CommunityViz, GeoPlanner, etc. In the era of Web 2.0 and collaboration, WikiGIS has been proposed as a new category of tools. In this paper, we present WikiGIS functionalities dealing mainly with the iteration and traceability management to support the collaboration of the Geodesign process. Actually, WikiGIS is built on GeoWeb 2.0 technologies —and primarily on wiki— and aims at managing the tracking of participants’ editing. This paper focuses on a simplified simulation to illustrate the strength of WikiGIS in the management of traceability and in the access to history in a Geodesign process. Indeed, a cartographic user interface has been implemented, and then a hypothetical use case has been imagined as proof of concept.Keywords: geodesign, history, traceability, tracking of participants’ editing, WikiGIS
Procedia PDF Downloads 24314850 Availability Analysis of Process Management in the Equipment Maintenance and Repair Implementation
Authors: Onur Ozveri, Korkut Karabag, Cagri Keles
Abstract:
It is an important issue that the occurring of production downtime and repair costs when machines fail in the machine intensive production industries. In the case of failure of more than one machine at the same time, which machines will have the priority to repair, how to determine the optimal repair time should be allotted for this machines and how to plan the resources needed to repair are the key issues. In recent years, Business Process Management (BPM) technique, bring effective solutions to different problems in business. The main feature of this technique is that it can improve the way the job done by examining in detail the works of interest. In the industries, maintenance and repair works are operating as a process and when a breakdown occurs, it is known that the repair work is carried out in a series of process. Maintenance main-process and repair sub-process are evaluated with process management technique, so it is thought that structure could bring a solution. For this reason, in an international manufacturing company, this issue discussed and has tried to develop a proposal for a solution. The purpose of this study is the implementation of maintenance and repair works which is integrated with process management technique and at the end of implementation, analyzing the maintenance related parameters like quality, cost, time, safety and spare part. The international firm that carried out the application operates in a free region in Turkey and its core business area is producing original equipment technologies, vehicle electrical construction, electronics, safety and thermal systems for the world's leading light and heavy vehicle manufacturers. In the firm primarily, a project team has been established. The team dealt with the current maintenance process again, and it has been revised again by the process management techniques. Repair process which is sub-process of maintenance process has been discussed again. In the improved processes, the ABC equipment classification technique was used to decide which machine or machines will be given priority in case of failure. This technique is a prioritization method of malfunctioned machine based on the effect of the production, product quality, maintenance costs and job security. Improved maintenance and repair processes have been implemented in the company for three months, and the obtained data were compared with the previous year data. In conclusion, breakdown maintenance was found to occur in a shorter time, with lower cost and lower spare parts inventory.Keywords: ABC equipment classification, business process management (BPM), maintenance, repair performance
Procedia PDF Downloads 19314849 The Role and Importance of Genome Sequencing in Prediction of Cancer Risk
Authors: M. Sadeghi, H. Pezeshk, R. Tusserkani, A. Sharifi Zarchi, A. Malekpour, M. Foroughmand, S. Goliaei, M. Totonchi, N. Ansari–Pour
Abstract:
The role and relative importance of intrinsic and extrinsic factors in the development of complex diseases such as cancer still remains a controversial issue. Determining the amount of variation explained by these factors needs experimental data and statistical models. These models are nevertheless based on the occurrence and accumulation of random mutational events during stem cell division, thus rendering cancer development a stochastic outcome. We demonstrate that not only individual genome sequencing is uninformative in determining cancer risk, but also assigning a unique genome sequence to any given individual (healthy or affected) is not meaningful. Current whole-genome sequencing approaches are therefore unlikely to realize the promise of personalized medicine. In conclusion, since genome sequence differs from cell to cell and changes over time, it seems that determining the risk factor of complex diseases based on genome sequence is somewhat unrealistic, and therefore, the resulting data are likely to be inherently uninformative.Keywords: cancer risk, extrinsic factors, genome sequencing, intrinsic factors
Procedia PDF Downloads 26814848 Designing Inventory System with Constrained by Reducing Ordering Cost, Lead Time and Lost Sale Rate and Considering Random Disturbance in Ordering Quantity
Authors: Arezoo Heidary, Abolfazl Mirzazadeh, Aref Gholami-Qadikolaei
Abstract:
In the business environment it is very common that a lot received may not be equal to quantity ordered. in this work, a random disturbance in a received quantity is considered. It is assumed a maximum allowable limit for storage space and inventory investment.The impact of lead time and ordering cost reductions once they act dependently is also investigated. Further, considering a mixture of back order and lost sales for allowable shortage system, the effect of investment on reducing lost sale rate is analyzed. For the proposed control system, a Lagrangian method is applied in order to solve the problem and an algorithmic procedure is utilized to achieve optimal solution with the global minimum expected cost. Finally, proves on concavity and convexity of the model in the decision variables are shown.Keywords: stochastic inventory system, lead time, ordering cost, lost sale rate, inventory constraints, random disturbance
Procedia PDF Downloads 417