Search results for: bilinear model
11415 Comprehensive Lifespan Support for Quality of Life
Authors: Joann Douziech
Abstract:
Individuals with intellectual and developmental disabilities (IDD) possess characteristics that present both challenges and gifts. Individuals with IDD require and are worthy of intentional, strategic, and specialized support throughout their lifespan to ensure optimum quality-of-life outcomes. The current global advocacy movement advancing the rights of individuals with IDD emphasizes a high degree of choice over life decisions. For some individuals, this degree of choice results in a variety of negative health and well-being outcomes. Improving the quality of life outcomes requires the combination of a commitment to the rights of the individual with a responsibility to provide support and choice commensurate with individual capacity. A belief that individuals with IDD are capable of learning and they are worthy of being taught provides the foundation for a holistic model of support throughout their lifespan. This model is based on three pillars of engineering the environment, promoting skill development and maintenance, and staff support. In an ever-changing world, supporting quality of life requires attention to moments, phases, and changes in stages throughout the lifespan. Balancing these complexities with strategic, responsive, and dynamic interventions enhances the quality of life of individuals with ID throughout their lifespan.Keywords: achieving optimum quality of life, comprehensive support, lifespan approach, philosophy and pedagogy
Procedia PDF Downloads 6811414 Spare Part Inventory Optimization Policy: A Study Literature
Authors: Zukhrof Romadhon, Nani Kurniati
Abstract:
Availability of Spare parts is critical to support maintenance tasks and the production system. Managing spare part inventory deals with some parameters and objective functions, as well as the tradeoff between inventory costs and spare parts availability. Several mathematical models and methods have been developed to optimize the spare part policy. Many researchers who proposed optimization models need to be considered to identify other potential models. This work presents a review of several pertinent literature on spare part inventory optimization and analyzes the gaps for future research. Initial investigation on scholars and many journal database systems under specific keywords related to spare parts found about 17K papers. Filtering was conducted based on five main aspects, i.e., replenishment policy, objective function, echelon network, lead time, model solving, and additional aspects of part classification. Future topics could be identified based on the number of papers that haven’t addressed specific aspects, including joint optimization of spare part inventory and maintenance.Keywords: spare part, spare part inventory, inventory model, optimization, maintenance
Procedia PDF Downloads 6611413 A Crystallization Kinetic Model for Long Fiber-Based Composite with Thermoplastic Semicrystalline Polymer Matrix
Authors: Nicolas Bigot, M'hamed Boutaous, Nahiene Hamila, Shihe Xin
Abstract:
Composite materials with polymer matrices are widely used in most industrial areas, particularly in aeronautical and automotive ones. Thanks to the development of a high-performance thermoplastic semicrystalline polymer matrix, those materials exhibit more and more efficient properties. The polymer matrix in composite materials can manifest a specific crystalline structure characteristic of crystallization in a fibrous medium. In order to guarantee a good mechanical behavior of structures and to optimize their performances, it is necessary to define realistic mechanical constitutive laws of such materials considering their physical structure. The interaction between fibers and matrix is a key factor in the mechanical behavior of composite materials. Transcrystallization phenomena which develops in the matrix around the fibers constitute the interphase which greatly affects and governs the nature of the fiber-matrix interaction. Hence, it becomes fundamental to quantify its impact on the thermo-mechanical behavior of composites material in relationship with processing conditions. In this work, we propose a numerical model coupling the thermal and crystallization kinetics in long fiber-based composite materials, considering both the spherulitic and transcrystalline types of the induced structures. After validation of the model with comparison to results from the literature and noticing a good correlation, a parametric study has been led on the effects of the thermal kinetics, the fibers volume fractions, the deformation, and the pressure on the crystallization rate in the material, under processing conditions. The ratio of the transcrystallinity is highlighted and analyzed with regard to the thermal kinetics and gradients in the material. Experimental results on the process are foreseen and pave the way to establish a mechanical constitutive law describing, with the introduction of the role on the crystallization rates and types on the thermo-mechanical behavior of composites materials.Keywords: composite materials, crystallization, heat transfer, modeling, transcrystallization
Procedia PDF Downloads 19511412 Drying Kinetics of Vacuum Dried Beef Meat Slices
Authors: Elif Aykin Dincer, Mustafa Erbas
Abstract:
The vacuum drying behavior of beef slices (10 x 4 x 0.2 cm3) was experimentally investigated at the temperature of 60, 70, and 80°C under 25 mbar ultimate vacuum pressure and the mathematical models (Lewis, Page, Midilli, Two-term, Wangh and Singh and Modified Henderson and Pabis) were used to fit the vacuum drying of beef slices. The increase in drying air temperature resulted in a decrease in drying time. It took approximately 206, 180 and 157 min to dry beef slices from an initial moisture content to a final moisture content of 0.05 kg water/kg dry matter at 60, 70 and 80 °C of vacuum drying, respectively. It is also observed that the drying rate increased with increasing drying temperature. The coefficients (R2), the reduced chi-square (x²) and root mean square error (RMSE) values were obtained by application of six models to the experimental drying data. The best model with the highest R2 and, the lowest x² and RMSE values was selected to describe the drying characteristics of beef slices. The Page model has shown a better fit to the experimental drying data as compared to other models. In addition, the effective moisture diffusivities of beef slices in the vacuum drying at 60 - 80 °C varied in the range of 1.05 – 1.09 x 10-10 m2/s. Consequently, this results can be used to simulate vacuum drying process of beef slices and improve efficiency of the drying process.Keywords: beef slice, drying models, effective diffusivity, vacuum
Procedia PDF Downloads 29211411 The Extent of Virgin Olive-Oil Prices' Distribution Revealing the Behavior of Market Speculators
Authors: Fathi Abid, Bilel Kaffel
Abstract:
The olive tree, the olive harvest during winter season and the production of olive oil better known by professionals under the name of the crushing operation have interested institutional traders such as olive-oil offices and private companies such as food industry refining and extracting pomace olive oil as well as export-import public and private companies specializing in olive oil. The major problem facing producers of olive oil each winter campaign, contrary to what is expected, it is not whether the harvest will be good or not but whether the sale price will allow them to cover production costs and achieve a reasonable margin of profit or not. These questions are entirely legitimate if we judge by the importance of the issue and the heavy complexity of the uncertainty and competition made tougher by a high level of indebtedness and the experience and expertise of speculators and producers whose objectives are sometimes conflicting. The aim of this paper is to study the formation mechanism of olive oil prices in order to learn about speculators’ behavior and expectations in the market, how they contribute by their industry knowledge and their financial alliances and the size the financial challenge that may be involved for them to build private information hoses globally to take advantage. The methodology used in this paper is based on two stages, in the first stage we study econometrically the formation mechanisms of olive oil price in order to understand the market participant behavior by implementing ARMA, SARMA, GARCH and stochastic diffusion processes models, the second stage is devoted to prediction purposes, we use a combined wavelet- ANN approach. Our main findings indicate that olive oil market participants interact with each other in a way that they promote stylized facts formation. The unstable participant’s behaviors create the volatility clustering, non-linearity dependent and cyclicity phenomena. By imitating each other in some periods of the campaign, different participants contribute to the fat tails observed in the olive oil price distribution. The best prediction model for the olive oil price is based on a back propagation artificial neural network approach with input information based on wavelet decomposition and recent past history.Keywords: olive oil price, stylized facts, ARMA model, SARMA model, GARCH model, combined wavelet-artificial neural network, continuous-time stochastic volatility mode
Procedia PDF Downloads 34011410 Planning of Construction Material Flow Using Hybrid Simulation Modeling
Authors: A. M. Naraghi, V. Gonzalez, M. O'Sullivan, C. G. Walker, M. Poshdar, F. Ying, M. Abdelmegid
Abstract:
Discrete Event Simulation (DES) and Agent Based Simulation (ABS) are two simulation approaches that have been proposed to support decision-making in the construction industry. Despite the wide use of these simulation approaches in the construction field, their applications for production and material planning is still limited. This is largely due to the dynamic and complex nature of construction material supply chain systems. Moreover, managing the flow of construction material is not well integrated with site logistics in traditional construction planning methods. This paper presents a hybrid of DES and ABS to simulate on-site and off-site material supply processes. DES is applied to determine the best production scenarios with information of on-site production systems, while ABS is used to optimize the supply chain network. A case study of a construction piling project in New Zealand is presented illustrating the potential benefits of using the proposed hybrid simulation model in construction material flow planning. The hybrid model presented can be used to evaluate the impact of different decisions on construction supply chain management.Keywords: construction supply-chain management, simulation modeling, decision-support tools, hybrid simulation
Procedia PDF Downloads 20811409 Effects of Education Equity Policy on Housing Prices: Evidence from Simultaneous Admission to Public and Private Schools Policy in Shanghai
Authors: Tianyu Chen
Abstract:
China's school district education policy has encouraged parents to purchase properties in school districts with high-quality education resources. Shanghai has implemented "Simultaneous Admission to Public and Private Schools" (SAPPS) since 2018, which has covered all nine-year compulsory education by 2020. This study examines the impact of SAPPS on the housing market, specifically the premium effect of houses located in dual-school districts. Based on the Hedonic Pricing Model and the Signaling Theory, data is collected from 585 second-hand house transactions in Pudong New Area, Shanghai, and it is analyzed with the Difference-in-Differences (DID) model. The results indicate that the implementation of SAPPS has exacerbated the premium of dual school district housing and weakened the effect of the policy to a certain degree. To ensure equal access to education for all students, the government should work both on the supply and demand sides of the education resource equation.Keywords: simultaneous admission to public and private schools, housing prices, education policy, education equity
Procedia PDF Downloads 7911408 Extending the Theory of Planned Behaviour to Predict Intention to Commute by Bicycle: Case Study of Mexico City
Authors: Magda Cepeda, Frances Hodgson, Ann Jopson
Abstract:
There are different barriers people face when choosing to cycle for commuting purposes. This study examined the role of psycho-social factors predicting the intention to cycle to commute in Mexico City. An extended version of the theory of planned behaviour was developed and utilized with a simple random sample of 401 road users. We applied exploratory and confirmatory factor analysis and after identifying five factors, a structural equation model was estimated to find the relationships among the variables. The results indicated that cycling attributes, attitudes to cycling, social comparison and social image and prestige were the most important factors influencing intention to cycle. Although the results from this study are specific to Mexico City, they indicate areas of interest to transportation planners in other regions especially in those cities where intention to cycle its linked to its perceived image and there is political ambition to instigate positive cycling cultures. Moreover, this study contributes to the current literature developing applications of the Theory of Planned Behaviour.Keywords: cycling, latent variable model, perception, theory of planned behaviour
Procedia PDF Downloads 35411407 Deep-Learning to Generation of Weights for Image Captioning Using Part-of-Speech Approach
Authors: Tiago do Carmo Nogueira, Cássio Dener Noronha Vinhal, Gélson da Cruz Júnior, Matheus Rudolfo Diedrich Ullmann
Abstract:
Generating automatic image descriptions through natural language is a challenging task. Image captioning is a task that consistently describes an image by combining computer vision and natural language processing techniques. To accomplish this task, cutting-edge models use encoder-decoder structures. Thus, Convolutional Neural Networks (CNN) are used to extract the characteristics of the images, and Recurrent Neural Networks (RNN) generate the descriptive sentences of the images. However, cutting-edge approaches still suffer from problems of generating incorrect captions and accumulating errors in the decoders. To solve this problem, we propose a model based on the encoder-decoder structure, introducing a module that generates the weights according to the importance of the word to form the sentence, using the part-of-speech (PoS). Thus, the results demonstrate that our model surpasses state-of-the-art models.Keywords: gated recurrent units, caption generation, convolutional neural network, part-of-speech
Procedia PDF Downloads 10411406 River Bank Erosion Studies: A Review on Investigation Approaches and Governing Factors
Authors: Azlinda Saadon
Abstract:
This paper provides detail review on river bank erosion studies with respect to their processes, methods of measurements and factors governing river bank erosion. Bank erosion processes are commonly associated with river changes initiation and development, through width adjustment and planform evolution. It consists of two main types of erosion processes; basal erosion due to fluvial hydraulic force and bank failure under the influence of gravity. Most studies had only focused on one factor rather than integrating both factors. Evidences of previous works have shown integration between both processes of fluvial hydraulic force and bank failure. Bank failure is often treated as probabilistic phenomenon without having physical characteristics and the geotechnical aspects of the bank. This review summarizes the findings of previous investigators with respect to measurement techniques and prediction rates of river bank erosion through field investigation, physical model and numerical model approaches. Factors governing river bank erosion considering physical characteristics of fluvial erosion are defined.Keywords: river bank erosion, bank erosion, dimensional analysis, geotechnical aspects
Procedia PDF Downloads 43611405 Simulation and Control of the Flywheel System in the Rotor of a Wind Turbine Using Simulink and OpenFAST for Assessing the Effect on the Mechanical Loads
Authors: Chinazo Onyeka Eziuzo
Abstract:
This work presents the simulation and control of the flywheel system in the rotor of a wind turbine using Simulink and OpenFAST for assessing the effect on the mechanical loads. This concept allows the flywheel system to serve two main tasks: supporting the power system and mitigating the mechanical loads in the wind turbine. These tasks are grouped into four control scenarios; scenario 1 represents steadying the power infeed in the Flywheel, scenario 2 represents steadying power with FW and grid loss, scenario 3 represents mitigating excitations from gravity, and scenario 4 represents damping in-plane blade vibrations. The s-function of the OpenFAST model was used to substitute the given 1st Eigen mode model of the WT. After that, the simulations were run for the above-listed scenarios. Additionally, the effects of the control options on the mechanical loads were assessed, and it was established that the FW system assists in steadying infeed power and mechanical load mitigation.Keywords: simulation, control, wind turbine, OpenFAST
Procedia PDF Downloads 13011404 Utilizing Google Earth for Internet GIS
Authors: Alireza Derambakhsh
Abstract:
The objective of this examination is to explore the capability of utilizing Google Earth for Internet GIS applications. The study particularly analyzes the utilization of vector and characteristic information and the capability of showing and preparing this information in new ways utilizing the Google Earth stage. It has progressively been perceived that future improvements in GIS will fixate on Internet GIS, and in three noteworthy territories: GIS information access, spatial data scattering and GIS displaying/preparing. Google Earth is one of the group of geobrowsers that offer a free and simple to utilize administration that empower information with a spatial part to be overlain on top of a 3-D model of the Earth. This examination makes a methodological structure to accomplish its objective that comprises of three noteworthy parts: A database level, an application level and a customer level. As verification of idea a web model has been produced, which incorporates a differing scope of datasets and lets clients direst inquiries and make perceptions of this custom information. The outcomes uncovered that both vector and property information can be successfully spoken to and imagined utilizing Google Earth. In addition, the usefulness to question custom information and envision results has been added to the Google Earth stage.Keywords: Google earth, internet GIS, vector, characteristic information
Procedia PDF Downloads 31011403 Validity and Reliability of the Iranian Version of the Self-Expansion Questionnaire
Authors: Mehravar Javid, James Sexton, Farzaneh Amani, Kainaz Patravala
Abstract:
Self-expansion is a procedure through which people expand the dimensions of their self-concept by incorporating novel content into their sense and experience of identity. Greater self-expansion predicts positive consequences for individuals and romantic relationships. The self-expansion questionnaire (SEQ) originally developed by Lewandowski & Aron (2002) assumes that self-expansion is constituted of key components from the self-expansion model. This study aimed to confirm the factor structure of SEQ and adapt the questions of the scale to the Iranian culture. The sample included 190 participants who responded to 14 items and were selected by simple random sampling. Using Amos-21 and SPSS-21, descriptive statistics, Pearson correlation and Confirmatory Factor Analysis (CFA) were calculated. Cronbach’s alpha coefficient for total SEQ items was 0.92. Results of CFA supported the factor structure SEQ [RMSEA=0.08, GFI=0.88 and CFI=0.92] that showed the model has a good fit and also all the items of SEQ, have a high correlation and have a direct and significant relationship. So, the SEQ demonstrated acceptable psychometric properties in Tehran University students. Looking forward, it would be interesting and exciting to see the implications of the scale as applied to romantic relationships.Keywords: validity, reliability, confirmatory factor analysis, self-expansion questionnaire
Procedia PDF Downloads 8411402 Application of Two Stages Adaptive Neuro-Fuzzy Inference System to Improve Dissolved Gas Analysis Interpretation Techniques
Authors: Kharisma Utomo Mulyodinoto, Suwarno, A. Abu-Siada
Abstract:
Dissolved Gas Analysis is one of impressive technique to detect and predict internal fault of transformers by using gas generated by transformer oil sample. A number of methods are used to interpret the dissolved gas from transformer oil sample: Doernenberg Ratio Method, IEC (International Electrotechnical Commission) Ratio Method, and Duval Triangle Method. While the assessment of dissolved gas within transformer oil samples has been standardized over the past two decades, analysis of the results is not always straight forward as it depends on personnel expertise more than mathematical formulas. To get over this limitation, this paper is aimed at improving the interpretation of Doernenberg Ratio Method, IEC Ratio Method, and Duval Triangle Method using Two Stages Adaptive Neuro-Fuzzy Inference System (ANFIS). Dissolved gas analysis data from 520 faulty transformers was analyzed to establish the proposed ANFIS model. Results show that the developed ANFIS model is accurate and can standardize the dissolved gas interpretation process with accuracy higher than 90%.Keywords: ANFIS, dissolved gas analysis, Doernenberg ratio method, Duval triangular method, IEC ratio method, transformer
Procedia PDF Downloads 15011401 Biodistribution Studies of 177Lu-DOTATOC in Mouse Tumor Model: Possible Utilization in Adenocarcinoma Breast Cancer Treatment
Authors: M. Mousavi-Daramoroudi, H. Yousefnia, F. Abbasi-Davani, S. Zolghadri, S. Kakaei
Abstract:
Despite the appropriate characteristics of 177Lu and DOTATOC, to our best knowledge, the therapeutic benefit of 177Lu-DOTATOC complex in breast cancer has not been reported until now. In this study, biodistribution of 177Lu-DOTA-TOC in mouse tumor model for evaluation of possible utilization of this complex in breast cancer treatment was investigated.177Lu was prepared with the specific activity of 2.6-3 GBq.mg-1 and radionuclidic purity higher than 99%. The radiolabeled complex was prepared in the optimized conditions with the radiochemical purity higher than 99%. The final solution was injected to the BALB/c mice with adenocarcinoma breast cancer. The biodistribution results showed major accumulation in the kidneys as the major excretion route and the somatostatin receptor-positive tissues such as pancreas compared with the other tissues. Also, significant uptake was observed in tumor even in longer time after injection. According to the results obtained in this research study, somatostatin receptors expressed in breast cancers can be targeted with DOTATOC analogues especially with 177Lu-DOTATOC as an ideal therapeutic agent.Keywords: ¹⁷⁷Lu, adenocarcinoma breast cancer, DOTATOC, BALB/c mice
Procedia PDF Downloads 22911400 Simulation of Multistage Extraction Process of Co-Ni Separation Using Ionic Liquids
Authors: Hongyan Chen, Megan Jobson, Andrew J. Masters, Maria Gonzalez-Miquel, Simon Halstead, Mayri Diaz de Rienzo
Abstract:
Ionic liquids offer excellent advantages over conventional solvents for industrial extraction of metals from aqueous solutions, where such extraction processes bring opportunities for recovery, reuse, and recycling of valuable resources and more sustainable production pathways. Recent research on the use of ionic liquids for extraction confirms their high selectivity and low volatility, but there is relatively little focus on how their properties can be best exploited in practice. This work addresses gaps in research on process modelling and simulation, to support development, design, and optimisation of these processes, focusing on the separation of the highly similar transition metals, cobalt, and nickel. The study exploits published experimental results, as well as new experimental results, relating to the separation of Co and Ni using trihexyl (tetradecyl) phosphonium chloride. This extraction agent is attractive because it is cheaper, more stable and less toxic than fluorinated hydrophobic ionic liquids. This process modelling work concerns selection and/or development of suitable models for the physical properties, distribution coefficients, for mass transfer phenomena, of the extractor unit and of the multi-stage extraction flowsheet. The distribution coefficient model for cobalt and HCl represents an anion exchange mechanism, supported by the literature and COSMO-RS calculations. Parameters of the distribution coefficient models are estimated by fitting the model to published experimental extraction equilibrium results. The mass transfer model applies Newman’s hard sphere model. Diffusion coefficients in the aqueous phase are obtained from the literature, while diffusion coefficients in the ionic liquid phase are fitted to dynamic experimental results. The mass transfer area is calculated from the surface to mean diameter of liquid droplets of the dispersed phase, estimated from the Weber number inside the extractor. New experiments measure the interfacial tension between the aqueous and ionic phases. The empirical models for predicting the density and viscosity of solutions under different metal loadings are also fitted to new experimental data. The extractor is modelled as a continuous stirred tank reactor with mass transfer between the two phases and perfect phase separation of the outlet flows. A multistage separation flowsheet simulation is set up to replicate a published experiment and compare model predictions with the experimental results. This simulation model is implemented in gPROMS software for dynamic process simulation. The results of single stage and multi-stage flowsheet simulations are shown to be in good agreement with the published experimental results. The estimated diffusion coefficient of cobalt in the ionic liquid phase is in reasonable agreement with published data for the diffusion coefficients of various metals in this ionic liquid. A sensitivity study with this simulation model demonstrates the usefulness of the models for process design. The simulation approach has potential to be extended to account for other metals, acids, and solvents for process development, design, and optimisation of extraction processes applying ionic liquids for metals separations, although a lack of experimental data is currently limiting the accuracy of models within the whole framework. Future work will focus on process development more generally and on extractive separation of rare earths using ionic liquids.Keywords: distribution coefficient, mass transfer, COSMO-RS, flowsheet simulation, phosphonium
Procedia PDF Downloads 19211399 Knowledge Sharing within a Team: Exploring the Antecedents and Role of Trust
Authors: Li Yan Hei, Au Wing Tung
Abstract:
Knowledge sharing is a process in which individuals mutually exchange existing knowledge and co-create new knowledge. Previous research has confirmed that trust is positively associated with knowledge sharing. However, only few studies systematically examined the antecedents of trust and these antecedents’ impacts on knowledge sharing. In order to explore and understand the relationships between trust and knowledge sharing in depth, this study proposed a relationship maintenance-based model to examine the antecedents of trust in knowledge sharing in project teams. Three critical elements within a project team were measured, including the environment, project team partner and interaction. It was hypothesized that the trust would lead to knowledge sharing and in turn result in perceived good team performance. With a sample of 200 Hong Kong employees, the proposed model was evaluated with structural equation modeling. Expected findings are trust will contribute to knowledge sharing, resulting in better team performance. The results will also offer insights into antecedents of trust that play a heavy role in the focal relationship. The present study contributes to a more holistic understanding of relationship between trust and knowledge sharing by linking the antecedents and outcomes. The findings will raise the awareness of project managers on ways to promote knowledge sharing.Keywords: knowledge sharing, project management, team, trust
Procedia PDF Downloads 61811398 Development of a Model for Predicting Radiological Risks in Interventional Cardiology
Authors: Stefaan Carpentier, Aya Al Masri, Fabrice Leroy, Thibault Julien, Safoin Aktaou, Malorie Martin, Fouad Maaloul
Abstract:
Introduction: During an 'Interventional Radiology (IR)' procedure, the patient's skin-dose may become very high for a burn, necrosis, and ulceration to appear. In order to prevent these deterministic effects, a prediction of the peak skin-dose for the patient is important in order to improve the post-operative care to be given to the patient. The objective of this study is to estimate, before the intervention, the patient dose for ‘Chronic Total Occlusion (CTO)’ procedures by selecting relevant clinical indicators. Materials and methods: 103 procedures were performed in the ‘Interventional Cardiology (IC)’ department using a Siemens Artis Zee image intensifier that provides the Air Kerma of each IC exam. Peak Skin Dose (PSD) was measured for each procedure using radiochromic films. Patient parameters such as sex, age, weight, and height were recorded. The complexity index J-CTO score, specific to each intervention, was determined by the cardiologist. A correlation method applied to these indicators allowed to specify their influence on the dose. A predictive model of the dose was created using multiple linear regressions. Results: Out of 103 patients involved in the study, 5 were excluded for clinical reasons and 2 for placement of radiochromic films outside the exposure field. 96 2D-dose maps were finally used. The influencing factors having the highest correlation with the PSD are the patient's diameter and the J-CTO score. The predictive model is based on these parameters. The comparison between estimated and measured skin doses shows an average difference of 0.85 ± 0.55 Gy for doses of less than 6 Gy. The mean difference between air-Kerma and PSD is 1.66 Gy ± 1.16 Gy. Conclusion: Using our developed method, a first estimate of the dose to the skin of the patient is available before the start of the procedure, which helps the cardiologist in carrying out its intervention. This estimation is more accurate than that provided by the Air-Kerma.Keywords: chronic total occlusion procedures, clinical experimentation, interventional radiology, patient's peak skin dose
Procedia PDF Downloads 13911397 Optimization and Retrofitting for an Egyptian Refinery Water Network
Authors: Mohamed Mousa
Abstract:
Sacristies in the supply of freshwater, strict regulations on discharging wastewater and the support to encourage sustainable development by water minimization techniques leads to raise the interest of water reusing, regeneration, and recycling. Water is considered a vital element in chemical industries. In this study, an optimization model will be developed to determine the optimal design of refinery’s water network system via source interceptor sink that involves several network alternatives, then a Mixed-Integer Non-Linear programming (MINLP) was used to obtain the optimal network superstructure based on flowrates, the concentration of contaminants, etc. The main objective of the model is to reduce the fixed cost of piping installation interconnections, reducing the operating cots of all streams within the refiner’s water network, and minimize the concentration of pollutants to comply with the environmental regulations. A real case study for one of the Egyptian refineries was studied by GAMS / BARON global optimization platform, and the water network had been retrofitted and optimized, leading to saving around 195 m³/ hr. of freshwater with a total reduction reaches to 26 %.Keywords: freshwater minimization, modelling, GAMS, BARON, water network design, wastewater reudction
Procedia PDF Downloads 23511396 Reversible and Irreversible Wrinkling in Tube Hydroforming Process
Authors: Ali Abd El-Aty, Ahmed Tauseef, Ahmad Farooq
Abstract:
This research aims at analyzing and optimizing the hydroforming process parameters to achieve a sound bulged tube without failure. Theoretical constitutive model is formulated to develop a working diagram including process window, which represents the optimize region to carry out the hydroforming process and predict the type of tube failure during the process accurately. The model is applied into different bulging ratios for low carbon steel (C1010). From this study, it is concluded that the tubes with bulging ratios up to 50% and 70% are successfully formed without defects. The tubes with bulging ratio of 90% are successfully formed by hydroforming with optimized the loading path (axial feed versus internal pressure) within the process window. The working diagram is modified due to different types of formation of wrinkling during the hydroforming process. The formation of wrinkles with increasing axial feed can be useful in terms of the achievement of higher bulging ratio and/or less thinning and this type of wrinkles can be overcome through the internal pressure in the later stage of the hydroforming process. On the other hand, the formation of wrinkles may be harmful, if it cannot be reversed.Keywords: finite element, hydroforming, process window, wrinkling
Procedia PDF Downloads 28011395 Hansen Solubility Parameter from Surface Measurements
Authors: Neveen AlQasas, Daniel Johnson
Abstract:
Membranes for water treatment are an established technology that attracts great attention due to its simplicity and cost effectiveness. However, membranes in operation suffer from the adverse effect of membrane fouling. Bio-fouling is a phenomenon that occurs at the water-membrane interface, and is a dynamic process that is initiated by the adsorption of dissolved organic material, including biomacromolecules, on the membrane surface. After initiation, attachment of microorganisms occurs, followed by biofilm growth. The biofilm blocks the pores of the membrane and consequently results in reducing the water flux. Moreover, the presence of a fouling layer can have a substantial impact on the membrane separation properties. Understanding the mechanism of the initiation phase of biofouling is a key point in eliminating the biofouling on membrane surfaces. The adhesion and attachment of different fouling materials is affected by the surface properties of the membrane materials. Therefore, surface properties of different polymeric materials had been studied in terms of their surface energies and Hansen solubility parameters (HSP). The difference between the combined HSP parameters (HSP distance) allows prediction of the affinity of two materials to each other. The possibilities of measuring the HSP of different polymer films via surface measurements, such as contact angle has been thoroughly investigated. Knowing the HSP of a membrane material and the HSP of a specific foulant, facilitate the estimation of the HSP distance between the two, and therefore the strength of attachment to the surface. Contact angle measurements using fourteen different solvents on five different polymeric films were carried out using the sessile drop method. Solvents were ranked as good or bad solvents using different ranking method and ranking was used to calculate the HSP of each polymeric film. Results clearly indicate the absence of a direct relation between contact angle values of each film and the HSP distance between each polymer film and the solvents used. Therefore, estimating HSP via contact angle alone is not sufficient. However, it was found if the surface tensions and viscosities of the used solvents are taken in to the account in the analysis of the contact angle values, a prediction of the HSP from contact angle measurements is possible. This was carried out via training of a neural network model. The trained neural network model has three inputs, contact angle value, surface tension and viscosity of solvent used. The model is able to predict the HSP distance between the used solvent and the tested polymer (material). The HSP distance prediction is further used to estimate the total and individual HSP parameters of each tested material. The results showed an accuracy of about 90% for all the five studied filmsKeywords: surface characterization, hansen solubility parameter estimation, contact angle measurements, artificial neural network model, surface measurements
Procedia PDF Downloads 9511394 SEM Image Classification Using CNN Architectures
Authors: Güzi̇n Ti̇rkeş, Özge Teki̇n, Kerem Kurtuluş, Y. Yekta Yurtseven, Murat Baran
Abstract:
A scanning electron microscope (SEM) is a type of electron microscope mainly used in nanoscience and nanotechnology areas. Automatic image recognition and classification are among the general areas of application concerning SEM. In line with these usages, the present paper proposes a deep learning algorithm that classifies SEM images into nine categories by means of an online application to simplify the process. The NFFA-EUROPE - 100% SEM data set, containing approximately 21,000 images, was used to train and test the algorithm at 80% and 20%, respectively. Validation was carried out using a separate data set obtained from the Middle East Technical University (METU) in Turkey. To increase the accuracy in the results, the Inception ResNet-V2 model was used in view of the Fine-Tuning approach. By using a confusion matrix, it was observed that the coated-surface category has a negative effect on the accuracy of the results since it contains other categories in the data set, thereby confusing the model when detecting category-specific patterns. For this reason, the coated-surface category was removed from the train data set, hence increasing accuracy by up to 96.5%.Keywords: convolutional neural networks, deep learning, image classification, scanning electron microscope
Procedia PDF Downloads 12711393 Ethical Decision-Making in AI and Robotics Research: A Proposed Model
Authors: Sylvie Michel, Emmanuelle Gagnou, Joanne Hamet
Abstract:
Researchers in the fields of AI and Robotics frequently encounter ethical dilemmas throughout their research endeavors. Various ethical challenges have been pinpointed in the existing literature, including biases and discriminatory outcomes, diffusion of responsibility, and a deficit in transparency within AI operations. This research aims to pinpoint these ethical quandaries faced by researchers and shed light on the mechanisms behind ethical decision-making in the research process. By synthesizing insights from existing literature and acknowledging prevalent shortcomings, such as overlooking the heterogeneous nature of decision-making, non-accumulative results, and a lack of consensus on numerous factors due to limited empirical research, the objective is to conceptualize and validate a model. This model will incorporate influences from individual perspectives and situational contexts, considering potential moderating factors in the ethical decision-making process. Qualitative analyses were conducted based on direct observation of an AI/Robotics research team focusing on collaborative robotics for several months. Subsequently, semi-structured interviews with 16 team members were conducted. The entire process took place during the first semester of 2023. Observations were analyzed using an analysis grid, and the interviews underwent thematic analysis using Nvivo software. An initial finding involves identifying the ethical challenges that AI/robotics researchers confront, underlining a disparity between practical applications and theoretical considerations regarding ethical dilemmas in the realm of AI. Notably, researchers in AI prioritize the publication and recognition of their work, sparking the genesis of these ethical inquiries. Furthermore, this article illustrated that researchers tend to embrace a consequentialist ethical framework concerning safety (for humans engaging with robots/AI), worker autonomy in relation to robots, and the societal implications of labor (can robots displace jobs?). A second significant contribution entails proposing a model for ethical decision-making within the AI/Robotics research sphere. The model proposed adopts a process-oriented approach, delineating various research stages (topic proposal, hypothesis formulation, experimentation, conclusion, and valorization). Across these stages and the ethical queries, they entail, a comprehensive four-point comprehension of ethical decision-making is presented: recognition of the moral quandary; moral judgment, signifying the decision-maker's aptitude to discern the morally righteous course of action; moral intention, reflecting the ability to prioritize moral values above others; and moral behavior, denoting the application of moral intention to the situation. Variables such as political inclinations ((anti)-capitalism, environmentalism, veganism) seem to wield significant influence. Moreover, age emerges as a noteworthy moderating factor. AI and robotics researchers are continually confronted with ethical dilemmas during their research endeavors, necessitating thoughtful decision-making. The contribution involves introducing a contextually tailored model, derived from meticulous observations and insightful interviews, enabling the identification of factors that shape ethical decision-making at different stages of the research process.Keywords: ethical decision making, artificial intelligence, robotics, research
Procedia PDF Downloads 8011392 Automated Adaptions of Semantic User- and Service Profile Representations by Learning the User Context
Authors: Nicole Merkle, Stefan Zander
Abstract:
Ambient Assisted Living (AAL) describes a technological and methodological stack of (e.g. formal model-theoretic semantics, rule-based reasoning and machine learning), different aspects regarding the behavior, activities and characteristics of humans. Hence, a semantic representation of the user environment and its relevant elements are required in order to allow assistive agents to recognize situations and deduce appropriate actions. Furthermore, the user and his/her characteristics (e.g. physical, cognitive, preferences) need to be represented with a high degree of expressiveness in order to allow software agents a precise evaluation of the users’ context models. The correct interpretation of these context models highly depends on temporal, spatial circumstances as well as individual user preferences. In most AAL approaches, model representations of real world situations represent the current state of a universe of discourse at a given point in time by neglecting transitions between a set of states. However, the AAL domain currently lacks sufficient approaches that contemplate on the dynamic adaptions of context-related representations. Semantic representations of relevant real-world excerpts (e.g. user activities) help cognitive, rule-based agents to reason and make decisions in order to help users in appropriate tasks and situations. Furthermore, rules and reasoning on semantic models are not sufficient for handling uncertainty and fuzzy situations. A certain situation can require different (re-)actions in order to achieve the best results with respect to the user and his/her needs. But what is the best result? To answer this question, we need to consider that every smart agent requires to achieve an objective, but this objective is mostly defined by domain experts who can also fail in their estimation of what is desired by the user and what not. Hence, a smart agent has to be able to learn from context history data and estimate or predict what is most likely in certain contexts. Furthermore, different agents with contrary objectives can cause collisions as their actions influence the user’s context and constituting conditions in unintended or uncontrolled ways. We present an approach for dynamically updating a semantic model with respect to the current user context that allows flexibility of the software agents and enhances their conformance in order to improve the user experience. The presented approach adapts rules by learning sensor evidence and user actions using probabilistic reasoning approaches, based on given expert knowledge. The semantic domain model consists basically of device-, service- and user profile representations. In this paper, we present how this semantic domain model can be used in order to compute the probability of matching rules and actions. We apply this probability estimation to compare the current domain model representation with the computed one in order to adapt the formal semantic representation. Our approach aims at minimizing the likelihood of unintended interferences in order to eliminate conflicts and unpredictable side-effects by updating pre-defined expert knowledge according to the most probable context representation. This enables agents to adapt to dynamic changes in the environment which enhances the provision of adequate assistance and affects positively the user satisfaction.Keywords: ambient intelligence, machine learning, semantic web, software agents
Procedia PDF Downloads 28211391 Rheological Characteristics of Ice Slurries Based on Propylene- and Ethylene-Glycol at High Ice Fractions
Authors: Senda Trabelsi, Sébastien Poncet, Michel Poirier
Abstract:
Ice slurries are considered as a promising phase-changing secondary fluids for air-conditioning, packaging or cooling industrial processes. An experimental study has been here carried out to measure the rheological characteristics of ice slurries. Ice slurries consist in a solid phase (flake ice crystals) and a liquid phase. The later is composed of a mixture of liquid water and an additive being here either (1) Propylene-Glycol (PG) or (2) Ethylene-Glycol (EG) used to lower the freezing point of water. Concentrations of 5%, 14% and 24% of both additives are investigated with ice mass fractions ranging from 5% to 85%. The rheological measurements are carried out using a Discovery HR-2 vane-concentric cylinder with four full-length blades. The experimental results show that the behavior of ice slurries is generally non-Newtonian with shear-thinning or shear-thickening behaviors depending on the experimental conditions. In order to determine the consistency and the flow index, the Herschel-Bulkley model is used to describe the behavior of ice slurries. The present results are finally validated against an experimental database found in the literature and the predictions of an Artificial Neural Network model.Keywords: ice slurry, propylene-glycol, ethylene-glycol, rheology
Procedia PDF Downloads 26411390 Investigation of Gas Tungsten Arc Welding Parameters on Residual Stress of Heat Affected Zone in Inconel X750 Super Alloy Welding Using Finite Element Method
Authors: Kimia Khoshdel Vajari, Saber Saffar
Abstract:
Reducing the residual stresses caused by welding is desirable for the industry. The effect of welding sequence, as well as the effect of yield stress on the number of residual stresses generated in Inconel X750 superalloy sheets and beams, have been investigated. The finite element model used in this research is a three-dimensional thermal and mechanical model, and the type of analysis is indirect coupling. This analysis is done in two stages. First, thermal analysis is performed, and then the thermal changes of the first analysis are used as the applied load in the second analysis. ABAQUS has been used for modeling, and the Dflux subroutine has been used in the Fortran programming environment to move the arc and the molten pool. The results of this study show that the amount of tensile residual stress in symmetric, discontinuous, and symmetric-discontinuous welds is reduced to a maximum of 27%, 54%, and 37% compared to direct welding, respectively. The results also show that the amount of residual stresses created by welding increases linearly with increasing yield stress with a slope of 40%.Keywords: residual stress, X750 superalloy, finite element, welding, thermal analysis
Procedia PDF Downloads 12011389 Extraction of Forest Plantation Resources in Selected Forest of San Manuel, Pangasinan, Philippines Using LiDAR Data for Forest Status Assessment
Authors: Mark Joseph Quinto, Roan Beronilla, Guiller Damian, Eliza Camaso, Ronaldo Alberto
Abstract:
Forest inventories are essential to assess the composition, structure and distribution of forest vegetation that can be used as baseline information for management decisions. Classical forest inventory is labor intensive and time-consuming and sometimes even dangerous. The use of Light Detection and Ranging (LiDAR) in forest inventory would improve and overcome these restrictions. This study was conducted to determine the possibility of using LiDAR derived data in extracting high accuracy forest biophysical parameters and as a non-destructive method for forest status analysis of San Manual, Pangasinan. Forest resources extraction was carried out using LAS tools, GIS, Envi and .bat scripts with the available LiDAR data. The process includes the generation of derivatives such as Digital Terrain Model (DTM), Canopy Height Model (CHM) and Canopy Cover Model (CCM) in .bat scripts followed by the generation of 17 composite bands to be used in the extraction of forest classification covers using ENVI 4.8 and GIS software. The Diameter in Breast Height (DBH), Above Ground Biomass (AGB) and Carbon Stock (CS) were estimated for each classified forest cover and Tree Count Extraction was carried out using GIS. Subsequently, field validation was conducted for accuracy assessment. Results showed that the forest of San Manuel has 73% Forest Cover, which is relatively much higher as compared to the 10% canopy cover requirement. On the extracted canopy height, 80% of the tree’s height ranges from 12 m to 17 m. CS of the three forest covers based on the AGB were: 20819.59 kg/20x20 m for closed broadleaf, 8609.82 kg/20x20 m for broadleaf plantation and 15545.57 kg/20x20m for open broadleaf. Average tree counts for the tree forest plantation was 413 trees/ha. As such, the forest of San Manuel has high percent forest cover and high CS.Keywords: carbon stock, forest inventory, LiDAR, tree count
Procedia PDF Downloads 39111388 Simulation Research of Diesel Aircraft Engine
Authors: Łukasz Grabowski, Michał Gęca, Mirosław Wendeker
Abstract:
This paper presents the simulation results of a new opposed piston diesel engine to power a light aircraft. Created in the AVL Boost, the model covers the entire charge passage, from the inlet up to the outlet. The model shows fuel injection into cylinders and combustion in cylinders. The calculation uses the module for two-stroke engines. The model was created using sub-models available in this software that structure the model. Each of the sub-models is complemented with parameters in line with the design premise. Since engine weight resulting from geometric dimensions is fundamental in aircraft engines, two configurations of stroke were studied. For each of the values, there were calculated selected operating conditions defined by crankshaft speed. The required power was achieved by changing air fuel ratio (AFR). There was also studied brake specific fuel consumption (BSFC). For stroke S1, the BSFC was lowest at all of the three operating points. This difference is approximately 1-2%, which means higher overall engine efficiency but the amount of fuel injected into cylinders is larger by several mg for S1. The cylinder maximum pressure is lower for S2 due to the fact that compressor gear driving remained the same and boost pressure was identical in the both cases. Calculations for various values of boost pressure were the next stage of the study. In each of the calculation case, the amount of fuel was changed to achieve the required engine power. In the former case, the intake system dimensions were modified, i.e. the duct connecting the compressor and the air cooler, so its diameter D = 40 mm was equal to the diameter of the compressor outlet duct. The impact of duct length was also examined to be able to reduce the flow pulsation during the operating cycle. For the so selected geometry of the intake system, there were calculations for various values of boost pressure. The boost pressure was changed by modifying the gear driving the compressor. To reach the required level of cruising power N = 68 kW. Due to the mechanical power consumed by the compressor, high pressure ratio results in a worsened overall engine efficiency. The figure on the change in BSFC from 210 g/kWh to nearly 270 g/kWh shows this correlation and the overall engine efficiency is reduced by about 8%. Acknowledgement: This work has been realized in the cooperation with The Construction Office of WSK "PZL-KALISZ" S.A." and is part of Grant Agreement No. POIR.01.02.00-00-0002/15 financed by the Polish National Centre for Research and Development.Keywords: aircraft, diesel, engine, simulation
Procedia PDF Downloads 20911387 Nonparametric Path Analysis with Truncated Spline Approach in Modeling Rural Poverty in Indonesia
Authors: Usriatur Rohma, Adji Achmad Rinaldo Fernandes
Abstract:
Nonparametric path analysis is a statistical method that does not rely on the assumption that the curve is known. The purpose of this study is to determine the best nonparametric truncated spline path function between linear and quadratic polynomial degrees with 1, 2, and 3-knot points and to determine the significance of estimating the best nonparametric truncated spline path function in the model of the effect of population migration and agricultural economic growth on rural poverty through the variable unemployment rate using the t-test statistic at the jackknife resampling stage. The data used in this study are secondary data obtained from statistical publications. The results showed that the best model of nonparametric truncated spline path analysis is quadratic polynomial degree with 3-knot points. In addition, the significance of the best-truncated spline nonparametric path function estimation using jackknife resampling shows that all exogenous variables have a significant influence on the endogenous variables.Keywords: nonparametric path analysis, truncated spline, linear, quadratic, rural poverty, jackknife resampling
Procedia PDF Downloads 5011386 Evolution of Performance Measurement Methods in Conditions of Uncertainty: The Implementation of Fuzzy Sets in Performance Measurement
Authors: E. A. Tkachenko, E. M. Rogova, V. V. Klimov
Abstract:
One of the basic issues of development management is connected with performance measurement as a prerequisite for identifying the achievement of development objectives. The aim of our research is to develop an improved model of assessing a company’s development results. The model should take into account the cyclical nature of development and the high degree of uncertainty in dealing with numerous management tasks. Our hypotheses may be formulated as follows: Hypothesis 1. The cycle of a company’s development may be studied from the standpoint of a project cycle. To do that, methods and tools of project analysis are to be used. Hypothesis 2. The problem of the uncertainty when justifying managerial decisions within the framework of a company’s development cycle can be solved through the use of the mathematical apparatus of fuzzy logic. The reasoned justification of the validity of the hypotheses made is given in the suggested article. The fuzzy logic toolkit applies to the case of technology shift within an enterprise. It is proven that some restrictions in performance measurement that are incurred to conventional methods could be eliminated by implementation of the fuzzy logic apparatus in performance measurement models.Keywords: logic, fuzzy sets, performance measurement, project analysis
Procedia PDF Downloads 383