Search results for: hybrid process
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 16254

Search results for: hybrid process

13194 Numerical Study of Jet Impingement Heat Transfer

Authors: A. M. Tiara, Sudipto Chakraborty, S. K. Pal

Abstract:

Impinging jets and their different configurations are important from the viewpoint of the fluid flow characteristics and their influence on heat transfer from metal surfaces due to their complex flow characteristics. Such flow characteristics results in highly variable heat transfer from the surface, resulting in varying cooling rates which affects the mechanical properties including hardness and strength. The overall objective of the current research is to conduct a fundamental investigation of the heat transfer mechanisms for an impinging coolant jet. Numerical simulation of the cooling process gives a detailed analysis of the different parameters involved even though employing Computational Fluid Dynamics (CFD) to simulate the real time process, being a relatively new research area, poses many challenges. The heat transfer mechanism in the current research is actuated by jet cooling. The computational tool used in the ongoing research for simulation of the cooling process is ANSYS Workbench software. The temperature and heat flux distribution along the steel strip with the effect of various flow parameters on the heat transfer rate can be observed in addition to determination of the jet impingement patterns, which is the major aim of the present analysis. Modelling both jet and air atomized cooling techniques using CFD methodology and validating with those obtained experimentally- including trial and error with different models and comparison of cooling rates from both the techniques have been included in this work. Finally some concluding remarks are made that identify some gaps in the available literature that have influenced the path of the current investigation.

Keywords: CFD, heat transfer, impinging jets, numerical simulation

Procedia PDF Downloads 223
13193 Effect of Molybdenum Addition to Aluminum Grain Refined by Titanium Plus Boron on Its Grain Size and Mechanical Characteristics in the Cast and After Pressing by the Equal Channel Angular Pressing Conditions

Authors: A. I. O. Zaid, A. M. Attieh, S. M. A. Al Qawabah

Abstract:

Aluminum and its alloys solidify in columnar structure with large grain size which tends to reduce their mechanical strength and surface quality. They are, therefore, grain refined by addition of either titanium or titanium plus boron to their melt before solidification. Equal channel angular pressing, ECAP, process is a recent forming method for producing heavy plastic deformation in materials. In this paper, the effect of molybdenum addition to aluminum grain refined by Ti+B on its metallurgical and mechanical characteristics are investigated in the as cast condition and after pressing by the ECAP process. It was found that addition of Mo or Ti+B alone or together to aluminum resulted in grain refining of its microstructure in the as cast condition, as the average grain size was reduced from 139 micron to 46 micron when Mo and Ti+B are added together. Pressing by the ECAP process resulted in further refinement of the microstructure where 32 micron of average grain size was achieved in Al and the Al-Mo microalloy. Regarding the mechanical strength, addition of Mo or Ti+B alone to Al resulted in deterioration of its mechanical behavior but resulted in enhancement of its mechanical behavior when added together, increase of 10% in flow stress was achieved at 20% strain. However, pressing by ECAP addition of Mo or Ti+B alone to Al resulted in enhancement of its mechanical strength but reduced its strength when added together.

Keywords: ECAP, aluminum, cast, mechanical characteristics, Mo grain refiner

Procedia PDF Downloads 458
13192 A Fresh Approach to Learn Evidence-Based Practice, a Prospective Interventional Study

Authors: Ebtehal Qulisy, Geoffrey Dougherty, Kholoud Hothan, Mylene Dandavino

Abstract:

Background: For more than 200 years, journal clubs (JCs) have been used to teach the fundamentals of critical appraisal and evidence-based practice (EBP). However, JCs curricula face important challenges, including poor sustainability, insufficient time to prepare for and conduct the activities, and lack of trainee skills and self-efficacy with critical appraisal. Andragogy principles and modern technology could help EBP be taught in more relevant, modern, and interactive ways. Method: We propose a fresh educational activity to teach EBP. Educational sessions are designed to encourage collaborative and experiential learning and do not require advanced preparation by the participants. Each session lasts 60 minutes and is adaptable to in-person, virtual, or hybrid contexts. Sessions are structured around a worksheet and include three educational objectives: “1. Identify a Clinical Conundrum”, “2. Compare and Contrast Current Guidelines”, and “3. Choose a Recent Journal Article”. Sessions begin with a short presentation by a facilitator of a clinical scenario highlighting a “grey-zone” in pediatrics. Trainees are placed in groups of two to four (based on the participants’ number) of varied training levels. The first task requires the identification of a clinical conundrum (a situation where there is no clear answer but only a reasonable solution) related to the scenario. For the second task, trainees must identify two or three clinical guidelines. The last task requires trainees to find a journal article published in the last year that reports an update regarding the scenario’s topic. Participants are allowed to use their electronic devices throughout the session. Our university provides full-text access to major journals, which facilitated this exercise. Results: Participants were a convenience sample of trainees in the inpatient services at the Montréal Children’s Hospital, McGill University. Sessions were conducted as a part of an existing weekly academic activity and facilitated by pediatricians with experience in critical appraisal. There were 28 participants in 4 sessions held during Spring 2022. Time was allocated at the end of each session to collect participants’ feedback via a self-administered online survey. There were 22 responses, were 41%(n=9) pediatric residents, 22.7%(n=5) family medicine residents, 31.8%(n=7) medical students, and 4.5%(n=1) nurse practitioner. Four respondents participated in more than one session. The “Satisfied” rates were 94.7% for session format, 100% for topic selection, 89.5% for time allocation, and 84.3% for worksheet structure. 60% of participants felt that including the sessions during the clinical ward rotation was “Feasible.” As per self-efficacy, participants reported being “Confident” for the tasks as follows: 89.5% for the ability to identify a relevant conundrum, 94.8% for the compare and contrast task, and 84.2% for the identification of a published update. The perceived effectiveness to learn EBP was reported as “Agreed” by all participants. All participants would recommend this session for further teaching. Conclusion: We developed a modern approach to teach EBP, enjoyed by all levels of participants, who also felt it was a useful learning experience. Our approach addresses known JCs challenges by being relevant to clinical care, fostering active engagement but not requiring any preparation, using available technology, and being adaptable to hybrid contexts.

Keywords: medical education, journal clubs, post-graduate teaching, andragogy, experiential learning, evidence-based practice

Procedia PDF Downloads 104
13191 A Nonstandard Finite Difference Method for Weather Derivatives Pricing Model

Authors: Clarinda Vitorino Nhangumbe, Fredericks Ebrahim, Betuel Canhanga

Abstract:

The price of an option weather derivatives can be approximated as a solution of the two-dimensional convection-diffusion dominant partial differential equation derived from the Ornstein-Uhlenbeck process, where one variable represents the weather dynamics and the other variable represent the underlying weather index. With appropriate financial boundary conditions, the solution of the pricing equation is approximated using a nonstandard finite difference method. It is shown that the proposed numerical scheme preserves positivity as well as stability and consistency. In order to illustrate the accuracy of the method, the numerical results are compared with other methods. The model is tested for real weather data.

Keywords: nonstandard finite differences, Ornstein-Uhlenbeck process, partial differential equations approach, weather derivatives

Procedia PDF Downloads 77
13190 Zinc Adsorption Determination of H2SO4 Activated Pomegranate Peel

Authors: S. N. Turkmen Koc, A. S. Kipcak, M. B. Piskin, E. Moroydor Derun, N. Tugrul

Abstract:

Active carbon can be obtained from agricultural sources. Due to the high surface area, the production of activated carbon from cheap resources is very important. Since the surface area of 1 g activated carbon is approximately between 300 and 2000 m2, it can be used to remove both organic and inorganic impurities. In this study, the adsorption of Zn metal was studied with the product of activated carbon, which is obtained from pomegranate peel by microwave and chemical activation methods. The microwave process of pomegranate peel was carried out under constant microwave power of 800 W and 1 to 4 minutes. After the microwave process, samples were treated with H2SO4 for 3 h. Then prepared product was used in synthetic waste water including 40 ppm Zn metal. As a result, removal of waste Zn in waste water ranged from 91% to 93%.

Keywords: activated carbon, chemical activation, H₂SO₄, microwave, pomegranate peel

Procedia PDF Downloads 149
13189 Information Tree: Establishment of Lifestyle-Based IT Visual Model

Authors: Chiung-Hui Chen

Abstract:

Traditional service channel is losing its edge due to emerging service technology. To establish interaction with the clients, the service industry is using effective mechanism to give clients direct access to services with emerging technologies. Thus, as service science receives attention, special and unique consumption pattern evolves; henceforth, leading to new market mechanism and influencing attitudes toward life and consumption patterns. The market demand for customized services is thus valued due to the emphasis of personal value, and is gradually changing the demand and supply relationship in the traditional industry. In respect of interior design service, in the process of traditional interior design, a designer converts to a concrete form the concept generated from the ideas and needs dictated by a user (client), by using his/her professional knowledge and drawing tool. The final product is generated through iterations of communication and modification, which is a very time-consuming process. Although this process has been accelerated with the help of computer graphics software today, repeated discussions and confirmations with users are still required to complete the task. In consideration of what is addressed above a space user’s life model is analyzed with visualization technique to create an interaction system modeled after interior design knowledge. The space user document intuitively personal life experience in a model requirement chart, allowing a researcher to analyze interrelation between analysis documents, identify the logic and the substance of data conversion. The repeated data which is documented are then transformed into design information for reuse and sharing. A professional interior designer may sort out the correlation among user’s preference, life pattern and design specification, thus deciding the critical design elements in the process of service design.

Keywords: information design, life model-based, aesthetic computing, communication

Procedia PDF Downloads 285
13188 "Gurza Incinerator" : Biomass Incinerator Powered by Empty Bunch of Palm Oil Fruits as Electrical Biomass Base Development

Authors: Andi Ismanto

Abstract:

Indonesia is the largest palm oil producer in the world. The increasing number of palm oil extensification in Indonesia started on 2000-2011. Based on preliminary figures from the Directorate General of Plantation, palm oil area in Indonesia until 2011 is about 8.91 million hectares.On 2011 production of palm oil CPO reaches 22.51 million tons. In the other hands, the increasing palm oil production has impact to environment. The Empty Bunch of Palm Oil (EBPO)waste was increased to 20 million tons in 2009. Utilization of waste EBPO currently only used as an organic fertilizer for plants. But, it was not a good solution, because TKKS that used as organic compost has high content of carbon and hydrogen compound. The EBPO waste has potential used as fuel by gasification because it has short time of decomposition. So, the process will be more efficient in time. Utilization of urban wastehas been created using an incinerator used as a source of electrical energy for household.Usually, waste burning process by incinerator is using diesel fuel and kerosene. It is certainly less effective and not environment friendly, considering the waste incineration process using Incinerator tools are continuously. Considering biomass is a renewable source of energy and the world's energy system must be switch from an energy based on fossil resources into the energy based on renewable resources, the "Gurza Incinerator": Design Build Powerful Biomass Incinerator Empty Bunch of Palm Oil (EBPO) as Elecrical Biomass Base Development, a renewable future technology. The tools is using EBPO waste as source of burning to burn garbage inside the Incinerator hopper. EBPO waste will be processed by means of gasification. Gasification isa process to produce gases that can be used as fuel for electrical power. Hopefully, this technology could be a renewable future energy and also as starting point of electrical biomass base development.

Keywords: incinerator, biomass, empty bunch palm oil, electrical energy

Procedia PDF Downloads 459
13187 Ultrasonic Micro Injection Molding: Manufacturing of Micro Plates of Biomaterials

Authors: Ariadna Manresa, Ines Ferrer

Abstract:

Introduction: Ultrasonic moulding process (USM) is a recent injection technology used to manufacture micro components. It is able to melt small amounts of material so the waste of material is certainly reduced comparing to microinjection molding. This is an important advantage when the materials are expensive like medical biopolymers. Micro-scaled components are involved in a variety of uses, such as biomedical applications. It is required replication fidelity so it is important to stabilize the process and minimize the variability of the responses. The aim of this research is to investigate the influence of the main process parameters on the filling behaviour, the dimensional accuracy and the cavity pressure when a micro-plate is manufactured by biomaterials such as PLA and PCL. Methodology or Experimental Procedure: The specimens are manufactured using a Sonorus 1G Ultrasound Micro Molding Machine. The used geometry is a rectangular micro-plate of 15x5mm and 1mm of thickness. The materials used for the investigation are PLA and PCL due to biocompatible and degradation properties. The experimentation is divided into two phases. Firstly, the influence of process parameters (vibration amplitude, sonotrodo velocity, ultrasound time and compaction force) on filling behavior is analysed, in Phase 1. Next, when filling cavity is assured, the influence of both cooling time and force compaction on the cavity pressure, part temperature and dimensional accuracy is instigated, which is done in Phase. Results and Discussion: Filling behavior depends on sonotrodo velocity and vibration amplitude. When the ultrasonic time is higher, more ultrasonic energy is applied and the polymer temperature increases. Depending on the cooling time, it is possible that when mold is opened, the micro-plate temperature is too warm. Consequently, the polymer relieve its stored internal energy (ultrasonic and thermal) expanding through the easier direction. This fact is reflected on dimensional accuracy, causing micro-plates thicker than the mold. It has also been observed the most important fact that affects cavity pressure is the compaction configuration during the manufacturing cycle. Conclusions: This research demonstrated the influence of process parameters on the final micro-plated manufactured. Future works will be focused in manufacturing other geometries and analysing the mechanical properties of the specimens.

Keywords: biomaterial, biopolymer, micro injection molding, ultrasound

Procedia PDF Downloads 271
13186 Effect of Aging Time on CeO2 Nanoparticle Size Distribution Synthesized via Sol-Gel Method

Authors: Navid Zanganeh, Hafez Balavi, Farbod Sharif, Mahla Zabet, Marzieh Bakhtiary Noodeh

Abstract:

Cerium oxide (CeO2) also known as cerium dioxide or ceria is a pale yellow-white powder with various applications in the industry from wood coating to cosmetics, filtration, fuel cell electrolytes, gas sensors, hybrid solar cells and catalysts. In this research, attempts were made to synthesize and characterization of CeO2 nano-particles via sol-gel method. In addition, the effect of aging time on the size of particles was investigated. For this purpose, the aging times adjusted 48, 56, 64, and 72 min. The obtained particles were characterized by x-ray diffraction spectroscopy (XRD), scanning electron microscopy (SEM), transmitted electron microscopy (TEM), and Brunauer–Emmett–Teller (BET). As a result, XRD patterns confirmed the formation of CeO2 nanoparticles. SEM and TEM images illustrated the nano-particles with cluster shape, spherical and a nano-size range which was in agreement with XRD results. The finest particles (7.3 nm) was obtained at the optimum condition which was aging time of 48 min, calcination temperature at 400 ⁰C, and cerium concentration of 0.004 mol. Average specific surface area of the particles at optimum condition was measured by BET analysis and recorded as 47.57 m2/g.

Keywords: aging time, CeO2 nanoparticles, size distribution, sol-gel

Procedia PDF Downloads 440
13185 Ground Motion Modeling Using the Least Absolute Shrinkage and Selection Operator

Authors: Yildiz Stella Dak, Jale Tezcan

Abstract:

Ground motion models that relate a strong motion parameter of interest to a set of predictive seismological variables describing the earthquake source, the propagation path of the seismic wave, and the local site conditions constitute a critical component of seismic hazard analyses. When a sufficient number of strong motion records are available, ground motion relations are developed using statistical analysis of the recorded ground motion data. In regions lacking a sufficient number of recordings, a synthetic database is developed using stochastic, theoretical or hybrid approaches. Regardless of the manner the database was developed, ground motion relations are developed using regression analysis. Development of a ground motion relation is a challenging process which inevitably requires the modeler to make subjective decisions regarding the inclusion criteria of the recordings, the functional form of the model and the set of seismological variables to be included in the model. Because these decisions are critically important to the validity and the applicability of the model, there is a continuous interest on procedures that will facilitate the development of ground motion models. This paper proposes the use of the Least Absolute Shrinkage and Selection Operator (LASSO) in selecting the set predictive seismological variables to be used in developing a ground motion relation. The LASSO can be described as a penalized regression technique with a built-in capability of variable selection. Similar to the ridge regression, the LASSO is based on the idea of shrinking the regression coefficients to reduce the variance of the model. Unlike ridge regression, where the coefficients are shrunk but never set equal to zero, the LASSO sets some of the coefficients exactly to zero, effectively performing variable selection. Given a set of candidate input variables and the output variable of interest, LASSO allows ranking the input variables in terms of their relative importance, thereby facilitating the selection of the set of variables to be included in the model. Because the risk of overfitting increases as the ratio of the number of predictors to the number of recordings increases, selection of a compact set of variables is important in cases where a small number of recordings are available. In addition, identification of a small set of variables can improve the interpretability of the resulting model, especially when there is a large number of candidate predictors. A practical application of the proposed approach is presented, using more than 600 recordings from the National Geospatial-Intelligence Agency (NGA) database, where the effect of a set of seismological predictors on the 5% damped maximum direction spectral acceleration is investigated. The set of candidate predictors considered are Magnitude, Rrup, Vs30. Using LASSO, the relative importance of the candidate predictors has been ranked. Regression models with increasing levels of complexity were constructed using one, two, three, and four best predictors, and the models’ ability to explain the observed variance in the target variable have been compared. The bias-variance trade-off in the context of model selection is discussed.

Keywords: ground motion modeling, least absolute shrinkage and selection operator, penalized regression, variable selection

Procedia PDF Downloads 317
13184 Design and Performance Evaluation of Hybrid Corrugated-GFRP Infill Panels

Authors: Woo Young Jung, Sung Min Park, Ho Young Son, Viriyavudh Sim

Abstract:

This study presents a way to reduce earthquake damage and emergency rehabilitation of critical structures such as schools, high-tech factories, and hospitals due to strong ground motions associated with climate changes. Regarding recent trend, a strong earthquake causes serious damage to critical structures and then the critical structure might be influenced by sequence aftershocks (or tsunami) due to fault plane adjustments. Therefore, in order to improve seismic performance of critical structures, retrofitted or strengthening study of the structures under aftershocks sequence after emergency rehabilitation of the structures subjected to strong earthquakes is widely carried out. Consequently, this study used composite material for emergency rehabilitation of the structure rather than concrete and steel materials because of high strength and stiffness, lightweight, rapid manufacturing, and dynamic performance. Also, this study was to develop or improve the seismic performance or seismic retrofit of critical structures subjected to strong ground motions and earthquake aftershocks, by utilizing GFRP-Corrugated Infill Panels (GCIP).

Keywords: aftershock, composite material, GFRP, infill panel

Procedia PDF Downloads 325
13183 Streamlining Cybersecurity Risk Assessment for Industrial Control and Automation Systems: Leveraging the National Institute of Standard and Technology’s Risk Management Framework (RMF) Using Model-Based System Engineering (MBSE)

Authors: Gampel Alexander, Mazzuchi Thomas, Sarkani Shahram

Abstract:

The cybersecurity landscape is constantly evolving, and organizations must adapt to the changing threat environment to protect their assets. The implementation of the NIST Risk Management Framework (RMF) has become critical in ensuring the security and safety of industrial control and automation systems. However, cybersecurity professionals are facing challenges in implementing RMF, leading to systems operating without authorization and being non-compliant with regulations. The current approach to RMF implementation based on business practices is limited and insufficient, leaving organizations vulnerable to cyberattacks resulting in the loss of personal consumer data and critical infrastructure details. To address these challenges, this research proposes a Model-Based Systems Engineering (MBSE) approach to implementing cybersecurity controls and assessing risk through the RMF process. The study emphasizes the need to shift to a modeling approach, which can streamline the RMF process and eliminate bloated structures that make it difficult to receive an Authorization-To-Operate (ATO). The study focuses on the practical application of MBSE in industrial control and automation systems to improve the security and safety of operations. It is concluded that MBSE can be used to solve the implementation challenges of the NIST RMF process and improve the security of industrial control and automation systems. The research suggests that MBSE provides a more effective and efficient method for implementing cybersecurity controls and assessing risk through the RMF process. The future work for this research involves exploring the broader applicability of MBSE in different industries and domains. The study suggests that the MBSE approach can be applied to other domains beyond industrial control and automation systems.

Keywords: authorization-to-operate (ATO), industrial control systems (ICS), model-based system’s engineering (MBSE), risk management framework (RMF)

Procedia PDF Downloads 72
13182 Colour Characteristics of Dried Cocoa Using Shallow Box Fermentation Technique

Authors: Khairul Bariah Sulaiman, Tajul Aris Yang

Abstract:

Fermentation is well known as an essential process in cocoa beans. Besides to develop the precursor of cocoa flavour, it also induce the colour changes in the beans.The fermentation process is reported to be influenced by duration of pod storage and fermentation. Therefore, this study was conducted to evaluate colour of Malaysian cocoa beans and how the pods storage and fermentation duration using shallow box technique will effect on it characteristics. There are two factors being studied ie duration of cocoa pod storage (0, 2, 4, and 6 days) and duration of cocoa fermentation (0, 1, 2, 3, 4 and 5 days). The experiment is arranged in 4 x 6 factorial design with 24 treatments and arrangement is in a Completely Randomised Design (CRD). The produced beans is inspected for colour changes under artificial light during cut test and divided into four groups of colour namely fully brown, purple brown, fully purple and slaty. Cut tests indicated that cocoa beans which are directly dried without undergone fermentation has the highest slaty percentage. However, application of pods storage before fermentation process is found to decrease the slaty percentage. In contrast, the percentages of fully brown beans start to dominate after two days of fermentation, especially from four and six days of pods storage batch. Whereas, almost all batch have percentage of fully purple less than 20%. Interestingly, the percentage of purple brown beans are scattered in the entire beans batch regardless any specific trend. Meanwhile, statistical analysis using General Linear Model showed that the pods storage has a significant effect on the colour characteristic of the Malaysian dried beans compared to fermentation duration.

Keywords: cocoa beans, colour, fermentation, shallow box

Procedia PDF Downloads 469
13181 Nationalization of the Social Life in Argentina: Accumulation of Capital, State Intervention, Labor Market, and System of Rights in the Last Decades

Authors: Mauro Cristeche

Abstract:

This work begins with a very simple question: How does the State spend? Argentina is witnessing a process of growing nationalization of social life, so it is necessary to find out the explanations of the phenomenon on the specific dynamic of the capitalist mode of production in Argentina and its transformations in the last decades. Then the new question is: what happened in Argentina that could explain this phenomenon? Since the seventies, the capital growth in Argentina faces deep competitive problems. Until that moment the agrarian wealth had worked as a compensation mechanism, but it began to find its limits. In the meantime, some important demographical and structural changes had happened. The strategy of the capitalist class had to become to seek in the cheapness of the labor force the main source of compensation of its weakness. As a result, a tendency to worsen the living conditions and fragmentation of the working class started to develop, manifested by unemployment, underemployment, and the fall of the purchasing power of the salary as a highlighted fact. As a consequence, it is suggested that the role of the State became stronger and public expenditure increased, as a historical trend, because it has to intervene to face the contradictions and constant growth problems posed by the development of capitalism in Argentina. On the one hand, the State has to guarantee the process of buying the cheapened workforce and at the same time the process of reproduction of the working class. On the other hand, it has to help to reproduce the individual capitals but needs to ‘attack’ them in different ways. This is why the role of the State is said to be the general political representative to the national portion of the total social capital. What will be studied is the dynamic of the intervention of the Argentine State in the context of the particular national process of capital growth, and its dynamics in the last decades. What this paper wants to show are the main general causes that could explain the phenomenon of nationalization of the social life and how it has impacted the life conditions of the working class and the system of rights.

Keywords: Argentina, nationalization, public policies, rights, state

Procedia PDF Downloads 122
13180 Simulation Model of Induction Heating in COMSOL Multiphysics

Authors: K. Djellabi, M. E. H. Latreche

Abstract:

The induction heating phenomenon depends on various factors, making the problem highly nonlinear. The mathematical analysis of this problem in most cases is very difficult and it is reduced to simple cases. Another knowledge of induction heating systems is generated in production environments, but these trial-error procedures are long and expensive. The numerical models of induction heating problem are another approach to reduce abovementioned drawbacks. This paper deals with the simulation model of induction heating problem. The simulation model of induction heating system in COMSOL Multiphysics is created. In this work we present results of numerical simulations of induction heating process in pieces of cylindrical shapes, in an inductor with four coils. The modeling of the inducting heating process was made with the software COMSOL Multiphysics Version 4.2a, for the study we present the temperature charts.

Keywords: induction heating, electromagnetic field, inductor, numerical simulation, finite element

Procedia PDF Downloads 299
13179 Performance Analysis of 180 nm Low Voltage Low Power CMOS OTA for High Frequency Application

Authors: D. J. Dahigaonkar, D. G. Wakde

Abstract:

The performance analysis of low voltage low power CMOS OTA is presented in this paper. The differential input single output OTA is simulated in 180nm CMOS process technology. The simulation results indicate high bandwidth of the order of 7.04GHz with 0.766mW power consumption and transconductance of -71.20dB. The total harmonic distortion for 100mV input at a frequency of 1MHz is found to be 2.3603%. In addition to this, to establish comparative analysis of designed OTA and analyze effect of technology scaling, the differential input single output OTA is further simulated using 350nm CMOS process technology and the comparative analysis is presented in this paper.

Keywords: Operational Transconductance Amplifier, Total Harmonic Distortions, low voltage/low power, power dissipation

Procedia PDF Downloads 392
13178 Developing Fault Tolerance Metrics of Web and Mobile Applications

Authors: Ahmad Mohsin, Irfan Raza Naqvi, Syda Fatima Usamn

Abstract:

Applications with higher fault tolerance index are considered more reliable and trustworthy to drive quality. In recent years application development has been shifted from traditional desktop and web to native and hybrid application(s) for the web and mobile platforms. With the emergence of Internet of things IOTs, cloud and big data trends, the need for measuring Fault Tolerance for these complex nature applications has increased to evaluate their performance. There is a phenomenal gap between fault tolerance metrics development and measurement. Classic quality metric models focused on metrics for traditional systems ignoring the essence of today’s applications software, hardware & deployment characteristics. In this paper, we have proposed simple metrics to measure fault tolerance considering general requirements for Web and Mobile Applications. We have aligned factors – subfactors, using GQM for metrics development considering the nature of mobile we apps. Systematic Mathematical formulation is done to measure metrics quantitatively. Three web mobile applications are selected to measure Fault Tolerance factors using formulated metrics. Applications are then analysed on the basis of results from observations in a controlled environment on different mobile devices. Quantitative results are presented depicting Fault tolerance in respective applications.

Keywords: web and mobile applications, reliability, fault tolerance metric, quality metrics, GQM based metrics

Procedia PDF Downloads 327
13177 A Combined Activated Sludge-Filtration-Ozonation Process for Abattoir Wastewater Treatment

Authors: Pello Alfonso-Muniozguren, Madeleine Bussemaker, Ralph Chadeesingh, Caryn Jones, David Oakley, Judy Lee, Devendra Saroj

Abstract:

Current industrialized livestock agriculture is growing every year leading to an increase in the generation of wastewater that varies considerably in terms of organic content and microbial population. Therefore, suitable wastewater treatment methods are required to ensure the wastewater quality meet regulations before discharge. In the present study, a combined lab scale activated sludge-filtration-ozonation system was used to treat a pre-treated abattoir wastewater. A hydraulic retention time of 24 hours and a solid retention time of 13 days were used for the activated sludge process, followed by a filtration step (4-7 µm) and using ozone as tertiary treatment. An average reduction of 93% and 98% was achieved for Chemical Oxygen Demand (COD) and Biological Oxygen Demand (BOD), respectively, obtaining final values of 128 mg/L COD and 12 mg/L BOD. For the Total Suspended Solids (TSS), the average reduction increased to 99% in the same system, reducing the final value down to 3 mg/L. Additionally, 98% reduction in Phosphorus (P) and a complete inactivation of Total Coliforms (TC) was obtained after 17 min ozonation time. For Total Viable Counts (TVC), a drastic reduction was observed with 30 min ozonation time (6 log inactivation) at an ozone dose of 71 mg O3/L. Overall, the combined process was sufficient to meet discharge requirements without further treatment for the measured parameters (COD, BOD, TSS, P, TC, and TVC).

Keywords: abattoir waste water, activated sludge, ozone, waste water treatment

Procedia PDF Downloads 259
13176 PaSA: A Dataset for Patent Sentiment Analysis to Highlight Patent Paragraphs

Authors: Renukswamy Chikkamath, Vishvapalsinhji Ramsinh Parmar, Christoph Hewel, Markus Endres

Abstract:

Given a patent document, identifying distinct semantic annotations is an interesting research aspect. Text annotation helps the patent practitioners such as examiners and patent attorneys to quickly identify the key arguments of any invention, successively providing a timely marking of a patent text. In the process of manual patent analysis, to attain better readability, recognising the semantic information by marking paragraphs is in practice. This semantic annotation process is laborious and time-consuming. To alleviate such a problem, we proposed a dataset to train machine learning algorithms to automate the highlighting process. The contributions of this work are: i) we developed a multi-class dataset of size 150k samples by traversing USPTO patents over a decade, ii) articulated statistics and distributions of data using imperative exploratory data analysis, iii) baseline Machine Learning models are developed to utilize the dataset to address patent paragraph highlighting task, and iv) future path to extend this work using Deep Learning and domain-specific pre-trained language models to develop a tool to highlight is provided. This work assists patent practitioners in highlighting semantic information automatically and aids in creating a sustainable and efficient patent analysis using the aptitude of machine learning.

Keywords: machine learning, patents, patent sentiment analysis, patent information retrieval

Procedia PDF Downloads 73
13175 Origin of Salinity Problems during Tsunami and Remedial Measures in Coastal Areas

Authors: N. K. Gupta, R. C. Bhattacharjee

Abstract:

In the aftermath of the tsunami in 2004 and terrible humanitarian disaster affecting thousands of kilometers of coastal south, the immediate priority is to begin the process of reconstruction of livelihoods including basic services. It is likely that many coastal wetlands would have been affected by the large inflow of salt-water and littoral sediments during the tsunami, with longer-term effects including changes in their hydrogeology caused by changes to coastlines and damage to sea-defenses. The reconstruction process is likely to provide opportunities to better integrate environmental protection and management with economic development in the region, including the opportunity to conserve and restore coastal habitats. Presented herein is a study pertaining to salinity problems encountered in coastal south during tsunami in 2004 and the consequent loss of fertility of agricultural land including remedial measures to revitalize economic growth in the region.

Keywords: tsunami, salinity, costal area, reconstruction

Procedia PDF Downloads 359
13174 Characterization of Leakage Current on the Surface of Porcelain Insulator under Contaminated Conditions

Authors: Hocine Terrab , Abdelhafid Bayadi, Adel Kara, Ayman El-Hag

Abstract:

Insulator flashover under polluted conditions has been a serious threat on the reliability of power systems. It is known that the flashover process is mainly affected by the environmental conditions such as; the pollution level and humidity. Those are the essential parameters influencing the wetting process. This paper presents an investigation of the characteristics of leakage current (LC) developed on the surface of porcelain insulator at contaminated conditions under AC voltage. The study is done in an artificial fog chamber and the LC is characterized for different stages; dry, wetted and presence of discharge activities. Time-frequency and spectral analysis are adopted to calculate the evolution of LC characteristics with various stages prior to flashover occurrence. The preliminary results could be used in analysing the LC to develop more effective diagnosis of early signs of dry band arcing as an indication for insulation washing.

Keywords: flashover, harmonic components, leakage current, phase angle, statistical analysis

Procedia PDF Downloads 412
13173 Investigation of the Effect of Plasticization Temperature on Polymer Thin Film Stability through Spin Coating Process

Authors: Bilge Bozdogan, Selda T. Sendogdular, Levent Sendogdular

Abstract:

We report a technique to control chain conformation during the plasticization process to achieve homogeneous and stable thin films, which allows to reduce post-process annealing times along with enhanced properties like controlled irreversible adsorbed layer (Guiselin brushes) formation. In this study, spin coating temperature was considered as a parameter; hence, all equipment, including the spin coater, substrate, vials, and the solution, was kept inside the same heated fume hood where solution was spin-coated after the temperature was stabilized at a desired value. AFM and SEM results revealed severe difference for solid and air interface between ambient and temperature-controlled samples, which suggest that enthalpic contribution dynamically helps to control film stability in a way where chain entanglements and conformational restrictions are avoided before film growing and allowing to control grafting density through spin coating temperature. The adsorbed layer was also characterized with SEM and Raman-spectroscopy technique right after seeding the adsorbed layer with gold nanoparticles. Stabilized gold nanoparticles and their surface distribution manifest the existence of a controllable polymer brush structure. Acknowledgments: This study was funded by Erciyes University Scientific Research Projects (BAP) Funding(Project ID:10058)

Keywords: chain stability, Guiselin brushes, polymer thin film, spin coating temperature

Procedia PDF Downloads 198
13172 Naphtha Catalytic Reform: Modeling and Simulation of Unity

Authors: Leal Leonardo, Pires Carlos Augusto de Moraes, Casiraghi Magela

Abstract:

In this work were realized the modeling and simulation of the catalytic reformer process, of ample form, considering all the equipment that influence the operation performance. Considered it a semi-regenerative reformer, with four reactors in series intercalated with four furnaces, two heat exchanges, one product separator and one recycle compressor. A simplified reactional system was considered, involving only ten chemical compounds related through five reactions. The considered process was the applied to aromatics production (benzene, toluene, and xylene). The models developed to diverse equipment were interconnecting in a simulator that consists of a computer program elaborate in FORTRAN 77. The simulation of the global model representative of reformer unity achieved results that are compatibles with the literature ones. It was then possible to study the effects of operational variables in the products concentration and in the performance of the unity equipment.

Keywords: catalytic reforming, modeling, simulation, petrochemical engineering

Procedia PDF Downloads 496
13171 Reconceptualizing “Best Practices” in Public Sector

Authors: Eftychia Kessopoulou, Styliani Xanthopoulou, Ypatia Theodorakioglou, George Tsiotras, Katerina Gotzamani

Abstract:

Public sector managers frequently herald that implementing best practices as a set of standards, may lead to superior organizational performance. However, recent research questions the objectification of best practices, highlighting: a) the inability of public sector organizations to develop innovative administrative practices, as well as b) the adoption of stereotypical renowned practices inculcated in the public sector by international governance bodies. The process through which organizations construe what a best practice is, still remains a black box that is yet to be investigated, given the trend of continuous changes in public sector performance, as well as the burgeoning interest of sharing popular administrative practices put forward by international bodies. This study aims to describe and understand how organizational best practices are constructed by public sector performance management teams, like benchmarkers, during the benchmarking-mediated performance improvement process and what mechanisms enable this construction. A critical realist action research methodology is employed, starting from a description of various approaches on best practice nature when a benchmarking-mediated performance improvement initiative, such as the Common Assessment Framework, is applied. Firstly, we observed the benchmarker’s management process of best practices in a public organization, so as to map their theories-in-use. As a second step we contextualized best administrative practices by reflecting the different perspectives emerged from the previous stage on the design and implementation of an interview protocol. We used this protocol to conduct 30 semi-structured interviews with “best practice” process owners, in order to examine their experiences and performance needs. Previous research on best practices has shown that needs and intentions of benchmarkers cannot be detached from the causal mechanisms of the various contexts in which they work. Such causal mechanisms can be found in: a) process owner capabilities, b) the structural context of the organization, and c) state regulations. Therefore, we developed an interview protocol theoretically informed in the first part to spot causal mechanisms suggested by previous research studies and supplemented it with questions regarding the provision of best practice support from the government. Findings of this work include: a) a causal account of the nature of best administrative practices in the Greek public sector that shed light on explaining their management, b) a description of the various contexts affecting best practice conceptualization, and c) a description of how their interplay changed the organization’s best practice management.

Keywords: benchmarking, action research, critical realism, best practices, public sector

Procedia PDF Downloads 112
13170 Human Performance Evaluating of Advanced Cardiac Life Support Procedure Using Fault Tree and Bayesian Network

Authors: Shokoufeh Abrisham, Seyed Mahmoud Hossieni, Elham Pishbin

Abstract:

In this paper, a hybrid method based on the fault tree analysis (FTA) and Bayesian networks (BNs) are employed to evaluate the team performance quality of advanced cardiac life support (ACLS) procedures in emergency department. According to American Heart Association (AHA) guidelines, a category relying on staff action leading to clinical incidents and also some discussions with emergency medicine experts, a fault tree model for ACLS procedure is obtained based on the human performance. The obtained FTA model is converted into BNs, and some different scenarios are defined to demonstrate the efficiency and flexibility of the presented model of BNs. Also, a sensitivity analysis is conducted to indicate the effects of team leader presence and uncertainty knowledge of experts on the quality of ACLS. The proposed model based on BNs shows that how the results of risk analysis can be closed to reality comparing to the obtained results based on only FTA in medical procedures.

Keywords: advanced cardiac life support, fault tree analysis, Bayesian belief networks, numan performance, healthcare systems

Procedia PDF Downloads 133
13169 Improvement of Mechanical Properties and Corrosion Resistance of AA7056 Aluminum Alloys by the Non-isothermal Aging Process

Authors: Tse-An Pan, Sheng-Long Lee

Abstract:

The effect of non-isothermal aging on the mechanical properties and corrosion resistance of Al-9Zn-2.3Mg-1.9Cu (AA7056) alloys was investigated. The results revealed that thick materials were limited to retrogression and re-aging treatment (RRA). It could not reach the retrogression temperature in the RRA treatment. Compared with the RRA treatment, the non-isothermal aging (NIA) treatment produced discontinuous precipitates at grain boundaries, while the intragranular precipitates were fine and dense. The strength was similar to that of the RRA treatment; the corrosion resistance of the alloy was significantly improved by NIA aging. NIA treatment was less affected by the thickness of the alloy. The difference between the actual temperature and the setting temperature of the alloy is minimal during the aging process. The combination of properties could overcome the fact that RRA treatment cannot handle thick materials.

Keywords: Al-Zn-Mg-Cu alloy, corrosion, retrogression, re-aging, non-isothermal aging

Procedia PDF Downloads 168
13168 Daylightophil Approach towards High-Performance Architecture for Hybrid-Optimization of Visual Comfort and Daylight Factor in BSk

Authors: Mohammadjavad Mahdavinejad, Hadi Yazdi

Abstract:

The greatest influence we have from the world is shaped through the visual form, thus light is an inseparable element in human life. The use of daylight in visual perception and environment readability is an important issue for users. With regard to the hazards of greenhouse gas emissions from fossil fuels, and in line with the attitudes on the reduction of energy consumption, the correct use of daylight results in lower levels of energy consumed by artificial lighting, heating and cooling systems. Windows are usually the starting points for analysis and simulations to achieve visual comfort and energy optimization; therefore, attention should be paid to the orientation of buildings to minimize electrical energy and maximize the use of daylight. In this paper, by using the Design Builder Software, the effect of the orientation of an 18m2(3m*6m) room with 3m height in city of Tehran has been investigated considering the design constraint limitations. In these simulations, the dimensions of the building have been changed with one degree and the window is located on the smaller face (3m*3m) of the building with 80% ratio. The results indicate that the orientation of building has a lot to do with energy efficiency to meet high-performance architecture and planning goals and objectives.

Keywords: daylight, window, orientation, energy consumption, design builder

Procedia PDF Downloads 215
13167 Outcomes of Pain Management for Patients in Srinagarind Hospital: Acute Pain Indicator

Authors: Chalermsri Sorasit, Siriporn Mongkhonthawornchai, Darawan Augsornwan, Sudthanom Kamollirt

Abstract:

Background: Although knowledge of pain and pain management is improving, they are still inadequate to patients. The Nursing Division of Srinagarind Hospital is responsible for setting the pain management system, including work instruction development and pain management indicators. We have developed an information technology program for monitoring pain quality indicators, which was implemented to all nursing departments in April 2013. Objective: To study outcomes of acute pain management in process and outcome indicators. Method: This is a retrospective descriptive study. The sample population was patients who had acute pain 24-48 hours after receiving a procedure, while admitted to Srinagarind Hospital in 2014. Data were collected from the information technology program. 2709 patients with acute pain from 10 Nursing Departments were recruited in the study. The research tools in this study were 1) the demographic questionnaire 2) the pain management questionnaire for process indicators, and 3) the pain management questionnaire for outcome indicators. Data were analyzed and presented by percentages and means. Results: The process indicators show that nurses used pain assessment tool and recorded 99.19%. The pain reassessment after the intervention was 96.09%. The 80.15% of the patients received opioid for pain medication and the most frequency of non-pharmacological intervention used was positioning (76.72%). For the outcome indicators, nearly half of them (49.90%) had moderate–severe pain, mean scores of worst pain was 6.48 and overall pain was 4.08. Patient satisfaction level with pain management was good (49.17%) and very good (46.62%). Conclusion: Nurses used pain assessment tools and pain documents which met the goal of the pain management process. Patient satisfaction with pain management was at high level. However the patients had still moderate to severe pain. Nurses should adhere more strictly to the guidelines of pain management, by using acute pain guidelines especially when pain intensity is particularly moderate-high. Nurses should also develop and practice a non-pharmacological pain management program to continually improve the quality of pain management. The information technology program should have more details about non-pharmacological pain techniques.

Keywords: outcome, pain management, acute pain, Srinagarind Hospital

Procedia PDF Downloads 217
13166 Developing Computational Thinking in Early Childhood Education

Authors: Kalliopi Kanaki, Michael Kalogiannakis

Abstract:

Nowadays, in the digital era, the early acquisition of basic programming skills and knowledge is encouraged, as it facilitates students’ exposure to computational thinking and empowers their creativity, problem-solving skills, and cognitive development. More and more researchers and educators investigate the introduction of computational thinking in K-12 since it is expected to be a fundamental skill for everyone by the middle of the 21st century, just like reading, writing and arithmetic are at the moment. In this paper, a doctoral research in the process is presented, which investigates the infusion of computational thinking into science curriculum in early childhood education. The whole attempt aims to develop young children’s computational thinking by introducing them to the fundamental concepts of object-oriented programming in an enjoyable, yet educational framework. The backbone of the research is the digital environment PhysGramming (an abbreviation of Physical Science Programming), which provides children the opportunity to create their own digital games, turning them from passive consumers to active creators of technology. PhysGramming deploys an innovative hybrid schema of visual and text-based programming techniques, with emphasis on object-orientation. Through PhysGramming, young students are familiarized with basic object-oriented programming concepts, such as classes, objects, and attributes, while, at the same time, get a view of object-oriented programming syntax. Nevertheless, the most noteworthy feature of PhysGramming is that children create their own digital games within the context of physical science courses, in a way that provides familiarization with the basic principles of object-oriented programming and computational thinking, even though no specific reference is made to these principles. Attuned to the ethical guidelines of educational research, interventions were conducted in two classes of second grade. The interventions were designed with respect to the thematic units of the curriculum of physical science courses, as a part of the learning activities of the class. PhysGramming was integrated into the classroom, after short introductory sessions. During the interventions, 6-7 years old children worked in pairs on computers and created their own digital games (group games, matching games, and puzzles). The authors participated in these interventions as observers in order to achieve a realistic evaluation of the proposed educational framework concerning its applicability in the classroom and its educational and pedagogical perspectives. To better examine if the objectives of the research are met, the investigation was focused on six criteria; the educational value of PhysGramming, its engaging and enjoyable characteristics, its child-friendliness, its appropriateness for the purpose that is proposed, its ability to monitor the user’s progress and its individualizing features. In this paper, the functionality of PhysGramming and the philosophy of its integration in the classroom are both described in detail. Information about the implemented interventions and the results obtained is also provided. Finally, several limitations of the research conducted that deserve attention are denoted.

Keywords: computational thinking, early childhood education, object-oriented programming, physical science courses

Procedia PDF Downloads 110
13165 Production Process for Diesel Fuel Components Polyoxymethylene Dimethyl Ethers from Methanol and Formaldehyde Solution

Authors: Xiangjun Li, Huaiyuan Tian, Wujie Zhang, Dianhua Liu

Abstract:

Polyoxymethylene dimethyl ethers (PODEn) as clean diesel additive can improve the combustion efficiency and quality of diesel fuel and alleviate the problem of atmospheric pollution. Considering synthetic routes, PODE production from methanol and formaldehyde is regarded as the most economical and promising synthetic route. However, methanol used for synthesizing PODE can produce water, which causes the loss of active center of catalyst and hydrolysis of PODEn in the production process. Macroporous strong acidic cation exchange resin catalyst was prepared, which has comparative advantages over other common solid acid catalysts in terms of stability and catalytic efficiency for synthesizing PODE. Catalytic reactions were carried out under 353 K, 1 MPa and 3mL·gcat-1·h-1 in a fixed bed reactor. Methanol conversion and PODE3-6 selectivity reached 49.91% and 23.43%, respectively. Catalyst lifetime evaluation showed that resin catalyst retained its catalytic activity for 20 days without significant changes and catalytic activity of completely deactivated resin catalyst can basically return to previous level by simple acid regeneration. The acid exchange capacities of original and deactivated catalyst were 2.5191 and 0.0979 mmol·g-1, respectively, while regenerated catalyst reached 2.0430 mmol·g-1, indicating that the main reason for resin catalyst deactivation is that Brønsted acid sites of original resin catalyst were temporarily replaced by non-hydrogen ion cations. A separation process consisting of extraction and distillation for PODE3-6 product was designed for separation of water and unreacted formaldehyde from reactive mixture and purification of PODE3-6, respectively. The concentration of PODE3-6 in final product can reach up to 97%. These results indicate that the scale-up production of PODE3-6 from methanol and formaldehyde solution is feasible.

Keywords: inactivation, polyoxymethylene dimethyl ethers, separation process, sulfonic cation exchange resin

Procedia PDF Downloads 127