Search results for: point process
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 18638

Search results for: point process

18248 Appropriate Depth of Needle Insertion during Rhomboid Major Trigger Point Block

Authors: Seongho Jang

Abstract:

Objective: To investigate an appropriate depth of needle insertion during trigger point injection into the rhomboid major muscle. Methods: Sixty-two patients who visited our department with shoulder or upper back pain participated in this study. The distance between the skin and the rhomboid major muscle (SM) and the distance between the skin and rib (SB) were measured using ultrasonography. The subjects were divided into 3 groups according to BMI: BMI less than 23 kg/m2 (underweight or normal group); 23 kg/m2 or more to less than 25 kg/m2 (overweight group); and 25 kg/m2 or more (obese group). The mean ±standard deviation (SD) of SM and SB of each group were calculated. A range between mean+1 SD of SM and the mean-1 SD of SB was defined as a safe margin. Results: The underweight or normal group’s SM, SB, and the safe margin were 1.2±0.2, 2.1±0.4, and 1.4 to 1.7 cm, respectively. The overweight group’s SM and SB were 1.4±0.2 and 2.4±0.9 cm, respectively. The safe margin could not be calculated for this group. The obese group’s SM, SB, and the safe margin were 1.8±0.3, 2.7±0.5, and 2.1 to 2.2 cm, respectively. Conclusion: This study will help us to set the standard depth of safe needle insertion into the rhomboid major muscle in an effective manner without causing any complications.

Keywords: pneumothorax, rhomboid major muscle, trigger point injection, ultrasound

Procedia PDF Downloads 268
18247 Clustering and Modelling Electricity Conductors from 3D Point Clouds in Complex Real-World Environments

Authors: Rahul Paul, Peter Mctaggart, Luke Skinner

Abstract:

Maintaining public safety and network reliability are the core objectives of all electricity distributors globally. For many electricity distributors, managing vegetation clearances from their above ground assets (poles and conductors) is the most important and costly risk mitigation control employed to meet these objectives. Light Detection And Ranging (LiDAR) is widely used by utilities as a cost-effective method to inspect their spatially-distributed assets at scale, often captured using high powered LiDAR scanners attached to fixed wing or rotary aircraft. The resulting 3D point cloud model is used by these utilities to perform engineering grade measurements that guide the prioritisation of vegetation cutting programs. Advances in computer vision and machine-learning approaches are increasingly applied to increase automation and reduce inspection costs and time; however, real-world LiDAR capture variables (e.g., aircraft speed and height) create complexity, noise, and missing data, reducing the effectiveness of these approaches. This paper proposes a method for identifying each conductor from LiDAR data via clustering methods that can precisely reconstruct conductors in complex real-world configurations in the presence of high levels of noise. It proposes 3D catenary models for individual clusters fitted to the captured LiDAR data points using a least square method. An iterative learning process is used to identify potential conductor models between pole pairs. The proposed method identifies the optimum parameters of the catenary function and then fits the LiDAR points to reconstruct the conductors.

Keywords: point cloud, LİDAR data, machine learning, computer vision, catenary curve, vegetation management, utility industry

Procedia PDF Downloads 68
18246 Biases in Macroprudential Supervision and Their Legal Implications

Authors: Anat Keller

Abstract:

Given that macro-prudential supervision is a relatively new policy area and its empirical and analytical research are still in their infancy, its theoretical foundations are also lagging behind. This paper contributes to the developing discussion on effective legal and institutional macroprudential supervision frameworks. In the first part of the paper, it is argued that effectiveness as a key benchmark poses some challenges in the context of macroprudential supervision such as the difficulty in proving causality between supervisory actions and the achievement of the supervisor’s mission. The paper suggests that effectiveness in the macroprudential context should, therefore, be assessed at the supervisory decision-making process (to be differentiated from the supervisory outcomes). The second part of the essay examines whether insights from behavioural economics can point to biases in the macroprudential decision-making process. These biases include, inter alia, preference bias, groupthink bias and inaction bias. It is argued that these biases are exacerbated in the multilateral setting of the macroprudential supervision framework in the EU. The paper then examines how legal and institutional frameworks should be designed to acknowledge and perhaps contain these identified biases. The paper suggests that the effectiveness of macroprudential policy will largely depend on the existence of clear and robust transparency and accountability arrangements. Accountability arrangements can be used as a vehicle for identifying and addressing potential biases in the macro-prudential framework, in particular, inaction bias. Inclusiveness of the public in the supervisory process in the form of transparency and awareness of the logic behind policy decisions may assist in minimising their potential unpopularity thus promoting their effectiveness. Furthermore, a governance structure which facilitates coordination of the macroprudential supervisor with other policymakers and incorporates outside perspectives and opinions could ‘break-down’ groupthink bias as well as inaction bias.

Keywords: behavioural economics and biases, effectiveness of macroprudential supervision, legal and institutional macroprudential frameworks, macroprudential decision-making process

Procedia PDF Downloads 252
18245 Reversible and Irreversible Wrinkling in Tube Hydroforming Process

Authors: Ali Abd El-Aty, Ahmed Tauseef, Ahmad Farooq

Abstract:

This research aims at analyzing and optimizing the hydroforming process parameters to achieve a sound bulged tube without failure. Theoretical constitutive model is formulated to develop a working diagram including process window, which represents the optimize region to carry out the hydroforming process and predict the type of tube failure during the process accurately. The model is applied into different bulging ratios for low carbon steel (C1010). From this study, it is concluded that the tubes with bulging ratios up to 50% and 70% are successfully formed without defects. The tubes with bulging ratio of 90% are successfully formed by hydroforming with optimized the loading path (axial feed versus internal pressure) within the process window. The working diagram is modified due to different types of formation of wrinkling during the hydroforming process. The formation of wrinkles with increasing axial feed can be useful in terms of the achievement of higher bulging ratio and/or less thinning and this type of wrinkles can be overcome through the internal pressure in the later stage of the hydroforming process. On the other hand, the formation of wrinkles may be harmful, if it cannot be reversed.

Keywords: finite element, hydroforming, process window, wrinkling

Procedia PDF Downloads 257
18244 Development and Validation of Research Process for Enhancing Humanities Competence of Medical Students

Authors: S. J. Yune, K. H. Park

Abstract:

The purpose of this study was to examine the validity of the research process for enhancing the humanities competence of the medical students. The research process was developed to be operated as a core subject course of 3 semesters. Among them, the research process for enhancing humanities capacity consisted of humanities and societies (6 teams) and education-psychology (2teams). The subjects of this study were 88-second grade students and 22 professors who participated in the research process. Among them, 13 professors participated in the study of humanities and 37 students. In the validity test, the professors were more likely to have more validity in the research process than the students in all areas of logic (p = .001), influence (p = .037), process (p = .001). The validity of the professor was higher than that of the students. The professors highly evaluated the students' learning outcomes and showed the most frequency to the prize group. As a result of analyzing the agreement between the students and the professors through the Kappa coefficient, the agreement degree of communication and cooperation competence was moderate to .430. Problem-solving ability was .340, which showed a fair degree of agreement. However, other factors showed only a slight degree of agreement of less than .20.

Keywords: research process, medical school, humanities competence, validity verification

Procedia PDF Downloads 162
18243 Numerical Investigation of Solid Subcooling on a Low Melting Point Metal in Latent Thermal Energy Storage Systems Based on Flat Slab Configuration

Authors: Cleyton S. Stampa

Abstract:

This paper addresses the perspectives of using low melting point metals (LMPMs) as phase change materials (PCMs) in latent thermal energy storage (LTES) units, through a numerical approach. This is a new class of PCMs that has been one of the most prospective alternatives to be considered in LTES, due to these materials present high thermal conductivity and elevated heat of fusion, per unit volume. The chosen type of LTES consists of several horizontal parallel slabs filled with PCM. The heat transfer fluid (HTF) circulates through the channel formed between each two consecutive slabs on a laminar regime through forced convection. The study deals with the LTES charging process (heat-storing) by using pure gallium as PCM, and it considers heat conduction in the solid phase during melting driven by natural convection in the melt. The transient heat transfer problem is analyzed in one arbitrary slab under the influence of the HTF. The mathematical model to simulate the isothermal phase change is based on a volume-averaged enthalpy method, which is successfully verified by comparing its predictions with experimental data from works available in the pertinent literature. Regarding the convective heat transfer problem in the HTF, it is assumed that the flow is thermally developing, whereas the velocity profile is already fully developed. The study aims to learn about the effect of the solid subcooling in the melting rate through comparisons with the melting process of the solid in which it starts to melt from its fusion temperature. In order to best understand this effect in a metallic compound, as it is the case of pure gallium, the study also evaluates under the same conditions established for the gallium, the melting process of commercial paraffin wax (organic compound) and of the calcium chloride hexahydrate (CaCl₂ 6H₂O-inorganic compound). In the present work, it is adopted the best options that have been established by several researchers in their parametric studies with respect to this type of LTES, which lead to high values of thermal efficiency. To do so, concerning with the geometric aspects, one considers a gap of the channel formed by two consecutive slabs, thickness and length of the slab. About the HTF, one considers the type of fluid, the mass flow rate, and inlet temperature.

Keywords: flat slab, heat storing, pure metal, solid subcooling

Procedia PDF Downloads 113
18242 Maximum Power Point Tracking Using Fuzzy Logic Control for a Stand-Alone PV System with PI Controller for Battery Charging Based on Evolutionary Technique

Authors: Mohamed A. Moustafa Hassan, Omnia S .S. Hussian, Hany M. Elsaved

Abstract:

This paper introduces the application of Fuzzy Logic Controller (FLC) to extract the Maximum Power Point Tracking (MPPT) from the PV panel. In addition, the proportional integral (PI) controller is used to be the strategy for battery charge control according to acceptable performance criteria. The parameters of the PI controller have been tuned via Modified Adaptive Accelerated Coefficient Particle Swarm Optimization (MAACPSO) technique. The simulation results, using MATLAB/Simulink tools, show that the FLC technique has advantages for use in the MPPT problem, as it provides a fast response under changes in environmental conditions such as radiation and temperature. In addition, the use of PI controller based on MAACPSO results in a good performance in terms of controlling battery charging with constant voltage and current to execute rapid charging.

Keywords: battery charging, fuzzy logic control, maximum power point tracking, PV system, PI controller, evolutionary technique

Procedia PDF Downloads 138
18241 Empirical Study of Correlation between the Cost Performance Index Stability and the Project Cost Forecast Accuracy in Construction Projects

Authors: Amin AminiKhafri, James M. Dawson-Edwards, Ryan M. Simpson, Simaan M. AbouRizk

Abstract:

Earned value management (EVM) has been introduced as an integrated method to combine schedule, budget, and work breakdown structure (WBS). EVM provides various indices to demonstrate project performance including the cost performance index (CPI). CPI is also used to forecast final project cost at completion based on the cost performance during the project execution. Knowing the final project cost during execution can initiate corrective actions, which can enhance project outputs. CPI, however, is not constant during the project, and calculating the final project cost using a variable index is an inaccurate and challenging task for practitioners. Since CPI is based on the cumulative progress values and because of the learning curve effect, CPI variation dampens and stabilizes as project progress. Although various definitions for the CPI stability have been proposed in literature, many scholars have agreed upon the definition that considers a project as stable if the CPI at 20% completion varies less than 0.1 from the final CPI. While 20% completion point is recognized as the stability point for military development projects, construction projects stability have not been studied. In the current study, an empirical study was first conducted using construction project data to determine the stability point for construction projects. Early findings have demonstrated that a majority of construction projects stabilize towards completion (i.e., after 70% completion point). To investigate the effect of CPI stability on cost forecast accuracy, the correlation between CPI stability and project cost at completion forecast accuracy was also investigated. It was determined that as projects progress closer towards completion, variation of the CPI decreases and final project cost forecast accuracy increases. Most projects were found to have 90% accuracy in the final cost forecast at 70% completion point, which is inlined with findings from the CPI stability findings. It can be concluded that early stabilization of the project CPI results in more accurate cost at completion forecasts.

Keywords: cost performance index, earned value management, empirical study, final project cost

Procedia PDF Downloads 133
18240 Numerical Simulation of Filtration Gas Combustion: Front Propagation Velocity

Authors: Yuri Laevsky, Tatyana Nosova

Abstract:

The phenomenon of filtration gas combustion (FGC) had been discovered experimentally at the beginning of 80’s of the previous century. It has a number of important applications in such areas as chemical technologies, fire-explosion safety, energy-saving technologies, oil production. From the physical point of view, FGC may be defined as the propagation of region of gaseous exothermic reaction in chemically inert porous medium, as the gaseous reactants seep into the region of chemical transformation. The movement of the combustion front has different modes, and this investigation is focused on the low-velocity regime. The main characteristic of the process is the velocity of the combustion front propagation. Computation of this characteristic encounters substantial difficulties because of the strong heterogeneity of the process. The mathematical model of FGC is formed by the energy conservation laws for the temperature of the porous medium and the temperature of gas and the mass conservation law for the relative concentration of the reacting component of the gas mixture. In this case the homogenization of the model is performed with the use of the two-temperature approach when at each point of the continuous medium we specify the solid and gas phases with a Newtonian heat exchange between them. The construction of a computational scheme is based on the principles of mixed finite element method with the usage of a regular mesh. The approximation in time is performed by an explicit–implicit difference scheme. Special attention was given to determination of the combustion front propagation velocity. Straight computation of the velocity as grid derivative leads to extremely unstable algorithm. It is worth to note that the term ‘front propagation velocity’ makes sense for settled motion when some analytical formulae linking velocity and equilibrium temperature are correct. The numerical implementation of one of such formulae leading to the stable computation of instantaneous front velocity has been proposed. The algorithm obtained has been applied in subsequent numerical investigation of the FGC process. This way the dependence of the main characteristics of the process on various physical parameters has been studied. In particular, the influence of the combustible gas mixture consumption on the front propagation velocity has been investigated. It also has been reaffirmed numerically that there is an interval of critical values of the interfacial heat transfer coefficient at which a sort of a breakdown occurs from a slow combustion front propagation to a rapid one. Approximate boundaries of such an interval have been calculated for some specific parameters. All the results obtained are in full agreement with both experimental and theoretical data, confirming the adequacy of the model and the algorithm constructed. The presence of stable techniques to calculate the instantaneous velocity of the combustion wave allows considering the semi-Lagrangian approach to the solution of the problem.

Keywords: filtration gas combustion, low-velocity regime, mixed finite element method, numerical simulation

Procedia PDF Downloads 281
18239 Modeling Average Paths Traveled by Ferry Vessels Using AIS Data

Authors: Devin Simmons

Abstract:

At the USDOT’s Bureau of Transportation Statistics, a biannual census of ferry operators in the U.S. is conducted, with results such as route mileage used to determine federal funding levels for operators. AIS data allows for the possibility of using GIS software and geographical methods to confirm operator-reported mileage for individual ferry routes. As part of the USDOT’s work on the ferry census, an algorithm was developed that uses AIS data for ferry vessels in conjunction with known ferry terminal locations to model the average route travelled for use as both a cartographic product and confirmation of operator-reported mileage. AIS data from each vessel is first analyzed to determine individual journeys based on the vessel’s velocity, and changes in velocity over time. These trips are then converted to geographic linestring objects. Using the terminal locations, the algorithm then determines whether the trip represented a known ferry route. Given a large enough dataset, routes will be represented by multiple trip linestrings, which are then filtered by DBSCAN spatial clustering to remove outliers. Finally, these remaining trips are ready to be averaged into one route. The algorithm interpolates the point on each trip linestring that represents the start point. From these start points, a centroid is calculated, and the first point of the average route is determined. Each trip is interpolated again to find the point that represents one percent of the journey’s completion, and the centroid of those points is used as the next point in the average route, and so on until 100 points have been calculated. Routes created using this algorithm have shown demonstrable improvement over previous methods, which included the implementation of a LOESS model. Additionally, the algorithm greatly reduces the amount of manual digitizing needed to visualize ferry activity.

Keywords: ferry vessels, transportation, modeling, AIS data

Procedia PDF Downloads 143
18238 Investigation on Machine Tools Energy Consumptions

Authors: Shiva Abdoli, Daniel T.Semere

Abstract:

Several researches have been conducted to study consumption of energy in cutting process. Most of these researches are focusing to measure the consumption and propose consumption reduction methods. In this work, the relation between the cutting parameters and the consumption is investigated in order to establish a generalized energy consumption model that can be used for process and production planning in real production lines. Using the generalized model, the process planning will be carried out by taking into account the energy as a function of the selected process parameters. Similarly, the generalized model can be used in production planning to select the right operational parameters like batch sizes, routing, buffer size, etc. in a production line. The description and derivation of the model as well as a case study are given in this paper to illustrate the applicability and validity of the model.

Keywords: process parameters, cutting process, energy efficiency, Material Removal Rate (MRR)

Procedia PDF Downloads 475
18237 Information Technology Service Management System Measurement Using ISO20000-1 and ISO15504-8

Authors: Imam Asrowardi, Septafiansyah Dwi Putra, Eko Subyantoro

Abstract:

Process assessments can improve IT service management system (IT SMS) processes but the assessment method is not always transparent. This paper outlines a project to develop a solution- mediated process assessment tool to enable transparent and objective SMS process assessment. Using the international standards for SMS and process assessment, the tool is being developed following the International standard approach in collaboration and evaluate by expert judgment from committee members and ITSM practitioners.

Keywords: SMS, tools evaluation, ITIL, ISO service

Procedia PDF Downloads 455
18236 Effect of Intrinsic Point Defects on the Structural and Optical Properties of SnO₂ Thin Films Grown by Ultrasonic Spray Pyrolysis Method

Authors: Fatiha Besahraoui, M'hamed Guezzoul, Kheira Chebbah, M'hamed Bouslama

Abstract:

SnO₂ thin film is characterized by Atomic Force Microscopy (AFM) and Photoluminescence Spectroscopies. AFM images show a dense surface of columnar grains with a roughness of 78.69 nm. The PL measurements at 7 K reveal the presence of PL peaks centered in IR and visible regions. They are attributed to radiative transitions via oxygen vacancies, Sn interstitials, and dangling bonds. A bands diagram model is presented with the approximate positions of intrinsic point defect levels in SnO₂ thin films. The integrated PL measurements demonstrate the good thermal stability of our sample, which makes it very useful in optoelectronic devices functioning at room temperature. The unusual behavior of the evolution of PL peaks and their full width at half maximum as a function of temperature indicates the thermal sensitivity of the point defects present in the band gap. The shallower energy levels due to dangling bonds and/or oxygen vacancies are more sensitive to the temperature. However, volume defects like Sn interstitials are thermally stable and constitute deep and stable energy levels for excited electrons. Small redshifting of PL peaks is observed with increasing temperature. This behavior is attributed to the reduction of oxygen vacancies.

Keywords: transparent conducting oxide, photoluminescence, intrinsic point defects, semiconductors, oxygen vacancies

Procedia PDF Downloads 60
18235 A Review of Applying Serious Games on Learning

Authors: Carlos Oliveira, Ulrick Pimentel

Abstract:

Digital games have conquered a growing space in the lives of children, adolescents and adults. In this perspective, the use of this resource has shown to be an important strategy that facilitates the learning process. This research is a literature review on the use of serious games in teaching, which shows the characteristics of these games, the benefits and possible harms that this resource can produce, in addition to the possible methods of evaluating the effectiveness of this resource in teaching. The results point out that Serious Games have significant potential as a tool for instruction. However, their effectiveness in terms of learning outcomes is still poorly studied, mainly due to the complexity involved in evaluating intangible measures.

Keywords: serious games, learning, application, literature review

Procedia PDF Downloads 278
18234 Numerical Simulation of a Three-Dimensional Framework under the Action of Two-Dimensional Moving Loads

Authors: Jia-Jang Wu

Abstract:

The objective of this research is to develop a general technique so that one may predict the dynamic behaviour of a three-dimensional scale crane model subjected to time-dependent moving point forces by means of conventional finite element computer packages. To this end, the whole scale crane model is divided into two parts: the stationary framework and the moving substructure. In such a case, the dynamic responses of a scale crane model can be predicted from the forced vibration responses of the stationary framework due to actions of the four time-dependent moving point forces induced by the moving substructure. Since the magnitudes and positions of the moving point forces are dependent on the relative positions between the trolley, moving substructure and the stationary framework, it can be found from the numerical results that the time histories for the moving speeds of the moving substructure and the trolley are the key factors affecting the dynamic responses of the scale crane model.

Keywords: moving load, moving substructure, dynamic responses, forced vibration responses

Procedia PDF Downloads 327
18233 Design and Advancement of Hybrid Multilevel Inverter Interface with PhotoVoltaic

Authors: P.Kiruthika, K. Ramani

Abstract:

This paper presented the design and advancement of a single-phase 27-level Hybrid Multilevel DC-AC Converter interfacing with Photo Voltaic. In this context, the Multicarrier Pulse Width Modulation method can be implemented in 27-level Hybrid Multilevel Inverter for generating a switching pulse. Perturb & Observer algorithm can be used in the Maximum Power Point Tracking method for the Photo Voltaic system. By implementing Maximum Power Point Tracking with three separate solar panels as an input source to the 27-level Hybrid Multilevel Inverter. This proposed method can be simulated by using MATLAB/simulink. The result shown that the proposed method can achieve silky output wave forms, more flexibility in voltage range, and to reduce Total Harmonic Distortion in medium-voltage drives.

Keywords: Multi Carrier Pulse Width Modulation Technique (MCPWM), Multi Level Inverter (MLI), Maximum Power Point Tracking (MPPT), Perturb and Observer (P&O)

Procedia PDF Downloads 556
18232 Online Learning for Modern Business Models: Theoretical Considerations and Algorithms

Authors: Marian Sorin Ionescu, Olivia Negoita, Cosmin Dobrin

Abstract:

This scientific communication reports and discusses learning models adaptable to modern business problems and models specific to digital concepts and paradigms. In the PAC (probably approximately correct) learning model approach, in which the learning process begins by receiving a batch of learning examples, the set of learning processes is used to acquire a hypothesis, and when the learning process is fully used, this hypothesis is used in the prediction of new operational examples. For complex business models, a lot of models should be introduced and evaluated to estimate the induced results so that the totality of the results are used to develop a predictive rule, which anticipates the choice of new models. In opposition, for online learning-type processes, there is no separation between the learning (training) and predictive phase. Every time a business model is approached, a test example is considered from the beginning until the prediction of the appearance of a model considered correct from the point of view of the business decision. After choosing choice a part of the business model, the label with the logical value "true" is known. Some of the business models are used as examples of learning (training), which helps to improve the prediction mechanisms for future business models.

Keywords: machine learning, business models, convex analysis, online learning

Procedia PDF Downloads 118
18231 A Geographical Study of Vindhyanchal in Mirzapur City, U.P. India

Authors: Akhilendra Nath Tiwary

Abstract:

Vindhyanchal is a very famous pilgrimage and tourism site in the west of Mirzapur city of Uttar Pradesh State in India. The city in east is a commercial center for cotton, metal ware and carpets. Among the Hindu population, it is believed that the primordial creative forces of the GOD and the power of the GODDESS make respective triangles which superimpose opposite to each other as hexagram at a point or node (Bindu (point) +Vasini (located) or Vindhyavasini, located in a point/node). Mirzapur city has served as a natural connecting point between north and south India. Before independence of India from Britain in 1947, it was a flourishing commercial center. Post-independence, the negligence of planning authorities and nexus of bureaucrats and politicians started affecting its development. In the meantime, emergence of new industrial cities as Kanpur, Agra, Moradabad, etc., nearer to the capital city of Delhi, posed serious challenges to the development of this small city as many commercial and business activities along with the skilled workforce started shifting to these new cities or to the relatively bigger neighboring cities of Varanasi in east and Allahabad in west. In the present paper, the significant causes, issues and challenges in development of Vindhyanchal is discussed with geographical perspective. An attempt has been made to find out the ways to restore the lost glory of the city as a center of pilgrimage, tourism, and commerce.

Keywords: cultural node, pilgrimage, sacred, Vindhyan triangle, ommercial centre

Procedia PDF Downloads 412
18230 Maximum Power Point Tracking for Small Scale Wind Turbine Using Multilayer Perceptron Neural Network Implementation without Mechanical Sensor

Authors: Piyangkun Kukutapan, Siridech Boonsang

Abstract:

The article proposes maximum power point tracking without mechanical sensor using Multilayer Perceptron Neural Network (MLPNN). The aim of article is to reduce the cost and complexity but still retain efficiency. The experimental is that duty cycle is generated maximum power, if it has suitable qualification. The measured data from DC generator, voltage (V), current (I), power (P), turnover rate of power (dP), and turnover rate of voltage (dV) are used as input for MLPNN model. The output of this model is duty cycle for driving the converter. The experiment implemented using Arduino Uno board. This diagram is compared to MPPT using MLPNN and P&O control (Perturbation and Observation control). The experimental results show that the proposed MLPNN based approach is more efficiency than P&O algorithm for this application.

Keywords: maximum power point tracking, multilayer perceptron netural network, optimal duty cycle, DC generator

Procedia PDF Downloads 304
18229 Cadmium Adsorption by Modified Magnetic Biochar

Authors: Chompoonut Chaiyaraksa, Chanida Singbubpha, Kliaothong Angkabkingkaew, Thitikorn Boonyasawin

Abstract:

Heavy metal contamination in an environment is an important problem in Thailand that needs to be addressed urgently, particularly contaminated with water. It can spread to other environments faster. This research aims to study the adsorption of cadmium ion by unmodified biochar and sodium dodecyl sulfate modified magnetic biochar derived from Eichhornia Crassipes. The determination of the adsorbent characteristics was by Scanning Electron Microscope, Fourier Transform Infrared Spectrometer, X-ray Diffractometer, and the pH drift method. This study also included the comparison of adsorption efficiency of both types of biochar, adsorption isotherms, and kinetics. The pH value at the point of zero charges of the unmodified biochar and modified magnetic biochar was 7.40 and 3.00, respectively. The maximum value of adsorption reached when using pH 8. The equilibrium adsorption time was 5 hours and 1 hour for unmodified biochar and modified magnetic biochar, respectively. The cadmium adsorption by both adsorbents followed Freundlich, Temkin, and Dubinin – Radushkevich isotherm model and the pseudo-second-order kinetic. The adsorption process was spontaneous at high temperatures and non-spontaneous at low temperatures. It was an endothermic process, physisorption in nature, and can occur naturally.

Keywords: Eichhornia crassipes, magnetic biochar, sodium dodecyl sulfate, water treatment

Procedia PDF Downloads 141
18228 A Randomised Controlled Trial and Process Evaluation of the Lifestart Parenting Programme

Authors: Sharon Millen, Sarah Miller, Laura Dunne, Clare McGeady, Laura Neeson

Abstract:

This paper presents the findings from a randomised controlled trial (RCT) and process evaluation of the Lifestart parenting programme. Lifestart is a structured child-centred programme of information and practical activity for parents of children aged from birth to five years of age. It is delivered to parents in their own homes by trained, paid family visitors and it is offered to parents regardless of their social, economic or other circumstances. The RCT evaluated the effectiveness of the programme and the process evaluation documented programme delivery and included a qualitative exploration of parent and child outcomes. 424 parents and children participated in the RCT: 216 in the intervention group and 208 in the control group across the island of Ireland. Parent outcomes included: parental knowledge of child development, parental efficacy, stress, social support, parenting skills and embeddedness in the community. Child outcomes included cognitive, language and motor development and social-emotional and behavioural development. Both groups were tested at baseline (when children were less than 1 year old), mid-point (aged 3) and at post-test (aged 5). Data were collected during a home visit, which took two hours. The process evaluation consisted of interviews with parents (n=16 at baseline and end-point), and focus groups with Lifestart Coordinators (n=9) and Family Visitors (n=24). Quantitative findings from the RCT indicated that, compared to the control group, parents who received the Lifestart programme reported reduced parenting-related stress, increased knowledge of their child’s development, and improved confidence in their parenting role. These changes were statistically significant and consistent with the hypothesised pathway of change depicted in the logic model. There was no evidence of any change in parents’ embeddedness in the community. Although four of the five child outcomes showed small positive change for children who took part in the programme, these were not statistically significant and there is no evidence that the programme improves child cognitive and non-cognitive skills by immediate post-test. The qualitative process evaluation highlighted important challenges related to conducting trials of this magnitude and design in the general population. Parents reported that a key incentive to take part in study was receiving feedback from the developmental assessment, which formed part of the data collection. This highlights the potential importance of appropriate incentives in relation to recruitment and retention of participants. The interviews with intervention parents indicated that one of the first changes they experienced as a result of the Lifestart programme was increased knowledge and confidence in their parenting ability. The outcomes and pathways perceived by parents and described in the interviews are also consistent with the findings of the RCT and the theory of change underpinning the programme. This hypothesises that improvement in parental outcomes, arising as a consequence of the programme, mediate the change in child outcomes. Parents receiving the Lifestart programme reported great satisfaction with and commitment to the programme, with the role of the Family Visitor being identified as one of the key components of the programme.

Keywords: parent-child relationship, parental self-efficacy, parental stress, school readiness

Procedia PDF Downloads 413
18227 Transport and Mixing Phenomena Developed by Vortex Formation in Flow around Airfoil Using Lagrangian Coherent Structures

Authors: Riaz Ahmad, Jiazhong Zhang, Asma Farooqi

Abstract:

In this study, mass transport between separation bubbles and the flow around a two-dimensional airfoil are numerically investigated using Lagrangian Coherent Structures (LCSs). Finite Time Lyapunov Exponent (FTLE) technique is used for the computation to identify invariant manifolds and LCSs. Moreover, the Characteristic Base Split (CBS) scheme combined with dual time stepping technique is applied to simulate such transient flow at low Reynolds number. We then investigate the evolution of vortex structures during the transport process with the aid of LCSs. To explore the vortex formation at the surface of the airfoil, the dynamics of separatrix is also taken into account which is formed by the combination of stable-unstable manifolds. The Lagrangian analysis gives a detailed understanding of vortex dynamics and separation bubbles which plays a significant role to explore the performance of the unsteady flow generated by the airfoil. Transport process and flow separation phenomena are studied extensively to analyze the flow pattern by Lagrangian point of view.

Keywords: transport phenomena, CBS Method, vortex formation, Lagrangian Coherent Structures

Procedia PDF Downloads 113
18226 Optimizing Performance of Tablet's Direct Compression Process Using Fuzzy Goal Programming

Authors: Abbas Al-Refaie

Abstract:

This paper aims at improving the performance of the tableting process using statistical quality control and fuzzy goal programming. The tableting process was studied. Statistical control tools were used to characterize the existing process for three critical responses including the averages of a tablet’s weight, hardness, and thickness. At initial process factor settings, the estimated process capability index values for the tablet’s averages of weight, hardness, and thickness were 0.58, 3.36, and 0.88, respectively. The L9 array was utilized to provide experimentation design. Fuzzy goal programming was then employed to find the combination of optimal factor settings. Optimization results showed that the process capability index values for a tablet’s averages of weight, hardness, and thickness were improved to 1.03, 4.42, and 1.42, respectively. Such improvements resulted in significant savings in quality and production costs.

Keywords: fuzzy goal programming, control charts, process capability, tablet optimization

Procedia PDF Downloads 245
18225 Linearization and Process Standardization of Construction Design Engineering Workflows

Authors: T. R. Sreeram, S. Natarajan, C. Jena

Abstract:

Civil engineering construction is a network of tasks involving varying degree of complexity and streamlining, and standardization is the only way to establish a systemic approach to design. While there are off the shelf tools such as AutoCAD that play a role in the realization of design, the repeatable process in which these tools are deployed often is ignored. The present paper addresses this challenge through a sustainable design process and effective standardizations at all stages in the design workflow. The same is demonstrated through a case study in the context of construction, and further improvement points are highlighted.

Keywords: syste, lean, value stream, process improvement

Procedia PDF Downloads 99
18224 Development of a Multi-Factorial Instrument for Accident Analysis Based on Systemic Methods

Authors: C. V. Pietreanu, S. E. Zaharia, C. Dinu

Abstract:

The present research is built on three major pillars, commencing by making some considerations on accident investigation methods and pointing out both defining aspects and differences between linear and non-linear analysis. The traditional linear focus on accident analysis describes accidents as a sequence of events, while the latest systemic models outline interdependencies between different factors and define the processes evolution related to a specific (normal) situation. Linear and non-linear accident analysis methods have specific limitations, so the second point of interest is mirrored by the aim to discover the drawbacks of systemic models which becomes a starting point for developing new directions to identify risks or data closer to the cause of incidents/accidents. Since communication represents a critical issue in the interaction of human factor and has been proved to be the answer of the problems made by possible breakdowns in different communication procedures, from this focus point, on the third pylon a new error-modeling instrument suitable for risk assessment/accident analysis will be elaborated.

Keywords: accident analysis, multi-factorial error modeling, risk, systemic methods

Procedia PDF Downloads 183
18223 The LMPA/Epoxy Mixture Encapsulation of OLED on Polyimide Substrate

Authors: Chuyi Ye, Minsang Kim, Cheol-Hee Moon

Abstract:

The organic light emitting diode(OLED), is a potential organic optical functional materials which is considered as the next generation display technology with the advantages such as all-solid state, ultra-thin thickness, active luminous and flexibility. Due to the development of polymer-inorganic substrate, it becomes possible to achieve the flexible OLED display. However the organic light-emitting material is very sensitive to the oxygen and water vapor, and the encapsulation requires water vapor transmission rate(WVTR) and oxygen transmission rate(OTR) as lower as 10-6 g/(m2.d) and 10-5 cm3/(m2.d) respectively. In current situation, the rigorous WVTR and OTR have restricted the application of the OLED display. Traditional epoxy/getter or glass frit approaches, which have been widely applied on glass-substrate-based devices, are not suitable for transparent flexible organic devices, and mechanically flexible thin-film approaches are required. To ensure the OLED’s lifetime, the encapsulation material of the OLED package is very important. In this paper, a low melting point alloy(LMPA)-epoxy mixture in the encapsulation process is introduced. There will be a phase separation when the mixture is heated to the melting of LMPA and the formation of the double line structure between two substrates: the alloy barrier has extremely low WVTR and OTR and the epoxy fills the potential tiny cracks. In our experiment, the PI film is chosen as a flexible transparent substrate, and Mo and Cu are deposited on the PI film successively. Then the two metal layers are photolithographied to the sealing pattern line. The Mo is a transition layer between the PI film and Cu, at the same time, the Cu has a good wettability with the LMPA(Sn-58Bi). At last, pattern is printed with LMPA layer and applied voltage, the gathering Joule heat melt the LMPA and form the double line structure and the OLED package is sealed in the same time. In this research, the double-line encapsulating structure of LMPA and epoxy on the PI film is manufactured for the flexible OLED encapsulation, and in this process it is investigated whether the encapsulation satisfies the requirement of WVTR and OTR for the flexible OLED.

Keywords: encapsulation, flexible, low melting point alloy, OLED

Procedia PDF Downloads 576
18222 Crude Distillation Process Simulation Using Unisim Design Simulator

Authors: C. Patrascioiu, M. Jamali

Abstract:

The paper deals with the simulation of the crude distillation process using the Unisim Design simulator. The necessity of simulating this process is argued both by considerations related to the design of the crude distillation column, but also by considerations related to the design of advanced control systems. In order to use the Unisim Design simulator to simulate the crude distillation process, the identification of the simulators used in Romania and an analysis of the PRO/II, HYSYS, and Aspen HYSYS simulators were carried out. Analysis of the simulators for the crude distillation process has allowed the authors to elaborate the conclusions of the success of the crude modelling. A first aspect developed by the authors is the implementation of specific problems of petroleum liquid-vapors equilibrium using Unisim Design simulator. The second major element of the article is the development of the methodology and the elaboration of the simulation program for the crude distillation process, using Unisim Design resources. The obtained results validate the proposed methodology and will allow dynamic simulation of the process.  

Keywords: crude oil, distillation, simulation, Unisim Design, simulators

Procedia PDF Downloads 219
18221 Experimental Study and Numerical Modelling of Failure of Rocks Typical for Kuzbass Coal Basin

Authors: Mikhail O. Eremin

Abstract:

Present work is devoted to experimental study and numerical modelling of failure of rocks typical for Kuzbass coal basin (Russia). The main goal was to define strength and deformation characteristics of rocks on the base of uniaxial compression and three-point bending loadings and then to build a mathematical model of failure process for both types of loading. Depending on particular physical-mechanical characteristics typical rocks of Kuzbass coal basin (sandstones, siltstones, mudstones, etc. of different series – Kolchuginsk, Tarbagansk, Balohonsk) manifest brittle and quasi-brittle character of failure. The strength characteristics for both tension and compression are found. Other characteristics are also found from the experiment or taken from literature reviews. On the base of obtained characteristics and structure (obtained from microscopy) the mathematical and structural models are built and numerical modelling of failure under different types of loading is carried out. Effective characteristics obtained from modelling and character of failure correspond to experiment and thus, the mathematical model was verified. An Instron 1185 machine was used to carry out the experiments. Mathematical model includes fundamental conservation laws of solid mechanics – mass, impulse, energy. Each rock has a sufficiently anisotropic structure, however, each crystallite might be considered as isotropic and then a whole rock model has a quasi-isotropic structure. This idea gives an opportunity to use the Hooke’s law inside of each crystallite and thus explicitly accounting for the anisotropy of rocks and the stress-strain state at loading. Inelastic behavior is described in frameworks of two different models: von Mises yield criterion and modified Drucker-Prager yield criterion. The damage accumulation theory is also implemented in order to describe a failure process. Obtained effective characteristics of rocks are used then for modelling of rock mass evolution when mining is carried out both by an open-pit or underground opening.

Keywords: damage accumulation, Drucker-Prager yield criterion, failure, mathematical modelling, three-point bending, uniaxial compression

Procedia PDF Downloads 149
18220 Prediction of Ionic Liquid Densities Using a Corresponding State Correlation

Authors: Khashayar Nasrifar

Abstract:

Ionic liquids (ILs) exhibit particular properties exemplified by extremely low vapor pressure and high thermal stability. The properties of ILs can be tailored by proper selection of cations and anions. As such, ILs are appealing as potential solvents to substitute traditional solvents with high vapor pressure. One of the IL properties required in chemical and process design is density. In developing corresponding state liquid density correlations, scaling hypothesis is often used. The hypothesis expresses the temperature dependence of saturated liquid densities near the vapor-liquid critical point as a function of reduced temperature. Extending the temperature dependence, several successful correlations were developed to accurately correlate the densities of normal liquids from the triple point to a critical point. Applying mixing rules, the liquid density correlations are extended to liquid mixtures as well. ILs are not molecular liquids, and they are not classified among normal liquids either. Also, ILs are often used where the condition is far from equilibrium. Nevertheless, in calculating the properties of ILs, the use of corresponding state correlations would be useful if no experimental data were available. With well-known generalized saturated liquid density correlations, the accuracy in predicting the density of ILs is not that good. An average error of 4-5% should be expected. In this work, a data bank was compiled. A simplified and concise corresponding state saturated liquid density correlation is proposed by phenomena-logically modifying reduced temperature using the temperature-dependence for an interacting parameter of the Soave-Redlich-Kwong equation of state. This modification improves the temperature dependence of the developed correlation. Parametrization was next performed to optimize the three global parameters of the correlation. The correlation was then applied to the ILs in our data bank with satisfactory predictions. The correlation of IL density applied at 0.1 MPa and was tested with an average uncertainty of around 2%. No adjustable parameter was used. The critical temperature, critical volume, and acentric factor were all required. Methods to extend the predictions to higher pressures (200 MPa) were also devised. Compared to other methods, this correlation was found more accurate. This work also presents the chronological order of developing such correlations dealing with ILs. The pros and cons are also expressed.

Keywords: correlation, corresponding state principle, ionic liquid, density

Procedia PDF Downloads 105
18219 Deep Learning Approach for Colorectal Cancer’s Automatic Tumor Grading on Whole Slide Images

Authors: Shenlun Chen, Leonard Wee

Abstract:

Tumor grading is an essential reference for colorectal cancer (CRC) staging and survival prognostication. The widely used World Health Organization (WHO) grading system defines histological grade of CRC adenocarcinoma based on the density of glandular formation on whole slide images (WSI). Tumors are classified as well-, moderately-, poorly- or un-differentiated depending on the percentage of the tumor that is gland forming; >95%, 50-95%, 5-50% and <5%, respectively. However, manually grading WSIs is a time-consuming process and can cause observer error due to subjective judgment and unnoticed regions. Furthermore, pathologists’ grading is usually coarse while a finer and continuous differentiation grade may help to stratifying CRC patients better. In this study, a deep learning based automatic differentiation grading algorithm was developed and evaluated by survival analysis. Firstly, a gland segmentation model was developed for segmenting gland structures. Gland regions of WSIs were delineated and used for differentiation annotating. Tumor regions were annotated by experienced pathologists into high-, medium-, low-differentiation and normal tissue, which correspond to tumor with clear-, unclear-, no-gland structure and non-tumor, respectively. Then a differentiation prediction model was developed on these human annotations. Finally, all enrolled WSIs were processed by gland segmentation model and differentiation prediction model. The differentiation grade can be calculated by deep learning models’ prediction of tumor regions and tumor differentiation status according to WHO’s defines. If multiple WSIs were possessed by a patient, the highest differentiation grade was chosen. Additionally, the differentiation grade was normalized into scale between 0 to 1. The Cancer Genome Atlas, project COAD (TCGA-COAD) project was enrolled into this study. For the gland segmentation model, receiver operating characteristic (ROC) reached 0.981 and accuracy reached 0.932 in validation set. For the differentiation prediction model, ROC reached 0.983, 0.963, 0.963, 0.981 and accuracy reached 0.880, 0.923, 0.668, 0.881 for groups of low-, medium-, high-differentiation and normal tissue in validation set. Four hundred and one patients were selected after removing WSIs without gland regions and patients without follow up data. The concordance index reached to 0.609. Optimized cut off point of 51% was found by “Maxstat” method which was almost the same as WHO system’s cut off point of 50%. Both WHO system’s cut off point and optimized cut off point performed impressively in Kaplan-Meier curves and both p value of logrank test were below 0.005. In this study, gland structure of WSIs and differentiation status of tumor regions were proven to be predictable through deep leaning method. A finer and continuous differentiation grade can also be automatically calculated through above models. The differentiation grade was proven to stratify CAC patients well in survival analysis, whose optimized cut off point was almost the same as WHO tumor grading system. The tool of automatically calculating differentiation grade may show potential in field of therapy decision making and personalized treatment.

Keywords: colorectal cancer, differentiation, survival analysis, tumor grading

Procedia PDF Downloads 115