Search results for: cone-beam computed tomography CBCT
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 1062

Search results for: cone-beam computed tomography CBCT

192 Real-Time Kinetic Analysis of Labor-Intensive Repetitive Tasks Using Depth-Sensing Camera

Authors: Sudip Subedi, Nipesh Pradhananga

Abstract:

The musculoskeletal disorders, also known as MSDs, are common in construction workers. MSDs include lower back injuries, knee injuries, spinal injuries, and joint injuries, among others. Since most construction tasks are still manual, construction workers often need to perform repetitive, labor-intensive tasks. And they need to stay in the same or an awkward posture for an extended time while performing such tasks. It induces significant stress to the joints and spines, increasing the risk of getting into MSDs. Manual monitoring of such tasks is virtually impossible with the handful of safety managers in a construction site. This paper proposes a methodology for performing kinetic analysis of the working postures while performing such tasks in real-time. Skeletal of different workers will be tracked using a depth-sensing camera while performing the task to create training data for identifying the best posture. For this, the kinetic analysis will be performed using a human musculoskeletal model in an open-source software system (OpenSim) to visualize the stress induced by essential joints. The “safe posture” inducing lowest stress on essential joints will be computed for different actions involved in the task. The identified “safe posture” will serve as a basis for real-time monitoring and identification of awkward and unsafe postural behaviors of construction workers. Besides, the temporal simulation will be carried out to find the associated long-term effect of repetitive exposure to such observed postures. This will help to create awareness in workers about potential future health hazards and encourage them to work safely. Furthermore, the collected individual data can then be used to provide need-based personalized training to the construction workers.

Keywords: construction workers’ safety, depth sensing camera, human body kinetics, musculoskeletal disorders, real time monitoring, repetitive labor-intensive tasks

Procedia PDF Downloads 130
191 Analysis of the Contribution of Coastal and Marine Physical Factors to Oil Slick Movement: Case Study of Misrata, Libya

Authors: Abduladim Maitieg, Mark Johnson

Abstract:

Developing a coastal oil spill management plan for the Misratah coast is the motivating factor for building a database for coastal and marine systems and energy resources. Wind direction and speed, currents, bathymetry, coastal topography and offshore dynamics influence oil spill deposition in coastal water. Therefore, oceanographic and climatological data can be used to understand oil slick movement and potential oil deposits on shoreline area and the behaviour of oil spill trajectories on the sea surface. The purpose of this study is to investigate the effects of the coastal and marine physical factors under strong wave conditions and various bathymetric and coastal topography gradients in the western coastal area of Libya on the movement of oil slicks. The movement of oil slicks was computed using a GNOME simulation model based on current and wind speed/direction. The results in this paper show that (1) Oil slick might reach the Misratah shoreline area in two days in the summer and winter. Seasons. (2 ) The North coast of Misratah is the potential oil deposit area on the Misratah coast. (3) Tarball pollution was observed along the North coast of Misratah. (4) Two scenarios for the summer and the winter season were run, along the western coast of Libya . (5) The eastern coast is at a lower potential risk due to the influence of wind and current energy in the Gulf of Sidra. (6) The Misratah coastline is more vulnerable to oil spill movement in the summer than in winter seasons. (7) Oil slick takes from 2 to 5 days to reach the saltmarsh in the eastern Misratah coast. (8) Oil slick moves 300 km in 30 days from the spill resource location near the Libyan western border to the Misratah coast.(9) Bathymetric features have a profound effect on oil spill movement. (9)Oil dispersion simulations using GNOME are carried out taking into account high-resolution wind and current data.

Keywords: oil spill movement, coastal and marine physical factors, coast area, Libyan

Procedia PDF Downloads 225
190 Evaluation of Weather Risk Insurance for Agricultural Products Using a 3-Factor Pricing Model

Authors: O. Benabdeljelil, A. Karioun, S. Amami, R. Rouger, M. Hamidine

Abstract:

A model for preventing the risks related to climate conditions in the agricultural sector is presented. It will determine the yearly optimum premium to be paid by a producer in order to reach his required turnover. The model is based on both climatic stability and 'soft' responses of usually grown species to average climate variations at the same place and inside a safety ball which can be determined from past meteorological data. This allows the use of linear regression expression for dependence of production result in terms of driving meteorological parameters, the main ones of which are daily average sunlight, rainfall and temperature. By simple best parameter fit from the expert table drawn with professionals, optimal representation of yearly production is determined from records of previous years, and yearly payback is evaluated from minimum yearly produced turnover. The model also requires accurate pricing of commodity at N+1. Therefore, a pricing model is developed using 3 state variables, namely the spot price, the difference between the mean-term and the long-term forward price, and the long-term structure of the model. The use of historical data enables to calibrate the parameters of state variables, and allows the pricing of commodity. Application to beet sugar underlines pricer precision. Indeed, the percentage of accuracy between computed result and real world is 99,5%. Optimal premium is then deduced and gives the producer a useful bound for negotiating an offer by insurance companies to effectively protect its harvest. The application to beet production in French Oise department illustrates the reliability of present model with as low as 6% difference between predicted and real data. The model can be adapted to almost any agricultural field by changing state parameters and calibrating their associated coefficients.

Keywords: agriculture, production model, optimal price, meteorological factors, 3-factor model, parameter calibration, forward price

Procedia PDF Downloads 376
189 Evaluating the Impact of Urban Green Spaces on Urban Microclimate of Lahore: A Rapidly Urbanizing Metropolis of the Punjab-Pakistan

Authors: Muhammad Nasar-U-Minallah, Dagmar Haase, Salman Qureshi, Safdar Ali Shirazi

Abstract:

Urban green spaces (UGS) play a key role in the urban ecology of an area since they provide significant ecological services to compensate for natural environment functions damaged by the rapid growth of urbanization. The transformation of urban green specs to impervious landscapes has been recognized as a key factor prompting the distinctive urban heat and associated microclimatic changes. There is no doubt that urban green spaces offer a range of ecosystem services that can help to mitigate the ill effects of urbanization, heat anomalies, and climate change. The present study attempts to appraise the impact of urban green spaces on the urban thermal environment for the development of the microclimatic conditions in Lahore, Pakistan. The influence of urban heat has been studied through Landsat 8 data. The land surface temperature (LST) of Lahore was computed through the Radiative transfer method (RTM). The spatial variation of land surface temperature is retrieved to describe their local heat effect on urban microclimate. The association between the LST, normalized difference vegetation index, and the normalized difference built-up index are investigated to explore the impact of the urban green spaces and impervious surfaces on urban microclimate. The results of this study show significant changes in (impervious land surface 18% increase) land use within the study area. However, conversion of natural green cover to commercial and residential uses considerably increases the LST. Furthermore, results show that green spaces were the major heat sinks while impervious landscapes were the major heat source in the study area. Urban green spaces reveal 1 to 3℃ lower LST associated with their surrounding urban built-up area. This study shows that urban green spaces will help to mitigate the effect of urban microclimate and it is significant for the sustainable urban environment as well as to improve the quality of life of the urban inhabitants.

Keywords: thermal environmental, urban green space, cooling effect, microclimate, Lahore

Procedia PDF Downloads 105
188 Comparison of Receiver Operating Characteristic Curve Smoothing Methods

Authors: D. Sigirli

Abstract:

The Receiver Operating Characteristic (ROC) curve is a commonly used statistical tool for evaluating the diagnostic performance of screening and diagnostic test with continuous or ordinal scale results which aims to predict the presence or absence probability of a condition, usually a disease. When the test results were measured as numeric values, sensitivity and specificity can be computed across all possible threshold values which discriminate the subjects as diseased and non-diseased. There are infinite numbers of possible decision thresholds along the continuum of the test results. The ROC curve presents the trade-off between sensitivity and the 1-specificity as the threshold changes. The empirical ROC curve which is a non-parametric estimator of the ROC curve is robust and it represents data accurately. However, especially for small sample sizes, it has a problem of variability and as it is a step function there can be different false positive rates for a true positive rate value and vice versa. Besides, the estimated ROC curve being in a jagged form, since the true ROC curve is a smooth curve, it underestimates the true ROC curve. Since the true ROC curve is assumed to be smooth, several smoothing methods have been explored to smooth a ROC curve. These include using kernel estimates, using log-concave densities, to fit parameters for the specified density function to the data with the maximum-likelihood fitting of univariate distributions or to create a probability distribution by fitting the specified distribution to the data nd using smooth versions of the empirical distribution functions. In the present paper, we aimed to propose a smooth ROC curve estimation based on the boundary corrected kernel function and to compare the performances of ROC curve smoothing methods for the diagnostic test results coming from different distributions in different sample sizes. We performed simulation study to compare the performances of different methods for different scenarios with 1000 repetitions. It is seen that the performance of the proposed method was typically better than that of the empirical ROC curve and only slightly worse compared to the binormal model when in fact the underlying samples were generated from the normal distribution.

Keywords: empirical estimator, kernel function, smoothing, receiver operating characteristic curve

Procedia PDF Downloads 152
187 Experimental Modeling of Spray and Water Sheet Formation Due to Wave Interactions with Vertical and Slant Bow-Shaped Model

Authors: Armin Bodaghkhani, Bruce Colbourne, Yuri S. Muzychka

Abstract:

The process of spray-cloud formation and flow kinematics produced from breaking wave impact on vertical and slant lab-scale bow-shaped models were experimentally investigated. Bubble Image Velocimetry (BIV) and Image Processing (IP) techniques were applied to study the various types of wave-model impacts. Different wave characteristics were generated in a tow tank to investigate the effects of wave characteristics, such as wave phase velocity, wave steepness on droplet velocities, and behavior of the process of spray cloud formation. The phase ensemble-averaged vertical velocity and turbulent intensity were computed. A high-speed camera and diffused LED backlights were utilized to capture images for further post processing. Various pressure sensors and capacitive wave probes were used to measure the wave impact pressure and the free surface profile at different locations of the model and wave-tank, respectively. Droplet sizes and velocities were measured using BIV and IP techniques to trace bubbles and droplets in order to measure their velocities and sizes by correlating the texture in these images. The impact pressure and droplet size distributions were compared to several previously experimental models, and satisfactory agreements were achieved. The distribution of droplets in front of both models are demonstrated. Due to the highly transient process of spray formation, the drag coefficient for several stages of this transient displacement for various droplet size ranges and different Reynolds number were calculated based on the ensemble average method. From the experimental results, the slant model produces less spray in comparison with the vertical model, and the droplet velocities generated from the wave impact with the slant model have a lower velocity as compared with the vertical model.

Keywords: spray charachteristics, droplet size and velocity, wave-body interactions, bubble image velocimetry, image processing

Procedia PDF Downloads 300
186 BiFormerDTA: Structural Embedding of Protein in Drug Target Affinity Prediction Using BiFormer

Authors: Leila Baghaarabani, Parvin Razzaghi, Mennatolla Magdy Mostafa, Ahmad Albaqsami, Al Warith Al Rushaidi, Masoud Al Rawahi

Abstract:

Predicting the interaction between drugs and their molecular targets is pivotal for advancing drug development processes. Due to the time and cost limitations, computational approaches have emerged as an effective approach to drug-target interaction (DTI) prediction. Most of the introduced computational based approaches utilize the drug molecule and protein sequence as input. This study does not only utilize these inputs, it also introduces a protein representation developed using a masked protein language model. In this representation, for every individual amino acid residue within the protein sequence, there exists a corresponding probability distribution that indicates the likelihood of each amino acid being present at that particular position. Then, the similarity between each pair of amino acids is computed to create a similarity matrix. To encode the knowledge of the similarity matrix, Bi-Level Routing Attention (BiFormer) is utilized, which combines aspects of transformer-based models with protein sequence analysis and represents a significant advancement in the field of drug-protein interaction prediction. BiFormer has the ability to pinpoint the most effective regions of the protein sequence that are responsible for facilitating interactions between the protein and drugs, thereby enhancing the understanding of these critical interactions. Thus, it appears promising in its ability to capture the local structural relationship of the proteins by enhancing the understanding of how it contributes to drugprotein interactions, thereby facilitating more accurate predictions. To evaluate the proposed method, it was tested on two widely recognized datasets: Davis and KIBA. A comprehensive series of experiments was conducted to illustrate its effectiveness in comparison to cutting edge techniques.

Keywords: BiFormer, transformer, protein language processing, self-attention mechanism, binding affinity, drug target interaction, similarity matrix, protein masked representation, protein language model

Procedia PDF Downloads 7
185 Milling Simulations with a 3-DOF Flexible Planar Robot

Authors: Hoai Nam Huynh, Edouard Rivière-Lorphèvre, Olivier Verlinden

Abstract:

Manufacturing technologies are becoming continuously more diversified over the years. The increasing use of robots for various applications such as assembling, painting, welding has also affected the field of machining. Machining robots can deal with larger workspaces than conventional machine-tools at a lower cost and thus represent a very promising alternative for machining applications. Furthermore, their inherent structure ensures them a great flexibility of motion to reach any location on the workpiece with the desired orientation. Nevertheless, machining robots suffer from a lack of stiffness at their joints restricting their use to applications involving low cutting forces especially finishing operations. Vibratory instabilities may also happen while machining and deteriorate the precision leading to scrap parts. Some researchers are therefore concerned with the identification of optimal parameters in robotic machining. This paper continues the development of a virtual robotic machining simulator in order to find optimized cutting parameters in terms of depth of cut or feed per tooth for example. The simulation environment combines an in-house milling routine (DyStaMill) achieving the computation of cutting forces and material removal with an in-house multibody library (EasyDyn) which is used to build a dynamic model of a 3-DOF planar robot with flexible links. The position of the robot end-effector submitted to milling forces is controlled through an inverse kinematics scheme while controlling the position of its joints separately. Each joint is actuated through a servomotor for which the transfer function has been computed in order to tune the corresponding controller. The output results feature the evolution of the cutting forces when the robot structure is deformable or not and the tracking errors of the end-effector. Illustrations of the resulting machined surfaces are also presented. The consideration of the links flexibility has highlighted an increase of the cutting forces magnitude. This proof of concept will aim to enrich the database of results in robotic machining for potential improvements in production.

Keywords: control, milling, multibody, robotic, simulation

Procedia PDF Downloads 248
184 Designing Stochastic Non-Invasively Applied DC Pulses to Suppress Tremors in Multiple Sclerosis by Computational Modeling

Authors: Aamna Lawrence, Ashutosh Mishra

Abstract:

Tremors occur in 60% of the patients who have Multiple Sclerosis (MS), the most common demyelinating disease that affects the central and peripheral nervous system, and are the primary cause of disability in young adults. While pharmacological agents provide minimal benefits, surgical interventions like Deep Brain Stimulation and Thalamotomy are riddled with dangerous complications which make non-invasive electrical stimulation an appealing treatment of choice for dealing with tremors. Hence, we hypothesized that if the non-invasive electrical stimulation parameters (mainly frequency) can be computed by mathematically modeling the nerve fibre to take into consideration the minutest details of the axon morphologies, tremors due to demyelination can be optimally alleviated. In this computational study, we have modeled the random demyelination pattern in a nerve fibre that typically manifests in MS using the High-Density Hodgkin-Huxley model with suitable modifications to account for the myelin. The internode of the nerve fibre in our model could have up to ten demyelinated regions each having random length and myelin thickness. The arrival time of action potentials traveling the demyelinated and the normally myelinated nerve fibre between two fixed points in space was noted, and its relationship with the nerve fibre radius ranging from 5µm to 12µm was analyzed. It was interesting to note that there were no overlaps between the arrival time for action potentials traversing the demyelinated and normally myelinated nerve fibres even when a single internode of the nerve fibre was demyelinated. The study gave us an opportunity to design DC pulses whose frequency of application would be a function of the random demyelination pattern to block only the delayed tremor-causing action potentials. The DC pulses could be delivered to the peripheral nervous system non-invasively by an electrode bracelet that would suppress any shakiness beyond it thus paving the way for wearable neuro-rehabilitative technologies.

Keywords: demyelination, Hodgkin-Huxley model, non-invasive electrical stimulation, tremor

Procedia PDF Downloads 128
183 The Role of Group Dynamics in Creativity: A Study Case from Italy

Authors: Sofya Komarova, Frashia Ndungu, Alessia Gavazzoli, Roberta Mineo

Abstract:

Modern society requires people to be flexible and to develop innovative solutions to unexpected situations. Creativity refers to the “interaction among aptitude, process, and the environment by which an individual or group produces a perceptible product that is both novel and useful as defined within a social context”. It allows humans to produce novel ideas, generate new solutions, and express themselves uniquely. Only a few scientific studies have examined group dynamics' influence on individuals' creativity. There exist some gaps in the research on creative thinking, such as the fact that collaborative effort frequently results in the enhanced production of new information and knowledge. Therefore, it is critical to evaluate creativity via social settings. The study aimed at exploring the group dynamics of young adults in small group settings and the influence of these dynamics on their creativity. The study included 30 participants aged 20 to 25 who were attending university after completing a bachelor's degree. The participants were divided into groups of three, in gender homogenous and heterogeneous groups. The groups’ creative task was tied to the Lego mosaic created for the Scintillae laboratory at the Reggio Children Foundation. Group dynamics were operationalized into patterns of behaviors classified into three major categories: 1) Social Interactions, 2) Play, and 3) Distraction. Data were collected through audio and video recording and observation. The qualitative data were converted into quantitative data using the observational coding system; then, they were analyzed, revealing correlations between behaviors using median points and averages. For each participant and group, the percentages of represented behavior signals were computed. The findings revealed a link between social interaction, creative thinking, and creative activities. Other findings revealed that the more intense the social interaction, the lower the amount of creativity demonstrated. This study bridges the research gap between group dynamics and creativity. The approach calls for further research on the relationship between creativity and social interaction.

Keywords: group dynamics, creative thinking, creative action, social interactions, group play

Procedia PDF Downloads 127
182 Delineating Floodplain along the Nasia River in Northern Ghana Using HAND Contour

Authors: Benjamin K. Ghansah, Richard K. Appoh, Iliya Nababa, Eric K. Forkuo

Abstract:

The Nasia River is an important source of water for domestic and agricultural purposes to the inhabitants of its catchment. Major farming activities takes place within the floodplain of the river and its network of tributaries. The actual inundation extent of the river system is; however, unknown. Reasons for this lack of information include financial constraints and inadequate human resources as flood modelling is becoming increasingly complex by the day. Knowledge of the inundation extent will help in the assessment of risk posed by the annual flooding of the river, and help in the planning of flood recession agricultural activities. This study used a simple terrain based algorithm, Height Above Nearest Drainage (HAND), to delineate the floodplain of the Nasia River and its tributaries. The HAND model is a drainage normalized digital elevation model, which has its height reference based on the local drainage systems rather than the average mean sea level (AMSL). The underlying principle guiding the development of the HAND model is that hillslope flow paths behave differently when the reference gradient is to the local drainage network as compared to the seaward gradient. The new terrain model of the catchment was created using the NASA’s SRTM Digital Elevation Model (DEM) 30m as the only data input. Contours (HAND Contour) were then generated from the normalized DEM. Based on field flood inundation survey, historical information of flooding of the area as well as satellite images, a HAND Contour of 2m was found to best correlates with the flood inundation extent of the river and its tributaries. A percentage accuracy of 75% was obtained when the surface area created by the 2m contour was compared with surface area of the floodplain computed from a satellite image captured during the peak flooding season in September 2016. It was estimated that the flooding of the Nasia River and its tributaries created a floodplain area of 1011 km².

Keywords: digital elevation model, floodplain, HAND contour, inundation extent, Nasia River

Procedia PDF Downloads 456
181 Relationship between the Level of Perceived Self-Efficacy of Children with Learning Disability and Their Mother’s Perception about the Efficacy of Their Child, and Children’s Academic Achievement

Authors: Payal Maheshwari, Maheaswari Brindavan

Abstract:

The present study aimed at studying the level of perceived self-efficacy of children with learning disability and their mother’s perception about the efficacy of the child and the relationship between the two. The study further aimed at finding out the relationship between the level of perceived self-efficacy of children with learning disability and their academic achievement and their mother’s perception about the Efficacy of the child and child’s Academic Achievement. The sample comprised of 80 respondents (40 children with learning disability and their mothers). Children with learning disability as their primary condition, belonging to middle or upper middle class, living with both the parents, residing in Mumbai and their mothers were selected. Purposive or judgmental and snowball sampling technique was used to select the sample for the present study. Proformas in the form of questionnaires were used to obtain the background information of the children with learning disability and their mother’s. A self-constructed Mother’s Perceived Efficacy of their Child Assessment Scale was used to measure mothers perceived level of efficacy of their child with learning disability. Self-constructed Child’s Perceived Self-Efficacy Assessment Scale was used to measure the level of child’s perceived self-efficacy. Academic scores of the child were collected from the child’s parents or teachers and were converted into percentage. The data were analyzed quantitatively using frequencies, mean and standard deviation. Correlations were computed to ascertain the relationships between the different variables. The findings revealed that majority of the mother’s perceived efficacy about their child with learning disability was above average as well as majority of the children with learning disability also perceived themselves as having above average level of self-efficacy. Further in the domains of self-regulated learning and emotional self-efficacy majority of the mothers perceived their child as having average or below average efficacy, 50% of the children also perceived their self-efficacy in the two domains at average or below average level. A significant (r=.322, p < .05) weak correlation (Spearman’s rho) was found between mother’s perceived efficacy about their child, and child’s perceived self-efficacy and a significant (r=.377, p < .01) weak correlation (Pearson Correlation) was also found between mother’s perceived efficacy about their child and child’s academic achievement. Significant weak positive correlation was found between child’s perceived self-efficacy and academic achievement (r=.332, p < .05). Based on the findings, the study discussed the need for intervention program for children in non-academic skills like self-regulation and emotional competence.

Keywords: learning disability, perceived self efficacy, academic achievement, mothers, children

Procedia PDF Downloads 321
180 Two-Level Graph Causality to Detect and Predict Random Cyber-Attacks

Authors: Van Trieu, Shouhuai Xu, Yusheng Feng

Abstract:

Tracking attack trajectories can be difficult, with limited information about the nature of the attack. Even more difficult as attack information is collected by Intrusion Detection Systems (IDSs) due to the current IDSs having some limitations in identifying malicious and anomalous traffic. Moreover, IDSs only point out the suspicious events but do not show how the events relate to each other or which event possibly cause the other event to happen. Because of this, it is important to investigate new methods capable of performing the tracking of attack trajectories task quickly with less attack information and dependency on IDSs, in order to prioritize actions during incident responses. This paper proposes a two-level graph causality framework for tracking attack trajectories in internet networks by leveraging observable malicious behaviors to detect what is the most probable attack events that can cause another event to occur in the system. Technically, given the time series of malicious events, the framework extracts events with useful features, such as attack time and port number, to apply to the conditional independent tests to detect the relationship between attack events. Using the academic datasets collected by IDSs, experimental results show that the framework can quickly detect the causal pairs that offer meaningful insights into the nature of the internet network, given only reasonable restrictions on network size and structure. Without the framework’s guidance, these insights would not be able to discover by the existing tools, such as IDSs. It would cost expert human analysts a significant time if possible. The computational results from the proposed two-level graph network model reveal the obvious pattern and trends. In fact, more than 85% of causal pairs have the average time difference between the causal and effect events in both computed and observed data within 5 minutes. This result can be used as a preventive measure against future attacks. Although the forecast may be short, from 0.24 seconds to 5 minutes, it is long enough to be used to design a prevention protocol to block those attacks.

Keywords: causality, multilevel graph, cyber-attacks, prediction

Procedia PDF Downloads 156
179 Effect of Halo Protection Device on the Aerodynamic Performance of Formula Racecar

Authors: Mark Lin, Periklis Papadopoulos

Abstract:

This paper explores the aerodynamics of the formula racecar when a ‘halo’ driver-protection device is added to the chassis. The halo protection device was introduced at the start of the 2018 racing season as a safety measure against foreign object impacts that a driver may encounter when driving an open-wheel racecar. In the one-year since its introduction, the device has received wide acclaim for protecting the driver on two separate occasions. The benefit of such a safety device certainly cannot be disputed. However, by adding the halo device to a car, it changes the airflow around the vehicle, and most notably, to the engine air-intake and the rear wing. These negative effects in the air supply to the engine, and equally to the downforce created by the rear wing are studied in this paper using numerical technique, and the resulting CFD outputs are presented and discussed. Comparing racecar design prior to and after the introduction of the halo device, it is shown that the design of the air intake and the rear wing has not followed suit since the addition of the halo device. The reduction of engine intake mass flow due to the halo device is computed and presented for various speeds the car may be going. Because of the location of the halo device in relation to the air intake, airflow is directed away from the engine, making the engine perform less than optimal. The reduction is quantified in this paper to show the correspondence to reduce the engine output when compared to a similar car without the halo device. This paper shows that through aerodynamic arguments, the engine in a halo car will not receive unobstructed, clean airflow that a non-halo car does. Another negative effect is on the downforce created by the rear wing. Because the amount of downforce created by the rear wing is influenced by every component that comes before it, when a halo device is added upstream to the rear wing, airflow is obstructed, and less is available for making downforce. This reduction in downforce is especially dramatic as the speed is increased. This paper presents a graph of downforce over a range of speeds for a car with and without the halo device. Acknowledging that although driver safety is paramount, the negative effect of this safety device on the performance of the car should still be well understood so that any possible redesign to mitigate these negative effects can be taken into account in next year’s rules regulation.

Keywords: automotive aerodynamics, halo device, downforce. engine intake

Procedia PDF Downloads 110
178 Factors Contributing to Farmers’ Attitude Towards Climate Adaptation Farming Practices: A Farm Level Study in Bangladesh

Authors: Md Rezaul Karim, Farha Taznin

Abstract:

The purpose of this study was to assess and describe the individual and household characteristics of farmers, to measure the attitude of farmers towards climate adaptation farming practices and to explore the individual and household factors contributing in predicting their attitude towards climate adaptation farming practices. Data were collected through personal interviews using a pre-tested interview schedule. The data collection was done at Biral Upazila under Dinajpur district in Bangladesh from 1st November to 15 December 2018. Besides descriptive statistical parameters, Pearson’s Product Moment Correlation Coefficient (r), multiple regression and step-wise multiple regression analysis were used for the statistical analysis. Findings indicated that the highest proportion (77.6 percent) of the farmers had moderately favorable attitudes, followed by only 11.2 percent with highly favorable attitudes and 11.2 percent with slightly favorable attitudes towards climate adaptation farming practices. According to the computed correlation coefficients (r), among the 10 selected factors, five of them, such as education of household head, farm size, annual household income, organizational participation, and information access by extension services, had a significant relationship with the attitude of farmers towards climate-smart practices. The step-wise multiple regression results showed that two characteristics as education of household head and information access by extension services, contributed 26.2% and 5.1%, respectively, in predicting farmers' attitudes towards climate adaptation farming practices. In addition, more than two-thirds of farmers cited their opinion to the problems in response to ‘price of vermi species is high and it is not easily available’ as 1st ranked problem, followed by ‘lack of information for innovative climate-smart technologies’. This study suggests that policy implications are necessary to promote extension education and information services and overcome the obstacles to climate adaptation farming practices. It further recommends that research study should be conducted in diverse contexts of nationally or globally.

Keywords: factors, attitude, climate adaptation, farming practices, Bangladesh

Procedia PDF Downloads 88
177 Geospatial Analysis of Hydrological Response to Forest Fires in Small Mediterranean Catchments

Authors: Bojana Horvat, Barbara Karleusa, Goran Volf, Nevenka Ozanic, Ivica Kisic

Abstract:

Forest fire is a major threat in many regions in Croatia, especially in coastal areas. Although they are often caused by natural processes, the most common cause is the human factor, intentional or unintentional. Forest fires drastically transform landscapes and influence natural processes. The main goal of the presented research is to analyse and quantify the impact of the forest fire on hydrological processes and propose the model that best describes changes in hydrological patterns in the analysed catchments. Keeping in mind the spatial component of the processes, geospatial analysis is performed to gain better insight into the spatial variability of the hydrological response to disastrous events. In that respect, two catchments that experienced severe forest fire were delineated, and various hydrological and meteorological data were collected both attribute and spatial. The major drawback is certainly the lack of hydrological data, common in small torrential karstic streams; hence modelling results should be validated with the data collected in the catchment that has similar characteristics and established hydrological monitoring. The event chosen for the modelling is the forest fire that occurred in July 2019 and burned nearly 10% of the analysed area. Surface (land use/land cover) conditions before and after the event were derived from the two Sentinel-2 images. The mapping of the burnt area is based on a comparison of the Normalized Burn Index (NBR) computed from both images. To estimate and compare hydrological behaviour before and after the event, curve number (CN) values are assigned to the land use/land cover classes derived from the satellite images. Hydrological modelling resulted in surface runoff generation and hence prediction of hydrological responses in the catchments to a forest fire event. The research was supported by the Croatian Science Foundation through the project 'Influence of Open Fires on Water and Soil Quality' (IP-2018-01-1645).

Keywords: Croatia, forest fire, geospatial analysis, hydrological response

Procedia PDF Downloads 136
176 Reliability Levels of Reinforced Concrete Bridges Obtained by Mixing Approaches

Authors: Adrián D. García-Soto, Alejandro Hernández-Martínez, Jesús G. Valdés-Vázquez, Reyna A. Vizguerra-Alvarez

Abstract:

Reinforced concrete bridges designed by code are intended to achieve target reliability levels adequate for the geographical environment where the code is applicable. Several methods can be used to estimate such reliability levels. Many of them require the establishment of an explicit limit state function (LSF). When such LSF is not available as a close-form expression, the simulation techniques are often employed. The simulation methods are computing intensive and time consuming. Note that if the reliability of real bridges designed by code is of interest, numerical schemes, the finite element method (FEM) or computational mechanics could be required. In these cases, it can be quite difficult (or impossible) to establish a close-form of the LSF, and the simulation techniques may be necessary to compute reliability levels. To overcome the need for a large number of simulations when no explicit LSF is available, the point estimate method (PEM) could be considered as an alternative. It has the advantage that only the probabilistic moments of the random variables are required. However, in the PEM, fitting of the resulting moments of the LSF to a probability density function (PDF) is needed. In the present study, a very simple alternative which allows the assessment of the reliability levels when no explicit LSF is available and without the need of extensive simulations is employed. The alternative includes the use of the PEM, and its applicability is shown by assessing reliability levels of reinforced concrete bridges in Mexico when a numerical scheme is required. Comparisons with results by using the Monte Carlo simulation (MCS) technique are included. To overcome the problem of approximating the probabilistic moments from the PEM to a PDF, a well-known distribution is employed. The approach mixes the PEM and other classic reliability method (first order reliability method, FORM). The results in the present study are in good agreement whit those computed with the MCS. Therefore, the alternative of mixing the reliability methods is a very valuable option to determine reliability levels when no close form of the LSF is available, or if numerical schemes, the FEM or computational mechanics are employed.

Keywords: structural reliability, reinforced concrete bridges, combined approach, point estimate method, monte carlo simulation

Procedia PDF Downloads 346
175 Effects of Effort and Water Quality on Productivity (CPUE) of Hampal (Hampala macrolepidota) Resources in Jatiluhur Dam, West Java

Authors: Ririn Marinasari, S. Pi

Abstract:

Hampal (Hampala macrolepidota) is one of Citarum river indigenous fishes that still find in Jatiluhur dam. IUCN at 2013 said that hampal listed on redlist species category, this species was rare in Jatiluhur dam. This species more and more decreasing because change of habitats characteristic such as water quality and fishing effort. This study aims to determine and identify the influence of fishing effort and the quality of water on the productivity of fish resources hampal (Hampala macrolepidota) in Jatiluhur. The study was conducted from October to November 2013. Zones of research include lacustrine zone, transition and Riverin. Hampal fish productivity value computed by Hampal’s CPUE values. The results showed that fish MSY hampal obtained from surplus production model of Schaefer is equal to 0.2045 tons / quarterly. In the years 2011-2012 have occurred over fishing in 2013 while still under fishing. Total catches have exceeded the MSY during the year 2011 and the third quarterly of 2012 tons of fish that exceed 0.2045 hampal. The rate of utilization of fish resources hampal is equal to 80% of MSY or equal to the allowable catch (Total Allowable Catch) for fish in Jatiluhur hampal based Schaefer surplus production theory. Fishing effort, water quality parameters such as DO, turbidity and negatively correlated sulfide as H2S, while the temperature and pH positively correlated to productivity or unit catches fish hampal efforts in quarterly time series in the period 2011-2013. Shows that the higher fishing effort, DO, turbidity and sulfide in H2S and diminishing the temperature and pH of the productivity decreases. Variables that affect the productivity of fishing hampal only H2S only factor beta coefficient -0.834 which indicates a negative effect. It can be caused by H2S levels are toxic and have already exceeded the quality standard, while for other water quality parameters are still below the maximum standards allowed in the waters. Result of the study can be a reference of fishing regulation for hampal conservation in Jatiluhur dam.

Keywords: effort, hampal, productivity, water quality

Procedia PDF Downloads 298
174 Utilization of Antenatal Care Services by Domestic Workers in Delhi

Authors: Meenakshi

Abstract:

Background: The complications during pregnancy are the major cause of morbidity and deaths among women in the reproductive age group. Childbearing is the most important phase in women’s lives that occur mainly in the adolescent and adult years. Maternal health, thus is an important issue as this as this is important phase is also productive time for women as they strive fulfill their capabilities as an individual, mothers, family members and also as a citizen. The objective of the study is to document the coverage of ANC and its determinants among domestic workers. Method: A survey of 300 domestic workers were carried in Delhi. Only respondents in the age group (15-49) and whose recent birth was of 5 years preceding the survey were included. Socio-demographic data and information on maternal health was collected from these respondents Information on ANC was collected from total 300 respondents. Standard of living index were composed based on households assists and similarly autonomy index was computed based on women decision making power in the households taking certain key variables. Cross tabulations were performed to obtain frequency and percentages. Potential socio-economic determinants of utilization of ANC among domestic workers were examined using binary logistic regressions. Results: Out of 300 domestic workers survey, only 70.7 per cent per cent received ANC. Domestic workers who married at age 18 years and above are 4 times more likely to utilize antenatal services during their last birth (***p< 0.01). Comparison to domestic workers with number of living children two or less, domestic workers with number of living children more than two are less likely to utilize antenatal care services (**p< 0.05). Domestic workers belonging to Other Backward Castes are more likely to utilize antenatal care services than domestic workers belonging to scheduled tribes ((**p< 0.05). Conclusion: The level of utilization of maternal health services are less among domestic workers is less, as they spend most of their time at the employers household. Though demonstration effect do have impact on their life styles but utilization of maternal health services is poor. Strategies and action are needed to improve the utilization of maternal health services among this section of workers as they are vulnerable because of no proper labour legislations.

Keywords: antenatal care, domestic workers, health services, maternal health, women’s health

Procedia PDF Downloads 198
173 Differentiated Surgical Treatment of Patients With Nontraumatic Intracerebral Hematomas

Authors: Mansur Agzamov, Valery Bersnev, Natalia Ivanova, Istam Agzamov, Timur Khayrullaev, Yulduz Agzamova

Abstract:

Objectives. Treatment of hypertensive intracerebral hematoma (ICH) is controversial. Advantage of one surgical method on other has not been established. Recent reports suggest a favorable effect of minimally invasive surgery. We conducted a small comparative study of different surgical methods. Methods. We analyzed the result of surgical treatment of 176 patients with intracerebral hematomas at the age from 41 to 78 years. Men were been113 (64.2%), women - 63 (35.8%). Level of consciousness: conscious -18, lethargy -63, stupor –55, moderate coma - 40. All patients on admission and in the dynamics underwent computer tomography (CT) of the brain. ICH was located in the putamen in 87 cases, thalamus in 19, in the mix area in 50, in the lobar area in 20. Ninety seven patients of them had an intraventricular hemorrhage component. The baseline volume of the ICH was measured according to a bedside method of measuring CT intracerebral hematomas volume. Depending on the intervention of the patients were divided into three groups. Group 1 patients, 90 patients, operated open craniotomy. Level of consciousness: conscious-11, lethargy-33, stupor–18, moderate coma -18. The hemorrhage was located in the putamen in 51, thalamus in 3, in the mix area in 25, in the lobar area in 11. Group 2 patients, 22 patients, underwent smaller craniotomy with endoscopic-assisted evacuation. Level of consciousness: conscious-4, lethargy-9, stupor–5, moderate coma -4. The hemorrhage was located in the putamen in 5, thalamus in 15, in the mix area in 2. Group 3 patients, 64 patients, was conducted minimally invasive removal of intracerebral hematomas using the original device (patent of Russian Federation № 65382). The device - funnel cannula - which after the special markings introduced into the hematoma cavity. Level of consciousness: conscious-3, lethargy-21, stupor–22, moderate coma -18. The hemorrhage was located in the putamen in 31, in the mix area in 23, thalamus in 1, in the lobar area in 9. Results of treatment were evaluated by Glasgow outcome scale. Results. The study showed that the results of surgical treatment in three groups depending on the degree of consciousness, the volume and localization of hematoma. In group 1, good recovery observed in 8 cases (8.9%), moderate disability in 22 (24.4%), severe disability - 17 (18.9%), death-43 (47.8%). In group 2, good recovery observed in 7 cases (31.8%), moderate disability in 7 (31.8%), severe disability - 5 (29.7%), death-7 (31.8%). In group 3, good recovery was observed in 9 cases (14.1%), moderate disability-17 (26.5%), severe disability-19 (29.7%), death-19 (29.7%). Conclusions. The method of using cannulae allowed to abandon from open craniotomy of the majority of patients with putaminal hematomas. Minimally invasive technique reduced the postoperative mortality and improves treatment outcomes of these patients.

Keywords: nontraumatic intracerebral hematoma, minimal invasive surgical technique, funnel canula, differentiated surcical treatment

Procedia PDF Downloads 83
172 Comparative Evaluation of Root Uptake Models for Developing Moisture Uptake Based Irrigation Schedules for Crops

Authors: Vijay Shankar

Abstract:

In the era of water scarcity, effective use of water via irrigation requires good methods for determining crop water needs. Implementation of irrigation scheduling programs requires an accurate estimate of water use by the crop. Moisture depletion from the root zone represents the consequent crop evapotranspiration (ET). A numerical model for simulating soil water depletion in the root zone has been developed by taking into consideration soil physical properties, crop and climatic parameters. The governing differential equation for unsaturated flow of water in the soil is solved numerically using the fully implicit finite difference technique. The water uptake by plants is simulated by using three different sink functions. The non-linear model predictions are in good agreement with field data and thus it is possible to schedule irrigations more effectively. The present paper describes irrigation scheduling based on moisture depletion from the different layers of the root zone, obtained using different sink functions for three cash, oil and forage crops: cotton, safflower and barley, respectively. The soil is considered at a moisture level equal to field capacity prior to planting. Two soil moisture regimes are then imposed for irrigated treatment, one wherein irrigation is applied whenever soil moisture content is reduced to 50% of available soil water; and other wherein irrigation is applied whenever soil moisture content is reduced to 75% of available soil water. For both the soil moisture regimes it has been found that the model incorporating a non-linear sink function which provides best agreement of computed root zone moisture depletion with field data, is most effective in scheduling irrigations. Simulation runs with this moisture uptake function result in saving 27.3 to 45.5% & 18.7 to 37.5%, 12.5 to 25% % &16.7 to 33.3% and 16.7 to 33.3% & 20 to 40% irrigation water for cotton, safflower and barley respectively, under 50 & 75% moisture depletion regimes over other moisture uptake functions considered in the study. Simulation developed can be used for an optimized irrigation planning for different crops, choosing a suitable soil moisture regime depending upon the irrigation water availability and crop requirements.

Keywords: irrigation water, evapotranspiration, root uptake models, water scarcity

Procedia PDF Downloads 330
171 High Heating Value Bio-Chars from a Bio-Oil Upgrading Process

Authors: Julius K. Gane, Mohamad N. Nahil, Paul T. Williams

Abstract:

In today’s world of rapid population growth and a changing climate, one way to mitigate various negative effects is via renewable energy solutions. Energy and power as basic requirements in almost all human endeavours are also the banes of the changing climate and the impacts thereof. Thus it is crucial to develop innovative and environmentally friendly energy options to ameliorate various negative repercussions. Upgrading of fast pyrolysis bio-oil via hydro-treatment offers such opportunities, as quality renewable liquid transportation fuels can be produced. The process, however, is typically accompanied by bio-char formation as a by-product. The goal of this work was to study the yield and some properties of bio-chars formed from a hydrotreatment process, with an overall aim to promote the valuable utilization of wastes or by-products from renewable energy technologies. It is assumed that bio-chars that have comparable energy contents with coals will be more desirable as solid energy materials due to renewability and environmental friendliness. Therefore, the analytical work in this study focused mainly on determining the higher heating value (HHV) of the chars. The method involved the reaction of bio-oil in an autoclave supplied by the Parr Instrument Company, IL, USA. Two main parameters (different temperatures and resident times) were investigated. The chars were characterized using a Thermo EA2000 CHNS analyser, then oxygen contents and HHVs computed based on the literature. From the results, these bio-chars can readily serve as feedstocks for the production of renewable solid fuels. Their HHVs ranged between 29.26-39.18 MJ/kg, affected by different temperatures and retention times. There was an inverse relationship between the oxygen content and the HHVs of the chars. It can, therefore, be concluded that it is possible to optimize the process efficiency of the hydrotreatment process used through the production of renewable energy materials from the 'waste’ char by-products. Future work should consider developing a suitable balance between the primary objective of bio-oil upgrading processes (which is to improve the quality of the liquid fuels) and the conversion of its solid wastes into value-added products such as smokeless briquettes.

Keywords: bio-char, renewable solid biofuels, valorisation, waste-to-energy

Procedia PDF Downloads 127
170 Prevention of Heart Failure Progression in Patients with Post-Infarction Cardiosclerosis After Coronavirus Infection

Authors: Sujayeva V. A., Karpova I. S., Koslataya O. V., Kolyadko M. G., Russkikh I. I., Vankovich E. A.

Abstract:

Objective: The goal of this study is to develop a method for the prevention of the progression of heart failure (HF) in patients with post-infarction cardiosclerosis who have suffered coronavirus infection. Methods: 135 patients with post-infarction cardiosclerosis were divided into 2 groups: Group I - patients who had suffered COVID-19 - 85 people, and Group II - patients who had not suffered COVID-19 - 50 people. Patients of group I, depending on the level of N-terminal fragment of natriuretic peptide (NTproBNP), were divided into 2 subgroups - subgroup A - with HF - 40 people, subgroup B - without HF - 45 people. All patients underwent a clinical examination, echocardiography, electrocardiotopography in 60 leads, computed angiography of the coronary arteries, heart magnetic resonance imaging, NTproBNP. Results: In the post-Covid period, in patients with post-infarction cardiosclerosis, remodeling of the left ventricle and right parts of the heart, deterioration of the systolic-diastolic function of both ventricles, increased pressure in the pulmonary artery, progression of coronary artery atherosclerosis, and an increase in the size of myocardial fibrosis were revealed. The consequence of these changes was the progression of heart failure. The developed method of medical prevention made it possible to improve the clinical course of coronary artery disease and prevent the progression of chronic heart failure in patients with post-infarction cardiosclerosis. Conclusions: In patients with post-infarction cardiosclerosis who initially had HF, after 1 year, according to laboratory and instrumental data, a slight decrease in its severity was revealed. In patients with post-infarction cardiosclerosis who did not have HF before COVID-19, HF developed 1 year after the coronavirus disease, which may be due to the identified process of myocardial fibrosis, which dictates the need to prevent the development of HF in patients with post-infarction cardiosclerosis, even those who did not initially have HF. The proposed method of medical prevention made it possible to improve the clinical course of coronary artery disease in patients with post-infarction cardiosclerosis after COVID-19, both in persons with and without HF, when included in the study. A method of medical prevention in people with post-infarction cardiosclerosis after COVID-19 infection, including spironolactone, loop diuretics, empagliflozin, sacubitril/valsartan, helped prevent the progression of HF.

Keywords: elderly, myocardial infarction, COVID-19, prevention

Procedia PDF Downloads 22
169 A Normalized Non-Stationary Wavelet Based Analysis Approach for a Computer Assisted Classification of Laryngoscopic High-Speed Video Recordings

Authors: Mona K. Fehling, Jakob Unger, Dietmar J. Hecker, Bernhard Schick, Joerg Lohscheller

Abstract:

Voice disorders origin from disturbances of the vibration patterns of the two vocal folds located within the human larynx. Consequently, the visual examination of vocal fold vibrations is an integral part within the clinical diagnostic process. For an objective analysis of the vocal fold vibration patterns, the two-dimensional vocal fold dynamics are captured during sustained phonation using an endoscopic high-speed camera. In this work, we present an approach allowing a fully automatic analysis of the high-speed video data including a computerized classification of healthy and pathological voices. The approach bases on a wavelet-based analysis of so-called phonovibrograms (PVG), which are extracted from the high-speed videos and comprise the entire two-dimensional vibration pattern of each vocal fold individually. Using a principal component analysis (PCA) strategy a low-dimensional feature set is computed from each phonovibrogram. From the PCA-space clinically relevant measures can be derived that quantify objectively vibration abnormalities. In the first part of the work it will be shown that, using a machine learning approach, the derived measures are suitable to distinguish automatically between healthy and pathological voices. Within the approach the formation of the PCA-space and consequently the extracted quantitative measures depend on the clinical data, which were used to compute the principle components. Therefore, in the second part of the work we proposed a strategy to achieve a normalization of the PCA-space by registering the PCA-space to a coordinate system using a set of synthetically generated vibration patterns. The results show that owing to the normalization step potential ambiguousness of the parameter space can be eliminated. The normalization further allows a direct comparison of research results, which bases on PCA-spaces obtained from different clinical subjects.

Keywords: Wavelet-based analysis, Multiscale product, normalization, computer assisted classification, high-speed laryngoscopy, vocal fold analysis, phonovibrogram

Procedia PDF Downloads 265
168 Physics-Based Earthquake Source Models for Seismic Engineering: Analysis and Validation for Dip-Slip Faults

Authors: Percy Galvez, Anatoly Petukhin, Paul Somerville, Ken Miyakoshi, Kojiro Irikura, Daniel Peter

Abstract:

Physics-based dynamic rupture modelling is necessary for estimating parameters such as rupture velocity and slip rate function that are important for ground motion simulation, but poorly resolved by observations, e.g. by seismic source inversion. In order to generate a large number of physically self-consistent rupture models, whose rupture process is consistent with the spatio-temporal heterogeneity of past earthquakes, we use multicycle simulations under the heterogeneous rate-and-state (RS) friction law for a 45deg dip-slip fault. We performed a parametrization study by fully dynamic rupture modeling, and then, a set of spontaneous source models was generated in a large magnitude range (Mw > 7.0). In order to validate rupture models, we compare the source scaling relations vs. seismic moment Mo for the modeled rupture area S, as well as average slip Dave and the slip asperity area Sa, with similar scaling relations from the source inversions. Ground motions were also computed from our models. Their peak ground velocities (PGV) agree well with the GMPE values. We obtained good agreement of the permanent surface offset values with empirical relations. From the heterogeneous rupture models, we analyzed parameters, which are critical for ground motion simulations, i.e. distributions of slip, slip rate, rupture initiation points, rupture velocities, and source time functions. We studied cross-correlations between them and with the friction weakening distance Dc value, the only initial heterogeneity parameter in our modeling. The main findings are: (1) high slip-rate areas coincide with or are located on an outer edge of the large slip areas, (2) ruptures have a tendency to initiate in small Dc areas, and (3) high slip-rate areas correlate with areas of small Dc, large rupture velocity and short rise-time.

Keywords: earthquake dynamics, strong ground motion prediction, seismic engineering, source characterization

Procedia PDF Downloads 144
167 A Relational Approach to Adverb Use in Interactions

Authors: Guillaume P. Fernandez

Abstract:

Individual language use is a matter of choice in particular interactions. The paper proposes a conceptual and theoretical framework with methodological consideration to develop how language produced in dyadic relations is to be considered and situated in the larger social configuration the interaction is embedded within. An integrated and comprehensive view is taken: social interactions are expected to be ruled by a normative context, defined by the chain of interdependences that structures the personal network. In this approach, the determinants of discursive practices are not only constrained by the moment of production and isolated from broader influences. Instead, the position the individual and the dyad have in the personal network influences the discursive practices in a twofold manner: on the one hand, the network limits the access to linguistic resources available within it, and, on the other hand, the structure of the network influences the agency of the individual, by the social control inherent to particular network characteristics. Concretely, we investigate how and to what extent consistent ego is from one interaction to another in his or her use of adverbs. To do so, social network analysis (SNA) methods are mobilized. Participants (N=130) are college students recruited in the french speaking part of Switzerland. The personal network of significant ones of each individual is created using name generators and edge interpreters, with a focus on social support and conflict. For the linguistic parts, respondents were asked to record themselves with five of their close relations. From the recordings, we computed an average similarity score based on the adverb used across interactions. In terms of analyses, two are envisaged: First, OLS regressions including network-level measures, such as density and reciprocity, and individual-level measures, such as centralities, are performed to understand the tenets of linguistic similarity from one interaction to another. The second analysis considers each social tie as nested within ego networks. Multilevel models are performed to investigate how the different types of ties may influence the likelihood to use adverbs, by controlling structural properties of the personal network. Primary results suggest that the more cohesive the network, the less likely is the individual to change his or her manner of speaking, and social support increases the use of adverbs in interactions. While promising results emerge, further research should consider a longitudinal approach to able the claim of causality.

Keywords: personal network, adverbs, interactions, social influence

Procedia PDF Downloads 67
166 Optimization of Tundish Geometry for Minimizing Dead Volume Using OpenFOAM

Authors: Prateek Singh, Dilshad Ahmad

Abstract:

Growing demand for high-quality steel products has inspired researchers to investigate the unit operations involved in the manufacturing of these products (slabs, rods, sheets, etc.). One such operation is tundish operation, in which a vessel (tundish) acts as a buffer of molten steel for the solidification operation in mold. It is observed that tundish also plays a crucial role in the quality and cleanliness of the steel produced, besides merely acting as a reservoir for the mold. It facilitates removal of dissolved oxygen (inclusions) from the molten steel thus improving its cleanliness. Inclusion removal can be enhanced by increasing the residence time of molten steel in the tundish by incorporation of flow modifiers like dams, weirs, turbo-pad, etc. These flow modifiers also help in reducing the dead or short circuit zones within the tundish which is significant for maintaining thermal and chemical homogeneity of molten steel. Thus, it becomes important to analyze the flow of molten steel in the tundish for different configuration of flow modifiers. In the present work, effect of varying positions and heights/depths of dam and weir on the dead volume in tundish is studied. Steady state thermal and flow profiles of molten steel within the tundish are obtained using OpenFOAM. Subsequently, Residence Time Distribution analysis is performed to obtain the percentage of dead volume in the tundish. Design of Experiment method is then used to configure different tundish geometries for varying positions and heights/depths of dam and weir, and dead volume for each tundish design is obtained. A second-degree polynomial with two-term interactions of independent variables to predict the dead volume in the tundish with positions and heights/depths of dam and weir as variables are computed using Multiple Linear Regression model. This polynomial is then used in an optimization framework to obtain the optimal tundish geometry for minimizing dead volume using Sequential Quadratic Programming optimization.

Keywords: design of experiments, multiple linear regression, OpenFOAM, residence time distribution, sequential quadratic programming optimization, steel, tundish

Procedia PDF Downloads 208
165 Numerical Solution of Steady Magnetohydrodynamic Boundary Layer Flow Due to Gyrotactic Microorganism for Williamson Nanofluid over Stretched Surface in the Presence of Exponential Internal Heat Generation

Authors: M. A. Talha, M. Osman Gani, M. Ferdows

Abstract:

This paper focuses on the study of two dimensional magnetohydrodynamic (MHD) steady incompressible viscous Williamson nanofluid with exponential internal heat generation containing gyrotactic microorganism over a stretching sheet. The governing equations and auxiliary conditions are reduced to a set of non-linear coupled differential equations with the appropriate boundary conditions using similarity transformation. The transformed equations are solved numerically through spectral relaxation method. The influences of various parameters such as Williamson parameter γ, power constant λ, Prandtl number Pr, magnetic field parameter M, Peclet number Pe, Lewis number Le, Bioconvection Lewis number Lb, Brownian motion parameter Nb, thermophoresis parameter Nt, and bioconvection constant σ are studied to obtain the momentum, heat, mass and microorganism distributions. Moment, heat, mass and gyrotactic microorganism profiles are explored through graphs and tables. We computed the heat transfer rate, mass flux rate and the density number of the motile microorganism near the surface. Our numerical results are in better agreement in comparison with existing calculations. The Residual error of our obtained solutions is determined in order to see the convergence rate against iteration. Faster convergence is achieved when internal heat generation is absent. The effect of magnetic parameter M decreases the momentum boundary layer thickness but increases the thermal boundary layer thickness. It is apparent that bioconvection Lewis number and bioconvection parameter has a pronounced effect on microorganism boundary. Increasing brownian motion parameter and Lewis number decreases the thermal boundary layer. Furthermore, magnetic field parameter and thermophoresis parameter has an induced effect on concentration profiles.

Keywords: convection flow, similarity, numerical analysis, spectral method, Williamson nanofluid, internal heat generation

Procedia PDF Downloads 180
164 Spectroscopy and Electron Microscopy for the Characterization of CdSxSe1-x Quantum Dots in a Glass Matrix

Authors: C. Fornacelli, P. Colomban, E. Mugnaioli, I. Memmi Turbanti

Abstract:

When semiconductor particles are reduced in scale to nanometer dimension, their optical and electro-optical properties strongly differ from those of bulk crystals of the same composition. Since sampling is often not allowed concerning cultural heritage artefacts, the potentialities of two non-invasive techniques, such as Raman and Fiber Optic Reflectance Spectroscopy (FORS), have been investigated and the results of the analysis on some original glasses of different colours (from yellow to orange and deep red) and periods (from the second decade of the 20th century to present days) are reported in the present study. In order to evaluate the potentialities of the application of non-invasive techniques to the investigation of the structure and distribution of nanoparticles dispersed in a glass matrix, Scanning Electron Microscopy (SEM) and energy-disperse spectroscopy (EDS) mapping, together with Transmission Electron Microscopy (TEM) and Electron Diffraction Tomography (EDT) have also been used. Raman spectroscopy allows a fast and non-destructive measure of the quantum dots composition and size, thanks to the evaluation of the frequencies and the broadening/asymmetry of the LO phonons bands, respectively, though the important role of the compressive strain arising from the glass matrix and the possible diffusion of zinc from the matrix to the nanocrystals should be taken into account when considering the optical-phonons frequency values. The incorporation of Zn has been assumed by an upward shifting of the LO band related to the most abundant anion (S or Se), while the role of the surface phonons as well as the confinement-induced scattering by phonons with a non-zero wavevectors on the Raman peaks broadening has been verified. The optical band gap varies from 2.42 eV (pure CdS) to 1.70 eV (CdSe). For the compositional range between 0.5≤x≤0.2, the presence of two absorption edges has been related to the contribution of both pure CdS and the CdSxSe1-x solid solution; this particular feature is probably due to the presence of unaltered cubic zinc blende structures of CdS that is not taking part to the formation of the solid solution occurring only between hexagonal CdS and CdSe. Moreover, the band edge tailing originating from the disorder due to the formation of weak bonds and characterized by the Urbach edge energy has been studied and, together with the FWHM of the Raman signal, has been assumed as a good parameter to evaluate the degree of topological disorder. SEM-EDS mapping showed a peculiar distribution of the major constituents of the glass matrix (fluxes and stabilizers), especially concerning those samples where a layered structure has been assumed thanks to the spectroscopic study. Finally, TEM-EDS and EDT were used to get high-resolution information about nanocrystals (NCs) and heterogeneous glass layers. The presence of ZnO NCs (< 4 nm) dispersed in the matrix has been verified for most of the samples, while, for those samples where a disorder due to a more complex distribution of the size and/or composition of the NCs has been assumed, the TEM clearly verified most of the assumption made by the spectroscopic techniques.

Keywords: CdSxSe1-x, EDT, glass, spectroscopy, TEM-EDS

Procedia PDF Downloads 299
163 Study of Morphological Changes of the River Ganga in Patna District, Bihar Using Remote Sensing and GIS Techniques

Authors: Bhawesh Kumar, A. P. Krishna

Abstract:

There are continuous changes upon earth’s surface by a variety of natural and anthropogenic agents cut, carry away and depositing of minerals from land. Running water has higher capacity of erosion than other geomorphologic agents. This research work has been carried out on Ganga River, whose channel is continuously changing under the influence of geomorphic agents and human activities in the surrounding regions. The main focus is to study morphological characteristics and sand dynamics of Ganga River with particular emphasis on bank lines and width changes using remote sensing and GIS techniques. The advance remote sensing data and topographical data were interpreted for obtaining 52 years of changes. For this, remote sensing data of different years (LANDSAT TM 1975, 1988, 1993, ETM 2005 and ETM 2012) and toposheet of SOI for the year 1960 were used as base maps for this study. Sinuosity ratio, braiding index and migratory activity index were also established. It was found to be 1.16 in 1975 and in 1988, 1993, 2005 and 2005 it was 1.09, 1.11, 1.1, 1.09 respectively. The analysis also shows that the minimum value found in 1960 was in reach 1 and maximum value is 4.8806 in 2012 found in reach 4 which suggests creation of number of islands in reach 4 for the year 2012. Migratory activity index (MAI), which is a standardized function of both length and time, was computed for the 8 representative reaches. MAI shows that maximum migration was in 1975-1988 in reach 6 and 7 and minimum migration was in 1993-2005. From the channel change analysis, it was found that the shifting of bank line was cyclic and the river Ganges showed a trend of southward maximum values. The advanced remote sensing data and topographical data helped in obtaining 52 years changes in the river due to various natural and manmade activities like flood, water velocity and excavation, removal of vegetation cover and fertile soil excavation for the various purposes of surrounding regions.

Keywords: braided index, migratory activity index (MAI), Ganga river, river morphology

Procedia PDF Downloads 346