Search results for: convenient facilities; distribution pattern
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 9018

Search results for: convenient facilities; distribution pattern

48 The Strategic Role of Accommodation Providers in Encouraging Travelers to Adopt Environmentally-Friendly Modes of Transportation: An Experiment from France

Authors: Luc Beal

Abstract:

Introduction. Among the stakeholders involved in the tourist decision-making process, the accommodation provider has the potential to play a crucial role in raising awareness, disseminating information, and thus influencing the tourists’ choice of transportation. Since the early days of tourism, the accommodation provider has consistently served as the primary point of contact with the destination, and consequently, as the primary source of information for visitors. By offering accommodation and hospitality, the accommodation provider has evolved into a trusted third party, functioning as an 'ambassador' capable of recommending the finest attractions and activities available at the destination. In contemporary times, when tourists plan their trips, they make a series of consecutive decisions, with the most important decision being to lock-in the accommodation reservation for the earliest days, so as to secure a safe arrival. Consequently, tourists place their trust in the accommodation provider not only for lodging but also for recommendations regarding restaurants, activities, and more. Thus, the latter has the opportunity to inform and influence tourists well in advance of their arrival, particularly during the booking phase, namely when it comes to selecting their mode of transportation. The pressing need to reduce greenhouse gas emissions within the tourism sector presents an opportunity to underscore the influence that accommodation providers have historically exerted on tourist decision-making . Methodology A participatory research, currently ongoing in south-western France, in collaboration with a nationwide hotel group and several destination management organizations, aims at examining the factors that determine the ability of accommodation providers to influence tourist transportation choices. Additionally, the research seeks to identify the conditions that motivate accommodation providers to assume a proactive role, such as fostering customer loyalty, reduced distribution costs, and financial compensation mechanisms. A panel of hotels participated in a series of focus group sessions with tourists, with the objective of modeling the decision-making process of tourists regarding their choice of transportation mode and to identify and quantify the types and levels of incentives liable to encourage environmentally responsible choices. Individual interviews were also conducted with hotel staff, including receptionists and guest relations officers, to develop a framework for interactions with tourists during crucial decision-making moments related to transportation choices. The primary finding of this research indicates that financial incentives significantly outweigh symbolic incentives in motivating tourists to opt for eco-friendly modes of transportation. Another noteworthy result underscores the crucial impact of organizational conditions governing interactions with tourists both before and during their stay. These conditions greatly influence the ability to raise awareness at key decision-making moments and the possibility of gathering data about the chosen transportation mode during the stay. In conclusion, this research has led to the formulation of practical recommendations for accommodation providers and Destination Marketing Organizations (DMOs). These recommendations pertain to communication protocols with tourists, the collection of evidences confirming chosen transportation modes, and the implementation of necessary incentives. Through these measures, accommodation provider can assume a central role in guiding tourists towards making responsible choices in terms of transportation.

Keywords: accommodation provider, trusted third party, environmentally-friendly transportation, green house gas, tourist decision-making process

Procedia PDF Downloads 36
47 Numerical Analysis of Mandible Fracture Stabilization System

Authors: Piotr Wadolowski, Grzegorz Krzesinski, Piotr Gutowski

Abstract:

The aim of the presented work is to recognize the impact of mini-plate application approach on the stress and displacement within the stabilization devices and surrounding bones. The mini-plate osteosynthesis technique is widely used by craniofacial surgeons as an improved replacement of wire connection approach. Many different types of metal plates and screws are used to the physical connection of fractured bones. Below investigation is based on a clinical observation of patient hospitalized with mini-plate stabilization system. Analysis was conducted on a solid mandible geometry, which was modeled basis on the computed tomography scan of the hospitalized patient. In order to achieve most realistic connected system behavior, the cortical and cancellous bone layers were assumed. The temporomandibular joint was simplified to the elastic element to allow physiological movement of loaded bone. The muscles of mastication system were reduced to three pairs, modeled as shell structures. Finite element grid was created by the ANSYS software, where hexahedral and tetrahedral variants of SOLID185 element were used. A set of nonlinear contact conditions were applied on connecting devices and bone common surfaces. Properties of particular contact pair depend on screw - mini-plate connection type and possible gaps between fractured bone around osteosynthesis region. Some of the investigated cases contain prestress introduced to the mini-plate during the application, what responds the initial bending of the connecting device to fit the retromolar fossa region. Assumed bone fracture occurs within the mandible angle zone. Due to the significant deformation of the connecting plate in some of the assembly cases the elastic-plastic model of titanium alloy was assumed. The bone tissues were covered by the orthotropic material. As a loading were used the gauge force of magnitude of 100N applied in three different locations. Conducted analysis shows significant impact of mini-plate application methodology on the stress distribution within the miniplate. Prestress effect introduces additional loading, which leads to locally exceed the titanium alloy yield limit. Stress in surrounding bone increases rapidly around the screws application region, exceeding assumed bone yield limit, what indicate the local bone destruction. Approach with the doubled mini-plate shows increased stress within the connector due to the too rigid connection, where the main path of loading leads through the mini-plates instead of plates and connected bones. Clinical observations confirm more frequent plate destruction of stiffer connections. Some of them could be an effect of decreased low cyclic fatigue capability caused by the overloading. The executed analysis prove that the mini-plate system provides sufficient support to mandible fracture treatment, however, many applicable solutions shifts the entire system to the allowable material limits. The results show that connector application with the initial loading needs to be carefully established due to the small material capability tolerances. Comparison to the clinical observations allows optimizing entire connection to prevent future incidents.

Keywords: mandible fracture, mini-plate connection, numerical analysis, osteosynthesis

Procedia PDF Downloads 249
46 A Multi-Model Approach to Assess Atlantic Bonito (Sarda Sarda, Bloch 1793) in the Eastern Atlantic Ocean: A Case Study of the Senegalese Exclusive Economic Zone

Authors: Ousmane Sarr

Abstract:

The Senegalese coasts have high productivity of fishery resources due to the frequency of intense up-welling system that occurs along its coast, caused by the maritime trade winds making its waters nutrients rich. Fishing plays a primordial role in Senegal's socioeconomic plans and food security. However, a global diagnosis of the Senegalese maritime fishing sector has highlighted the challenges this sector encounters. Among these concerns, some significant stocks, a priority target for artisanal fishing, need further assessment. If no efforts are made in this direction, most stock will be overexploited or even in decline. It is in this context that this research was initiated. This investigation aimed to apply a multi-modal approach (LBB, Catch-only-based CMSY model and its most recent version (CMSY++); JABBA, and JABBA-Select) to assess the stock of Atlantic bonito, Sarda sarda (Bloch, 1793) in the Senegalese Exclusive Economic Zone (SEEZ). Available catch, effort, and size data from Atlantic bonito over 15 years (2004-2018) were used to calculate the nominal and standardized CPUE, size-frequency distribution, and length at retentions (50 % and 95 % selectivity) of the species. These relevant results were employed as input parameters for stock assessment models mentioned above to define the stock status of this species in this region of the Atlantic Ocean. The LBB model indicated an Atlantic bonito healthy stock status with B/BMSY values ranging from 1.3 to 1.6 and B/B0 values varying from 0.47 to 0.61 of the main scenarios performed (BON_AFG_CL, BON_GN_Length, and BON_PS_Length). The results estimated by LBB are consistent with those obtained by CMSY. The CMSY model results demonstrate that the SEEZ Atlantic bonito stock is in a sound condition in the final year of the main scenarios analyzed (BON, BON-bt, BON-GN-bt, and BON-PS-bt) with sustainable relative stock biomass (B2018/BMSY = 1.13 to 1.3) and fishing pressure levels (F2018/FMSY= 0.52 to 1.43). The B/BMSY and F/FMSY results for the JABBA model ranged between 2.01 to 2.14 and 0.47 to 0.33, respectively. In contrast, The estimated B/BMSY and F/FMSY for JABBA-Select ranged from 1.91 to 1.92 and 0.52 to 0.54. The Kobe plots results of the base case scenarios ranged from 75% to 89% probability in the green area, indicating sustainable fishing pressure and an Atlantic bonito healthy stock size capable of producing high yields close to the MSY. Based on the stock assessment results, this study highlighted scientific advice for temporary management measures. This study suggests an improvement of the selectivity parameters of longlines and purse seines and a temporary prohibition of the use of sleeping nets in the fishery for the Atlantic bonito stock in the SEEZ based on the results of the length-base models. Although these actions are temporary, they can be essential to reduce or avoid intense pressure on the Atlantic bonito stock in the SEEZ. However, it is necessary to establish harvest control rules to provide coherent and solid scientific information that leads to appropriate decision-making for rational and sustainable exploitation of Atlantic bonito in the SEEZ and the Eastern Atlantic Ocean.

Keywords: multi-model approach, stock assessment, atlantic bonito, healthy stock, sustainable, SEEZ, temporary management measures

Procedia PDF Downloads 38
45 Harnessing the Benefits and Mitigating the Challenges of Neurosensitivity for Learners: A Mixed Methods Study

Authors: Kaaryn Cater

Abstract:

People vary in how they perceive, process, and react to internal, external, social, and emotional environmental factors; some are more sensitive than others. Compassionate people have a highly reactive nervous system and are more impacted by positive and negative environmental conditions (Differential Susceptibility). Further, some sensitive individuals are disproportionately able to benefit from positive and supportive environments without necessarily suffering negative impacts in less supportive environments (Vantage Sensitivity). Environmental sensitivity is underpinned by physiological, genetic, and personality/temperamental factors, and the phenotypic expression of high sensitivity is Sensory Processing Sensitivity. The hallmarks of Sensory Processing Sensitivity are deep cognitive processing, emotional reactivity, high levels of empathy, noticing environmental subtleties, a tendency to observe new and novel situations, and a propensity to become overwhelmed when over-stimulated. Several educational advantages associated with high sensitivity include creativity, enhanced memory, divergent thinking, giftedness, and metacognitive monitoring. High sensitivity can also lead to some educational challenges, particularly managing multiple conflicting demands and negotiating low sensory thresholds. A mixed methods study was undertaken. In the first quantitative study, participants completed the Perceived Success in Study Survey (PSISS) and the Highly Sensitive Person Scale (HSPS-12). Inclusion criteria were current or previous postsecondary education experience. The survey was presented on social media, and snowball recruitment was employed (n=365). The Excel spreadsheets were uploaded to the statistical package for the social sciences (SPSS)26, and descriptive statistics found normal distribution. T-tests and analysis of variance (ANOVA) calculations found no difference in the responses of demographic groups, and Principal Components Analysis and the posthoc Tukey calculations identified positive associations between high sensitivity and three of the five PSISS factors. Further ANOVA calculations found positive associations between the PSISS and two of the three sensitivity subscales. This study included a response field to register interest in further research. Respondents who scored in the 70th percentile on the HSPS-12 were invited to participate in a semi-structured interview. Thirteen interviews were conducted remotely (12 female). Reflexive inductive thematic analysis was employed to analyse data, and a descriptive approach was employed to present data reflective of participant experience. The results of this study found that compassionate students prioritize work-life balance; employ a range of practical metacognitive study and self-care strategies; value independent learning; connect with learning that is meaningful; and are bothered by aspects of the physical learning environment, including lighting, noise, and indoor environmental pollutants. There is a dearth of research investigating sensitivity in the educational context, and these studies highlight the need to promote widespread education sector awareness of environmental sensitivity, and the need to include sensitivity in sector and institutional diversity and inclusion initiatives.

Keywords: differential susceptibility, highly sensitive person, learning, neurosensitivity, sensory processing sensitivity, vantage sensitivity

Procedia PDF Downloads 44
44 EGF Serum Level in Diagnosis and Prediction of Mood Disorder in Adolescents and Young Adults

Authors: Monika Dmitrzak-Weglarz, Aleksandra Rajewska-Rager, Maria Skibinska, Natalia Lepczynska, Piotr Sibilski, Joanna Pawlak, Pawel Kapelski, Joanna Hauser

Abstract:

Epidermal growth factor (EGF) is a well-known neurotrophic factor that involves in neuronal growth and synaptic plasticity. The proteomic research provided in order to identify novel candidate biological markers for mood disorders focused on elevated EGF serum level in patients during depression episode. However, the EGF association with mood disorder spectrum among adolescents and young adults has not been studied extensively. In this study, we aim to investigate the serum levels of EGF in adolescents and young adults during hypo/manic, depressive episodes and in remission compared to healthy control group. In our study, we involved 80 patients aged 12-24 years in 2-year follow-up study with a primary diagnosis of mood disorder spectrum, and 35 healthy volunteers matched by age and gender. Diagnoses were established according to DSM-IV-TR criteria using structured clinical interviews: K-SADS for child and adolescents, and SCID for young adults. Clinical and biological evaluations were made at baseline and euthymic mood (at 3th or 6th month of treatment and after 1 and 2 years). The Young Mania Rating Scale and Hamilton Rating Scale for Depression were used for assessment. The study protocols were approved by the relevant ethics committee. Serum protein concentration was determined by Enzyme-Linked Immunosorbent Assays (ELISA) method. Human EGF (cat. no DY 236) DuoSet ELISA kit was used (R&D Systems). Serum EGF levels were analysed with following variables: age, age under 18 and above 18 years old, sex, family history of affective disorders, drug-free vs. medicated. Shapiro-Wilk test was used to test the normality of the data. The homogeneity of variance was calculated with Levene’s test. EGF levels showed non-normal distribution and the homogeneity of variance was violated. Non-parametric tests: Mann-Whitney U test, Kruskall-Wallis ANOVA, Friedman’s ANOVA, Wilcoxon signed rank test, Spearman correlation coefficient was applied in the analyses The statistical significance level was set at p<0.05. Elevated EGF level at baseline (p=0.001) and at month 24 (p=0.02) was detected in study subjects compared with controls. Increased EGF level in women at month 12 (p=0.02) compared to men in study group have been observed. Using Wilcoxon signed rank test differences in EGF levels were detected: decrease from baseline to month 3 (p=0.014) and increase comparing: month 3 vs. 24 (p=0.013); month 6 vs. 12 (p=0.021) and vs. 24 (p=0.008). EGF level at baseline was negatively correlated with depression and mania occurrence at 24 months. EGF level at 24 months was positively correlated with depression and mania occurrence at 12 months. No other correlations of EGF levels with clinical and demographical variables have been detected. The findings of the present study indicate that EGF serum level is significantly elevated in the study group of patients compared to the controls. We also observed fluctuations in EGF levels during two years of disease observation. EGF seems to be useful as an early marker for prediction of diagnosis, course of illness and treatment response in young patients during first episode od mood disorders, which requires further investigation. Grant was founded by National Science Center in Poland no 2011/03/D/NZ5/06146.

Keywords: biological marker, epidermal growth factor, mood disorders, prediction

Procedia PDF Downloads 166
43 Mapping Iron Content in the Brain with Magnetic Resonance Imaging and Machine Learning

Authors: Gabrielle Robertson, Matthew Downs, Joseph Dagher

Abstract:

Iron deposition in the brain has been linked with a host of neurological disorders such as Alzheimer’s, Parkinson’s, and Multiple Sclerosis. While some treatment options exist, there are no objective measurement tools that allow for the monitoring of iron levels in the brain in vivo. An emerging Magnetic Resonance Imaging (MRI) method has been recently proposed to deduce iron concentration through quantitative measurement of magnetic susceptibility. This is a multi-step process that involves repeated modeling of physical processes via approximate numerical solutions. For example, the last two steps of this Quantitative Susceptibility Mapping (QSM) method involve I) mapping magnetic field into magnetic susceptibility and II) mapping magnetic susceptibility into iron concentration. Process I involves solving an ill-posed inverse problem by using regularization via injection of prior belief. The end result from Process II highly depends on the model used to describe the molecular content of each voxel (type of iron, water fraction, etc.) Due to these factors, the accuracy and repeatability of QSM have been an active area of research in the MRI and medical imaging community. This work aims to estimate iron concentration in the brain via a single step. A synthetic numerical model of the human head was created by automatically and manually segmenting the human head on a high-resolution grid (640x640x640, 0.4mm³) yielding detailed structures such as microvasculature and subcortical regions as well as bone, soft tissue, Cerebral Spinal Fluid, sinuses, arteries, and eyes. Each segmented region was then assigned tissue properties such as relaxation rates, proton density, electromagnetic tissue properties and iron concentration. These tissue property values were randomly selected from a Probability Distribution Function derived from a thorough literature review. In addition to having unique tissue property values, different synthetic head realizations also possess unique structural geometry created by morphing the boundary regions of different areas within normal physical constraints. This model of the human brain is then used to create synthetic MRI measurements. This is repeated thousands of times, for different head shapes, volume, tissue properties and noise realizations. Collectively, this constitutes a training-set that is similar to in vivo data, but larger than datasets available from clinical measurements. This 3D convolutional U-Net neural network architecture was used to train data-driven Deep Learning models to solve for iron concentrations from raw MRI measurements. The performance was then tested on both synthetic data not used in training as well as real in vivo data. Results showed that the model trained on synthetic MRI measurements is able to directly learn iron concentrations in areas of interest more effectively than other existing QSM reconstruction methods. For comparison, models trained on random geometric shapes (as proposed in the Deep QSM method) are less effective than models trained on realistic synthetic head models. Such an accurate method for the quantitative measurement of iron deposits in the brain would be of important value in clinical studies aiming to understand the role of iron in neurological disease.

Keywords: magnetic resonance imaging, MRI, iron deposition, machine learning, quantitative susceptibility mapping

Procedia PDF Downloads 101
42 Microstructural Characterization of Bitumen/Montmorillonite/Isocyanate Composites by Atomic Force Microscopy

Authors: Francisco J. Ortega, Claudia Roman, Moisés García-Morales, Francisco J. Navarro

Abstract:

Asphaltic bitumen has been largely used in both industrial and civil engineering, mostly in pavement construction and roofing membrane manufacture. However, bitumen as such is greatly susceptible to temperature variations, and dramatically changes its in-service behavior from a viscoelastic liquid, at medium-high temperatures, to a brittle solid at low temperatures. Bitumen modification prevents these problems and imparts improved performance. Isocyanates like polymeric MDI (mixture of 4,4′-diphenylmethane di-isocyanate, 2,4’ and 2,2’ isomers, and higher homologues) have shown to remarkably enhance bitumen properties at the highest in-service temperatures expected. This comes from the reaction between the –NCO pendant groups of the oligomer and the most polar groups of asphaltenes and resins in bitumen. In addition, oxygen diffusion and/or UV radiation may provoke bitumen hardening and ageing. With the purpose of minimizing these effects, nano-layered-silicates (nanoclays) are increasingly being added to bitumen formulations. Montmorillonites, a type of naturally occurring mineral, may produce a nanometer scale dispersion which improves bitumen thermal, mechanical and barrier properties. In order to increase their lipophilicity, these nanoclays are normally treated so that organic cations substitute the inorganic cations located in their intergallery spacing. In the present work, the combined effect of polymeric MDI and the commercial montmorillonite Cloisite® 20A was evaluated. A selected bitumen with penetration within the range 160/220 was modified with 10 wt.% Cloisite® 20A and 2 wt.% polymeric MDI, and the resulting ternary composites were characterized by linear rheology, X-ray diffraction (XRD) and Atomic Force Microscopy (AFM). The rheological tests evidenced a notable solid-like behavior at the highest temperatures studied when bitumen was just loaded with 10 wt.% Cloisite® 20A and high-shear blended for 20 minutes. However, if polymeric MDI was involved, the sequence of addition exerted a decisive control on the linear rheology of the final ternary composites. Hence, in bitumen/Cloisite® 20A/polymeric MDI formulations, the previous solid-like behavior disappeared. By contrast, an inversion in the order of addition (bitumen/polymeric MDI/ Cloisite® 20A) enhanced further the solid-like behavior imparted by the nanoclay. In order to gain a better understanding of the factors that govern the linear rheology of these ternary composites, a morphological and microstructural characterization based on XRD and AFM was conducted. XRD demonstrated the existence of clay stacks intercalated by bitumen molecules to some degree. However, the XRD technique cannot provide detailed information on the extent of nanoclay delamination, unless the entire fraction has effectively been fully delaminated (situation in which no peak is observed). Furthermore, XRD was unable to provide precise knowledge neither about the spatial distribution of the intercalated/exfoliated platelets nor about the presence of other structures at larger length scales. In contrast, AFM proved its power at providing conclusive information on the morphology of the composites at the nanometer scale and at revealing the structural modification that yielded the rheological properties observed. It was concluded that high-shear blending brought about a nanoclay-reinforced network. As for the bitumen/Cloisite® 20A/polymeric MDI formulations, the solid-like behavior was destroyed as a result of the agglomeration of the nanoclay platelets promoted by chemical reactions.

Keywords: Atomic Force Microscopy, bitumen, composite, isocyanate, montmorillonite.

Procedia PDF Downloads 237
41 A Microwave Heating Model for Endothermic Reaction in the Cement Industry

Authors: Sofia N. Gonçalves, Duarte M. S. Albuquerque, José C. F. Pereira

Abstract:

Microwave technology has been gaining importance in contributing to decarbonization processes in high energy demand industries. Despite the several numerical models presented in the literature, a proper Verification and Validation exercise is still lacking. This is important and required to evaluate the physical process model accuracy and adequacy. Another issue addresses impedance matching, which is an important mechanism used in microwave experiments to increase electromagnetic efficiency. Such mechanism is not available in current computational tools, thus requiring an external numerical procedure. A numerical model was implemented to study the continuous processing of limestone with microwave heating. This process requires the material to be heated until a certain temperature that will prompt a highly endothermic reaction. Both a 2D and 3D model were built in COMSOL Multiphysics to solve the two-way coupling between Maxwell and Energy equations, along with the coupling between both heat transfer phenomena and limestone endothermic reaction. The 2D model was used to study and evaluate the required numerical procedure, being also a benchmark test, allowing other authors to implement impedance matching procedures. To achieve this goal, a controller built in MATLAB was used to continuously matching the cavity impedance and predicting the required energy for the system, thus successfully avoiding energy inefficiencies. The 3D model reproduces realistic results and therefore supports the main conclusions of this work. Limestone was modeled as a continuous flow under the transport of concentrated species, whose material and kinetics properties were taken from literature. Verification and Validation of the coupled model was taken separately from the chemical kinetic model. The chemical kinetic model was found to correctly describe the chosen kinetic equation by comparing numerical results with experimental data. A solution verification was made for the electromagnetic interface, where second order and fourth order accurate schemes were found for linear and quadratic elements, respectively, with numerical uncertainty lower than 0.03%. Regarding the coupled model, it was demonstrated that the numerical error would diverge for the heat transfer interface with the mapped mesh. Results showed numerical stability for the triangular mesh, and the numerical uncertainty was less than 0.1%. This study evaluated limestone velocity, heat transfer, and load influence on thermal decomposition and overall process efficiency. The velocity and heat transfer coefficient were studied with the 2D model, while different loads of material were studied with the 3D model. Both models demonstrated to be highly unstable when solving non-linear temperature distributions. High velocity flows exhibited propensity to thermal runways, and the thermal efficiency showed the tendency to stabilize for the higher velocities and higher filling ratio. Microwave efficiency denoted an optimal velocity for each heat transfer coefficient, pointing out that electromagnetic efficiency is a consequence of energy distribution uniformity. The 3D results indicated the inefficient development of the electric field for low filling ratios. Thermal efficiencies higher than 90% were found for the higher loads and microwave efficiencies up to 75% were accomplished. The 80% fill ratio was demonstrated to be the optimal load with an associated global efficiency of 70%.

Keywords: multiphysics modeling, microwave heating, verification and validation, endothermic reactions modeling, impedance matching, limestone continuous processing

Procedia PDF Downloads 120
40 Information Pollution: Exploratory Analysis of Subs-Saharan African Media’s Capabilities to Combat Misinformation and Disinformation

Authors: Muhammed Jamiu Mustapha, Jamiu Folarin, Stephen Obiri Agyei, Rasheed Ademola Adebiyi, Mutiu Iyanda Lasisi

Abstract:

The role of information in societal development and growth cannot be over-emphasized. It has remained an age-long strategy to adopt the information flow to make an egalitarian society. The same has become a tool for throwing society into chaos and anarchy. It has been adopted as a weapon of war and a veritable instrument of psychological warfare with a variety of uses. That is why some scholars posit that information could be deployed as a weapon to wreak “Mass Destruction" or promote “Mass Development". When used as a tool for destruction, the effect on society is like an atomic bomb which when it is released, pollutes the air and suffocates the people. Technological advancement has further exposed the latent power of information and many societies seem to be overwhelmed by its negative effect. While information remains one of the bedrock of democracy, the information ecosystem across the world is currently facing a more difficult battle than ever before due to information pluralism and technological advancement. The more the agents involved try to combat its menace, the difficult and complex it is proving to be curbed. In a region like Africa with dangling democracy enfolds with complexities of multi-religion, multi-cultures, inter-tribes, ongoing issues that are yet to be resolved, it is important to pay critical attention to the case of information disorder and find appropriate ways to curb or mitigate its effects. The media, being the middleman in the distribution of information, needs to build capacities and capabilities to separate the whiff of misinformation and disinformation from the grains of truthful data. From quasi-statistical senses, it has been observed that the efforts aimed at fighting information pollution have not considered the built resilience of media organisations against this disorder. Apparently, the efforts, resources and technologies adopted for the conception, production and spread of information pollution are much more sophisticated than approaches to suppress and even reduce its effects on society. Thus, this study seeks to interrogate the phenomenon of information pollution and the capabilities of select media organisations in Sub-Saharan Africa. In doing this, the following questions are probed; what are the media actions to curb the menace of information pollution? Which of these actions are working and how effective are they? And which of the actions are not working and why they are not working? Adopting quantitative and qualitative approaches and anchored on the Dynamic Capability Theory, the study aims at digging up insights to further understand the complexities of information pollution, media capabilities and strategic resources for managing misinformation and disinformation in the region. The quantitative approach involves surveys and the use of questionnaires to get data from journalists on their understanding of misinformation/disinformation and their capabilities to gate-keep. Case Analysis of select media and content analysis of their strategic resources to manage misinformation and disinformation is adopted in the study while the qualitative approach will involve an In-depth Interview to have a more robust analysis is also considered. The study is critical in the fight against information pollution for a number of reasons. One, it is a novel attempt to document the level of media capabilities to fight the phenomenon of information disorder. Two, the study will enable the region to have a clear understanding of the capabilities of existing media organizations to combat misinformation and disinformation in the countries that make up the region. Recommendations emanating from the study could be used to initiate, intensify or review existing approaches to combat the menace of information pollution in the region.

Keywords: disinformation, information pollution, misinformation, media capabilities, sub-Saharan Africa

Procedia PDF Downloads 143
39 Xen45 Gel Implant in Open Angle Glaucoma: Efficacy, Safety and Predictors of Outcome

Authors: Fossarello Maurizio, Mattana Giorgio, Tatti Filippo.

Abstract:

The most widely performed surgical procedure in Open-Angle Glaucoma (OAG) is trabeculectomy. Although this filtering procedure is extremely effective, surgical failure and postoperative complications are reported. Due to the its invasive nature and possible complications, trabeculectomy is usually reserved, in practice, for patients who are refractory to medical and laser therapy. Recently, a number of micro-invasive surgical techniques (MIGS: Micro-Invasive Glaucoma Surgery), have been introduced in clinical practice. They meet the criteria of micro-incisional approach, minimal tissue damage, short surgical time, reliable IOP reduction, extremely high safety profile and rapid post-operative recovery. Xen45 Gel Implant (Allergan, Dublin, Ireland) is one of the MIGS alternatives, and consists in a porcine gelatin tube designed to create an aqueous flow from the anterior chamber to the subconjunctival space, bypassing the resistance of the trabecular meshwork. In this study we report the results of this technique as a favorable option in the treatment of OAG for its benefits in term of efficacy and safety, either alone or in combination with cataract surgery. This is a retrospective, single-center study conducted in consecutive OAG patients, who underwent Xen45 Gel Stent implantation alone or in combination with phacoemulsification, from October 2018 to June 2019. The primary endpoint of the study was to evaluate the reduction of both IOP and number of antiglaucoma medications at 12 months. The secondary endpoint was to correlate filtering bleb morphology evaluated by means of anterior segment OCT with efficacy in IOP lowering and eventual further procedures requirement. Data were recorded on Microsoft Excel and study analysis was performed using Microsoft Excel and SPSS (IBM). Mean values with standard deviations were calculated for IOPs and number of antiglaucoma medications at all points. Kolmogorov-Smirnov test showed that IOP followed a normal distribution at all time, therefore the paired Student’s T test was used to compare baseline and postoperative mean IOP. Correlation between postoperative Day 1 IOP and Month 12 IOP was evaluated using Pearson coefficient. Thirty-six eyes of 36 patients were evaluated. As compared to baseline, mean IOP and the mean number of antiglaucoma medications significantly decreased from 27,33 ± 7,67 mmHg to 16,3 ± 2,89 mmHg (38,8% reduction) and from 2,64 ± 1,39 to 0,42 ± 0,8 (84% reduction), respectively, at 12 months after surgery (both p < 0,001). According to bleb morphology, eyes were divided in uniform group (n=8, 22,2%), subconjunctival separation group (n=5, 13,9%), microcystic multiform group (n=9, 25%) and multiple internal layer group (n=14, 38,9%). Comparing to baseline, there was no significative difference in IOP between the 4 groups at month 12 follow-up visit. Adverse events included bleb function decrease (n=14, 38,9%), hypotony (n=8, 22,2%) and choroidal detachment (n=2, 5,6%). All eyes presenting bleb flattening underwent needling and MMC injection. The higher percentage of patients that required secondary needling was in the uniform group (75%), with a significant difference between the groups (p=0,03). Xen45 gel stent, either alone or in combination with phacoemulsification, provided a significant lowering in both IOP and medical antiglaucoma treatment and an elevated safety profile.

Keywords: anterior segment OCT, bleb morphology, micro-invasive glaucoma surgery, open angle glaucoma, Xen45 gel implant

Procedia PDF Downloads 109
38 A Review on Biological Control of Mosquito Vectors

Authors: Asim Abbasi, Muhammad Sufyan, Iqra, Hafiza Javaria Ashraf

Abstract:

The share of vector-borne diseases (VBDs) in the global burden of infectious diseases is almost 17%. The advent of new drugs and latest research in medical science helped mankind to compete with these lethal diseases but still diseases transmitted by different mosquito species, including filariasis, malaria, viral encephalitis and dengue are serious threats for people living in disease endemic areas. Injudicious and repeated use of pesticides posed selection pressure on mosquitoes leading to development of resistance. Hence biological control agents are under serious consideration of scientific community to be used in vector control programmes. Fish have a history of predating immature stages of different aquatic insects including mosquitoes. The noteworthy examples in Africa and Asia includes, Aphanius discolour and a fish in the Panchax group. Moreover, common mosquito fish, Gambusia affinis predates mostly on temporary water mosquitoes like anopheline as compared to permanent water breeders like culicines. Mosquitoes belonging to genus Toxorhynchites have a worldwide distribution and are mostly associated with the predation of other mosquito larvae habituating with them in natural and artificial water containers. These species are harmless to humans as their adults do not suck human blood but feeds on floral nectar. However, their activity is mostly temperature dependent as Toxorhynchites brevipalpis consume 359 Aedes aegypti larvae at 30-32 ºC in contrast to 154 larvae at 20-26 ºC. Although many bacterial species were isolated from mosquito cadavers but those belonging to genus Bacillus are found highly pathogenic against them. The successful species of this genus include Bacillus thuringiensis and Bacillus sphaericus. The prime targets of B. thuringiensis are mostly the immatures of genus Aedes, Culex, Anopheles and Psorophora while B. sphaericus is specifically toxic against species of Culex, Psorophora and Culiseta. The entomopathogenic nematodes belonging to family, mermithidae are also pathogenic to different mosquito species. Eighty different species of mosquitoes including Anopheles, Aedes and Culex proved to be highly vulnerable to the attack of two mermithid species, Romanomermis culicivorax and R. iyengari. Cytoplasmic polyhedrosis virus was the first described pathogenic virus, isolated from the cadavers of mosquito specie, Culex tarsalis. Other viruses which are pathogenic to culicine includes, iridoviruses, cytopolyhedrosis viruses, entomopoxviruses and parvoviruses. Protozoa species belonging to division microsporidia are the common pathogenic protozoans in mosquito populations which kill their host by the chronic effects of parasitism. Moreover, due to their wide prevalence in anopheline mosquitoes and transversal and horizontal transmission from infected to healthy host, microsporidia of the genera Nosema and Amblyospora have received much attention in various mosquito control programmes. Fungal based mycopesticides are used in biological control of insect pests with 47 species reported virulent against different stages of mosquitoes. These include both aquatic fungi i.e. species of Coelomomyces, Lagenidium giganteum and Culicinomyces clavosporus, and the terrestrial fungi Metarhizium anisopliae and Beauveria bassiana. Hence, it was concluded that the integrated use of all these biological control agents can be a healthy contribution in mosquito control programmes and become a dire need of the time to avoid repeated use of pesticides.

Keywords: entomopathogenic nematodes, protozoa, Toxorhynchites, vector-borne

Procedia PDF Downloads 242
37 Influence of Thermal Annealing on Phase Composition and Structure of Quartz-Sericite Minerale

Authors: Atabaev I. G., Fayziev Sh. A., Irmatova Sh. K.

Abstract:

Raw materials with high content of Kalium oxide widely used in ceramic technology for prevention or decreasing of deformation of ceramic goods during drying process and under thermal annealing. Becouse to low melting temperature it is also used to decreasing of the temperature of thermal annealing during fabrication of ceramic goods [1,2]. So called “Porceline or China stones” - quartz-sericite (muscovite) minerals is also can be used for prevention of deformation as the content of Kalium oxide in muscovite is rather high (SiO2, + KAl2[AlSi3O10](OH)2). [3] . To estimation of possibility of use of this mineral for ceramic manufacture, in the presented article the influence of thermal processing on phase and a chemical content of this raw material is investigated. As well as to other ceramic raw materials (kaoline, white burning clays) the basic requirements of the industry to quality of "a porcelain stone» are following: small size of particles, relative high uniformity of disrtribution of components and phase, white color after burning, small content of colorant oxides or chromophores (Fe2O3, FeO, TiO2, etc) [4,5]. In the presented work natural minerale from the Boynaksay deposit (Uzbekistan) is investigated. The samples was mechanically polished for investigation by Scanning Electron Microscope. Powder with size of particle up to 63 μm was used to X-ray diffractometry and chemical analysis. The annealing of samples was performed at 900, 1120, 1350oC during 1 hour. Chemical composition of Boynaksay raw material according to chemical analysis presented in the table 1. For comparison the composition of raw materials from Russia and USA are also presented. In the Boynaksay quartz – sericite the average parity of quartz and sericite makes 55-60 and 30-35 % accordingly. The distribution of quartz and sericite phases in raw material was investigated using electron probe scanning electronic microscope «JEOL» JXA-8800R. In the figure 1 the scanning electron microscope (SEM) micrograps of the surface and the distributions of Al, Si and K atoms in the sample are presented. As it seen small granular, white and dense mineral includes quartz, sericite and small content of impurity minerals. Basically, crystals of quartz have the sizes from 80 up to 500 μm. Between quartz crystals the sericite inclusions having a tablet form with radiant structure are located. The size of sericite crystals is ~ 40-250 μm. Using data on interplanar distance [6,7] and ASTM Powder X-ray Diffraction Data it is shown that natural «a porcelain stone» quartz – sericite consists the quartz SiO2, sericite (muscovite type) KAl2[AlSi3O10](OH)2 and kaolinite Al203SiO22Н2О (See Figure 2 and Table 2). As it seen in the figure 3 and table 3a after annealing at 900oC the quartz – sericite contains quartz – SiO2 and muscovite - KAl2[AlSi3O10](OH)2, the peaks related with Kaolinite are absent. After annealing at 1120oC the full disintegration of muscovite and formation of mullite phase Al203 SiO2 is observed (the weak peaks of mullite appears in fig 3b and table 3b). After annealing at 1350oC the samples contains crystal phase of quartz and mullite (figure 3c and table 3с). Well known Mullite gives to ceramics high density, abrasive and chemical stability. Thus the obtained experimental data on formation of various phases during thermal annealing can be used for development of fabrication technology of advanced materials. Conclusion: The influence of thermal annealing in the interval 900-1350oC on phase composition and structure of quartz-sericite minerale is investigated. It is shown that during annealing the phase content of raw material is changed. After annealing at 1350oC the samples contains crystal phase of quartz and mullite (which gives gives to ceramics high density, abrasive and chemical stability).

Keywords: quartz-sericite, kaolinite, mullite, thermal processing

Procedia PDF Downloads 386
36 Reverse Logistics Network Optimization for E-Commerce

Authors: Albert W. K. Tan

Abstract:

This research consolidates a comprehensive array of publications from peer-reviewed journals, case studies, and seminar reports focused on reverse logistics and network design. By synthesizing this secondary knowledge, our objective is to identify and articulate key decision factors crucial to reverse logistics network design for e-commerce. Through this exploration, we aim to present a refined mathematical model that offers valuable insights for companies seeking to optimize their reverse logistics operations. The primary goal of this research endeavor is to develop a comprehensive framework tailored to advising organizations and companies on crafting effective networks for their reverse logistics operations, thereby facilitating the achievement of their organizational goals. This involves a thorough examination of various network configurations, weighing their advantages and disadvantages to ensure alignment with specific business objectives. The key objectives of this research include: (i) Identifying pivotal factors pertinent to network design decisions within the realm of reverse logistics across diverse supply chains. (ii) Formulating a structured framework designed to offer informed recommendations for sound network design decisions applicable to relevant industries and scenarios. (iii) Propose a mathematical model to optimize its reverse logistics network. A conceptual framework for designing a reverse logistics network has been developed through a combination of insights from the literature review and information gathered from company websites. This framework encompasses four key stages in the selection of reverse logistics operations modes: (1) Collection, (2) Sorting and testing, (3) Processing, and (4) Storage. Key factors to consider in reverse logistics network design: I) Centralized vs. decentralized processing: Centralized processing, a long-standing practice in reverse logistics, has recently gained greater attention from manufacturing companies. In this system, all products within the reverse logistics pipeline are brought to a central facility for sorting, processing, and subsequent shipment to their next destinations. Centralization offers the advantage of efficiently managing the reverse logistics flow, potentially leading to increased revenues from returned items. Moreover, it aids in determining the most appropriate reverse channel for handling returns. On the contrary, a decentralized system is more suitable when products are returned directly from consumers to retailers. In this scenario, individual sales outlets serve as gatekeepers for processing returns. Considerations encompass the product lifecycle, product value and cost, return volume, and the geographic distribution of returns. II) In-house vs. third-party logistics providers: The decision between insourcing and outsourcing in reverse logistics network design is pivotal. In insourcing, a company handles the entire reverse logistics process, including material reuse. In contrast, outsourcing involves third-party providers taking on various aspects of reverse logistics. Companies may choose outsourcing due to resource constraints or lack of expertise, with the extent of outsourcing varying based on factors such as personnel skills and cost considerations. Based on the conceptual framework, the authors have constructed a mathematical model that optimizes reverse logistics network design decisions. The model will consider key factors identified in the framework, such as transportation costs, facility capacities, and lead times. The authors have employed mixed LP to find the optimal solutions that minimize costs while meeting organizational objectives.

Keywords: reverse logistics, supply chain management, optimization, e-commerce

Procedia PDF Downloads 11
35 Anajaa-Visual Substitution System: A Navigation Assistive Device for the Visually Impaired

Authors: Juan Pablo Botero Torres, Alba Avila, Luis Felipe Giraldo

Abstract:

Independent navigation and mobility through unknown spaces pose a challenge for the autonomy of visually impaired people (VIP), who have relied on the use of traditional assistive tools like the white cane and trained dogs. However, emerging visually assistive technologies (VAT) have proposed several human-machine interfaces (HMIs) that could improve VIP’s ability for self-guidance. Hereby, we introduce the design and implementation of a visually assistive device, Anajaa – Visual Substitution System (AVSS). This system integrates ultrasonic sensors with custom electronics, and computer vision models (convolutional neural networks), in order to achieve a robust system that acquires information of the surrounding space and transmits it to the user in an intuitive and efficient manner. AVSS consists of two modules: the sensing and the actuation module, which are fitted to a chest mount and belt that communicate via Bluetooth. The sensing module was designed for the acquisition and processing of proximity signals provided by an array of ultrasonic sensors. The distribution of these within the chest mount allows an accurate representation of the surrounding space, discretized in three different levels of proximity, ranging from 0 to 6 meters. Additionally, this module is fitted with an RGB-D camera used to detect potentially threatening obstacles, like staircases, using a convolutional neural network specifically trained for this purpose. Posteriorly, the depth data is used to estimate the distance between the stairs and the user. The information gathered from this module is then sent to the actuation module that creates an HMI, by the means of a 3x2 array of vibration motors that make up the tactile display and allow the system to deliver haptic feedback. The actuation module uses vibrational messages (tactones); changing both in amplitude and frequency to deliver different awareness levels according to the proximity of the obstacle. This enables the system to deliver an intuitive interface. Both modules were tested under lab conditions, and the HMI was additionally tested with a focal group of VIP. The lab testing was conducted in order to establish the processing speed of the computer vision algorithms. This experimentation determined that the model can process 0.59 frames per second (FPS); this is considered as an adequate processing speed taking into account that the walking speed of VIP is 1.439 m/s. In order to test the HMI, we conducted a focal group composed of two females and two males between the ages of 35-65 years. The subject selection was aided by the Colombian Cooperative of Work and Services for the Sightless (COOTRASIN). We analyzed the learning process of the haptic messages throughout five experimentation sessions using two metrics: message discrimination and localization success. These correspond to the ability of the subjects to recognize different tactones and locate them within the tactile display. Both were calculated as the mean across all subjects. Results show that the focal group achieved message discrimination of 70% and a localization success of 80%, demonstrating how the proposed HMI leads to the appropriation and understanding of the feedback messages, enabling the user’s awareness of its surrounding space.

Keywords: computer vision on embedded systems, electronic trave aids, human-machine interface, haptic feedback, visual assistive technologies, vision substitution systems

Procedia PDF Downloads 53
34 Renewable Energy Micro-Grid Control Using Microcontroller in LabVIEW

Authors: Meena Agrawal, Chaitanya P. Agrawal

Abstract:

The power systems are transforming and becoming smarter with innovations in technologies to enable embark simultaneously upon the sustainable energy needs, rising environmental concerns, economic benefits and quality requirements. The advantages provided by inter-connection of renewable energy resources are becoming more viable and dependable with the smart controlling technologies. The limitation of most renewable resources have their diversity and intermittency causing problems in power quality, grid stability, reliability, security etc. is being cured by these efforts. A necessitate of optimal energy management by intelligent Micro-Grids at the distribution end of the power system has been accredited to accommodate sustainable renewable Distributed Energy Resources on large scale across the power grid. All over the world Smart Grids are emerging now as foremost concern infrastructure upgrade programs. The hardware setup includes NI cRIO 9022, Compact Reconfigurable Input Output microcontroller board connected to the PC on a LAN router with three hardware modules. The Real-Time Embedded Controller is reconfigurable controller device consisting of an embedded real-time processor controller for communication and processing, a reconfigurable chassis housing the user-programmable FPGA, Eight hot-swappable I/O modules, and graphical LabVIEW system design software. It has been employed for signal analysis, controls and acquisition and logging of the renewable sources with the LabVIEW Real-Time applications. The employed cRIO chassis controls the timing for the module and handles communication with the PC over the USB, Ethernet, or 802.11 Wi-Fi buses. It combines modular I/O, real-time processing, and NI LabVIEW programmable. In the presented setup, the Analog Input Module NI 9205 five channels have been used for input analog voltage signals from renewable energy sources and NI 9227 four channels have been used for input analog current signals of the renewable sources. For switching actions based on the programming logic developed in software, a module having Electromechanical Relays (single-pole single throw) with 4-Channels, electrically isolated and LED indicating the state of that channel have been used for isolating the renewable Sources on fault occurrence, which is decided by the logic in the program. The module for Ethernet based Data Acquisition Interface ENET 9163 Ethernet Carrier, which is connected on the LAN Router for data acquisition from a remote source over Ethernet also has the module NI 9229 installed. The LabVIEW platform has been employed for efficient data acquisition, monitoring and control. Control logic utilized in program for operation of the hardware switching Related to Fault Relays has been portrayed as a flowchart. A communication system has been successfully developed amongst the sources and loads connected on different computers using Hypertext transfer protocol, HTTP or Ethernet Local Stacked area Network TCP/IP protocol. There are two main I/O interfacing clients controlling the operation of the switching control of the renewable energy sources over internet or intranet. The paper presents experimental results of the briefed setup for intelligent control of the micro-grid for renewable energy sources, besides the control of Micro-Grid with data acquisition and control hardware based on a microcontroller with visual program developed in LabVIEW.

Keywords: data acquisition and control, LabVIEW, microcontroller cRIO, Smart Micro-Grid

Procedia PDF Downloads 301
33 Design, Control and Implementation of 3.5 kW Bi-Directional Energy Harvester for Intelligent Green Energy Management System

Authors: P. Ramesh, Aby Joseph, Arya G. Lal, U. S. Aji

Abstract:

Integration of distributed green renewable energy sources in addition with battery energy storage is an inevitable requirement in a smart grid environment. To achieve this, an Intelligent Green Energy Management System (i-GEMS) needs to be incorporated to ensure coordinated operation between supply and load demand based on the hierarchy of Renewable Energy Sources (RES), battery energy storage and distribution grid. A bi-directional energy harvester is an integral component facilitating Intelligent Green Energy Management System (i-GEMS) and it is required to meet the technical challenges mentioned as follows: (1) capability for bi-directional mode of operation (buck/boost) (2) reduction of circuit parasitic to suppress voltage spikes (3) converter startup problem (4) high frequency magnetics (5) higher power density (6) mode transition issues during battery charging and discharging. This paper is focused to address the above mentioned issues and targeted to design, develop and implement a bi-directional energy harvester with galvanic isolation. In this work, the hardware architecture for bi-directional energy harvester rated 3.5 kW is developed with Isolated Full Bridge Boost Converter (IFBBC) as well as Dual Active Bridge (DAB) Converter configuration using modular power electronics hardware which is identical for both solar PV array and battery energy storage. In IFBBC converter, the current fed full bridge circuit is enabled and voltage fed full bridge circuit is disabled through Pulse Width Modulation (PWM) pulses for boost mode of operation and vice-versa for buck mode of operation. In DAB converter, all the switches are in active state so as to adjust the phase shift angle between primary full bridge and secondary full bridge which in turn decides the power flow directions depending on modes (boost/buck) of operation. Here, the control algorithm is developed to ensure the regulation of the common DC link voltage and maximum power extraction from the renewable energy sources depending on the selected mode (buck/boost) of operation. The circuit analysis and simulation study are conducted using PSIM 9.0 in three scenarios which are - 1.IFBBC with passive clamp, 2. IFBBC with active clamp, 3. DAB converter. In this work, a common hardware prototype for bi-directional energy harvester with 3.5 kW rating is built for IFBBC and DAB converter configurations. The power circuit is equipped with right choice of MOSFETs, gate drivers with galvanic isolation, high frequency transformer, filter capacitors, and filter boost inductor. The experiment was conducted for IFBBC converter with passive clamp under boost mode and the prototype confirmed the simulation results showing the measured efficiency as 88% at 2.5 kW output power. The digital controller hardware platform is developed using floating point microcontroller TMS320F2806x from Texas Instruments. The firmware governing the operation of the bi-directional energy harvester is written in C language and developed using code composer studio. The comprehensive analyses of the power circuit design, control strategy for battery charging/discharging under buck/boost modes and comparative performance evaluation using simulation and experimental results will be presented.

Keywords: bi-directional energy harvester, dual active bridge, isolated full bridge boost converter, intelligent green energy management system, maximum power point tracking, renewable energy sources

Procedia PDF Downloads 107
32 Biocellulose as Platform for the Development of Multifunctional Materials

Authors: Junkal Gutierrez, Hernane S. Barud, Sidney J. L. Ribeiro, Agnieszka Tercjak

Abstract:

Nowadays the interest on green nanocomposites and on the development of more environmental friendly products has been increased. Bacterial cellulose has been recently investigated as an attractive environmentally friendly material for the preparation of low-cost nanocomposites. The formation of cellulose by laboratory bacterial cultures is an interesting and attractive biomimetic access to obtain pure cellulose with excellent properties. Additionally, properties as molar mass, molar mass distribution, and the supramolecular structure could be control using different bacterial strain, culture mediums and conditions, including the incorporation of different additives. This kind of cellulose is a natural nanomaterial, and therefore, it has a high surface-to-volume ratio which is highly advantageous in composites production. Such property combined with good biocompatibility, high tensile strength, and high crystallinity makes bacterial cellulose a potential material for applications in different fields. The aim of this investigation work was the fabrication of novel hybrid inorganic-organic composites based on bacterial cellulose, cultivated in our laboratory, as a template. This kind of biohybrid nanocomposites gathers together excellent properties of bacterial cellulose with the ones displayed by typical inorganic nanoparticles like optical, magnetic and electrical properties, luminescence, ionic conductivity and selectivity, as well as chemical or biochemical activity. In addition, the functionalization of cellulose with inorganic materials opens new pathways for the fabrication of novel multifunctional hybrid materials with promising properties for a wide range of applications namely electronic paper, flexible displays, solar cells, sensors, among others. In this work, different pathways for fabrication of multifunctional biohybrid nanopapers with tunable properties based on BC modified with amphiphilic poly(ethylene oxide-b-propylene oxide-b-ethylene oxide) (EPE) block copolymer, sol-gel synthesized nanoparticles (titanium, vanadium and a mixture of both oxides) and functionalized iron oxide nanoparticles will be presented. In situ (biosynthesized) and ex situ (at post-production level) approaches were successfully used to modify BC membranes. Bacterial cellulose based biocomposites modified with different EPE block copolymer contents were developed by in situ technique. Thus, BC growth conditions were manipulated to fabricate EPE/BC nanocomposite during the biosynthesis. Additionally, hybrid inorganic/organic nanocomposites based on BC membranes and inorganic nanoparticles were designed via ex-situ method, by immersion of never-dried BC membranes into different nanoparticle solutions. On the one hand, sol-gel synthesized nanoparticles (titanium, vanadium and a mixture of both oxides) and on the other hand superparamagnetic iron oxide nanoparticles (SPION), Fe2O3-PEO solution. The morphology of designed novel bionanocomposites hybrid materials was investigated by atomic force microscopy (AFM) and scanning electron microscopy (SEM). In order to characterized obtained materials from the point of view of future applications different techniques were employed. On the one hand, optical properties were analyzed by UV-vis spectroscopy and spectrofluorimetry and on the other hand electrical properties were studied at nano and macroscale using electric force microscopy (EFM), tunneling atomic force microscopy (TUNA) and Keithley semiconductor analyzer, respectively. Magnetic properties were measured by means of magnetic force microscopy (MFM). Additionally, mechanical properties were also analyzed.

Keywords: bacterial cellulose, block copolymer, advanced characterization techniques, nanoparticles

Procedia PDF Downloads 198
31 A Quantitative Case Study Analysis of Store Format Contributors to U.S. County Obesity Prevalence in Virginia

Authors: Bailey Houghtaling, Sarah Misyak

Abstract:

Food access; the availability, affordability, convenience, and desirability of food and beverage products within communities, is influential on consumers’ purchasing and consumption decisions. These variables may contribute to lower dietary quality scores and a higher obesity prevalence documented among rural and disadvantaged populations in the United States (U.S.). Current research assessing linkages between food access and obesity outcomes has primarily focused on distance to a traditional grocery/supermarket store as a measure of optimality. However, low-income consumers especially, including U.S. Department of Agriculture’s Supplemental Nutrition Assistance Program (SNAP) participants, seem to utilize non-traditional food store formats with greater frequency for household dietary needs. Non-traditional formats have been associated with less nutritious food and beverage options and consumer purchases that are high in saturated fats, added sugars, and sodium. Authors’ formative research indicated differences by U.S. region and rurality in the distribution of traditional and non-traditional SNAP-authorized food store formats. Therefore, using Virginia as a case study, the purpose of this research was to determine if a relationship between store format, rurality, and obesity exists. This research applied SNAP-authorized food store data (food access points for SNAP as well as non-SNAP consumers) and obesity prevalence data by Virginia county using publicly available databases: (1) SNAP Retailer Locator, and; (2) U.S. County Health Rankings. The alpha level was set a priori at 0.05. All Virginia SNAP-authorized stores (n=6,461) were coded by format – grocery, drug, mass merchandiser, club, convenience, dollar, supercenter, specialty, farmers market, independent grocer, and non-food store. Simple linear regression was applied primarily to assess the relationship between store format and obesity. Thereafter, multiple variables were added to the regression to account for potential moderating relationships (e.g., county income, rurality). Convenience, dollar, non-food or restaurant, mass merchandiser, farmers market, and independent grocer formats were significantly, positively related to obesity prevalence. Upon controlling for urban-rural status and income, results indicated the following formats to be significantly related to county obesity prevalence with a small, positive effect: convenience (p=0.010), accounting for 0.3% of the variance in obesity prevalence; dollar (p=0.005; 0.5% of the variance), and; non-food (p=0.030; 1.3% of the variance) formats. These results align with current literature on consumer behavior at non-traditional formats. For example, consumers’ food and beverage purchases at convenience and dollar stores are documented to be high in saturated fats, added sugars, and sodium. Further, non-food stores (i.e., quick-serve restaurants) often contribute to a large portion of U.S. consumers’ dietary intake and thus poor dietary quality scores. Current food access research investigates grocery/supermarket access and obesity outcomes. These results suggest more research is needed that focuses on non-traditional food store formats. Nutrition interventions within convenience, dollar, and non-food stores, for example, that aim to enhance not only healthy food access but the affordability, convenience, and desirability of nutritious food and beverage options may impact obesity rates in Virginia. More research is warranted utilizing the presented investigative framework in other U.S. and global regions to explore the role and the potential of non-traditional food store formats to prevent and reduce obesity.

Keywords: food access, food store format, non-traditional food stores, obesity prevalence

Procedia PDF Downloads 111
30 The Prospects of Optimized KOH/Cellulose 'Papers' as Hierarchically Porous Electrode Materials for Supercapacitor Devices

Authors: Dina Ibrahim Abouelamaiem, Ana Jorge Sobrido, Magdalena Titirici, Paul R. Shearing, Daniel J. L. Brett

Abstract:

Global warming and scarcity of fossil fuels have had a radical impact on the world economy and ecosystem. The urgent need for alternative energy sources has hence elicited an extensive research for exploiting efficient and sustainable means of energy conversion and storage. Among various electrochemical systems, supercapacitors attracted significant attention in the last decade due to their high power supply, long cycle life compared to batteries and simple mechanism. Recently, the performance of these devices has drastically improved, as tuning of nanomaterials provided efficient charge and storage mechanisms. Carbon materials, in various forms, are believed to pioneer the next generation of supercapacitors due to their attractive properties that include high electronic conductivities, high surface areas and easy processing and functionalization. Cellulose has eco-friendly attributes that are feasible to replace man-made fibers. The carbonization of cellulose yields carbons, including activated carbon and graphite fibers. Activated carbons successively are the most exploited candidates for supercapacitor electrode materials that can be complemented with pseudocapacitive materials to achieve high energy and power densities. In this work, the optimum functionalization conditions of cellulose have been investigated for supercapacitor electrode materials. The precursor was treated with potassium hydroxide (KOH) at different KOH/cellulose ratios prior to the carbonization process in an inert nitrogen atmosphere at 850 °C. The chalky products were washed, dried and characterized with different techniques including transmission electron microscopy (TEM), x-ray tomography and nitrogen adsorption-desorption isotherms. The morphological characteristics and their effect on the electrochemical performances were investigated in two and three-electrode systems. The KOH/cellulose ratios of 0.5:1 and 1:1 exhibited the highest performances with their unique hierarchal porous network structure, high surface areas and low cell resistances. Both samples acquired the best results in three-electrode systems and coin cells with specific gravimetric capacitances as high as 187 F g-1 and 20 F g-1 at a current density of 1 A g-1 and retention rates of 72% and 70%, respectively. This is attributed to the morphology of the samples that constituted of a well-balanced micro-, meso- and macro-porosity network structure. This study reveals that the electrochemical performance doesn’t solely depend on high surface areas but also an optimum pore size distribution, specifically at low current densities. The micro- and meso-pore contribution to the final pore structure was found to dominate at low KOH loadings, reaching ‘equilibrium’ with macropores at the optimum KOH loading, after which macropores dictate the porous network. The wide range of pore sizes is detrimental for the mobility and penetration of electrolyte ions in the porous structures. These findings highlight the influence of various morphological factors on the double-layer capacitances and high performance rates. In addition, they open a platform for the investigation of the optimized conditions for double-layer capacitance that can be coupled with pseudocapacitive materials to yield higher energy densities and capacities.

Keywords: carbon, electrochemical performance, electrodes, KOH/cellulose optimized ratio, morphology, supercapacitor

Procedia PDF Downloads 196
29 Deep Learning Based on Image Decomposition for Restoration of Intrinsic Representation

Authors: Hyohun Kim, Dongwha Shin, Yeonseok Kim, Ji-Su Ahn, Kensuke Nakamura, Dongeun Choi, Byung-Woo Hong

Abstract:

Artefacts are commonly encountered in the imaging process of clinical computed tomography (CT) where the artefact refers to any systematic discrepancy between the reconstructed observation and the true attenuation coefficient of the object. It is known that CT images are inherently more prone to artefacts due to its image formation process where a large number of independent detectors are involved, and they are assumed to yield consistent measurements. There are a number of different artefact types including noise, beam hardening, scatter, pseudo-enhancement, motion, helical, ring, and metal artefacts, which cause serious difficulties in reading images. Thus, it is desired to remove nuisance factors from the degraded image leaving the fundamental intrinsic information that can provide better interpretation of the anatomical and pathological characteristics. However, it is considered as a difficult task due to the high dimensionality and variability of data to be recovered, which naturally motivates the use of machine learning techniques. We propose an image restoration algorithm based on the deep neural network framework where the denoising auto-encoders are stacked building multiple layers. The denoising auto-encoder is a variant of a classical auto-encoder that takes an input data and maps it to a hidden representation through a deterministic mapping using a non-linear activation function. The latent representation is then mapped back into a reconstruction the size of which is the same as the size of the input data. The reconstruction error can be measured by the traditional squared error assuming the residual follows a normal distribution. In addition to the designed loss function, an effective regularization scheme using residual-driven dropout determined based on the gradient at each layer. The optimal weights are computed by the classical stochastic gradient descent algorithm combined with the back-propagation algorithm. In our algorithm, we initially decompose an input image into its intrinsic representation and the nuisance factors including artefacts based on the classical Total Variation problem that can be efficiently optimized by the convex optimization algorithm such as primal-dual method. The intrinsic forms of the input images are provided to the deep denosing auto-encoders with their original forms in the training phase. In the testing phase, a given image is first decomposed into the intrinsic form and then provided to the trained network to obtain its reconstruction. We apply our algorithm to the restoration of the corrupted CT images by the artefacts. It is shown that our algorithm improves the readability and enhances the anatomical and pathological properties of the object. The quantitative evaluation is performed in terms of the PSNR, and the qualitative evaluation provides significant improvement in reading images despite degrading artefacts. The experimental results indicate the potential of our algorithm as a prior solution to the image interpretation tasks in a variety of medical imaging applications. This work was supported by the MISP(Ministry of Science and ICT), Korea, under the National Program for Excellence in SW (20170001000011001) supervised by the IITP(Institute for Information and Communications Technology Promotion).

Keywords: auto-encoder neural network, CT image artefact, deep learning, intrinsic image representation, noise reduction, total variation

Procedia PDF Downloads 169
28 Fabrication of Zeolite Modified Cu Doped ZnO Films and Their Response towards Nitrogen Monoxide

Authors: Irmak Karaduman, Tugba Corlu, Sezin Galioglu, Burcu Akata, M. Ali Yildirim, Aytunç Ateş, Selim Acar

Abstract:

Breath analysis represents a promising non-invasive, fast and cost-effective alternative to well-established diagnostic and monitoring techniques such as blood analysis, endoscopy, ultrasonic and tomographic monitoring. Portable, non-invasive, and low-cost breath analysis devices are becoming increasingly desirable for monitoring different diseases, especially asthma. Beacuse of this, NO gas sensing at low concentrations has attracted progressive attention for clinical analysis in asthma. Recently, nanomaterials based sensors are considered to be a promising clinical and laboratory diagnostic tool, because its large surface–to–volume ratio, controllable structure, easily tailored chemical and physical properties, which bring high sensitivity, fast dynamic processand even the increasing specificity. Among various nanomaterials, semiconducting metal oxides are extensively studied gas-sensing materials and are potential sensing elements for breathanalyzer due to their high sensitivity, simple design, low cost and good stability.The sensitivities of metal oxide semiconductor gas sensors can be enhanced by adding noble metals. Doping contents, distribution, and size of metallic or metal oxide catalysts are key parameters for enhancing gas selectivity as well as sensitivity. By manufacturing doping MOS structures, it is possible to develop more efficient sensor sensing layers. Zeolites are perhaps the most widely employed group of silicon-based nanoporous solids. Their well-defined pores of sub nanometric size have earned them the name of molecular sieves, meaning that operation in the size exclusion regime is possible by selecting, among over 170 structures available, the zeolite whose pores allow the pass of the desired molecule, while keeping larger molecules outside.In fact it is selective adsorption, rather than molecular sieving, the mechanism that explains most of the successful gas separations achieved with zeolite membranes. In view of their molecular sieving and selective adsorption properties, it is not surprising that zeolites have found use in a number of works dealing with gas sensing devices. In this study, the Cu doped ZnO nanostructure film was produced by SILAR method and investigated the NO gas sensing properties. To obtain the selectivity of the sample, the gases including CO,NH3,H2 and CH4 were detected to compare with NO. The maximum response is obtained at 85 C for 20 ppb NO gas. The sensor shows high response to NO gas. However, acceptable responses are calculated for CO and NH3 gases. Therefore, there are no responses obtain for H2 and CH4 gases. Enhanced to selectivity, Cu doped ZnO nanostructure film was coated with zeolite A thin film. It is found that the sample possess an acceptable response towards NO hardly respond to CO, NH3, H2 and CH4 at room temperature. This difference in the response can be expressed in terms of differences in the molecular structure, the dipole moment, strength of the electrostatic interaction and the dielectric constant. The as-synthesized thin film is considered to be one of the extremely promising candidate materials in electronic nose applications. This work is supported by The Scientific and Technological Research Council of Turkey (TUBİTAK) under Project No, 115M658 and Gazi University Scientific Research Fund under project no 05/2016-21.

Keywords: Cu doped ZnO, electrical characterization, gas sensing, zeolite

Procedia PDF Downloads 260
27 Reflection of Landscape Agrogenization in the Soil Cover Structure and Profile Morphology: Example of Lithuania Agroecosystem

Authors: Jonas Volungevicius, Kristina Amaleviciute, Rimantas Vaisvalavicius, Alvyra Slepetiene, Darijus Veteikis

Abstract:

Lithuanian territory is characterized by landscape with prevailing morain hills and clayey lowlands. The largest part of it has endured agrogenization of various degrees which was the cause of changes both in the structure of landscape and soil cover, transformations of soil profile and degradation of natural background soils. These changes influence negatively geoecological potential of landscape and soil and contribute to the weakening of the sustainability of agroecosystems. Research objective: to reveal the landscape agrogenization induced alterations of catenae and their appendant soil profiles in Lithuanian morain hills and clayey lowlands. Methods: Soil cover analysis and catenae charting was conducted using landscape profiling; soil morphology detected and soil type identified following WRB 2014. Granulometric composition of soil profiles was obtained by laser diffraction method (lazer diffractometer Mastersizer 2000). pH was measured in H2O extraction using potentiometric titration; SOC was determined by the Tyurin method modified by Nikitin, measuring with spectrometer Cary 50 (VARIAN) in 590 nm wavelength using glucose standards. Results: analysis showed that the decrease of forest vegetation and the other natural landscape components following the agrogenization of the research area influenced differently but significantly the structural alterations in soil cover and vertical soil profile. The research detected that due to landscape agrogenization, the suppression of zone-specific processes and the intensification of inter-zone processes determined by agrogenic factors take place in Lithuanian agroecosystems. In forested hills historically prevailing Retisols and Histosols territorial complex is transforming into the territorial complex of Regosols, Deluvial soils and drained Histosols. Processes taking place are simplification of vertical profile structure, intensive rejuvenation of profile, disappearance of the features of zone-specific soil-forming processes (podzolization, lessivage, gley formation). Erosion and deluvial processes manifest more intensively and weakly accumulating organic material more intensively spread in a vertical soil profile. The territorial soil complex of Gleyic Luvisols and Gleysols dominating in forested clayey lowlands subjected to agrogenization is transformed into the catena of drained Luvisols and pseudo Cambisols. The best expressed are their changes in moisture regime (morphological features of gley and stagnic properties are on decline) together with alterations of pH and distribution and intensity of accumulation of organic matter in profile. A specific horizon, antraquic, uncharacteristic to natural soil formation is appearing. Important to note that due to deep ploughing and other agrotechnical measures, the natural vertical differentiation of clay particles in a soil profile is destroyed which leads not only to alterations of physical qualities of soil, but also encumbers the identification of Luvisols by creating presumptions to misidentify them as Cambisols. The latter have never developed in these ecosystems under the present climatic conditions. Acknowledgements: This work was supported by the National Science Program: The effect of long-term, different-intensity management of resources on the soils of different genesis and on other components of the agro-ecosystems [grant number SIT-9/2015] funded by the Research Council of Lithuania.

Keywords: agroecosystems, landscape agrogenization, luvisols, retisols, transformation of soil profile

Procedia PDF Downloads 234
26 Image Segmentation with Deep Learning of Prostate Cancer Bone Metastases on Computed Tomography

Authors: Joseph M. Rich, Vinay A. Duddalwar, Assad A. Oberai

Abstract:

Prostate adenocarcinoma is the most common cancer in males, with osseous metastases as the commonest site of metastatic prostate carcinoma (mPC). Treatment monitoring is based on the evaluation and characterization of lesions on multiple imaging studies, including Computed Tomography (CT). Monitoring of the osseous disease burden, including follow-up of lesions and identification and characterization of new lesions, is a laborious task for radiologists. Deep learning algorithms are increasingly used to perform tasks such as identification and segmentation for osseous metastatic disease and provide accurate information regarding metastatic burden. Here, nnUNet was used to produce a model which can segment CT scan images of prostate adenocarcinoma vertebral bone metastatic lesions. nnUNet is an open-source Python package that adds optimizations to deep learning-based UNet architecture but has not been extensively combined with transfer learning techniques due to the absence of a readily available functionality of this method. The IRB-approved study data set includes imaging studies from patients with mPC who were enrolled in clinical trials at the University of Southern California (USC) Health Science Campus and Los Angeles County (LAC)/USC medical center. Manual segmentation of metastatic lesions was completed by an expert radiologist Dr. Vinay Duddalwar (20+ years in radiology and oncologic imaging), to serve as ground truths for the automated segmentation. Despite nnUNet’s success on some medical segmentation tasks, it only produced an average Dice Similarity Coefficient (DSC) of 0.31 on the USC dataset. DSC results fell in a bimodal distribution, with most scores falling either over 0.66 (reasonably accurate) or at 0 (no lesion detected). Applying more aggressive data augmentation techniques dropped the DSC to 0.15, and reducing the number of epochs reduced the DSC to below 0.1. Datasets have been identified for transfer learning, which involve balancing between size and similarity of the dataset. Identified datasets include the Pancreas data from the Medical Segmentation Decathlon, Pelvic Reference Data, and CT volumes with multiple organ segmentations (CT-ORG). Some of the challenges of producing an accurate model from the USC dataset include small dataset size (115 images), 2D data (as nnUNet generally performs better on 3D data), and the limited amount of public data capturing annotated CT images of bone lesions. Optimizations and improvements will be made by applying transfer learning and generative methods, including incorporating generative adversarial networks and diffusion models in order to augment the dataset. Performance with different libraries, including MONAI and custom architectures with Pytorch, will be compared. In the future, molecular correlations will be tracked with radiologic features for the purpose of multimodal composite biomarker identification. Once validated, these models will be incorporated into evaluation workflows to optimize radiologist evaluation. Our work demonstrates the challenges of applying automated image segmentation to small medical datasets and lays a foundation for techniques to improve performance. As machine learning models become increasingly incorporated into the workflow of radiologists, these findings will help improve the speed and accuracy of vertebral metastatic lesions detection.

Keywords: deep learning, image segmentation, medicine, nnUNet, prostate carcinoma, radiomics

Procedia PDF Downloads 69
25 Environmental Fate and Toxicity of Aged Titanium Dioxide Nano-Composites Used in Sunscreen

Authors: Danielle Slomberg, Jerome Labille, Riccardo Catalano, Jean-Claude Hubaud, Alexandra Lopes, Alice Tagliati, Teresa Fernandes

Abstract:

In the assessment and management of cosmetics and personal care products, sunscreens are of emerging concern regarding both human and environmental health. Organic UV blockers in many sunscreens have been evidenced to undergo rapid photodegradation, induce dermal allergic reactions due to skin penetration, and to cause adverse effects on marine systems. While mineral UV-blockers may offer a safer alternative, their fate and impact and resulting regulation are still under consideration, largely related to the potential influence of nanotechnology-based products on both consumers and the environment. Nanometric titanium dioxide (TiO₂) UV-blockers have many advantages in terms of sun protection and asthetics (i.e., transparency). These UV-blockers typically consist of rutile nanoparticles coated with a primary mineral layer (silica or alumina) aimed at blocking the nanomaterial photoactivity and can include a secondary organic coating (e.g., stearic acid, methicone) aimed at favouring dispersion of the nanomaterial in the sunscreen formulation. The nanomaterials contained in the sunscreen can leave the skin either through a bathing of everyday usage, with subsequent release into rivers, lakes, seashores, and/or sewage treatment plants. The nanomaterial behaviour, fate and impact in these different systems is largely determined by its surface properties, (e.g. the nanomaterial coating type) and lifetime. The present work aims to develop the eco-design of sunscreens through the minimisation of risks associated with nanomaterials incorporated into the formulation. All stages of the sunscreen’s life cycle must be considered in this aspect, from its manufacture to its end-of-life, through its use by the consumer to its impact on the exposed environment. Reducing the potential release and/or toxicity of the nanomaterial from the sunscreen is a decisive criterion for its eco-design. TiO₂ UV-blockers of varied size and surface coating (e.g., stearic acid and silica) have been selected for this study. Hydrophobic TiO₂ UV-blockers (i.e., stearic acid-coated) were incorporated into a typical water-in-oil (w/o) formulation while hydrophilic, silica-coated TiO₂ UV-blockers were dispersed into an oil-in-water (o/w) formulation. The resulting sunscreens were characterised in terms of nanomaterial localisation, sun protection factor, and photo-passivation. The risk to the direct aquatic environment was assessed by evaluating the release of nanomaterials from the sunscreen through a simulated laboratory aging procedure. The size distribution, surface charge, and degradation state of the nano-composite by-products, as well as their nanomaterial concentration and colloidal behaviour were determined in a variety of aqueous environments (e.g., seawater and freshwater). Release of the hydrophobic nanocomposites into the aqueous environment was driven by oil droplet formation while hydrophilic nano-composites were readily dispersed. Ecotoxicity of the sunscreen by-products (from both w/o and o/w formulations) and their risk to marine organisms were assessed using coral symbiotes and tropical corals, evaluating both lethal and sublethal toxicities. The data dissemination and provided risk knowledge from the present work will help guide regulation related to nanomaterials in sunscreen, provide better information for consumers, and allow for easier decision-making for manufacturers.

Keywords: alteration, environmental fate, sunscreens, titanium dioxide nanoparticles

Procedia PDF Downloads 238
24 A Comparative Evaluation of Cognitive Load Management: Case Study of Postgraduate Business Students

Authors: Kavita Goel, Donald Winchester

Abstract:

In a world of information overload and work complexities, academics often struggle to create an online instructional environment enabling efficient and effective student learning. Research has established that students’ learning styles are different, some learn faster when taught using audio and visual methods. Attributes like prior knowledge and mental effort affect their learning. ‘Cognitive load theory’, opines learners have limited processing capacity. Cognitive load depends on the learner’s prior knowledge, the complexity of content and tasks, and instructional environment. Hence, the proper allocation of cognitive resources is critical for students’ learning. Consequently, a lecturer needs to understand the limits and strengths of the human learning processes, various learning styles of students, and accommodate these requirements while designing online assessments. As acknowledged in the cognitive load theory literature, visual and auditory explanations of worked examples potentially lead to a reduction of cognitive load (effort) and increased facilitation of learning when compared to conventional sequential text problem solving. This will help learner to utilize both subcomponents of their working memory. Instructional design changes were introduced at the case site for the delivery of the postgraduate business subjects. To make effective use of auditory and visual modalities, video recorded lectures, and key concept webinars were delivered to students. Videos were prepared to free up student limited working memory from irrelevant mental effort as all elements in a visual screening can be viewed simultaneously, processed quickly, and facilitates greater psychological processing efficiency. Most case study students in the postgraduate programs are adults, working full-time at higher management levels, and studying part-time. Their learning style and needs are different from other tertiary students. The purpose of the audio and visual interventions was to lower the students cognitive load and provide an online environment supportive to their efficient learning. These changes were expected to impact the student’s learning experience, their academic performance and retention favourably. This paper posits that these changes to instruction design facilitates students to integrate new knowledge into their long-term memory. A mixed methods case study methodology was used in this investigation. Primary data were collected from interviews and survey(s) of students and academics. Secondary data were collected from the organisation’s databases and reports. Some evidence was found that the academic performance of students does improve when new instructional design changes are introduced although not statistically significant. However, the overall grade distribution of student’s academic performance has changed and skewed higher which shows deeper understanding of the content. It was identified from feedback received from students that recorded webinars served as better learning aids than material with text alone, especially with more complex content. The recorded webinars on the subject content and assessments provides flexibility to students to access this material any time from repositories, many times, and this enhances students learning style. Visual and audio information enters student’s working memory more effectively. Also as each assessment included the application of the concepts, conceptual knowledge interacted with the pre-existing schema in the long-term memory and lowered student’s cognitive load.

Keywords: cognitive load theory, learning style, instructional environment, working memory

Procedia PDF Downloads 119
23 Evolution of Plio/Pleistocene Sedimentary Processes in Patraikos Gulf, Offshore Western Greece

Authors: E. K. Tripsanas, D. Spanos, I. Oikonomopoulos, K. Stathopoulou, A. S. Abdelsamad, A. Pagoulatos

Abstract:

Patraikos Gulf is located offshore western Greece, and it is limited to the west by the Zante, Cephalonia, and Lefkas islands. The Plio/Pleistocene sequence is characterized by two depocenters, the east and west Patraikos basins separated from each other by a prominent sill. This study is based on the Plio/Pleistocene seismic stratigraphy analysis of a newly acquired 3D PSDM (Pre-Stack depth migration) seismic survey in the west Patraikos Basin and few 2D seismic profiles throughout the entire Patraikos Gulf. The eastern Patraikos Basin, although completely buried today with water depths less than 100 m, it was a deep basin during Pliocene ( > 2 km of Pliocene-Pleistocene sediments) and appears to have gathered most of Achelous River discharges. The west Patraikos Gulf was shallower ( < 1300 m of Pliocene-Pleistocene sediments) and characterized by a hummocky relief due to thrust-belt tectonics and Miocene to Pleistocene halokinetic processes. The transition from Pliocene to Miocene is expressed by a widespread erosional unconformity with evidence of fluvial drainage patterns. This indicates that west Patraikos Basin was aerially exposed during the Messinian Salinity Crisis. Continuous to semi-continuous, parallel reflections in the lower, early- to mid-Pliocene seismic packet provides evidence that the re-connection of the Mediterranean Sea with the Atlantic Ocean during Zanclean resulted in the flooding of the west Patraikos basin and the domination of hemipelagic sedimentation interrupted by occasional gravity flows. This is evident in amplitude and semblance horizon slices, which clearly show the presence of long-running, meandering submarine channels sourced from the southeast (northwest Peloponnese) and north. The long-running nature of the submarine channels suggests mobile efficient turbidity currents, probably due to the participation of a sufficient amount of clay minerals in their suspended load. The upper seismic section in the study area mainly consists of several successions of clinoforms, interpreted as progradational delta complexes of Achelous River. This sudden change from marine to shallow marine sedimentary processes is attributed to climatic changes and eustatic perturbations since late Pliocene onwards (~ 2.6 Ma) and/or a switch of Achelous River from the east Patraikos Basin to the west Patraikos Basin. The deltaic seismic unit consists of four delta complexes. The first two complexes result in the infill of topographic depressions and smoothing of an initial hummocky bathymetry. The distribution of the upper two delta complexes is controlled by compensational stacking. Amplitude and semblance horizon slices depict the development of several almost straight and short (a few km long) distributary submarine channels at the delta slopes and proximal prodeltaic plains with lobate sand-sheet deposits at their mouths. Such channels are interpreted to result from low-efficiency turbidity currents with low content in clay minerals. Such a differentiation in the nature of the gravity flows is attributed to the switch of the sediment supply from clay-rich sediments derived from the draining of flysch formations of the Ionian and Gavrovo zones, to the draining of poor in clay minerals carbonate formations of Gavrovo zone through the Achelous River.

Keywords: sequence stratigraphy, basin analysis, river deltas, submarine channels

Procedia PDF Downloads 300
22 Fold and Thrust Belts Seismic Imaging and Interpretation

Authors: Sunjay

Abstract:

Plate tectonics is of very great significance as it represents the spatial relationships of volcanic rock suites at plate margins, the distribution in space and time of the conditions of different metamorphic facies, the scheme of deformation in mountain belts, or orogens, and the association of different types of economic deposit. Orogenic belts are characterized by extensive thrust faulting, movements along large strike-slip fault zones, and extensional deformation that occur deep within continental interiors. Within oceanic areas there also are regions of crustal extension and accretion in the backarc basins that are located on the landward sides of many destructive plate margins.Collisional orogens develop where a continent or island arc collides with a continental margin as a result of subduction. collisional and noncollisional orogens can be explained by differences in the strength and rheology of the continental lithosphere and by processes that influence these properties during orogenesis.Seismic Imaging Difficulties-In triangle zones, several factors reduce the effectiveness of seismic methods. The topography in the central part of the triangle zone is usually rugged and is associated with near-surface velocity inversions which degrade the quality of the seismic image. These characteristics lead to low signal-to-noise ratio, inadequate penetration of energy through overburden, poor geophone coupling with the surface and wave scattering. Depth Seismic Imaging Techniques-Seismic processing relates to the process of altering the seismic data to suppress noise, enhancing the desired signal (higher signal-to-noise ratio) and migrating seismic events to their appropriate location in space and depth. Processing steps generally include analysis of velocities, static corrections, moveout corrections, stacking and migration. Exploration seismology Bow-tie effect -Shadow Zones-areas with no reflections (dead areas). These are called shadow zones and are common in the vicinity of faults and other discontinuous areas in the subsurface. Shadow zones result when energy from a reflector is focused on receivers that produce other traces. As a result, reflectors are not shown in their true positions. Subsurface Discontinuities-Diffractions occur at discontinuities in the subsurface such as faults and velocity discontinuities (as at “bright spot” terminations). Bow-tie effect caused by the two deep-seated synclines. Seismic imaging of thrust faults and structural damage-deepwater thrust belts, Imaging deformation in submarine thrust belts using seismic attributes,Imaging thrust and fault zones using 3D seismic image processing techniques, Balanced structural cross sections seismic interpretation pitfalls checking, The seismic pitfalls can originate due to any or all of the limitations of data acquisition, processing, interpretation of the subsurface geology,Pitfalls and limitations in seismic attribute interpretation of tectonic features, Seismic attributes are routinely used to accelerate and quantify the interpretation of tectonic features in 3D seismic data. Coherence (or variance) cubes delineate the edges of megablocks and faulted strata, curvature delineates folds and flexures, while spectral components delineate lateral changes in thickness and lithology. Carbon capture and geological storage leakage surveillance because fault behave as a seal or a conduit for hydrocarbon transportation to a trap,etc.

Keywords: tectonics, seismic imaging, fold and thrust belts, seismic interpretation

Procedia PDF Downloads 43
21 Addressing Educational Injustice through Collective Teacher Professional Development

Authors: Wenfan Yan, Yumei Han

Abstract:

Objectives: Educational inequality persists between China's ethnic minority regions and the mainland. The key to rectifying this disparity lies in enhancing the quality of educators. This paper delves into the Chinese government's innovative policy, "Group Educators Supporting Tibet" (GEST), designed to bridge the shortage of high-quality teachers in Tibet, a representative underprivileged ethnic minority area. GEST aims to foster collective action by networking provincial expert educators with Tibetan counterparts and collaborating between supporting provincial educational entities and Tibetan education entities. Theoretical Framework: The unequal distribution of social capital contributes significantly to the educational gap between ethnic minority areas and other regions in China. Within the framework of social network theory, motivated GEST educators take action to foster resources and relationships. This study captures grassroots perspectives to outline how social networking contributes to the policy objective of enhancing Tibetan teachers' quality and eradicating educational injustice. Methodology: A sequential mixed-methods approach was adopted to scrutinize policy impacts from the vantage point of social networking. Quantitative research involved surveys for GEST and Tibetan teachers, exploring demographics, perceptions of policy significance, motivations, actions, and networking habits. Qualitative research included focus group interviews with GEST educators, local teachers, and students from program schools. The findings were meticulously analyzed to provide comprehensive insights into stakeholders' experiences and the impacts of the GEST policy. Key Findings: The policy empowers individuals to impact Tibetan education significantly. Motivated GEST educators with prior educational support experiences contribute to its success. Supported by a collective -school, city, province, and government- the new social structure fosters higher efficiency. GEST's approach surpasses conventional methods. The individual, backed by educators, realizes the potential of transformative class design. Collective activities -pedagogy research, teaching, mentoring, training, and partnerships- equip Tibetan teachers, enhancing educational quality and equity. This collaborative effort establishes a robust foundation for the policy's success, emphasizing the collective impact on Tibetan education. Contributions: This study contributes to international policy studies focused on educational equity through collective teacher action. Using a mixed-methods approach and guided by social networking theory, it accentuates stakeholders' perspectives, elucidating the genuine impacts of the GEST policy. The study underscores the advancement of social networking, the reinforcement of local teacher quality, and the transformative potential of cultivating a more equitable and adept teaching workforce in Tibet. Limitations of the Study and Suggestions for Future Research Directions: While the study emphasizes the positive impacts of motivated GEST educators, there might be aspects or challenges not fully explored. A more comprehensive understanding of potential drawbacks or obstacles would provide a more balanced view. For future studies, investigating the long-term impact of the GEST policy on educational quality could provide insights into the sustainability of the improvements observed. Also, understanding the perspectives of Tibetan teachers who may not have directly benefited from GEST could reveal potential disparities in policy implementation.

Keywords: teacher development, social networking, teacher quality, mixed research method

Procedia PDF Downloads 33
20 Analysis of Composite Health Risk Indicators Built at a Regional Scale and Fine Resolution to Detect Hotspot Areas

Authors: Julien Caudeville, Muriel Ismert

Abstract:

Analyzing the relationship between environment and health has become a major preoccupation for public health as evidenced by the emergence of the French national plans for health and environment. These plans have identified the following two priorities: (1) to identify and manage geographic areas, where hotspot exposures are suspected to generate a potential hazard to human health; (2) to reduce exposure inequalities. At a regional scale and fine resolution of exposure outcome prerequisite, environmental monitoring networks are not sufficient to characterize the multidimensionality of the exposure concept. In an attempt to increase representativeness of spatial exposure assessment approaches, risk composite indicators could be built using additional available databases and theoretical framework approaches to combine factor risks. To achieve those objectives, combining data process and transfer modeling with a spatial approach is a fundamental prerequisite that implies the need to first overcome different scientific limitations: to define interest variables and indicators that could be built to associate and describe the global source-effect chain; to link and process data from different sources and different spatial supports; to develop adapted methods in order to improve spatial data representativeness and resolution. A GIS-based modeling platform for quantifying human exposure to chemical substances (PLAINE: environmental inequalities analysis platform) was used to build health risk indicators within the Lorraine region (France). Those indicators combined chemical substances (in soil, air and water) and noise risk factors. Tools have been developed using modeling, spatial analysis and geostatistic methods to build and discretize interest variables from different supports and resolutions on a 1 km2 regular grid within the Lorraine region. By example, surface soil concentrations have been estimated by developing a Kriging method able to integrate surface and point spatial supports. Then, an exposure model developed by INERIS was used to assess the transfer from soil to individual exposure through ingestion pathways. We used distance from polluted soil site to build a proxy for contaminated site. Air indicator combined modeled concentrations and estimated emissions to take in account 30 polluants in the analysis. For water, drinking water concentrations were compared to drinking water standards to build a score spatialized using a distribution unit serve map. The Lden (day-evening-night) indicator was used to map noise around road infrastructures. Aggregation of the different factor risks was made using different methodologies to discuss weighting and aggregation procedures impact on the effectiveness of risk maps to take decisions for safeguarding citizen health. Results permit to identify pollutant sources, determinants of exposure, and potential hotspots areas. A diagnostic tool was developed for stakeholders to visualize and analyze the composite indicators in an operational and accurate manner. The designed support system will be used in many applications and contexts: (1) mapping environmental disparities throughout the Lorraine region; (2) identifying vulnerable population and determinants of exposure to set priorities and target for pollution prevention, regulation and remediation; (3) providing exposure database to quantify relationships between environmental indicators and cancer mortality data provided by French Regional Health Observatories.

Keywords: health risk, environment, composite indicator, hotspot areas

Procedia PDF Downloads 224
19 EcoTeka, an Open-Source Software for Urban Ecosystem Restoration through Technology

Authors: Manon Frédout, Laëtitia Bucari, Mathias Aloui, Gaëtan Duhamel, Olivier Rovellotti, Javier Blanco

Abstract:

Ecosystems must be resilient to ensure cleaner air, better water and soil quality, and thus healthier citizens. Technology can be an excellent tool to support urban ecosystem restoration projects, especially when based on Open Source and promoting Open Data. This is the goal of the ecoTeka application: one single digital tool for tree management which allows decision-makers to improve their urban forestry practices, enabling more responsible urban planning and climate change adaptation. EcoTeka provides city councils with three main functionalities tackling three of their challenges: easier biodiversity inventories, better green space management, and more efficient planning. To answer the cities’ need for reliable tree inventories, the application has been first built with open data coming from the websites OpenStreetMap and OpenTrees, but it will also include very soon the possibility of creating new data. To achieve this, a multi-source algorithm will be elaborated, based on existing artificial intelligence Deep Forest, integrating open-source satellite images, 3D representations from LiDAR, and street views from Mapillary. This data processing will permit identifying individual trees' position, height, crown diameter, and taxonomic genus. To support urban forestry management, ecoTeka offers a dashboard for monitoring the city’s tree inventory and trigger alerts to inform about upcoming due interventions. This tool was co-constructed with the green space departments of the French cities of Alès, Marseille, and Rouen. The third functionality of the application is a decision-making tool for urban planning, promoting biodiversity and landscape connectivity metrics to drive ecosystem restoration roadmap. Based on landscape graph theory, we are currently experimenting with new methodological approaches to scale down regional ecological connectivity principles to local biodiversity conservation and urban planning policies. This methodological framework will couple graph theoretic approach and biological data, mainly biodiversity occurrences (presence/absence) data available on both international (e.g., GBIF), national (e.g., Système d’Information Nature et Paysage) and local (e.g., Atlas de la Biodiversté Communale) biodiversity data sharing platforms in order to help reasoning new decisions for ecological networks conservation and restoration in urban areas. An experiment on this subject is currently ongoing with Montpellier Mediterranee Metropole. These projects and studies have shown that only 26% of tree inventory data is currently geo-localized in France - the rest is still being done on paper or Excel sheets. It seems that technology is not yet used enough to enrich the knowledge city councils have about biodiversity in their city and that existing biodiversity open data (e.g., occurrences, telemetry, or genetic data), species distribution models, landscape graph connectivity metrics are still underexploited to make rational decisions for landscape and urban planning projects. This is the goal of ecoTeka: to support easier inventories of urban biodiversity and better management of urban spaces through rational planning and decisions relying on open databases. Future studies and projects will focus on the development of tools for reducing the artificialization of soils, selecting plant species adapted to climate change, and highlighting the need for ecosystem and biodiversity services in cities.

Keywords: digital software, ecological design of urban landscapes, sustainable urban development, urban ecological corridor, urban forestry, urban planning

Procedia PDF Downloads 40