Search results for: adaptive thermal comfort model
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 20436

Search results for: adaptive thermal comfort model

9936 Forced-Choice Measurement Models of Behavioural, Social, and Emotional Skills: Theory, Research, and Development

Authors: Richard Roberts, Anna Kravtcova

Abstract:

Introduction: The realisation that personality can change over the course of a lifetime has led to a new companion model to the Big Five, the behavioural, emotional, and social skills approach (BESSA). BESSA hypothesizes that this set of skills represents how the individual is thinking, feeling, and behaving when the situation calls for it, as opposed to traits, which represent how someone tends to think, feel, and behave averaged across situations. The five major skill domains share parallels with the Big Five Factor (BFF) model creativity and innovation (openness), self-management (conscientiousness), social engagement (extraversion), cooperation (agreeableness), and emotional resilience (emotional stability) skills. We point to noteworthy limitations in the current operationalisation of BESSA skills (i.e., via Likert-type items) and offer up a different measurement approach: forced choice. Method: In this forced-choice paradigm, individuals were given three skill items (e.g., managing my time) and asked to select one response they believed they were “worst at” and “best at”. The Thurstonian IRT models allow these to be placed on a normative scale. Two multivariate studies (N = 1178) were conducted with a 22-item forced-choice version of the BESSA, a published measure of the BFF, and various criteria. Findings: Confirmatory factor analysis of the forced-choice assessment showed acceptable model fit (RMSEA<0.06), while reliability estimates were reasonable (around 0.70 for each construct). Convergent validity evidence was as predicted (correlations between 0.40 and 0.60 for corresponding BFF and BESSA constructs). Notable was the extent the forced-choice BESSA assessment improved upon test-criterion relationships over and above the BFF. For example, typical regression models find BFF personality accounting for 25% of the variance in life satisfaction scores; both studies showed incremental gains over the BFF exceeding 6% (i.e., BFF and BESSA together accounted for over 31% of the variance in both studies). Discussion: Forced-choice measurement models offer up the promise of creating equated test forms that may unequivocally measure skill gains and are less prone to fakability and reference bias effects. Implications for practitioners are discussed, especially those interested in selection, succession planning, and training and development. We also discuss how the forced choice method can be applied to other constructs like emotional immunity, cross-cultural competence, and self-estimates of cognitive ability.

Keywords: Big Five, forced-choice method, BFF, methods of measurements

Procedia PDF Downloads 92
9935 Ultradrawing and Ultimate Tensile Properties of Ultrahigh Molecular Weight Polyethylene Composite Fibers Filled with Activated Nanocarbon Particles with Varying Specific Surface Areas

Authors: Wang-Xi Fan, Yi Ding, Zhong-Dan Tu, Kuo-Shien Huang, Chao-Ming Huang, Jen-Taut Yeh

Abstract:

Original and/or functionalized activated nanocarbon particles with a quoted specific surface area of 100, 500, 1000 and 1400 m2/g, respectively, were used to investigate the influence of specific surface areas of activated nanocarbon on ultra drawing and ultimate tensile properties of ultrahigh molecular weight polyethylene (UHMWPE), UHMWPE/activated nanocarbon and UHMWPE/ functionalized activated nanocarbon fibers. The specific surface areas of well dispersed functionalized activated nanocarbon in UHMWPE/functionalized activated nanocarbon fibers can positively affect their ultra drawing, orientation, ultimate tensile properties and “micro-fibril” characteristics. Excellent orientation and ultimate tensile properties of UHMWPE/nanofiller fibers can be prepared by ultra drawing the UHMWPE/functionalized activated nanocarbon as-prepared fibers with optimal contents and compositions of functionalized activated nanocarbon. The ultimate tensile strength value of the best prepared UHMWPE/functionalized activated nanocarbon drawn fiber reached 8.0 GPa, which was about 2.86 times of that of the best-prepared UHMWPE drawn fiber prepared in this study. Specific surface area, morphological and Fourier transform infrared analyses of original and functionalized activated nanocarbon and/or investigations of thermal, orientation factor and ultimate tensile properties of as-prepared and/or drawn UHMWPE/functionalized activated nanocarbon fibers were performed to understand the above-improved ultra drawing and ultimate tensile properties of the UHMWPE/functionalized activated nanocarbon fibers.

Keywords: activated nanocarbon, specific surface areas, ultradrawing, ultrahigh molecular weight polyethylene

Procedia PDF Downloads 363
9934 Conceptualizing Personalized Learning: Review of Literature 2007-2017

Authors: Ruthanne Tobin

Abstract:

As our data-driven, cloud-based, knowledge-centric lives become ever more global, mobile, and digital, educational systems everywhere are struggling to keep pace. Schools need to prepare students to become critical-thinking, tech-savvy, life-long learners who are engaged and adaptable enough to find their unique calling in a post-industrial world of work. Recognizing that no nation can afford poor achievement or high dropout rates without jeopardizing its social and economic future, the thirty-two nations of the OECD are launching initiatives to redesign schools, generally under the banner of Personalized Learning or 21st Century Learning. Their intention is to transform education by situating students as co-enquirers and co-contributors with their teachers of what, when, and how learning happens for each individual. In this focused review of the 2007-2017 literature on personalized learning, the author sought answers to two main questions: “What are the theoretical frameworks that guide personalized learning?” and “What is the conceptual understanding of the model?” Ultimately, the review reveals that, although the research area is overly theorized and under-substantiated, it does provide a significant body of knowledge about this potentially transformative educational restructuring. For example, it addresses the following questions: a) What components comprise a PL model? b) How are teachers facilitating agency (voice & choice) in their students? c) What kinds of systems, processes and procedures are being used to guide the innovation? d) How is learning organized, monitored and assessed? e) What role do inquiry based models play? f) How do teachers integrate the three types of knowledge: Content, pedagogical and technological? g) Which kinds of forces enable, and which impede, personalizing learning? h) What is the nature of the collaboration among teachers? i) How do teachers co-regulate differentiated tasks? One finding of the review shows that while technology can dramatically expand access to information, expectations of its impact on teaching and learning are often disappointing unless the technologies are paired with excellent pedagogies in order to address students’ needs, interests and aspirations. This literature review fills a significant gap in this emerging field of research, as it serves to increase conceptual clarity that has hampered both the theorizing and the classroom implementation of a personalized learning model.

Keywords: curriculum change, educational innovation, personalized learning, school reform

Procedia PDF Downloads 217
9933 Characterization and Degradation of 3D Printed Polycaprolactone-Freeze Dried Bone Matrix Constructs for Use in Critical Sized Bone Defects

Authors: Samantha Meyr, Eman Mirdamadi, Martha Wang, Tao Lowe, Ryan Smith, Quinn Burke

Abstract:

Critical-sized bone defects (CSD) treatment options remain a major clinical orthopedic challenge. They are uniquely contoured diseased or damaged bones and can be defined as those that will not heal spontaneously and require surgical intervention. Autografts are the current gold standard CSD treatment, which are histocompatible and provoke a minimal immunogenic response; however, they can cause donor site morbidity and will not suffice for the size required for replacement. As an alternative to traditional surgical methods, bone tissue engineering will be implemented via 3D printing methods. A freeze-dried bone matrix (FDBM) is a type of graft material available but will only function as desired when in the presence of bone growth factors. Polycaprolactone (PCL) is a known biodegradable material with good biocompatibility that has been proven manageable in 3D printing as a medical device. A 3D-extrusion printing strategy is introduced to print these materials into scaffolds for bone grafting purposes, which could be more accessible and rapid than the current standard. Mechanical, thermal, cytotoxic, and physical properties were investigated throughout a degradation period of 6 months using fibroblasts and dental pulp stem cells. PCL-FDBM scaffolds were successfully printed with high print fidelity in their respective pore sizes and allograft content. Additionally, we have created a method for evaluating PCL using differential scanning calorimetry (DSC) and have evaluated PCL degradation over roughly 6 months.

Keywords: 3D printing, bone tissue engineering, cytotoxicity, degradation, scaffolds

Procedia PDF Downloads 98
9932 Study of Superconducting Patch Printed on Electric-Magnetic Substrates Materials

Authors: Fortaki Tarek, S. Bedra

Abstract:

In this paper, the effects of both uniaxial anisotropy in the substrate and high Tc superconducting patch on the resonant frequency, half-power bandwidth, and radiation patterns are investigated using an electric field integral equation and the spectral domain Green’s function. The analysis has been based on a full electromagnetic wave model with London’s equations and the Gorter-Casimir two-fluid model has been improved to investigate the resonant and radiation characteristics of high Tc superconducting rectangular microstrip patch in the case where the patch is printed on electric-magnetic uniaxially anisotropic substrate materials. The stationary phase technique has been used for computing the radiation electric field. The obtained results demonstrate a considerable improvement in the half-power bandwidth, of the rectangular microstrip patch, by using a superconductor patch instead of a perfect conductor one. Further results show that high Tc superconducting rectangular microstrip patch on the uniaxial substrate with properly selected electric and magnetic anisotropy ratios is more advantageous than the one on the isotropic substrate by exhibiting wider bandwidth and radiation characteristic. This behavior agrees with that discovered experimentally for superconducting patches on isotropic substrates. The calculated results have been compared with measured one available in the literature and excellent agreement has been found.

Keywords: high Tc superconducting microstrip patch, electric-magnetic anisotropic substrate, Galerkin method, surface complex impedance with boundary conditions, radiation patterns

Procedia PDF Downloads 440
9931 Rocket Launch Simulation for a Multi-Mode Failure Prediction Analysis

Authors: Mennatallah M. Hussein, Olivier de Weck

Abstract:

The advancement of space exploration demands a robust space launch services program capable of reliably propelling payloads into orbit. Despite rigorous testing and quality assurance, launch failures still occur, leading to significant financial losses and jeopardizing mission objectives. Traditional failure prediction methods often lack the sophistication to account for multi-mode failure scenarios, as well as the predictive capability in complex dynamic systems. Traditional approaches also rely on expert judgment, leading to variability in risk prioritization and mitigation strategies. Hence, there is a pressing need for robust approaches that enhance launch vehicle reliability from lift-off until it reaches its parking orbit through comprehensive simulation techniques. In this study, the developed model proposes a multi-mode launch vehicle simulation framework for predicting failure scenarios when incorporating new technologies, such as new propulsion systems or advanced staging separation mechanisms in the launch system. To this end, the model combined a 6-DOF system dynamics with comprehensive data analysis to simulate multiple failure modes impacting launch performance. The simulator utilizes high-fidelity physics-based simulations to capture the complex interactions between different subsystems and environmental conditions.

Keywords: launch vehicle, failure prediction, propulsion anomalies, rocket launch simulation, rocket dynamics

Procedia PDF Downloads 21
9930 Performance Evaluation of Reinforced Concrete Framed Structure with Steel Bracing and Supplemental Energy Dissipation

Authors: Swanand Patil, Pankaj Agarwal

Abstract:

In past few decades, seismic performance objectives have shifted from earthquake resistance to earthquake resilience of the structures, especially for the lifeline buildings. Features such as negligible post-earthquake damage and replaceable damaged components, makes energy dissipating systems a valid choice for a seismically resilient building. In this study, various energy dissipation devices are applied on an eight-storey moment resisting RC building model. The energy dissipating devices include both hysteresis-based and viscous type of devices. The seismic response of the building is obtained for different positioning and mechanical properties of the devices. The investigation is carried forward to the deficiently ductile RC frame also. The performance assessment is done on the basis of drift ratio, mode shapes and displacement response of the model structures. Nonlinear dynamic analysis shows largely improved displacement response. The damping devices improve displacement response more efficiently in the deficient ductile frames than that in the perfectly moment resisting frames. This finding is important considering the number of deficient buildings in India and the world. The placement and mechanical properties of the dampers prove to be a crucial part in modelling, analyzing and designing of the structures with supplemental energy dissipation.

Keywords: earthquake resilient structures, lifeline buildings, retrofitting of structures, supplemental energy dissipation

Procedia PDF Downloads 344
9929 Wear Map for Cu-Based Friction Materials with Different Contents of Fe Reinforcement

Authors: Haibin Zhou, Pingping Yao, Kunyang Fan

Abstract:

Copper-based sintered friction materials are widely used in the brake system of different applications such as engineering machinery or high-speed train, due to the excellent mechanical, thermal and tribological performance. Considering the diversity of the working conditions of brake system, it is necessary to identify well and understand the tribological performance and wear mechanisms of friction materials for different conditions. Fe has been a preferred reinforcement for copper-based friction materials, due to its ability to improve the wear resistance and mechanical properties of material. Wear map is well accepted as a useful research method for evaluation of wear performances and wear mechanisms over a wider range of working conditions. Therefore, it is significantly important to construct a wear map which can give out the effects of work condition and Fe reinforcement on tribological performance of Cu-based friction materials. In this study, the copper-based sintered friction materials with the different addition of Fe reinforcement (0-20 vol. %) were studied. The tribological tests were performed against stainless steel in a ring-on-ring braking tester with varying braking energy density (0-5000 J/cm2). The linear wear and friction coefficient were measured. The worn surface, cross section and debris were analyzed to determine the dominant wear mechanisms for different testing conditions. On the basis of experimental results, the wear map and wear mechanism map were established, in terms of braking energy density and the addition of Fe. It was found that with low contents of Fe and low braking energy density, adhesive wear was the dominant wear mechanism of friction materials. Oxidative wear and abrasive wear mainly occurred under moderate braking energy density. In the condition of high braking energy density, with both high and low addition of Fe, delamination appeared as the main wear mechanism.

Keywords: Cu-based friction materials, Fe reinforcement, wear map, wear mechanism

Procedia PDF Downloads 270
9928 Learning from Dendrites: Improving the Point Neuron Model

Authors: Alexander Vandesompele, Joni Dambre

Abstract:

The diversity in dendritic arborization, as first illustrated by Santiago Ramon y Cajal, has always suggested a role for dendrites in the functionality of neurons. In the past decades, thanks to new recording techniques and optical stimulation methods, it has become clear that dendrites are not merely passive electrical components. They are observed to integrate inputs in a non-linear fashion and actively participate in computations. Regardless, in simulations of neural networks dendritic structure and functionality are often overlooked. Especially in a machine learning context, when designing artificial neural networks, point neuron models such as the leaky-integrate-and-fire (LIF) model are dominant. These models mimic the integration of inputs at the neuron soma, and ignore the existence of dendrites. In this work, the LIF point neuron model is extended with a simple form of dendritic computation. This gives the LIF neuron increased capacity to discriminate spatiotemporal input sequences, a dendritic functionality as observed in another study. Simulations of the spiking neurons are performed using the Bindsnet framework. In the common LIF model, incoming synapses are independent. Here, we introduce a dependency between incoming synapses such that the post-synaptic impact of a spike is not only determined by the weight of the synapse, but also by the activity of other synapses. This is a form of short term plasticity where synapses are potentiated or depressed by the preceding activity of neighbouring synapses. This is a straightforward way to prevent inputs from simply summing linearly at the soma. To implement this, each pair of synapses on a neuron is assigned a variable,representing the synaptic relation. This variable determines the magnitude ofthe short term plasticity. These variables can be chosen randomly or, more interestingly, can be learned using a form of Hebbian learning. We use Spike-Time-Dependent-Plasticity (STDP), commonly used to learn synaptic strength magnitudes. If all neurons in a layer receive the same input, they tend to learn the same through STDP. Adding inhibitory connections between the neurons creates a winner-take-all (WTA) network. This causes the different neurons to learn different input sequences. To illustrate the impact of the proposed dendritic mechanism, even without learning, we attach five input neurons to two output neurons. One output neuron isa regular LIF neuron, the other output neuron is a LIF neuron with dendritic relationships. Then, the five input neurons are allowed to fire in a particular order. The membrane potentials are reset and subsequently the five input neurons are fired in the reversed order. As the regular LIF neuron linearly integrates its inputs at the soma, the membrane potential response to both sequences is similar in magnitude. In the other output neuron, due to the dendritic mechanism, the membrane potential response is different for both sequences. Hence, the dendritic mechanism improves the neuron’s capacity for discriminating spa-tiotemporal sequences. Dendritic computations improve LIF neurons even if the relationships between synapses are established randomly. Ideally however, a learning rule is used to improve the dendritic relationships based on input data. It is possible to learn synaptic strength with STDP, to make a neuron more sensitive to its input. Similarly, it is possible to learn dendritic relationships with STDP, to make the neuron more sensitive to spatiotemporal input sequences. Feeding structured data to a WTA network with dendritic computation leads to a significantly higher number of discriminated input patterns. Without the dendritic computation, output neurons are less specific and may, for instance, be activated by a sequence in reverse order.

Keywords: dendritic computation, spiking neural networks, point neuron model

Procedia PDF Downloads 127
9927 Planning Quality and Maintenance Activities in a Closed-Loop Serial Multi-Stage Manufacturing System under Constant Degradation

Authors: Amauri Josafat Gomez Aguilar, Jean Pierre Kenné

Abstract:

This research presents the development of a self-sustainable manufacturing system from a circular economy perspective, structured by a multi-stage serial production system consisting of a series of machines under deterioration in charge of producing a single product and a reverse remanufacturing system constituted by the same productive systems of the first scheme and different tooling, fed by-products collected at the end of their life cycle, and non-conforming elements of the first productive scheme. Since the advanced production manufacturing system is unable to satisfy the customer's quality expectations completely, we propose the development of a mixed integer linear mathematical model focused on the optimal search and assignment of quality stations and preventive maintenance operation to the machines over a time horizon, intending to segregate the correct number of non-conforming parts for reuse in the remanufacturing system and thereby minimizing production, quality, maintenance, and customer non-conformance penalties. Numerical experiments are performed to analyze the solutions found by the model under different scenarios. The results showed that the correct implementation of a closed manufacturing system and allocation of quality inspection and preventive maintenance operations generate better levels of customer satisfaction and an efficient manufacturing system.

Keywords: closed loop, mixed integer linear programming, preventive maintenance, quality inspection

Procedia PDF Downloads 76
9926 Comparison of Spiral Circular Coil and Helical Coil Structures for Wireless Power Transfer System

Authors: Zhang Kehan, Du Luona

Abstract:

Wireless power transfer (WPT) systems have been widely investigated for advantages of convenience and safety compared to traditional plug-in charging systems. The research contents include impedance matching, circuit topology, transfer distance et al. for improving the efficiency of WPT system, which is a decisive factor in the practical application. What is more, coil structures such as spiral circular coil and helical coil with variable distance between two turns also have indispensable effects on the efficiency of WPT systems. This paper compares the efficiency of WPT systems utilizing spiral or helical coil with variable distance between two turns, and experimental results show that efficiency of spiral circular coil with an optimum distance between two turns is the highest. According to efficiency formula of resonant WPT system with series-series topology, we introduce M²/R₋₁ to measure the efficiency of spiral circular coil and helical coil WPT system. If the distance between two turns s is too close, proximity effect theory shows that the induced current in the conductor, caused by a variable flux created by the current flows in the skin of vicinity conductor, is the opposite direction of source current and has assignable impart on coil resistance. Thus in two coil structures, s affects coil resistance. At the same time, when the distance between primary and secondary coils is not variable, s can also make the influence on M to some degrees. The aforementioned study proves that s plays an indispensable role in changing M²/R₋₁ and then can be adjusted to find the optimum value with which WPT system achieves the highest efficiency. In actual application situations of WPT systems especially in underwater vehicles, miniaturization is one vital issue in designing WPT system structures. Limited by system size, the largest external radius of spiral circular coil is 100 mm, and the largest height of helical coil is 40 mm. In other words, the turn of coil N changes with s. In spiral circular and helical structures, the distance between each two turns in secondary coil is set as a constant value 1 mm to guarantee that the R2 is not variable. Based on the analysis above, we set up spiral circular coil and helical coil model using COMSOL to analyze the value of M²/R₋₁ when the distance between each two turns in primary coil sp varies from 0 mm to 10 mm. In the two structure models, the distance between primary and secondary coils is 50 mm and wire diameter is chosen as 1.5 mm. The turn of coil in secondary coil are 27 in helical coil model and 20 in spiral circular coil model. The best value of s in helical coil structure and spiral circular coil structure are 1 mm and 2 mm respectively, in which the value of M²/R₋₁ is the largest. It is obviously to select spiral circular coil as the first choice to design the WPT system for that the value of M²/R₋₁ in spiral circular coil is larger than that in helical coil under the same condition.

Keywords: distance between two turns, helical coil, spiral circular coil, wireless power transfer

Procedia PDF Downloads 339
9925 Non-Coplanar Nuclei in Heavy-Ion Reactions

Authors: Sahila Chopra, Hemdeep, Arshdeep Kaur, Raj K. Gupta

Abstract:

In recent times, we noticed an interesting and important role of non-coplanar degree-of-freedom (Φ = 00) in heavy ion reactions. Using the dynamical cluster-decay model (DCM) with Φ degree-of-freedom included, we have studied three compound systems 246Bk∗, 164Yb∗ and 105Ag∗. Here, within the DCM with pocket formula for nuclear proximity potential, we look for the effects of including compact, non-coplanar configurations (Φc = 00) on the non-compound nucleus (nCN) contribution in total fusion cross section σfus. For 246Bk∗, formed in 11B+235U and 14N+232Th reaction channels, the DCM with coplanar nuclei (Φc = 00) shows an nCN contribution for 11B+235U channel, but none for 14N+232Th channel, which on including Φ gives both reaction channels as pure compound nucleus decays. In the case of 164Yb∗, formed in 64Ni+100Mo, the small nCN effects for Φ=00 are reduced to almost zero for Φ = 00. Interestingly, however, 105Ag∗ for Φ = 00 shows a small nCN contribution, which gets strongly enhanced for Φ = 00, such that the characteristic property of PCN presents a change of behaviour, like that of a strongly fissioning superheavy element to a weakly fissioning nucleus; note that 105Ag∗ is a weakly fissioning nucleus and Psurv behaves like one for a weakly fissioning nucleus for both Φ = 00 and Φ = 00. Apparently, Φ is presenting itself like a good degree-of-freedom in the DCM.

Keywords: dynamical cluster-decay model, fusion cross sections, non-compound nucleus effects, non-coplanarity

Procedia PDF Downloads 301
9924 Production of Hydrophilic PVC Surfaces with Microwave Treatment for its Separation from Mixed Plastics by Froth Floatation

Authors: Srinivasa Reddy Mallampati, Chi-Hyeon Lee, Nguyen Thanh Truc, Byeong-Kyu Lee

Abstract:

Organic polymeric materials (plastics) are widely used in our daily life and various industrial fields. The separation of waste plastics is important for its feedstock and mechanical recycling. One of the major problems in incineration for thermal recycling or heat melting for material recycling is the polyvinyl chloride (PVC) contained in waste plastics. This is due to the production of hydrogen chloride, chlorine gas, dioxins, and furans originated from PVC. Therefore, the separation of PVC from waste plastics is necessary before recycling. The separation of heavy polymers (PVC 1.42, PMMA 1.12, PC 1.22 and PET 1.27 g/cm3 ) from light ones (PE and PP 0.99 g/cm3) can be achieved on the basis of their density. However it is difficult to separate PVC from other heavy polymers basis of density. There are no simple and inexpensive techniques to separate PVC from others. If hydrophobic the PVC surface is selectively changed into hydrophilic, where other polymers still have hydrophobic surface, flotation process can separate PVC from others. In the present study, the selective surface hydrophilization of polyvinyl chloride (PVC) by microwave treatment after alkaline/acid washing and with activated carbon was studied as the pre-treatment of its separation by the following froth flotation. In presence of activated carbon as absorbent, the microwave treatment could selectively increase the hydrophilicity of the PVC surface (i.e. PVC contact angle decreased about 19o) among other plastics mixture. At this stage, 100% PVC separation from other plastics could be achieved by the combination of the pre- microwave treatment with activated carbon and the following froth floatation. The hydrophilization of PVC by surface analysis would be due to the hydrophilic groups produced by microwave treatment with activated carbon. The effect of optimum condition and detailed mechanism onto separation efficiency in the froth floatation was also investigated.

Keywords: Hydrophilic, PVC, contact angle, additive, microwave, froth floatation, waste plastics

Procedia PDF Downloads 618
9923 The Optimization of Topical Antineoplastic Therapy Using Controlled Release Systems Based on Amino-functionalized Mesoporous Silica

Authors: Lacramioara Ochiuz, Aurelia Vasile, Iulian Stoleriu, Cristina Ghiciuc, Maria Ignat

Abstract:

Topical administration of chemotherapeutic agents (eg. carmustine, bexarotene, mechlorethamine etc.) in local treatment of cutaneous T-cell lymphoma (CTCL) is accompanied by multiple side effects, such as contact hypersensitivity, pruritus, skin atrophy or even secondary malignancies. A known method of reducing the side effects of anticancer agent is the development of modified drug release systems using drug incapsulation in biocompatible nanoporous inorganic matrices, such as mesoporous MCM-41 silica. Mesoporous MCM-41 silica is characterized by large specific surface, high pore volume, uniform porosity, and stable dispersion in aqueous medium, excellent biocompatibility, in vivo biodegradability and capacity to be functionalized with different organic groups. Therefore, MCM-41 is an attractive candidate for a wide range of biomedical applications, such as controlled drug release, bone regeneration, protein immobilization, enzymes, etc. The main advantage of this material lies in its ability to host a large amount of the active substance in uniform pore system with adjustable size in a mesoscopic range. Silanol groups allow surface controlled functionalization leading to control of drug loading and release. This study shows (I) the amino-grafting optimization of mesoporous MCM-41 silica matrix by means of co-condensation during synthesis and post-synthesis using APTES (3-aminopropyltriethoxysilane); (ii) loading the therapeutic agent (carmustine) obtaining a modified drug release systems; (iii) determining the profile of in vitro carmustine release from these systems; (iv) assessment of carmustine release kinetics by fitting on four mathematical models. Obtained powders have been described in terms of structure, texture, morphology thermogravimetric analysis. The concentration of the therapeutic agent in the dissolution medium has been determined by HPLC method. In vitro dissolution tests have been done using cell Enhancer in a 12 hours interval. Analysis of carmustine release kinetics from mesoporous systems was made by fitting to zero-order model, first-order model Higuchi model and Korsmeyer-Peppas model, respectively. Results showed that both types of highly ordered mesoporous silica (amino grafted by co-condensation process or post-synthesis) are thermally stable in aqueous medium. In what regards the degree of loading and efficiency of loading with the therapeutic agent, there has been noticed an increase of around 10% in case of co-condensation method application. This result shows that direct co-condensation leads to even distribution of amino groups on the pore walls while in case of post-synthesis grafting many amino groups are concentrated near the pore opening and/or on external surface. In vitro dissolution tests showed an extended carmustine release (more than 86% m/m) both from systems based on silica functionalized directly by co-condensation and after synthesis. Assessment of carmustine release kinetics revealed a release through diffusion from all studied systems as a result of fitting to Higuchi model. The results of this study proved that amino-functionalized mesoporous silica may be used as a matrix for optimizing the anti-cancer topical therapy by loading carmustine and developing prolonged-release systems.

Keywords: carmustine, silica, controlled, release

Procedia PDF Downloads 257
9922 Identifying Confirmed Resemblances in Problem-Solving Engineering, Both in the Past and Present

Authors: Colin Schmidt, Adrien Lecossier, Pascal Crubleau, Philippe Blanchard, Simon Richir

Abstract:

Introduction:The widespread availability of artificial intelligence, exemplified by Generative Pre-trained Transformers (GPT) relying on large language models (LLM), has caused a seismic shift in the realm of knowledge. Everyone now has the capacity to swiftly learn how these models can either serve them well or not. Today, conversational AI like ChatGPT is grounded in neural transformer models, a significant advance in natural language processing facilitated by the emergence of renowned LLMs constructed using neural transformer architecture. Inventiveness of an LLM : OpenAI's GPT-3 stands as a premier LLM, capable of handling a broad spectrum of natural language processing tasks without requiring fine-tuning, reliably producing text that reads as if authored by humans. However, even with an understanding of how LLMs respond to questions asked, there may be lurking behind OpenAI’s seemingly endless responses an inventive model yet to be uncovered. There may be some unforeseen reasoning emerging from the interconnection of neural networks here. Just as a Soviet researcher in the 1940s questioned the existence of Common factors in inventions, enabling an Under standing of how and according to what principles humans create them, it is equally legitimate today to explore whether solutions provided by LLMs to complex problems also share common denominators. Theory of Inventive Problem Solving (TRIZ) : We will revisit some fundamentals of TRIZ and how Genrich ALTSHULLER was inspired by the idea that inventions and innovations are essential means to solve societal problems. It's crucial to note that traditional problem-solving methods often fall short in discovering innovative solutions. The design team is frequently hampered by psychological barriers stemming from confinement within a highly specialized knowledge domain that is difficult to question. We presume ChatGPT Utilizes TRIZ 40. Hence, the objective of this research is to decipher the inventive model of LLMs, particularly that of ChatGPT, through a comparative study. This will enhance the efficiency of sustainable innovation processes and shed light on how the construction of a solution to a complex problem was devised. Description of the Experimental Protocol : To confirm or reject our main hypothesis that is to determine whether ChatGPT uses TRIZ, we will follow a stringent protocol that we will detail, drawing on insights from a panel of two TRIZ experts. Conclusion and Future Directions : In this endeavor, we sought to comprehend how an LLM like GPT addresses complex challenges. Our goal was to analyze the inventive model of responses provided by an LLM, specifically ChatGPT, by comparing it to an existing standard model: TRIZ 40. Of course, problem solving is our main focus in our endeavours.

Keywords: artificial intelligence, Triz, ChatGPT, inventiveness, problem-solving

Procedia PDF Downloads 60
9921 Classification of Business Models of Italian Bancassurance by Balance Sheet Indicators

Authors: Andrea Bellucci, Martina Tofi

Abstract:

The aim of paper is to analyze business models of bancassurance in Italy for life business. The life insurance business is very developed in the Italian market and banks branches have 80% of the market share. Given its maturity, the life insurance market needs to consolidate its organizational form to allow for the development of non-life business, which nowadays collects few premiums but represents a great opportunity to enlarge the market share of bancassurance using its strength in the distribution channel while the market share of independent agents is decreasing. Starting with the main business model of bancassurance for life business, this paper will analyze the performances of life companies in the Italian market by balance sheet indicators and by main discriminant variables of business models. The study will observe trends from 2013 to 2015 for the Italian market by exploiting a database managed by Associazione Nazionale delle Imprese di Assicurazione (ANIA). The applied approach is based on a bottom-up analysis starting with variables and indicators to define business models’ classification. The statistical classification algorithm proposed by Ward is employed to design business models’ profiles. Results from the analysis will be a representation of the main business models built by their profile related to indicators. In that way, an unsupervised analysis is developed that has the limit of its judgmental dimension based on research opinion, but it is possible to obtain a design of effective business models.

Keywords: bancassurance, business model, non life bancassurance, insurance business value drivers

Procedia PDF Downloads 294
9920 Numerical Simulation of a Point Absorber Wave Energy Converter Using OpenFOAM in Indian Scenario

Authors: Pooja Verma, Sumana Ghosh

Abstract:

There is a growing need for alternative way of power generation worldwide. The reason can be attributed to limited resources of fossil fuels, environmental pollution, increasing cost of conventional fuels, and lower efficiency of conversion of energy in existing systems. In this context, one of the potential alternatives for power generation is wave energy. However, it is difficult to estimate the amount of electrical energy generation in an irregular sea condition by experiment and or analytical methods. Therefore in this work, a numerical wave tank is developed using the computational fluid dynamics software Open FOAM. In this software a specific utility known as waves2Foam utility is being used to carry out the simulation work. The computational domain is a tank of dimension: 5m*1.5m*1m with a floating object of dimension: 0.5m*0.2m*0.2m. Regular waves are generated at the inlet of the wave tank according to Stokes second order theory. The main objective of the present study is to validate the numerical model against existing experimental data. It shows a good matching with the existing experimental data of floater displacement. Later the model is exploited to estimate energy extraction due to the movement of such a point absorber in real sea conditions. Scale down the wave properties like wave height, wave length, etc. are used as input parameters. Seasonal variations are also considered.

Keywords: OpenFOAM, numerical wave tank, regular waves, floating object, point absorber

Procedia PDF Downloads 349
9919 Exposing The Invisible

Authors: Kimberley Adamek

Abstract:

According to the Council on Tall Buildings, there has been a rapid increase in the construction of tall or “megatall” buildings over the past two decades. Simultaneously, the New England Journal of Medicine has reported that there has been a steady increase in climate related natural disasters since the 1970s; the eastern expansion of the USA's infamous Tornado Alley being just one of many current issues. In the future, this could mean that tall buildings, which already guide high speed winds down to pedestrian levels would have to withstand stronger forces and protect pedestrians in more extreme ways. Although many projects are required to be verified within wind tunnels and a handful of cities such as San Francisco have included wind testing within building code standards, there are still many examples where wind is only considered for basic loading. This typically results in and an increase of structural expense and unwanted mitigation strategies that are proposed late within a project. When building cities, architects rarely consider how each building alters the invisible patterns of wind and how these alterations effect other areas in different ways later on. It is not until these forces move, overpower and even destroy cities that people take notice. For example, towers have caused winds to blow objects into people (Walkie-Talkie Tower, Leeds, England), cause building parts to vibrate and produce loud humming noises (Beetham Tower, Manchester), caused wind tunnels in streets as well as many other issues. Alternatively, there exist towers which have used their form to naturally draw in air and ventilate entire facilities in order to eliminate the needs for costly HVAC systems (The Met, Thailand) and used their form to increase wind speeds to generate electricity (Bahrain Tower, Dubai). Wind and weather exist and effect all parts of the world in ways such as: Science, health, war, infrastructure, catastrophes, tourism, shopping, media and materials. Working in partnership with a leading wind engineering company RWDI, a series of tests, images and animations documenting discovered interactions of different building forms with wind will be collected to emphasize the possibilities for wind use to architects. A site within San Francisco (due to its increasing tower development, consistently wind conditions and existing strict wind comfort criteria) will host a final design. Iterations of this design will be tested within the wind tunnel and computational fluid dynamic systems which will expose, utilize and manipulate wind flows to create new forms, technologies and experiences. Ultimately, this thesis aims to question the amount which the environment is allowed to permeate building enclosures, uncover new programmatic possibilities for wind in buildings, and push the boundaries of working with the wind to ensure the development and safety of future cities. This investigation will improve and expand upon the traditional understanding of wind in order to give architects, wind engineers as well as the general public the ability to broaden their scope in order to productively utilize this living phenomenon that everyone constantly feels but cannot see.

Keywords: wind engineering, climate, visualization, architectural aerodynamics

Procedia PDF Downloads 356
9918 A Systematic Review Examining the Experimental methodology behind in vivo testing of hiatus hernia and Diaphragmatic Hernia Mesh

Authors: Whitehead-Clarke T., Beynon V., Banks J., Karanjia R., Mudera V., Windsor A., Kureshi A.

Abstract:

Introduction: Mesh implants are regularly used to help repair both hiatus hernias (HH) and diaphragmatic hernias (DH). In vivo studies are used to test not only mesh safety but increasingly comparative efficacy. Our work examines the field of in vivo mesh testing for HH and DH models to establish current practices and standards. Method: This systematic review was registered with PROSPERO. Medline and Embase databases were searched for relevant in vivo studies. 44 articles were identified and underwent abstract review, where 22 were excluded. 4 further studies were excluded after full text review – leaving 18 to undergo data extraction. Results: Of 18 studies identified, 9 used an in vivo HH model and 9 a DH model. 5 studies undertook mechanical testing on tissue samples – all uniaxial in nature. Testing strip widths ranged from 1-20mm (median 3mm). Testing speeds varied from 1.5-60mm/minute. Upon histology, the most commonly assessed structural and cellular factors were neovascularization and macrophages, respectively (n=9 each). Structural analysis was mostly qualitative, where cellular analysis was equally likely to be quantitative. 11 studies assessed adhesion formation, of which 8 used one of four scoring systems. 8 studies measured mesh shrinkage. Discussion: In vivo studies assessing mesh for HH and DH repair are uncommon. Within this relatively young field, we encourage surgical and materials testing institutions to discuss its standardisation.

Keywords: hiatus, diaphragmatic, hernia, mesh, materials testing, in vivo

Procedia PDF Downloads 207
9917 Radio Labeling and Characterization of Cysteine and Its Derivatives with Tc99m and Their Bio-Distribution

Authors: Rabia Ashfaq, Saeed Iqbal, Atiq ur Rehman, Irfanullah Khan

Abstract:

An extensive series of radiopharmaceuticals have been explored in order to discover a better brain tumour diagnostic agent. Tc99m labelling with cysteine and its derivatives in liposomes shows effective tagging of about 70% to 80 %. Due to microscopic size it successfully crossed the brain barrier in 2 minutes which gradually decreases in 5 to 15 minutes. HMPAO labelled with Tc99m is another important radiopharmaceutical used to study brain perfusion but it comes with a flaw that it’s only functional during epilepsy. 1, 1 ECD is purely used in Tc99m ECD formulation; because it not only tends to cross the blood brain barrier but it can be metabolized which can be easily entrapped in human brain. Radio labelling of Cysteine with Tc99m at room temperature was performed which yielded no good results. Hence cysteine derivatives with salicylaldehyde were prepared that produced about 75 % yield for ligand. In order to perform it’s radio labelling a suitable solvent DMSO was selected and physical parameters were performed. Elemental analyser produced remarkably similar results for ligand as reported in literature. IR spectra of Ligand in DMSO concluded in the absence of SH stretch and presence of N-H vibration. Thermal analysis of the ligand further suggested its decomposition pattern with no distinct curve for a melting point. Radio labelling of ligand was performed which produced excellent results giving up to 88% labelling at pH 5.0. Clinical trials using Rabbit were performed after validating the products reproducibility. The radiopharmaceutical prepared was injected into the rabbit. Dynamic as well as static study was performed under the SPECT. It showed considerable uptake in the kidneys and liver considering it suitable for the Hypatobilliary study.

Keywords: marcapto compounds, 99mTc - radiolabeling, salicylaldicysteine, thiozolidine

Procedia PDF Downloads 340
9916 Household Climate-Resilience Index Development for the Health Sector in Tanzania: Use of Demographic and Health Surveys Data Linked with Remote Sensing

Authors: Heribert R. Kaijage, Samuel N. A. Codjoe, Simon H. D. Mamuya, Mangi J. Ezekiel

Abstract:

There is strong evidence that climate has changed significantly affecting various sectors including public health. The recommended feasible solution is adopting development trajectories which combine both mitigation and adaptation measures for improving resilience pathways. This approach demands a consideration for complex interactions between climate and social-ecological systems. While other sectors such as agriculture and water have developed climate resilience indices, the public health sector in Tanzania is still lagging behind. The aim of this study was to find out how can we use Demographic and Health Surveys (DHS) linked with Remote Sensing (RS) technology and metrological information as tools to inform climate change resilient development and evaluation for the health sector. Methodological review was conducted whereby a number of studies were content analyzed to find appropriate indicators and indices for climate resilience household and their integration approach. These indicators were critically reviewed, listed, filtered and their sources determined. Preliminary identification and ranking of indicators were conducted using participatory approach of pairwise weighting by selected national stakeholders from meeting/conferences on human health and climate change sciences in Tanzania. DHS datasets were retrieved from Measure Evaluation project, processed and critically analyzed for possible climate change indicators. Other sources for indicators of climate change exposure were also identified. For the purpose of preliminary reporting, operationalization of selected indicators was discussed to produce methodological approach to be used in resilience comparative analysis study. It was found that household climate resilient index depends on the combination of three indices namely Household Adaptive and Mitigation Capacity (HC), Household Health Sensitivity (HHS) and Household Exposure Status (HES). It was also found that, DHS alone cannot complement resilient evaluation unless integrated with other data sources notably flooding data as a measure of vulnerability, remote sensing image of Normalized Vegetation Index (NDVI) and Metrological data (deviation from rainfall pattern). It can be concluded that if these indices retrieved from DHS data sets are computed and scientifically integrated can produce single climate resilience index and resilience maps could be generated at different spatial and time scales to enhance targeted interventions for climate resilient development and evaluations. However, further studies are need to test for the sensitivity of index in resilience comparative analysis among selected regions.

Keywords: climate change, resilience, remote sensing, demographic and health surveys

Procedia PDF Downloads 161
9915 Evaluation of Kabul BRT Route Network with Application of Integrated Land-use and Transportation Model

Authors: Mustafa Mutahari, Nao Sugiki, Kojiro Matsuo

Abstract:

The four decades of war, lack of job opportunities, poverty, lack of services, and natural disasters in different provinces of Afghanistan have contributed to a rapid increase in the population of Kabul, the capital city of Afghanistan. Population census has not been conducted since 1979, the first and last population census in Afghanistan. However, according to population estimations by Afghan authorities, the population of Kabul has been estimated at more than 4 million people, whereas the city was designed for two million people. Although the major transport mode of Kabul residents is public transport, responsible authorities within the country failed to supply the required means of transportation systems for the city. Besides, informal resettlement, lack of intersection control devices, presence of illegal vendors on streets, presence of illegal and unstandardized on-street parking and bus stops, driver`s unprofessional behavior, weak traffic law enforcement, and blocked roads and sidewalks have contributed to the extreme traffic congestion of Kabul. In 2018, the government of Afghanistan approved the Kabul city Urban Design Framework (KUDF), a vision towards the future of Kabul, which provides strategies and design guidance at different scales to direct urban development. Considering traffic congestion of the city and its budget limitations, the KUDF proposes a BRT route network with seven lines to reduce the traffic congestion, and it is said to facilitate more than 50% of Kabul population to benefit from this service. Based on the KUDF, it is planned to increase the BRT mode share from 0% to 17% and later to 30% in medium and long-term planning scenarios, respectively. Therefore, a detailed research study is needed to evaluate the proposed system before the implementation stage starts. The integrated land-use transport model is an effective tool to evaluate the Kabul BRT because of its future assessment capabilities that take into account the interaction between land use and transportation. This research aims to analyze and evaluate the proposed BRT route network with the application of an integrated land-use and transportation model. The research estimates the population distribution and travel behavior of Kabul within small boundary scales. The actual road network and land-use detailed data of the city are used to perform the analysis. The BRT corridors are evaluated not only considering its impacts on the spatial interactions in the city`s transportation system but also on the spatial developments. Therefore, the BRT are evaluated with the scenarios of improving the Kabul transportation system based on the distribution of land-use or spatial developments, planned development typology and population distribution of the city. The impacts of the new improved transport system on the BRT network are analyzed and the BRT network is evaluated accordingly. In addition, the research also focuses on the spatial accessibility of BRT stops, corridors, and BRT line beneficiaries, and each BRT stop and corridor are evaluated in terms of both access and geographic coverage, as well.

Keywords: accessibility, BRT, integrated land-use and transport model, travel behavior, spatial development

Procedia PDF Downloads 209
9914 A Triad Pedagogy for Increased Digital Competence of Human Resource Management Students: Reflecting on Human Resource Information Systems at a South African University

Authors: Esther Pearl Palmer

Abstract:

Driven by the increased pressure on Higher Education Institutions (HEIs) to produce work-ready graduates for the modern world of work, this study reflects on triad teaching and learning practices to increase student engagement and employability. In the South African higher education context, the employability of graduates is imperative in strengthening the country’s economy and in increasing competitiveness. Within this context, the field of Human Resource Management (HRM) calls for innovative methods and approaches to teaching and learning and assessing the skills and competencies of graduates to render them employable. Digital competency in Human Resource Information Systems (HRIS) is an important component and prerequisite for employment in HRM. The purpose of this research is to reflect on the subject HRIS developed by lecturers at the Central University of Technology, Free State (CUT), with the intention to actively engage students in real-world learning activities and increase their employability. The Enrichment Triad Model (ETM) was used as theoretical framework to develop the subject as it supports a triad teaching and learning approach to education. It is, furthermore, an inter-structured model that supports collaboration between industry, academics and students. The study follows a mixed-method approach to reflect on the learning experiences of the industry, academics and students in the subject field over the past three years. This paper is a work in progress and seeks to broaden the scope of extant studies about student engagement in work-related learning to increase employability. Based on the ETM as theoretical framework and pedagogical practice, this paper proposes that following a triad teaching and learning approach will increase work-related skills of students. Findings from the study show that students, academics and industry alike regard educational opportunities that incorporate active learning experiences with the world of work enhances student engagement in learning and renders them more employable.

Keywords: digital competence, enriched triad model, human resource information systems, student engagement, triad pedagogy.

Procedia PDF Downloads 85
9913 The Post-Hegemony of Post-Capitalism: Towards a Political Theory of Open Cooperativism

Authors: Vangelis Papadimitropoulos

Abstract:

The paper is part of the research project “Techno-Social Innovation in the Collaborative Economy'', funded by the Hellenic Foundation of Research and Innovation for the years 2022-2024. The research project examines the normative and empirical conditions of grassroots technologically driven innovation, potentially enabling the transition towards a commons-oriented post-capitalist economy. The project carries out a conceptually led and empirically grounded multi-case study of the digital commons, open-source technologies, platform cooperatives, open cooperatives and Distributed Autonomous Organizations (DAOs) on the Blockchain. The methodological scope of research is interdisciplinary inasmuch as it comprises political theory, economics, sustainability science and computer science, among others. The research draws specifically on Michel Bauwens and Vasilis Kostakis' model of open cooperativism between the commons, ethical market entities and a partner state. Bauwens and Kostakis advocate for a commons-based counter-hegemonic post-capitalist transition beyond and against neoliberalism. The research further employs Laclau and Mouffe's discourse theory of hegemony to introduce a post-hegemonic conceptualization of the model of open cooperativism. Thus, the paper aims to outline the theoretical contribution of the research project to contemporary political theory debates on post-capitalism and the collaborative economy.

Keywords: open cooperativism, techno-social innovation, post-hegemony, post-capitalism

Procedia PDF Downloads 62
9912 Comparative Parametric and Emission Characteristics of Single Cylinder Spark Ignition Engine Using Gasoline, Ethanol, and H₂O as Micro Emulsion Fuels

Authors: Ufaith Qadri, M Marouf Wani

Abstract:

In this paper, the performance and emission characteristics of a Single Cylinder Spark Ignition engine have been investigated. The research is based on micro emulsion application as fuel in a gasoline engine. We have analyzed many micro emulsion compositions in various proportions, for predicting the performance of the Spark Ignition engine. This new technology of fuel modifications is emerging very rapidly as lot of research is going on in the field of micro emulsion fuels in Compression Ignition engines, but the micro emulsion fuel used in a Gasoline engine is very rare. The use of micro emulsion as fuel in a Spark Ignition engine is virtually unexplored. So, our main goal is to see the performance and emission characteristics of micro emulsions as fuel, in Spark Ignition engines, and finding which composition is more efficient. In this research, we have used various micro emulsion fuels whose composition varies for all the three blends, and their performance and emission characteristic were predicted in AVL Boost software. Conventional Gasoline fuel 90%, 80% and 85% were blended with co-surfactant Ethanol in different compositions, and water was used as an additive for making it crystal clear transparent micro emulsion fuel, which is thermodynamically stable. By comparing the performances of engines, the power has shown similarity for micro emulsion fuel and conventional Gasoline fuel. On the other hand, Torque and BMEP shows increase for all the micro emulsion fuels. Micro emulsion fuel shows higher thermal efficiency and lower Specific Fuel Consumption for all the compositions as compared to the Gasoline fuel. Carbon monoxide and Hydro carbon emissions were also measured. The result shows that emissions decrease for all the composition of micro emulsion fuels, and proved to be the most efficient fuel both in terms of performance and emission characteristics.

Keywords: AVL Boost, emissions, microemulsions, performance, Spark Ignition (SI) engine

Procedia PDF Downloads 257
9911 Potential and Techno-Economic Analysis of Hydrogen Production from Portuguese Solid Recovered Fuels

Authors: A. Ribeiro, N. Pacheco, M. Soares, N. Valério, L. Nascimento, A. Silva, C. Vilarinho, J. Carvalho

Abstract:

Hydrogen will play a key role in changing the current global energy paradigm, associated with the high use of fossil fuels and the release of greenhouse gases. This work intended to identify and quantify the potential of Solid Recovered Fuels (SFR) existing in Portugal and project the cost of hydrogen, produced through its steam gasification in different scenarios, associated with the size or capacity of the plant and the existence of carbon capture and storage (CCS) systems. Therefore, it was performed a techno-economic analysis simulation using an ASPEN base model, the H2A Hydrogen Production Model Version 3.2018. Regarding the production of SRF, it was possible to verify the annual production of more than 200 thousand tons of SRF in Portugal in 2019. The results of the techno-economic analysis simulations showed that in the scenarios containing a high (200,000 tons/year) and medium (40,000 tons/year) amount of SFR, the cost of hydrogen production was competitive concerning the current prices of hydrogen. The results indicate that scenarios 1 and 2, which use 200,000 tons of SRF per year, have lower hydrogen production values, 1.22 USD/kg H2 and 1.63 USD/kg H2, respectively. The cost of producing hydrogen without carbon capture and storage (CCS) systems in an average amount of SFR (40,000 tons/year) was 1.70 USD/kg H2. In turn, scenarios 5 (without CCS) and 6 (with CCS), which use only 683 tons of SFR from urban sources, have the highest costs, 6.54 USD/kg H2 and 908.97 USD/kg H2, respectively. Therefore, it was possible to conclude that there is a huge potential for the use of SRF for the production of hydrogen through steam gasification in Portugal.

Keywords: gasification, hydrogen, solid recovered fuels, techno-economic analysis, waste-to-energy

Procedia PDF Downloads 119
9910 Controlled Release of Glucosamine from Pluronic-Based Hydrogels for the Treatment of Osteoarthritis

Authors: Papon Thamvasupong, Kwanchanok Viravaidya-Pasuwat

Abstract:

Osteoarthritis affects a lot of people worldwide. Local injection of glucosamine is one of the alternative treatment methods to replenish the natural lubrication of cartilage. However, multiple injections can potentially lead to possible bacterial infection. Therefore, a drug delivery system is desired to reduce the frequencies of injections. A hydrogel is one of the delivery systems that can control the release of drugs. Thermo-reversible hydrogels can be beneficial to the drug delivery system especially in the local injection route because this formulation can change from liquid to gel after getting into human body. Once the gel is in the body, it will slowly release the drug in a controlled manner. In this study, various formulations of Pluronic-based hydrogels were synthesized for the controlled release of glucosamine. One of the challenges of the Pluronic controlled release system is its fast dissolution rate. To overcome this problem, alginate and calcium sulfate (CaSO4) were added to the polymer solution. The characteristics of the hydrogels were investigated including the gelation temperature, gelation time, hydrogel dissolution and glucosamine release mechanism. Finally, a mathematical model of glucosamine release from Pluronic-alginate-hyaluronic acid hydrogel was developed. Our results have shown that crosslinking Pluronic gel with alginate did not significantly extend the dissolution rate of the gel. Moreover, the gel dissolution profiles and the glucosamine release mechanisms were best described using the zeroth-order kinetic model, indicating that the release of glucosamine was primarily governed by the gel dissolution.

Keywords: controlled release, drug delivery system, glucosamine, pluronic, thermoreversible hydrogel

Procedia PDF Downloads 265
9909 Modified Silicates as Dissolved Oxygen Sensors in Water: Structural and Optical Properties

Authors: Andile Mkhohlakali, Tien-Chien Jen, James Tshilongo, Happy Mabowa

Abstract:

Among different parameters, oxygen is one of the most important analytes of interest, dissolved oxygen (DO) concentration is very crucial and significant for various areas of physical, chemical, and environmental monitoring. Herein we report oxygen-sensitive luminophores -based lanthanum(III) trifluoromethanesulfonate), [La]³⁺ was encapsulated into SiO₂-based xerogel matrix. The nanosensor is composed of organically modified silica nanoparticles, doped with the luminescent oxygen–sensitive lanthanum(III) trifluoromethanesulfonate complex. The precursor materials used for sensing film were triethyl ethoxy silane (TEOS) and (3-Mercaptopropyltriethoxysilane) (MPTMS- TEOS) used for SiO2-baed matrices. Brunauer–Emmett–Teller (BET), and BJH indicate that the SiO₂ transformed from microporous to mesoporous upon the addition of La³⁺ luminophore with increased surface area (SBET). The typical amorphous SiO₂ based xerogels were revealed with X-Ray diffraction (XRD) and Selected Area Electron Diffraction (SAED) analysis. Scanning electron microscope- (SEM) and transmission electron microscope (TEM) showed the porous morphology and reduced particle for SiO₂ and La-SiO₂ xerogels respectively. The existence of elements, siloxane networks, and thermal stability of xerogel was confirmed by energy dispersive spectroscopy (EDS), Fourier-transform infrared spectroscopy (FTIR), and Thermographic analysis (TGA). UV-Vis spectroscopy and photoluminescence (PL) have been used to characterize the optical properties of xerogels. La-SiO₂ demonstrates promising characteristic features of an active sensing film for dissolved oxygen in the water. Keywords: Sol-gel, ORMOSILs, encapsulation, Luminophores quenching, O₂-sensing

Keywords: sol-gel, ORMOSILs, luminophores quenching, O₂-sensing

Procedia PDF Downloads 120
9908 Mining User-Generated Contents to Detect Service Failures with Topic Model

Authors: Kyung Bae Park, Sung Ho Ha

Abstract:

Online user-generated contents (UGC) significantly change the way customers behave (e.g., shop, travel), and a pressing need to handle the overwhelmingly plethora amount of various UGC is one of the paramount issues for management. However, a current approach (e.g., sentiment analysis) is often ineffective for leveraging textual information to detect the problems or issues that a certain management suffers from. In this paper, we employ text mining of Latent Dirichlet Allocation (LDA) on a popular online review site dedicated to complaint from users. We find that the employed LDA efficiently detects customer complaints, and a further inspection with the visualization technique is effective to categorize the problems or issues. As such, management can identify the issues at stake and prioritize them accordingly in a timely manner given the limited amount of resources. The findings provide managerial insights into how analytics on social media can help maintain and improve their reputation management. Our interdisciplinary approach also highlights several insights by applying machine learning techniques in marketing research domain. On a broader technical note, this paper illustrates the details of how to implement LDA in R program from a beginning (data collection in R) to an end (LDA analysis in R) since the instruction is still largely undocumented. In this regard, it will help lower the boundary for interdisciplinary researcher to conduct related research.

Keywords: latent dirichlet allocation, R program, text mining, topic model, user generated contents, visualization

Procedia PDF Downloads 185
9907 Valorization of Surveillance Data and Assessment of the Sensitivity of a Surveillance System for an Infectious Disease Using a Capture-Recapture Model

Authors: Jean-Philippe Amat, Timothée Vergne, Aymeric Hans, Bénédicte Ferry, Pascal Hendrikx, Jackie Tapprest, Barbara Dufour, Agnès Leblond

Abstract:

The surveillance of infectious diseases is necessary to describe their occurrence and help the planning, implementation and evaluation of risk mitigation activities. However, the exact number of detected cases may remain unknown whether surveillance is based on serological tests because identifying seroconversion may be difficult. Moreover, incomplete detection of cases or outbreaks is a recurrent issue in the field of disease surveillance. This study addresses these two issues. Using a viral animal disease as an example (equine viral arteritis), the goals were to establish suitable rules for identifying seroconversion in order to estimate the number of cases and outbreaks detected by a surveillance system in France between 2006 and 2013, and to assess the sensitivity of this system by estimating the total number of outbreaks that occurred during this period (including unreported outbreaks) using a capture-recapture model. Data from horses which exhibited at least one positive result in serology using viral neutralization test between 2006 and 2013 were used for analysis (n=1,645). Data consisted of the annual antibody titers and the location of the subjects (towns). A consensus among multidisciplinary experts (specialists in the disease and its laboratory diagnosis, epidemiologists) was reached to consider seroconversion as a change in antibody titer from negative to at least 32 or as a three-fold or greater increase. The number of seroconversions was counted for each town and modeled using a unilist zero-truncated binomial (ZTB) capture-recapture model with R software. The binomial denominator was the number of horses tested in each infected town. Using the defined rules, 239 cases located in 177 towns (outbreaks) were identified from 2006 to 2013. Subsequently, the sensitivity of the surveillance system was estimated as the ratio of the number of detected outbreaks to the total number of outbreaks that occurred (including unreported outbreaks) estimated using the ZTB model. The total number of outbreaks was estimated at 215 (95% credible interval CrI95%: 195-249) and the surveillance sensitivity at 82% (CrI95%: 71-91). The rules proposed for identifying seroconversion may serve future research. Such rules, adjusted to the local environment, could conceivably be applied in other countries with surveillance programs dedicated to this disease. More generally, defining ad hoc algorithms for interpreting the antibody titer could be useful regarding other human and animal diseases and zoonosis when there is a lack of accurate information in the literature about the serological response in naturally infected subjects. This study shows how capture-recapture methods may help to estimate the sensitivity of an imperfect surveillance system and to valorize surveillance data. The sensitivity of the surveillance system of equine viral arteritis is relatively high and supports its relevance to prevent the disease spreading.

Keywords: Bayesian inference, capture-recapture, epidemiology, equine viral arteritis, infectious disease, seroconversion, surveillance

Procedia PDF Downloads 292