Search results for: RBF neural network modelling
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 6766

Search results for: RBF neural network modelling

2896 Synthesis and Characterization of Fibrin/Polyethylene Glycol-Based Interpenetrating Polymer Networks for Dermal Tissue Engineering

Authors: O. Gsib, U. Peirera, C. Egles, S. A. Bencherif

Abstract:

In skin regenerative medicine, one of the critical issues is to produce a three-dimensional scaffold with optimized porosity for dermal fibroblast infiltration and neovascularization, which exhibits high mechanical properties and displays sufficient wound healing characteristics. In this study, we report on the synthesis and characterization of macroporous sequential interpenetrating polymer networks (IPNs) combining skin wound healing properties of fibrin with the excellent physical properties of polyethylene glycol (PEG). Fibrin fibers serve as a provisional biologically active network to promote cell adhesion and proliferation while PEG provides the mechanical stability to maintain the entire 3D construct. After having modified both PEG and Serum Albumin (used for promoting enzymatic degradability) by adding methacrylate residues (PEGDM and SAM, respectively), Fibrin/PEGDM-SAM sequential IPNs were synthesized as follows: Macroporous sponges were first produced from PEGDM-SAM hydrogels by a freeze-drying technique and then rehydrated by adding the fibrin precursors. Environmental Scanning Electron Microscopy (ESEM) and Confocal Laser Scanning Microscopy (CLSM) were used to characterize their microstructure. Human dermal fibroblasts were cultivated during one week in the constructs and different cell culture parameters (viability, morphology, proliferation) were evaluated. Subcutaneous implantations of the scaffolds were conducted on five-week old male nude mice to investigate their biocompatibility in vivo. We successfully synthesized interconnected and macroporous Fibrin/PEGDM-SAM sequential IPNs. The viability of primary dermal fibroblasts was well maintained (above 90%) after 2 days of culture. Cells were able to adhere, spread and proliferate in the scaffolds suggesting the suitable porosity and intrinsic biologic properties of the constructs. The fibrin network adopted a spider web shape that covered partially the pores allowing easier cell infiltration into the macroporous structure. To further characterize the in vitro cell behavior, cell proliferation (EdU incorporation, MTS assay) is being studied. Preliminary histological analysis of animal studies indicated the persistence of hydrogels even after one-month post implantation and confirmed the absence of inflammation response, good biocompatibility and biointegration of our scaffolds within the surrounding tissues. These results suggest that our Fibrin/PEGDM-SAM IPNs could be considered as potential candidates for dermis regenerative medicine. Histological analysis will be completed to further assess scaffold remodeling including de novo extracellular matrix protein synthesis and early stage angiogenesis analysis. Compression measurements will be conducted to investigate the mechanical properties.

Keywords: fibrin, hydrogels for dermal reconstruction, polyethylene glycol, semi-interpenetrating polymer network

Procedia PDF Downloads 221
2895 Modelling the Photovoltaic Pump Output Using Empirical Data from Local Conditions in the Vhembe District

Authors: C. Matasane, C. Dwarika, R. Naidoo

Abstract:

The mathematical analysis on radiation obtained and the development of the solar photovoltaic (PV) array groundwater pumping is needed in the rural areas of Thohoyandou, Limpopo Province for sizing and power performance subject to the climate conditions within the area. A simple methodology approach is developed for the directed coupled solar, controller and submersible ground water pump system. The system consists of a PV array, pump controller and submerged pump, battery backup and charger controller. For this reason, the theoretical solar radiation obtained for optimal predictions and system performance in order to achieve different design and operating parameters. Here the examination of the PV schematic module in a Direct Current (DC) application is used for obtainable maximum solar power energy for water pumping. In this paper, a simple efficient photovoltaic water pumping system is presented with its theoretical studies and mathematical modeling of photovoltaics (PV) system.

Keywords: renewable energy sources, solar groundwater pumping, theoretical and mathematical analysis of photovoltaic (PV) system, theoretical solar radiation

Procedia PDF Downloads 358
2894 Predictive Modelling Approach to Identify Spare Parts Inventory Obsolescence

Authors: Madhu Babu Cherukuri, Tamoghna Ghosh

Abstract:

Factory supply chain management spends billions of dollars every year to procure and manage equipment spare parts. Due to technology -and processes changes some of these spares become obsolete/dead inventory. Factories have huge dead inventory worth millions of dollars accumulating over time. This is due to lack of a scientific methodology to identify them and send the inventory back to the suppliers on a timely basis. The standard approach followed across industries to deal with this is: if a part is not used for a set pre-defined period of time it is declared dead. This leads to accumulation of dead parts over time and these parts cannot be sold back to the suppliers as it is too late as per contract agreement. Our main idea is the time period for identifying a part as dead cannot be a fixed pre-defined duration across all parts. Rather, it should depend on various properties of the part like historical consumption pattern, type of part, how many machines it is being used in, whether it- is a preventive maintenance part etc. We have designed a predictive algorithm which predicts part obsolescence well in advance with reasonable accuracy and which can help save millions.

Keywords: obsolete inventory, machine learning, big data, supply chain analytics, dead inventory

Procedia PDF Downloads 306
2893 The Facilitators and Barriers to the Implementation of Educational Neuroscience: Teachers’ Perspectives

Authors: S. Kawther, C. Marshall

Abstract:

Educational neuroscience has the intention of transforming research findings of the underpinning neural processes of learning to educational practices. A main criticism of the field, hitherto, is that less focus has been put on studying the in-progress practical application of these findings. Therefore, this study aims to gain a better understanding of teachers’ perceptions of the practical application and utilization of brain knowledge. This was approached by investigating the answer to 'What are the facilitators and barriers for bringing research from neuroscience to bear on education?'. Following a qualitative design, semi-structured interviews were conducted with 12 teachers who had a proficient course in educational neuroscience. Thematic analysis was performed on the transcribed data applying Braun & Clark’s steps. Findings emerged with four main themes: time, knowledge, teacher’s involvement, and system. These themes revealed that some effective brain-based practices are being engaged in by the teachers. However, the lack of guidance and challenges regarding this implementation were also found. This study discusses findings in light of the development of educational neuroscience implementation.

Keywords: brain-based, educational neuroscience, neuroeducation, neuroscience-informed

Procedia PDF Downloads 149
2892 Multimodal Data Fusion Techniques in Audiovisual Speech Recognition

Authors: Hadeer M. Sayed, Hesham E. El Deeb, Shereen A. Taie

Abstract:

In the big data era, we are facing a diversity of datasets from different sources in different domains that describe a single life event. These datasets consist of multiple modalities, each of which has a different representation, distribution, scale, and density. Multimodal fusion is the concept of integrating information from multiple modalities in a joint representation with the goal of predicting an outcome through a classification task or regression task. In this paper, multimodal fusion techniques are classified into two main classes: model-agnostic techniques and model-based approaches. It provides a comprehensive study of recent research in each class and outlines the benefits and limitations of each of them. Furthermore, the audiovisual speech recognition task is expressed as a case study of multimodal data fusion approaches, and the open issues through the limitations of the current studies are presented. This paper can be considered a powerful guide for interested researchers in the field of multimodal data fusion and audiovisual speech recognition particularly.

Keywords: multimodal data, data fusion, audio-visual speech recognition, neural networks

Procedia PDF Downloads 92
2891 Mapping Iron Content in the Brain with Magnetic Resonance Imaging and Machine Learning

Authors: Gabrielle Robertson, Matthew Downs, Joseph Dagher

Abstract:

Iron deposition in the brain has been linked with a host of neurological disorders such as Alzheimer’s, Parkinson’s, and Multiple Sclerosis. While some treatment options exist, there are no objective measurement tools that allow for the monitoring of iron levels in the brain in vivo. An emerging Magnetic Resonance Imaging (MRI) method has been recently proposed to deduce iron concentration through quantitative measurement of magnetic susceptibility. This is a multi-step process that involves repeated modeling of physical processes via approximate numerical solutions. For example, the last two steps of this Quantitative Susceptibility Mapping (QSM) method involve I) mapping magnetic field into magnetic susceptibility and II) mapping magnetic susceptibility into iron concentration. Process I involves solving an ill-posed inverse problem by using regularization via injection of prior belief. The end result from Process II highly depends on the model used to describe the molecular content of each voxel (type of iron, water fraction, etc.) Due to these factors, the accuracy and repeatability of QSM have been an active area of research in the MRI and medical imaging community. This work aims to estimate iron concentration in the brain via a single step. A synthetic numerical model of the human head was created by automatically and manually segmenting the human head on a high-resolution grid (640x640x640, 0.4mm³) yielding detailed structures such as microvasculature and subcortical regions as well as bone, soft tissue, Cerebral Spinal Fluid, sinuses, arteries, and eyes. Each segmented region was then assigned tissue properties such as relaxation rates, proton density, electromagnetic tissue properties and iron concentration. These tissue property values were randomly selected from a Probability Distribution Function derived from a thorough literature review. In addition to having unique tissue property values, different synthetic head realizations also possess unique structural geometry created by morphing the boundary regions of different areas within normal physical constraints. This model of the human brain is then used to create synthetic MRI measurements. This is repeated thousands of times, for different head shapes, volume, tissue properties and noise realizations. Collectively, this constitutes a training-set that is similar to in vivo data, but larger than datasets available from clinical measurements. This 3D convolutional U-Net neural network architecture was used to train data-driven Deep Learning models to solve for iron concentrations from raw MRI measurements. The performance was then tested on both synthetic data not used in training as well as real in vivo data. Results showed that the model trained on synthetic MRI measurements is able to directly learn iron concentrations in areas of interest more effectively than other existing QSM reconstruction methods. For comparison, models trained on random geometric shapes (as proposed in the Deep QSM method) are less effective than models trained on realistic synthetic head models. Such an accurate method for the quantitative measurement of iron deposits in the brain would be of important value in clinical studies aiming to understand the role of iron in neurological disease.

Keywords: magnetic resonance imaging, MRI, iron deposition, machine learning, quantitative susceptibility mapping

Procedia PDF Downloads 118
2890 Mathematical Modelling and Numerical Simulation of Maisotsenko Cycle

Authors: Rasikh Tariq, Fatima Z. Benarab

Abstract:

Evaporative coolers has a minimum potential to reach the wet-bulb temperature of intake air which is not enough to handle a large cooling load; therefore, it is not a feasible option to overcome cooling requirement of a building. The invention of Maisotsenko (M) cycle has led evaporative cooling technology to reach the sub-wet-bulb temperature of the intake air; therefore, it brings an innovation in evaporative cooling techniques. In this work, we developed a mathematical model of the Maisotsenko based air cooler by applying energy and mass balance laws on different air channels. The governing ordinary differential equations are discretized and simulated on MATLAB. The temperature and the humidity plots are shown in the simulation results. A parametric study is conducted by varying working air inlet conditions (temperature and humidity), inlet air velocity, geometric parameters and water temperature. The influence of these aforementioned parameters on the cooling effectiveness of the HMX is reported.  Results have shown that the effectiveness of the M-Cycle is increased by increasing the ambient temperature and decreasing absolute humidity. An air velocity of 0.5 m/sec and a channel height of 6-8mm is recommended.

Keywords: HMX, maisotsenko cycle, mathematical modeling, numerical simulation, parametric study

Procedia PDF Downloads 135
2889 Linearization and Process Standardization of Construction Design Engineering Workflows

Authors: T. R. Sreeram, S. Natarajan, C. Jena

Abstract:

Civil engineering construction is a network of tasks involving varying degree of complexity and streamlining, and standardization is the only way to establish a systemic approach to design. While there are off the shelf tools such as AutoCAD that play a role in the realization of design, the repeatable process in which these tools are deployed often is ignored. The present paper addresses this challenge through a sustainable design process and effective standardizations at all stages in the design workflow. The same is demonstrated through a case study in the context of construction, and further improvement points are highlighted.

Keywords: syste, lean, value stream, process improvement

Procedia PDF Downloads 109
2888 Methodology for Developing an Intelligent Tutoring System Based on Marzano’s Taxonomy

Authors: Joaquin Navarro Perales, Ana Lidia Franzoni Velázquez, Francisco Cervantes Pérez

Abstract:

The Mexican educational system faces diverse challenges related with the quality and coverage of education. The development of Intelligent Tutoring Systems (ITS) may help to solve some of them by helping teachers to customize their classes according to the performance of the students in online courses. In this work, we propose the adaptation of a functional ITS based on Bloom’s taxonomy called Sistema de Apoyo Generalizado para la Enseñanza Individualizada (SAGE), to measure student’s metacognition and their emotional response based on Marzano’s taxonomy. The students and the system will share the control over the advance in the course, so they can improve their metacognitive skills. The system will not allow students to get access to subjects not mastered yet. The interaction between the system and the student will be implemented through Natural Language Processing techniques, thus avoiding the use of sensors to evaluate student’s response. The teacher will evaluate student’s knowledge utilization, which is equivalent to the last cognitive level in Marzano’s taxonomy.

Keywords: intelligent tutoring systems, student modelling, metacognition, affective computing, natural language processing

Procedia PDF Downloads 179
2887 Assessing the Effectiveness of Machine Learning Algorithms for Cyber Threat Intelligence Discovery from the Darknet

Authors: Azene Zenebe

Abstract:

Deep learning is a subset of machine learning which incorporates techniques for the construction of artificial neural networks and found to be useful for modeling complex problems with large dataset. Deep learning requires a very high power computational and longer time for training. By aggregating computing power, high performance computer (HPC) has emerged as an approach to resolving advanced problems and performing data-driven research activities. Cyber threat intelligence (CIT) is actionable information or insight an organization or individual uses to understand the threats that have, will, or are currently targeting the organization. Results of review of literature will be presented along with results of experimental study that compares the performance of tree-based and function-base machine learning including deep learning algorithms using secondary dataset collected from darknet.

Keywords: deep-learning, cyber security, cyber threat modeling, tree-based machine learning, function-based machine learning, data science

Procedia PDF Downloads 137
2886 Developing a New Relationship between Undrained Shear Strength and Over-Consolidation Ratio

Authors: Wael M Albadri, Hassnen M Jafer, Ehab H Sfoog

Abstract:

Relationship between undrained shear strength (Su) and over consolidation ratio (OCR) of clay soil (marine clay) is very important in the field of geotechnical engineering to estimate the settlement behaviour of clay and to prepare a small scale physical modelling test. In this study, a relationship between shear strength and OCR parameters was determined using the laboratory vane shear apparatus and the fully automatic consolidated apparatus. The main objective was to establish non-linear correlation formula between shear strength and OCR and comparing it with previous studies. Therefore, in order to achieve this objective, three points were chosen to obtain 18 undisturbed samples which were collected with an increasing depth of 1.0 m to 3.5 m each 0.5 m. Clay samples were prepared under undrained condition for both tests. It was found that the OCR and shear strength are inversely proportional at similar depth and at same undrained conditions. However, a good correlation was obtained from the relationships where the R2 values were very close to 1.0 using polynomial equations. The comparison between the experimental result and previous equation from other researchers produced a non-linear correlation which has a similar pattern with this study.

Keywords: shear strength, over-consolidation ratio, vane shear test, clayey soil

Procedia PDF Downloads 260
2885 Improved Rare Species Identification Using Focal Loss Based Deep Learning Models

Authors: Chad Goldsworthy, B. Rajeswari Matam

Abstract:

The use of deep learning for species identification in camera trap images has revolutionised our ability to study, conserve and monitor species in a highly efficient and unobtrusive manner, with state-of-the-art models achieving accuracies surpassing the accuracy of manual human classification. The high imbalance of camera trap datasets, however, results in poor accuracies for minority (rare or endangered) species due to their relative insignificance to the overall model accuracy. This paper investigates the use of Focal Loss, in comparison to the traditional Cross Entropy Loss function, to improve the identification of minority species in the “255 Bird Species” dataset from Kaggle. The results show that, although Focal Loss slightly decreased the accuracy of the majority species, it was able to increase the F1-score by 0.06 and improve the identification of the bottom two, five and ten (minority) species by 37.5%, 15.7% and 10.8%, respectively, as well as resulting in an improved overall accuracy of 2.96%.

Keywords: convolutional neural networks, data imbalance, deep learning, focal loss, species classification, wildlife conservation

Procedia PDF Downloads 170
2884 Risk Assessment on Construction Management with “Fuzzy Logy“

Authors: Mehrdad Abkenari, Orod Zarrinkafsh, Mohsen Ramezan Shirazi

Abstract:

Construction projects initiate in complicated dynamic environments and, due to the close relationships between project parameters and the unknown outer environment, they are faced with several uncertainties and risks. Success in time, cost and quality in large scale construction projects is uncertain in consequence of technological constraints, large number of stakeholders, too much time required, great capital requirements and poor definition of the extent and scope of the project. Projects that are faced with such environments and uncertainties can be well managed through utilization of the concept of risk management in project’s life cycle. Although the concept of risk is dependent on the opinion and idea of management, it suggests the risks of not achieving the project objectives as well. Furthermore, project’s risk analysis discusses the risks of development of inappropriate reactions. Since evaluation and prioritization of construction projects has been a difficult task, the network structure is considered to be an appropriate approach to analyze complex systems; therefore, we have used this structure for analyzing and modeling the issue. On the other hand, we face inadequacy of data in deterministic circumstances, and additionally the expert’s opinions are usually mathematically vague and are introduced in the form of linguistic variables instead of numerical expression. Owing to the fact that fuzzy logic is used for expressing the vagueness and uncertainty, formulation of expert’s opinion in the form of fuzzy numbers can be an appropriate approach. In other words, the evaluation and prioritization of construction projects on the basis of risk factors in real world is a complicated issue with lots of ambiguous qualitative characteristics. In this study, evaluated and prioritization the risk parameters and factors with fuzzy logy method by combination of three method DEMATEL (Decision Making Trial and Evaluation), ANP (Analytic Network Process) and TOPSIS (Technique for Order-Preference by Similarity Ideal Solution) on Construction Management.

Keywords: fuzzy logy, risk, prioritization, assessment

Procedia PDF Downloads 573
2883 Modeling and Controlling the Rotational Degree of a Quadcopter Using Proportional Integral and Derivative Controller

Authors: Sanjay Kumar, Lillie Dewan

Abstract:

The study of complex dynamic systems has advanced through various scientific approaches with the help of computer modeling. The common design trends in aerospace system design can be applied to quadcopter design. A quadcopter is a nonlinear, under-actuated system with complex aerodynamics parameters and creates challenges that demand new, robust, and effective control approaches. The flight control stability can be improved by planning and tracking the trajectory and reducing the effect of sensors and the operational environment. This paper presents a modern design Simmechanics visual modeling approach for a mechanical model of a quadcopter with three degrees of freedom. The Simmechanics model, considering inertia, mass, and geometric properties of a dynamic system, produces multiple translation and rotation maneuvers. The proportional, integral, and derivative (PID) controller is integrated with the Simmechanics model to follow a predefined quadcopter rotational trajectory for a fixed time interval. The results presented are satisfying. The simulation of the quadcopter control performed operations successfully.

Keywords: nonlinear system, quadcopter model, simscape modelling, proportional-integral-derivative controller

Procedia PDF Downloads 178
2882 Tram Track Deterioration Modeling

Authors: Mohammad Yousefikia, Sara Moridpour, Ehsan Mazloumi

Abstract:

Perceiving track geometry deterioration decisively influences the optimization of track maintenance operations. The effective management of this deterioration and increasingly utilized system with limited financial resources is a significant challenge. This paper provides a review of degradation models relevant for railroad tracks. Furthermore, due to the lack of long term information on the condition development of tram infrastructures, presents the methodology which will be used to derive degradation models from the data of Melbourne tram network.

Keywords: deterioration modeling, asset management, railway, tram

Procedia PDF Downloads 360
2881 Material Flow Modeling in Friction Stir Welding of AA6061-T6 Alloy and Study of the Effect of Process Parameters

Authors: B. SahaRoy, T. Medhi, S. C. Saha

Abstract:

To understand the friction stir welding process, it is very important to know the nature of the material flow in and around the tool. The process is a combination of both thermal as well as mechanical work i.e it is a coupled thermo-mechanical process. Numerical simulations are very much essential in order to obtain a complete knowledge of the process as well as the physics underlying it. In the present work a model based approach is adopted in order to study material flow. A thermo-mechanical based CFD model is developed using a Finite Element package, Comsol Multiphysics. The fluid flow analysis is done. The model simultaneously predicts shear strain fields, shear strain rates and shear stress over the entire workpiece for the given conditions. The flow fields generated by the streamline plot give an idea of the material flow. The variation of dynamic viscosity, velocity field and shear strain fields with various welding parameters is studied. Finally the result obtained from the above mentioned conditions is discussed elaborately and concluded.

Keywords: AA6061-T6, CFD modelling, friction stir welding, material flow

Procedia PDF Downloads 503
2880 Building a Parametric Link between Mapping and Planning: A Sunlight-Adaptive Urban Green System Plan Formation Process

Authors: Chenhao Zhu

Abstract:

Quantitative mapping is playing a growing role in guiding urban planning, such as using a heat map created by CFX, CFD2000, or Envi-met, to adjust the master plan. However, there is no effective quantitative link between the mappings and planning formation. So, in many cases, the decision-making is still based on the planner's subjective interpretation and understanding of these mappings, which limits the improvement of scientific and accuracy brought by the quantitative mapping. Therefore, in this paper, an effort has been made to give a methodology of building a parametric link between the mapping and planning formation. A parametric planning process based on radiant mapping has been proposed for creating an urban green system. In the first step, a script is written in Grasshopper to build a road network and form the block, while the Ladybug Plug-in is used to conduct a radiant analysis in the form of mapping. Then, the research creatively transforms the radiant mapping from a polygon into a data point matrix, because polygon is hard to engage in the design formation. Next, another script is created to select the main green spaces from the road network based on the criteria of radiant intensity and connect the green spaces' central points to generate a green corridor. After that, a control parameter is introduced to adjust the corridor's form based on the radiant intensity. Finally, a green system containing greenspace and green corridor is generated under the quantitative control of the data matrix. The designer only needs to modify the control parameter according to the relevant research results and actual conditions to realize the optimization of the green system. This method can also be applied to much other mapping-based analysis, such as wind environment analysis, thermal environment analysis, and even environmental sensitivity analysis. The parameterized link between the mapping and planning will bring about a more accurate, objective, and scientific planning.

Keywords: parametric link, mapping, urban green system, radiant intensity, planning strategy, grasshopper

Procedia PDF Downloads 126
2879 The Organizational Justice-Citizenship Behavior Link in Hotels: Does Customer Orientation Matter?

Authors: Pablo Zoghbi-Manrique-de-Lara, Miguel A. Suárez-Acosta

Abstract:

The goal of the present paper is to model two classic lines of research in which employees starred, organizational justice and citizenship behaviour (OCB), but that have never been studied together when targeting customers. The suggestion is made that a hotel’s fair treatment (in terms of distributive, procedural, and interactional justice) toward customers will be appreciated by the employees, who will reciprocate in kind by favouring the hotel with increased customer-oriented behaviours (COBs). Data were collected from 204 employees at eight upscale hotels in the Canary Islands (Spain). Unlike in the case of perceptions of distributive justice, results of structural equation modelling demonstrate that employees substantively react to interactional and procedural justice toward guests by engaging in customer-oriented behaviours (COBs). The findings offer new reasons why employees decide to engage in COBs, and they highlight potentially beneficial effects of fair treatment toward guests bring to hospitality through promoting COBs.

Keywords: hotel guests’ (mis) treatment, customer-oriented behaviours, employee citizenship, organizational justice, third-party observers, third-party intervention

Procedia PDF Downloads 250
2878 Investigation of Delivery of Triple Play Data in GE-PON Fiber to the Home Network

Authors: Ashima Anurag Sharma

Abstract:

Optical fiber based networks can deliver performance that can support the increasing demands for high speed connections. One of the new technologies that have emerged in recent years is Passive Optical Networks. This research paper is targeted to show the simultaneous delivery of triple play service (data, voice, and video). The comparison between various data rates is presented. It is demonstrated that as we increase the data rate, number of users to be decreases due to increase in bit error rate.

Keywords: BER, PON, TDMPON, GPON, CWDM, OLT, ONT

Procedia PDF Downloads 512
2877 Theoretical Modelling of Molecular Mechanisms in Stimuli-Responsive Polymers

Authors: Catherine Vasnetsov, Victor Vasnetsov

Abstract:

Context: Thermo-responsive polymers are materials that undergo significant changes in their physical properties in response to temperature changes. These polymers have gained significant attention in research due to their potential applications in various industries and medicine. However, the molecular mechanisms underlying their behavior are not well understood, particularly in relation to cosolvency, which is crucial for practical applications. Research Aim: This study aimed to theoretically investigate the phenomenon of cosolvency in long-chain polymers using the Flory-Huggins statistical-mechanical framework. The main objective was to understand the interactions between the polymer, solvent, and cosolvent under different conditions. Methodology: The research employed a combination of Monte Carlo computer simulations and advanced machine-learning methods. The Flory-Huggins mean field theory was used as the basis for the simulations. Spinodal graphs and ternary plots were utilized to develop an initial computer model for predicting polymer behavior. Molecular dynamic simulations were conducted to mimic real-life polymer systems. Machine learning techniques were incorporated to enhance the accuracy and reliability of the simulations. Findings: The simulations revealed that the addition of very low or very high volumes of cosolvent molecules resulted in smaller radii of gyration for the polymer, indicating poor miscibility. However, intermediate volume fractions of cosolvent led to higher radii of gyration, suggesting improved miscibility. These findings provide a possible microscopic explanation for the cosolvency phenomenon in polymer systems. Theoretical Importance: This research contributes to a better understanding of the behavior of thermo-responsive polymers and the role of cosolvency. The findings provide insights into the molecular mechanisms underlying cosolvency and offer specific predictions for future experimental investigations. The study also presents a more rigorous analysis of the Flory-Huggins free energy theory in the context of polymer systems. Data Collection and Analysis Procedures: The data for this study was collected through Monte Carlo computer simulations and molecular dynamic simulations. The interactions between the polymer, solvent, and cosolvent were analyzed using the Flory-Huggins mean field theory. Machine learning techniques were employed to enhance the accuracy of the simulations. The collected data was then analyzed to determine the impact of cosolvent volume fractions on the radii of gyration of the polymer. Question Addressed: The research addressed the question of how cosolvency affects the behavior of long-chain polymers. Specifically, the study aimed to investigate the interactions between the polymer, solvent, and cosolvent under different volume fractions and understand the resulting changes in the radii of gyration. Conclusion: In conclusion, this study utilized theoretical modeling and computer simulations to investigate the phenomenon of cosolvency in long-chain polymers. The findings suggest that moderate cosolvent volume fractions can lead to improved miscibility, as indicated by higher radii of gyration. These insights contribute to a better understanding of the molecular mechanisms underlying cosolvency in polymer systems and provide predictions for future experimental studies. The research also enhances the theoretical analysis of the Flory-Huggins free energy theory.

Keywords: molecular modelling, flory-huggins, cosolvency, stimuli-responsive polymers

Procedia PDF Downloads 55
2876 Overview About Sludge Produced From Treatment Plant of Bahr El-Baqar Drain and Reusing It With Cement in Outdoor Paving

Authors: Khaled M.Naguib, Ahmed M.Noureldin

Abstract:

This paper aims to achieve many goals such as knowing (quantities produced- main properties- characteristics) of sludge produced from Bahr EL-Baqar drains treatment plant. This prediction or projection was made by laboratory analysis and modelling of Model samples from sludge depending on many studies that have previously done, second check the feasibility and do a risk analysis to know the best alternatives for reuse in producing secondary products that add value to sludge. Also, to know alternatives that have no value to add. All recovery methods are relatively very expensive and challenging to be done in this mega plant, so the recommendation from this study is to use the sludge as a coagulant to reduce some compounds or in secondary products. The study utilized sludge-cement replacement percentages of 10%, 20%, 30%, 40% and 50%. Produced tiles were tested for water absorption and breaking (bending) strength. The study showed that all produced tiles exhibited a water absorption ratio of around 10%. The study concluded that produced tiles, except for 50% sludge-cement replacement, comply with the breaking strength requirements of 2.8 MPa for tiles for external use.

Keywords: cement, tiles, water treatment sludge, breaking strength, absorption, heavy metals, risk analysis

Procedia PDF Downloads 92
2875 Named Entity Recognition System for Tigrinya Language

Authors: Sham Kidane, Fitsum Gaim, Ibrahim Abdella, Sirak Asmerom, Yoel Ghebrihiwot, Simon Mulugeta, Natnael Ambassager

Abstract:

The lack of annotated datasets is a bottleneck to the progress of NLP in low-resourced languages. The work presented here consists of large-scale annotated datasets and models for the named entity recognition (NER) system for the Tigrinya language. Our manually constructed corpus comprises over 340K words tagged for NER, with over 118K of the tokens also having parts-of-speech (POS) tags, annotated with 12 distinct classes of entities, represented using several types of tagging schemes. We conducted extensive experiments covering convolutional neural networks and transformer models; the highest performance achieved is 88.8% weighted F1-score. These results are especially noteworthy given the unique challenges posed by Tigrinya’s distinct grammatical structure and complex word morphologies. The system can be an essential building block for the advancement of NLP systems in Tigrinya and other related low-resourced languages and serve as a bridge for cross-referencing against higher-resourced languages.

Keywords: Tigrinya NER corpus, TiBERT, TiRoBERTa, BiLSTM-CRF

Procedia PDF Downloads 96
2874 An Investigation on Overstrength Factor (Ω) of Reinforced Concrete Buildings in Turkish Earthquake Draft Code (TEC-2016)

Authors: M. Hakan Arslan, I. Hakkı Erkan

Abstract:

Overstrength factor is an important parameter of load reduction factor. In this research, the overstrength factor (Ω) of reinforced concrete (RC) buildings and the parameters of Ω in TEC-2016 draft version have been explored. For this aim, 48 RC buildings have been modeled according to the current seismic code TEC-2007 and Turkish Building Code-500-2000 criteria. After modelling step, nonlinear static pushover analyses have been applied to these buildings by using TEC-2007 Section 7. After the nonlinear pushover analyses, capacity curves (lateral load-lateral top displacement curves) have been plotted for 48 RC buildings. Using capacity curves, overstrength factors (Ω) have been derived for each building. The obtained overstrength factor (Ω) values have been compared with TEC-2016 values for related building types, and the results have been interpreted. According to the obtained values from the study, overstrength factor (Ω) given in TEC-2016 draft code is found quite suitable.

Keywords: reinforced concrete buildings, overstrength factor, earthquake, static pushover analysis

Procedia PDF Downloads 344
2873 Critical Appraisal of Different Drought Indices of Drought Predection and Their Application in KBK Districts of Odisha

Authors: Bibhuti Bhusan Sahoo, Ramakar Jha

Abstract:

Mapping of the extreme events (droughts) is one of the adaptation strategies to consequences of increasing climatic inconsistency and climate alterations. There is no operational practice to forecast the drought. One of the suggestions is to update mapping of drought prone areas for developmental planning. Drought indices play a significant role in drought mitigation. Many scientists have worked on different statistical analysis in drought and other climatological hazards. Many researchers have studied droughts individually for different sub-divisions or for India. Very few workers have studied district wise probabilities over large scale. In the present study, district wise drought probabilities over KBK (Kalahandi-Balangir-Koraput) districts of Odisha, India, Which are seriously prone to droughts, has been established using Hydrological drought index and Meteorological drought index along with the remote sensing drought indices to develop a multidirectional approach in the field of drought mitigation. Mapping for moderate and severe drought probabilities for KBK districts has been done and regions belonging different class intervals of probabilities of drought have been demarcated. Such type of information would be a good tool for planning purposes, for input in modelling and better promising results can be achieved.

Keywords: drought indices, KBK districts, proposed drought severity index, SPI

Procedia PDF Downloads 432
2872 Home Range and Spatial Interaction Modelling of Black Bears

Authors: Fekadu L. Bayisa, Elvan Ceyhan, Todd D. Steury

Abstract:

Interaction between individuals within the same species is an important component of population dynamics. An interaction can be either static (based on spatial overlap) or dynamic (based on movement interactions). Using GPS collar data, we can quantify both static and dynamic interactions between black bears. The goal of this work is to determine the level of black bear interactions using the 95% and 50% home ranges, as well as to model black bear spatial interactions, which could be attraction, avoidance/repulsion, or a lack of interaction at all, to gain new insights and improve our understanding of ecological processes. Recent methodological developments in home range estimation, inhomogeneous multitype/cross-type summary statistics, and envelope testing methods are explored to study the nature of black bear interactions. Our findings, in general, indicate that the black bears of one type in our data set tend to cluster around another type.

Keywords: autocorrelated kernel density estimator, cross-type summary function, inhomogeneous multitype Poisson process, kernel density estimator, minimum convex polygon, pointwise and global envelope tests

Procedia PDF Downloads 66
2871 Natural Frequency Analysis of a Porous Functionally Graded Shaft System

Authors: Natural Frequency Analysis of a Porous Functionally Graded Shaft System

Abstract:

The vibration characteristics of a functionally graded (FG) rotor model having porosities and micro-voids is investigated using three-dimensional finite element analysis. The FG shaft is mounted with a steel disc located at the midspan. The shaft ends are supported on isotropic bearings. The FG material is composed of a metallic (stainless-steel) and ceramic phase (zirconium oxide) as its constituent phases. The layer wise material property variation is governed by power law. Material property equations are developed for the porosity modelling. Python code is developed to assign the material properties to each layer including the effect of porosities. ANSYS commercial software is used to extract the natural frequencies and whirl frequencies for the FG shaft system. The obtained results show the influence of porosity volume fraction and power-law index, on the vibration characteristics of the ceramic-based FG shaft system.

Keywords: Finite element method, Functionally graded material, Porosity volume fraction, Power law

Procedia PDF Downloads 185
2870 Effect of Different Porous Media Models on Drug Delivery to Solid Tumors: Mathematical Approach

Authors: Mostafa Sefidgar, Sohrab Zendehboudi, Hossein Bazmara, Madjid Soltani

Abstract:

Based on findings from clinical applications, most drug treatments fail to eliminate malignant tumors completely even though drug delivery through systemic administration may inhibit their growth. Therefore, better understanding of tumor formation is crucial in developing more effective therapeutics. For this purpose, nowadays, solid tumor modeling and simulation results are used to predict how therapeutic drugs are transported to tumor cells by blood flow through capillaries and tissues. A solid tumor is investigated as a porous media for fluid flow simulation. Most of the studies use Darcy model for porous media. In Darcy model, the fluid friction is neglected and a few simplified assumptions are implemented. In this study, the effect of these assumptions is studied by considering Brinkman model. A multi scale mathematical method which calculates fluid flow to a solid tumor is used in this study to investigate how neglecting fluid friction affects the solid tumor simulation. In this work, the mathematical model in our previous studies is developed by considering two model of momentum equation for porous media: Darcy and Brinkman. The mathematical method involves processes such as fluid flow through solid tumor as porous media, extravasation of blood flow from vessels, blood flow through vessels and solute diffusion, convective transport in extracellular matrix. The sprouting angiogenesis model is used for generating capillary network and then fluid flow governing equations are implemented to calculate blood flow through the tumor-induced capillary network. Finally, the two models of porous media are used for modeling fluid flow in normal and tumor tissues in three different shapes of tumors. Simulations of interstitial fluid transport in a solid tumor demonstrate that the simplifications used in Darcy model affect the interstitial velocity and Brinkman model predicts a lower value for interstitial velocity than the values that Darcy model does.

Keywords: solid tumor, porous media, Darcy model, Brinkman model, drug delivery

Procedia PDF Downloads 283
2869 The Concept of Neurostatistics as a Neuroscience

Authors: Igwenagu Chinelo Mercy

Abstract:

This study is on the concept of Neurostatistics in relation to neuroscience. Neuroscience also known as neurobiology is the scientific study of the nervous system. In the study of neuroscience, it has been noted that brain function and its relations to the process of acquiring knowledge and behaviour can be better explained by the use of various interrelated methods. The scope of neuroscience has broadened over time to include different approaches used to study the nervous system at different scales. On the other hand, Neurostatistics based on this study is viewed as a statistical concept that uses similar techniques of neuron mechanisms to solve some problems especially in the field of life science. This study is imperative in this era of Artificial intelligence/Machine leaning in the sense that clear understanding of the technique and its proper application could assist in solving some medical disorder that are mainly associated with the nervous system. This will also help in layman’s understanding of the technique of the nervous system in order to overcome some of the health challenges associated with it. For this concept to be well understood, an illustrative example using a brain associated disorder was used for demonstration. Structural equation modelling was adopted in the analysis. The results clearly show the link between the techniques of statistical model and nervous system. Hence, based on this study, the appropriateness of Neurostatistics application in relation to neuroscience could be based on the understanding of the behavioural pattern of both concepts.

Keywords: brain, neurons, neuroscience, neurostatistics, structural equation modeling

Procedia PDF Downloads 56
2868 Sediment Patterns from Fluid-Bed Interactions: A Direct Numerical Simulations Study on Fluvial Turbulent Flows

Authors: Nadim Zgheib, Sivaramakrishnan Balachandar

Abstract:

We present results on the initial formation of ripples from an initially flattened erodible bed. We use direct numerical simulations (DNS) of turbulent open channel flow over a fixed sinusoidal bed coupled with hydrodynamic stability analysis. We use the direct forcing immersed boundary method to account for the presence of the sediment bed. The resolved flow provides the bed shear stress and consequently the sediment transport rate, which is needed in the stability analysis of the Exner equation. The approach is different from traditional linear stability analysis in the sense that the phase lag between the bed topology, and the sediment flux is obtained from the DNS. We ran 11 simulations at a fixed shear Reynolds number of 180, but for different sediment bed wavelengths. The analysis allows us to sweep a large range of physical and modelling parameters to predict their effects on linear growth. The Froude number appears to be the critical controlling parameter in the early linear development of ripples, in contrast with the dominant role of particle Reynolds number during the equilibrium stage.

Keywords: direct numerical simulation, immersed boundary method, sediment-bed interactions, turbulent multiphase flow, linear stability analysis

Procedia PDF Downloads 166
2867 Electronic Nose Based on Metal Oxide Semiconductor Sensors as an Alternative Technique for the Spoilage Classification of Oat Milk

Authors: A. Deswal, N. S. Deora, H. N. Mishra

Abstract:

The aim of the present study was to develop a rapid method for electronic nose for online quality control of oat milk. Analysis by electronic nose and bacteriological measurements were performed to analyse spoilage kinetics of oat milk samples stored at room temperature and refrigerated conditions for up to 15 days. Principal component analysis (PCA), discriminant factorial analysis (DFA) and soft independent modelling by class analogy (SIMCA) classification techniques were used to differentiate the samples of oat milk at different days. The total plate count (bacteriological method) was selected as the reference method to consistently train the electronic nose system. The e-nose was able to differentiate between the oat milk samples of varying microbial load. The results obtained by the bacteria total viable counts showed that the shelf-life of oat milk stored at room temperature and refrigerated conditions were 20 hours and 13 days, respectively. The models built classified oat milk samples based on the total microbial population into “unspoiled” and “spoiled”.

Keywords: electronic-nose, bacteriological, shelf-life, classification

Procedia PDF Downloads 245