Search results for: lumped parameter method
12925 Advanced Exergetic Analysis: Decomposition Method Applied to a Membrane-Based Hard Coal Oxyfuel Power Plant
Authors: Renzo Castillo, George Tsatsaronis
Abstract:
High-temperature ceramic membranes for air separation represents an important option to reduce the significant efficiency drops incurred in state-of-the-art cryogenic air separation for high tonnage oxygen production required in oxyfuel power stations. This study is focused on the thermodynamic analysis of two power plant model designs: the state-of-the-art supercritical 600ᵒC hard coal plant (reference power plant Nordrhein-Westfalen) and the membrane-based oxyfuel concept implemented in this reference plant. In the latter case, the oxygen is separated through a mixed-conducting hollow fiber perovskite membrane unit in the three-end operation mode, which has been simulated under vacuum conditions on the permeate side and at high-pressure conditions on the feed side. The thermodynamic performance of each plant concept is assessed by conventional exergetic analysis, which determines location, magnitude and sources of efficiency losses, and advanced exergetic analysis, where endogenous/exogenous and avoidable/unavoidable parts of exergy destruction are calculated at the component and full process level. These calculations identify thermodynamic interdependencies among components and reveal the real potential for efficiency improvements. The endogenous and exogenous exergy destruction portions are calculated by the decomposition method, a recently developed straightforward methodology, which is suitable for complex power stations with a large number of process components. Lastly, an improvement priority ranking for relevant components, as well as suggested changes in process layouts are presented for both power stations.Keywords: exergy, carbon capture and storage, ceramic membranes, perovskite, oxyfuel combustion
Procedia PDF Downloads 18512924 Multi Biomertric Personal Identification System Based On Hybird Intellegence Method
Authors: Laheeb M. Ibrahim, Ibrahim A. Salih
Abstract:
Biometrics is a technology that has been widely used in many official and commercial identification applications. The increased concerns in security during recent years (especially during the last decades) have essentially resulted in more attention being given to biometric-based verification techniques. Here, a novel fusion approach of palmprint, dental traits has been suggested. These traits which are authentication techniques have been employed in a range of biometric applications that can identify any postmortem PM person and antemortem AM. Besides improving the accuracy, the fusion of biometrics has several advantages such as increasing, deterring spoofing activities and reducing enrolment failure. In this paper, a first unimodel biometric system has been made by using (palmprint and dental) traits, for each one classification applying an artificial neural network and a hybrid technique that combines swarm intelligence and neural network together, then attempt has been made to combine palmprint and dental biometrics. Principally, the fusion of palmprint and dental biometrics and their potential application has been explored as biometric identifiers. To address this issue, investigations have been carried out about the relative performance of several statistical data fusion techniques for integrating the information in both unimodal and multimodal biometrics. Also the results of the multimodal approach have been compared with each one of these two traits authentication approaches. This paper studies the features and decision fusion levels in multimodal biometrics. To determine the accuracy of GAR to parallel system decision-fusion including (AND, OR, Majority fating) has been used. The backpropagation method has been used for classification and has come out with result (92%, 99%, 97%) respectively for GAR, while the GAR) for this algorithm using hybrid technique for classification (95%, 99%, 98%) respectively. To determine the accuracy of the multibiometric system for feature level fusion has been used, while the same preceding methods have been used for classification. The results have been (98%, 99%) respectively while to determine the GAR of feature level different methods have been used and have come out with (98%).Keywords: back propagation neural network BP ANN, multibiometric system, parallel system decision-fusion, practical swarm intelligent PSO
Procedia PDF Downloads 53312923 Reliability of Movement Assessment Battery for Children-2 Age Band 3 Using Multiple Testers
Authors: Jernice S. Y. Tan
Abstract:
Introduction: Reliability within and between testers is vital to ensure the accuracy of any motor assessment instrument. However, reliability checks of the Movement Assessment Battery for Children-2 (MABC-2) age band 3 using multiple testers assigned to different MABC-2 tasks for the same group of participants are uncommon. Multiple testers were not stated as a choice in the MABC-2 manual. Therefore, the purpose of this study was to determine the inter- and intra-tester reliability for using multiple testers to administer the test protocols of MABC-2 age band 3. Methods: Thirty volunteered adolescents (n = 30; 15 males, 15 females; age range: 13 – 16 years) performed the eight tasks in a randomised sequence at three different test stations for the MABC-2 task components (Manual Dexterity, Aiming and Catching, Balance). Ethics approval and parental consent were obtained. The participants were videotaped while performing the test protocols of MABC-2 age band 3. Five testers were involved in the data collection process. They were Sports Science graduating students doing their final year project and were supervised by experienced motor assessor. Inter- and intra-tester reliability checks using intra-class coefficient (ICC) were carried out using the videotaped data. Results: The inter-tester reliability between the five testers for the eight tasks ranged from rᵢcc = 0.705 to rᵢcc = 0.995. This suggests that the average agreement between them was considered good to excellent. With the exception of one tester who had rᵢcc = 0.687 for one of the eight tasks (i.e. zip-zap hopping), the intra-tester reliability within each tester ranged from rᵢcc = 0.728 to rᵢcc = 1.000, and this also suggested good to excellent consistency within testers. Discussion: The use of multiple testers with good intra-tester reliability for different test stations is feasible. This method allows several participants to be assessed concurrently at different test stations and saves overall data collection time. Therefore, it is recommended that the administering of MABC-2 with multiple testers should be extended to other age bands ensuring the feasibility of such method for other age bands.Keywords: adolescents, MABC, motor assessment, motor skills, reliability
Procedia PDF Downloads 32312922 Linguistic Cyberbullying, a Legislative Approach
Authors: Simona Maria Ignat
Abstract:
Bullying online has been an increasing studied topic during the last years. Different approaches, psychological, linguistic, or computational, have been applied. To our best knowledge, a definition and a set of characteristics of phenomenon agreed internationally as a common framework are still waiting for answers. Thus, the objectives of this paper are the identification of bullying utterances on Twitter and their algorithms. This research paper is focused on the identification of words or groups of words, categorized as “utterances”, with bullying effect, from Twitter platform, extracted on a set of legislative criteria. This set is the result of analysis followed by synthesis of law documents on bullying(online) from United States of America, European Union, and Ireland. The outcome is a linguistic corpus with approximatively 10,000 entries. The methods applied to the first objective have been the following. The discourse analysis has been applied in identification of keywords with bullying effect in texts from Google search engine, Images link. Transcription and anonymization have been applied on texts grouped in CL1 (Corpus linguistics 1). The keywords search method and the legislative criteria have been used for identifying bullying utterances from Twitter. The texts with at least 30 representations on Twitter have been grouped. They form the second corpus linguistics, Bullying utterances from Twitter (CL2). The entries have been identified by using the legislative criteria on the the BoW method principle. The BoW is a method of extracting words or group of words with same meaning in any context. The methods applied for reaching the second objective is the conversion of parts of speech to alphabetical and numerical symbols and writing the bullying utterances as algorithms. The converted form of parts of speech has been chosen on the criterion of relevance within bullying message. The inductive reasoning approach has been applied in sampling and identifying the algorithms. The results are groups with interchangeable elements. The outcomes convey two aspects of bullying: the form and the content or meaning. The form conveys the intentional intimidation against somebody, expressed at the level of texts by grammatical and lexical marks. This outcome has applicability in the forensic linguistics for establishing the intentionality of an action. Another outcome of form is a complex of graphemic variations essential in detecting harmful texts online. This research enriches the lexicon already known on the topic. The second aspect, the content, revealed the topics like threat, harassment, assault, or suicide. They are subcategories of a broader harmful content which is a constant concern for task forces and legislators at national and international levels. These topic – outcomes of the dataset are a valuable source of detection. The analysis of content revealed algorithms and lexicons which could be applied to other harmful contents. A third outcome of content are the conveyances of Stylistics, which is a rich source of discourse analysis of social media platforms. In conclusion, this corpus linguistics is structured on legislative criteria and could be used in various fields.Keywords: corpus linguistics, cyberbullying, legislation, natural language processing, twitter
Procedia PDF Downloads 8612921 Inclusion Body Refolding at High Concentration for Large-Scale Applications
Authors: J. Gabrielczyk, J. Kluitmann, T. Dammeyer, H. J. Jördening
Abstract:
High-level expression of proteins in bacteria often causes production of insoluble protein aggregates, called inclusion bodies (IB). They contain mainly one type of protein and offer an easy and efficient way to get purified protein. On the other hand, proteins in IB are normally devoid of function and therefore need a special treatment to become active. Most refolding techniques aim at diluting the solubilizing chaotropic agents. Unfortunately, optimal refolding conditions have to be found empirically for every protein. For large-scale applications, a simple refolding process with high yields and high final enzyme concentrations is still missing. The constructed plasmid pASK-IBA63b containing the sequence of fructosyltransferase (FTF, EC 2.4.1.162) from Bacillus subtilis NCIMB 11871 was transformed into E. coli BL21 (DE3) Rosetta. The bacterium was cultivated in a fed-batch bioreactor. The produced FTF was obtained mainly as IB. For refolding experiments, five different amounts of IBs were solubilized in urea buffer with protein concentration of 0.2-8.5 g/L. Solubilizates were refolded with batch or continuous dialysis. The refolding yield was determined by measuring the protein concentration of the clear supernatant before and after the dialysis. Particle size was measured by dynamic light scattering. We tested the solubilization properties of fructosyltransferase IBs. The particle size measurements revealed that the solubilization of the aggregates is achieved at urea concentration of 5M or higher and confirmed by absorption spectroscopy. All results confirm previous investigations that refolding yields are dependent upon initial protein concentration. In batch dialysis, the yields dropped from 67% to 12% and 72% to 19% for continuous dialysis, in relation to initial concentrations from 0.2 to 8.5 g/L. Often used additives such as sucrose and glycerol had no effect on refolding yields. Buffer screening indicated a significant increase in activity but also temperature stability of FTF with citrate/phosphate buffer. By adding citrate to the dialysis buffer, we were able to increase the refolding yields to 82-47% in batch and 90-74% in the continuous process. Further experiments showed that in general, higher ionic strength of buffers had major impact on refolding yields; doubling the buffer concentration increased the yields up to threefold. Finally, we achieved corresponding high refolding yields by reducing the chamber volume by 75% and the amount of buffer needed. The refolded enzyme had an optimal activity of 12.5±0.3 x104 units/g. However, detailed experiments with native FTF revealed a reaggregation of the molecules and loss in specific activity depending on the enzyme concentration and particle size. For that reason, we actually focus on developing a process of simultaneous enzyme refolding and immobilization. The results of this study show a new approach in finding optimal refolding conditions for inclusion bodies at high concentrations. Straightforward buffer screening and increase of the ionic strength can optimize the refolding yield of the target protein by 400%. Gentle removal of chaotrope with continuous dialysis increases the yields by an additional 65%, independent of the refolding buffer applied. In general time is the crucial parameter for successful refolding of solubilized proteins.Keywords: dialysis, inclusion body, refolding, solubilization
Procedia PDF Downloads 29412920 Research on the Spatio-Temporal Evolution Pattern of Traffic Dominance in Shaanxi Province
Authors: Leng Jian-Wei, Wang Lai-Jun, Li Ye
Abstract:
In order to measure and analyze the transportation situation within the counties of Shaanxi province over a certain period of time and to promote the province's future transportation planning and development, this paper proposes a reasonable layout plan and compares model rationality. The study uses entropy weight method to measure the transportation advantages of 107 counties in Shaanxi province from three dimensions: road network density, trunk line influence and location advantage in 2013 and 2021, and applies spatial autocorrelation analysis method to analyze the spatial layout and development trend of county-level transportation, and conducts ordinary least square (OLS)regression on transportation impact factors and other influencing factors. The paper also compares the regression fitting degree of the Geographically weighted regression(GWR) model and the OLS model. The results show that spatially, the transportation advantages of Shaanxi province generally show a decreasing trend from the Weihe Plain to the surrounding areas and mainly exhibit high-high clustering phenomenon. Temporally, transportation advantages show an overall upward trend, and the phenomenon of spatial imbalance gradually decreases. People's travel demands have changed to some extent, and the demand for rapid transportation has increased overall. The GWR model regression fitting degree of transportation advantages is 0.74, which is higher than the OLS regression model's fitting degree of 0.64. Based on the evolution of transportation advantages, it is predicted that this trend will continue for a period of time in the future. To improve the transportation advantages of Shaanxi province increasing the layout of rapid transportation can effectively enhance the transportation advantages of Shaanxi province. When analyzing spatial heterogeneity, geographic factors should be considered to establish a more reliable modelKeywords: traffic dominance, GWR model, spatial autocorrelation analysis, temporal and spatial evolution
Procedia PDF Downloads 8912919 Effect of Baffles on the Cooling of Electronic Components
Authors: O. Bendermel, C. Seladji, M. Khaouani
Abstract:
In this work, we made a numerical study of the thermal and dynamic behaviour of air in a horizontal channel with electronic components. The influence to use baffles on the profiles of velocity and temperature is discussed. The finite volume method and the algorithm Simple are used for solving the equations of conservation of mass, momentum and energy. The results found show that baffles improve heat transfer between the cooling air and electronic components. The velocity will increase from 3 times per rapport of the initial velocity.Keywords: electronic components, baffles, cooling, fluids engineering
Procedia PDF Downloads 29712918 AFM Probe Sensor Designed for Cellular Membrane Components
Authors: Sarmiza Stanca, Wolfgang Fritzsche, Christoph Krafft, Jürgen Popp
Abstract:
Independent of the cell type a thin layer of a few nanometers thickness surrounds the cell interior as the cellular membrane. The transport of ions and molecules through the membrane is achieved in a very precise way by pores. Understanding the process of opening and closing the pores due to an electrochemical gradient across the membrane requires knowledge of the pore constitutive proteins. Recent reports prove the access to the molecular level of the cellular membrane by atomic force microscopy (AFM). This technique also permits an electrochemical study in the immediate vicinity of the tip. Specific molecules can be electrochemically localized in the natural cellular membrane. Our work aims to recognize the protein domains of the pores using an AFM probe as a miniaturized amperometric sensor, and to follow the protein behavior while changing the applied potential. The intensity of the current produced between the surface and the AFM probe is amplified and detected simultaneously with the surface imaging. The AFM probe plays the role of the working electrode and the substrate, a conductive glass on which the cells are grown, represent the counter electrode. For a better control of the electric potential on the probe, a third electrode Ag/AgCl wire is mounted in the circuit as a reference electrode. The working potential is applied between the electrodes with a programmable source and the current intensity in the circuit is recorded with a multimeter. The applied potential considers the overpotential at the electrode surface and the potential drop due to the current flow through the system. The reported method permits a high resolved electrochemical study of the protein domains on the living cell membrane. The amperometric map identifies areas of different current intensities on the pore depending on the applied potential. The reproducibility of this method is limited by the tip shape, the uncontrollable capacitance, which occurs at the apex and a potential local charge separation.Keywords: AFM, sensor, membrane, pores, proteins
Procedia PDF Downloads 30812917 Interface Fracture of Sandwich Composite Influenced by Multiwalled Carbon Nanotube
Authors: Alak Kumar Patra, Nilanjan Mitra
Abstract:
Higher strength to weight ratio is the main advantage of sandwich composite structures. Interfacial delamination between the face sheet and core is a major problem in these structures. Many research works are devoted to improve the interfacial fracture toughness of composites majorities of which are on nano and laminated composites. Work on influence of multiwalled carbon nano-tubes (MWCNT) dispersed resin system on interface fracture of glass-epoxy PVC core sandwich composite is extremely limited. Finite element study is followed by experimental investigation on interface fracture toughness of glass-epoxy (G/E) PVC core sandwich composite with and without MWCNT. Results demonstrate an improvement in interface fracture toughness values (Gc) of samples with a certain percentages of MWCNT. In addition, dispersion of MWCNT in epoxy resin through sonication followed by mixing of hardener and vacuum resin infusion (VRI) technology used in this study is an easy and cost effective methodology in comparison to previously adopted other methods limited to laminated composites. The study also identifies the optimum weight percentage of MWCNT addition in the resin system for maximum performance gain in interfacial fracture toughness. The results agree with finite element study, high-resolution transmission electron microscope (HRTEM) analysis and fracture micrograph of field emission scanning electron microscope (FESEM) investigation. Interface fracture toughness (GC) of the DCB sandwich samples is calculated using the compliance calibration (CC) method considering the modification due to shear. Compliance (C) vs. crack length (a) data of modified sandwich DCB specimen is fitted to a power function of crack length. The calculated mean value of the exponent n from the plots of experimental results is 2.22 and is different from the value (n=3) prescribed in ASTM D5528-01for mode 1 fracture toughness of laminate composites (which is the basis for modified compliance calibration method). Differentiating C with respect to crack length (a) and substituting it in the expression GC provides its value. The research demonstrates improvement of 14.4% in peak load carrying capacity and 34.34% in interface fracture toughness GC for samples with 1.5 wt% MWCNT (weight % being taken with respect to weight of resin) in comparison to samples without MWCNT. The paper focuses on significant improvement in experimentally determined interface fracture toughness of sandwich samples with MWCNT over the samples without MWCNT using much simpler method of sonication. Good dispersion of MWCNT was observed in HRTEM with 1.5 wt% MWCNT addition in comparison to other percentages of MWCNT. FESEM studies have also demonstrated good dispersion and fiber bridging of MWCNT in resin system. Ductility is also observed to be higher for samples with MWCNT in comparison to samples without.Keywords: carbon nanotube, epoxy resin, foam, glass fibers, interfacial fracture, sandwich composite
Procedia PDF Downloads 30312916 Classification Using Worldview-2 Imagery of Giant Panda Habitat in Wolong, Sichuan Province, China
Authors: Yunwei Tang, Linhai Jing, Hui Li, Qingjie Liu, Xiuxia Li, Qi Yan, Haifeng Ding
Abstract:
The giant panda (Ailuropoda melanoleuca) is an endangered species, mainly live in central China, where bamboos act as the main food source of wild giant pandas. Knowledge of spatial distribution of bamboos therefore becomes important for identifying the habitat of giant pandas. There have been ongoing studies for mapping bamboos and other tree species using remote sensing. WorldView-2 (WV-2) is the first high resolution commercial satellite with eight Multi-Spectral (MS) bands. Recent studies demonstrated that WV-2 imagery has a high potential in classification of tree species. The advanced classification techniques are important for utilising high spatial resolution imagery. It is generally agreed that object-based image analysis is a more desirable method than pixel-based analysis in processing high spatial resolution remotely sensed data. Classifiers that use spatial information combined with spectral information are known as contextual classifiers. It is suggested that contextual classifiers can achieve greater accuracy than non-contextual classifiers. Thus, spatial correlation can be incorporated into classifiers to improve classification results. The study area is located at Wuyipeng area in Wolong, Sichuan Province. The complex environment makes it difficult for information extraction since bamboos are sparsely distributed, mixed with brushes, and covered by other trees. Extensive fieldworks in Wuyingpeng were carried out twice. The first one was on 11th June, 2014, aiming at sampling feature locations for geometric correction and collecting training samples for classification. The second fieldwork was on 11th September, 2014, for the purposes of testing the classification results. In this study, spectral separability analysis was first performed to select appropriate MS bands for classification. Also, the reflectance analysis provided information for expanding sample points under the circumstance of knowing only a few. Then, a spatially weighted object-based k-nearest neighbour (k-NN) classifier was applied to the selected MS bands to identify seven land cover types (bamboo, conifer, broadleaf, mixed forest, brush, bare land, and shadow), accounting for spatial correlation within classes using geostatistical modelling. The spatially weighted k-NN method was compared with three alternatives: the traditional k-NN classifier, the Support Vector Machine (SVM) method and the Classification and Regression Tree (CART). Through field validation, it was proved that the classification result obtained using the spatially weighted k-NN method has the highest overall classification accuracy (77.61%) and Kappa coefficient (0.729); the producer’s accuracy and user’s accuracy achieve 81.25% and 95.12% for the bamboo class, respectively, also higher than the other methods. Photos of tree crowns were taken at sample locations using a fisheye camera, so the canopy density could be estimated. It is found that it is difficult to identify bamboo in the areas with a large canopy density (over 0.70); it is possible to extract bamboos in the areas with a median canopy density (from 0.2 to 0.7) and in a sparse forest (canopy density is less than 0.2). In summary, this study explores the ability of WV-2 imagery for bamboo extraction in a mountainous region in Sichuan. The study successfully identified the bamboo distribution, providing supporting knowledge for assessing the habitats of giant pandas.Keywords: bamboo mapping, classification, geostatistics, k-NN, worldview-2
Procedia PDF Downloads 31312915 Calculation of Fractal Dimension and Its Relation to Some Morphometric Characteristics of Iranian Landforms
Authors: Mitra Saberi, Saeideh Fakhari, Amir Karam, Ali Ahmadabadi
Abstract:
Geomorphology is the scientific study of the characteristics of form and shape of the Earth's surface. The existence of types of landforms and their variation is mainly controlled by changes in the shape and position of land and topography. In fact, the interest and application of fractal issues in geomorphology is due to the fact that many geomorphic landforms have fractal structures and their formation and transformation can be explained by mathematical relations. The purpose of this study is to identify and analyze the fractal behavior of landforms of macro geomorphologic regions of Iran, as well as studying and analyzing topographic and landform characteristics based on fractal relationships. In this study, using the Iranian digital elevation model in the form of slopes, coefficients of deposition and alluvial fan, the fractal dimensions of the curves were calculated through the box counting method. The morphometric characteristics of the landforms and their fractal dimension were then calculated for 4criteria (height, slope, profile curvature and planimetric curvature) and indices (maximum, Average, standard deviation) using ArcMap software separately. After investigating their correlation with fractal dimension, two-way regression analysis was performed and the relationship between fractal dimension and morphometric characteristics of landforms was investigated. The results show that the fractal dimension in different pixels size of 30, 90 and 200m, topographic curves of different landform units of Iran including mountain, hill, plateau, plain of Iran, from1.06in alluvial fans to1.17in The mountains are different. Generally, for all pixels of different sizes, the fractal dimension is reduced from mountain to plain. The fractal dimension with the slope criterion and the standard deviation index has the highest correlation coefficient, with the curvature of the profile and the mean index has the lowest correlation coefficient, and as the pixels become larger, the correlation coefficient between the indices and the fractal dimension decreases.Keywords: box counting method, fractal dimension, geomorphology, Iran, landform
Procedia PDF Downloads 8312914 Photocatalytic Properties of Pt/Er-KTaO3
Authors: Anna Krukowska, Tomasz Klimczuk, Adriana Zaleska-Medynska
Abstract:
Photoactive materials have attracted attention due to their potential application in the degradation of environmental pollutants to non-hazardous compounds in an eco-friendly route. Among semiconductor photocatalysts, tantalates such as potassium tantalate (KTaO3) is one of the excellent functional photomaterial. However, tantalates-based materials are less active under visible-light irradiation, the enhancement in photoactivity could be improved with the modification of opto-eletronic properties of KTaO3 by doping rare earth metal (Er) and further photodeposition of noble metal nanoparticles (Pt). Inclusion of rare earth element in orthorhombic structure of tantalate can generate one high-energy photon by absorbing two or more incident low-energy photons, which convert visible-light and infrared-light into the ultraviolet-light to satisfy the requirement of KTaO3 photocatalysts. On the other hand, depositions of noble metal nanoparticles on the surface of semiconductor strongly absorb visible-light due to their surface plasmon resonance, in which their conducting electrons undergo a collective oscillation induced by electric field of visible-light. Furthermore, the high dispersion of Pt nanoparticles, which will be obtained by photodeposition process is additional important factor to improve the photocatalytic activity. The present work is aimed to study the effect of photocatalytic process of the prepared Er-doped KTaO3 and further incorporation of Pt nanoparticles by photodeposition. Moreover, the research is also studied correlations between photocatalytic activity and physico-chemical properties of obtained Pt/Er-KTaO3 samples. The Er-doped KTaO3 microcomposites were synthesized by a hydrothermal method. Then photodeposition method was used for Pt loading over Er-KTaO3. The structural and optical properties of Pt/Er-KTaO3 photocatalytic were characterized using scanning electron microscope (SEM), X-ray diffraction (XRD), volumetric adsorption method (BET), UV-Vis absorption measurement, Raman spectroscopy and luminescence spectroscopy. The photocatalytic properties of Pt/Er-KTaO3 microcomposites were investigated by degradation of phenol in aqueous phase as model pollutant under visible and ultraviolet-light irradiation. Results of this work show that all the prepared photocatalysis exhibit low BET surface area, although doping of the bare KTaO3 with rare earth element (Er) presents a slight increase in this value. The crystalline structure of Pt/Er-KTaO3 powders exhibited nearly identical positions for the main peak at about 22,8o and the XRD pattern could be assigned to an orthorhombic distorted perovskite structure. The Raman spectra of obtained semiconductors confirmed demonstrating perovskite-like structure. The optical absorption spectra of Pt nanoparticles exhibited plasmon absorption band for main peaks at about 216 and 264 nm. The addition of Pt nanoparticles increased photoactivity compared to Er-KTaO3 and pure KTaO3. Summary optical properties of KTaO3 change with its doping Er-element and further photodeposition of Pt nanoparticles.Keywords: heterogeneous photocatalytic, KTaO3 photocatalysts, Er3+ ion doping, Pt photodeposition
Procedia PDF Downloads 36012913 The Contribution of Sanitation Practices to Marine Pollution and the Prevalence of Water-Borne Diseases in Prampram Coastal Area, Greater Accra-Ghana
Authors: Precious Roselyn Obuobi
Abstract:
Background: In Ghana, water-borne diseases remain a public health concern due to its impact. While marine pollution has been linked to outbreak of diseases especially in communities along the coast, associated risks such as oil spillage, marine debris, erosion, improper waste disposal and management practices persist. Objective: The study seeks to investigate sanitation practices that contribute to marine pollution in Prampram and the prevalence of selected water-borne diseases (diarrhea and typhoid fever). Method: This study used a descriptive cross-sectional design, employing the mix-method (qualitative and quantitative) approach. Twenty-two (22) participants were selected and semistructured questionnaire were administered to them. Additionally, interviews were conducted to collect more information. Further, an observation check-list was used to aid the data collection process. Secondary data comprising information on water-borne diseases in the district was acquired from the district health directorate to determine the prevalence of selected water-borne diseases in the community. Data Analysis: The qualitative data was analyzed using NVIVO® software by adapting the six steps thematic analysis by Braun and Clarke whiles STATA® version 16 was used to analyze the secondary data collected from the district health directorate. A descriptive statistic employed using mean, standard deviation, frequencies and proportions were used to summarize the results. Results: The results showed that open defecation and indiscriminate waste disposal were the main practices contributing to marine pollution in Prampram and its effect on public health. Conclusion: These findings have implications on public health and the environment, thus effort needs to be stepped up in educating the community on best sanitation practices.Keywords: environment, sanitation, marine pollution, water-borne diseases
Procedia PDF Downloads 7612912 Micromechanical Modelling of Ductile Damage with a Cohesive-Volumetric Approach
Authors: Noe Brice Nkoumbou Kaptchouang, Pierre-Guy Vincent, Yann Monerie
Abstract:
The present work addresses the modelling and the simulation of crack initiation and propagation in ductile materials which failed by void nucleation, growth, and coalescence. One of the current research frameworks on crack propagation is the use of cohesive-volumetric approach where the crack growth is modelled as a decohesion of two surfaces in a continuum material. In this framework, the material behavior is characterized by two constitutive relations, the volumetric constitutive law relating stress and strain, and a traction-separation law across a two-dimensional surface embedded in the three-dimensional continuum. Several cohesive models have been proposed for the simulation of crack growth in brittle materials. On the other hand, the application of cohesive models in modelling crack growth in ductile material is still a relatively open field. One idea developed in the literature is to identify the traction separation for ductile material based on the behavior of a continuously-deforming unit cell failing by void growth and coalescence. Following this method, the present study proposed a semi-analytical cohesive model for ductile material based on a micromechanical approach. The strain localization band prior to ductile failure is modelled as a cohesive band, and the Gurson-Tvergaard-Needleman plasticity model (GTN) is used to model the behavior of the cohesive band and derived a corresponding traction separation law. The numerical implementation of the model is realized using the non-smooth contact method (NSCD) where cohesive models are introduced as mixed boundary conditions between each volumetric finite element. The present approach is applied to the simulation of crack growth in nuclear ferritic steel. The model provides an alternative way to simulate crack propagation using the numerical efficiency of cohesive model with a traction separation law directly derived from porous continuous model.Keywords: ductile failure, cohesive model, GTN model, numerical simulation
Procedia PDF Downloads 14912911 Application of a Lighting Design Method Using Mean Room Surface Exitance
Authors: Antonello Durante, James Duff, Kevin Kelly
Abstract:
The visual needs of people in modern work based buildings are changing. Self-illuminated screens of computers, televisions, tablets and smart phones have changed the relationship between people and the lit environment. In the past, lighting design practice was primarily based on providing uniform horizontal illuminance on the working plane, but this has failed to ensure good quality lit environments. Lighting standards of today continue to be set based upon a 100 year old approach that at its core, considers the task illuminance of the utmost importance, with this task typically being located on a horizontal plane. An alternative method focused on appearance has been proposed, as opposed to the traditional performance based approach. Mean Room Surface Exitance (MRSE) and Target-Ambient Illuminance Ratio (TAIR) are two new metrics proposed to assess illumination adequacy in interiors. The hypothesis is that these factors will be superior to the existing metrics used, which are horizontal illuminance led. For the six past years, research has examined this, within the Dublin Institute of Technology, with a view to determining the suitability of this approach for application to general lighting practice. Since the start of this research, a number of key findings have been produced that centered on how occupants will react to various levels of MRSE. This paper provides a broad update on how this research has progressed. More specifically, this paper will: i) Demonstrate how MRSE can be measured using HDR images technology, ii) Illustrate how MRSE can be calculated using scripting and an open source lighting computation engine, iii) Describe experimental results that demonstrate how occupants have reacted to various levels of MRSE within experimental office environments.Keywords: illumination hierarchy (IH), mean room surface exitance (MRSE), perceived adequacy of illumination (PAI), target-ambient illumination ratio (TAIR)
Procedia PDF Downloads 18712910 Review of Concepts and Tools Applied to Assess Risks Associated with Food Imports
Authors: A. Falenski, A. Kaesbohrer, M. Filter
Abstract:
Introduction: Risk assessments can be performed in various ways and in different degrees of complexity. In order to assess risks associated with imported foods additional information needs to be taken into account compared to a risk assessment on regional products. The present review is an overview on currently available best practise approaches and data sources used for food import risk assessments (IRAs). Methods: A literature review has been performed. PubMed was searched for articles about food IRAs published in the years 2004 to 2014 (English and German texts only, search string “(English [la] OR German [la]) (2004:2014 [dp]) import [ti] risk”). Titles and abstracts were screened for import risks in the context of IRAs. The finally selected publications were analysed according to a predefined questionnaire extracting the following information: risk assessment guidelines followed, modelling methods used, data and software applied, existence of an analysis of uncertainty and variability. IRAs cited in these publications were also included in the analysis. Results: The PubMed search resulted in 49 publications, 17 of which contained information about import risks and risk assessments. Within these 19 cross references were identified to be of interest for the present study. These included original articles, reviews and guidelines. At least one of the guidelines of the World Organisation for Animal Health (OIE) and the Codex Alimentarius Commission were referenced in any of the IRAs, either for import of animals or for imports concerning foods, respectively. Interestingly, also a combination of both was used to assess the risk associated with the import of live animals serving as the source of food. Methods ranged from full quantitative IRAs using probabilistic models and dose-response models to qualitative IRA in which decision trees or severity tables were set up using parameter estimations based on expert opinions. Calculations were done using @Risk, R or Excel. Most heterogeneous was the type of data used, ranging from general information on imported goods (food, live animals) to pathogen prevalence in the country of origin. These data were either publicly available in databases or lists (e.g., OIE WAHID and Handystatus II, FAOSTAT, Eurostat, TRACES), accessible on a national level (e.g., herd information) or only open to a small group of people (flight passenger import data at national airport customs office). In the IRAs, an uncertainty analysis has been mentioned in some cases, but calculations have been performed only in a few cases. Conclusion: The current state-of-the-art in the assessment of risks of imported foods is characterized by a great heterogeneity in relation to general methodology and data used. Often information is gathered on a case-by-case basis and reformatted by hand in order to perform the IRA. This analysis therefore illustrates the need for a flexible, modular framework supporting the connection of existing data sources with data analysis and modelling tools. Such an infrastructure could pave the way to IRA workflows applicable ad-hoc, e.g. in case of a crisis situation.Keywords: import risk assessment, review, tools, food import
Procedia PDF Downloads 30212909 Investigating the Relationship Between Alexithymia and Mobile Phone Addiction Along with the Mediating Role of Anxiety, Stress and Depression: A Path Analysis Study and Structural Model Testing
Authors: Pouriya Darabiyan, Hadis Nazari, Kourosh Zarea, Saeed Ghanbari, Zeinab Raiesifar, Morteza Khafaie, Hanna Tuvesson
Abstract:
Introduction Since the beginning of mobile phone addiction, alexithymia, depression, anxiety and stress have been stated as risk factors for Internet addiction, so this study was conducted with the aim of investigating the relationship between Alexithymia and Mobile phone addiction along with the mediating role of anxiety, stress and depression. Materials and methods In this descriptive-analytical and cross-sectional study in 2022, 412 students School of Nursing & Midwifery of Ahvaz Jundishapur University of Medical Sciences were included in the study using available sampling method. Data collection tools were: Demographic Information Questionnaire, Toronto Alexithymia Scale (TAS-20), Depression, Anxiety, Stress Scale (DASS-21) and Mobile Phone Addiction Index (MPAI). Frequency, Pearson correlation coefficient test and linear regression were used to describe and analyze the data. Also, structural equation models and path analysis method were used to investigate the direct and indirect effects as well as the total effect of each dimension of Alexithymia on Mobile phone addiction with the mediating role of stress, depression and anxiety. Statistical analysis was done by SPSS version 22 and Amos version 16 software. Results Alexithymia was a predictive factor for mobile phone addiction. Also, Alexithymia had a positive and significant effect on depression, anxiety and stress. Depression, anxiety and stress had a positive and significant effect on mobile phone addiction. Depression, anxiety and stress variables played the role of a relative mediating variable between Alexithymia and mobile phone addiction. Alexithymia through depression, anxiety and stress also has an indirect effect on Internet addiction. Conclusion Alexithymia is a predictive factor for mobile phone addiction; And the variables of depression, anxiety and stress play the role of a relative mediating variable between Alexithymia and mobile phone addiction.Keywords: alexithymia, mobile phone, depression, anxiety, stress
Procedia PDF Downloads 9912908 Shear Strength and Consolidation Behavior of Clayey Soil with Vertical and Radial Drainage
Authors: R. Pillai Aparna, S. R. Gandhi
Abstract:
Soft clay deposits having low strength and high compressibility are found all over the world. Preloading with vertical drains is a widely used method for improving such type of soils. The coefficient of consolidation, irrespective of the drainage type, plays an important role in the design of vertical drains and it controls accurate prediction of the rate of consolidation of soil. Also, the increase in shear strength of soil with consolidation is another important factor considered in preloading or staged construction. To our best knowledge no clear guidelines are available to estimate the increase in shear strength for a particular degree of consolidation (U) at various stages during the construction. Various methods are available for finding out the consolidation coefficient. This study mainly focuses on the variation of, consolidation coefficient which was found out using different methods and shear strength with pressure intensity. The variation of shear strength with the degree of consolidation was also studied. The consolidation test was done using two types of highly compressible clays with vertical, radial and a few with combined drainage. The test was carried out at different pressures intensities and for each pressure intensity, once the target degree of consolidation is achieved, vane shear test was done at different locations in the sample, in order to determine the shear strength. The shear strength of clayey soils under the application of vertical stress with vertical and radial drainage with target U value of 70% and 90% was studied. It was found that there is not much variation in cv or cr value beyond 80kPa pressure intensity. Correlations were developed between shear strength ratio and consolidation pressure based on laboratory testing under controlled condition. It was observed that the shear strength of sample with target U value of 90% is about 1.4 to 2 times than that of 70% consolidated sample. Settlement analysis was done using Asaoka’s and hyperbolic method. The variation of strength with respect to the depth of sample was also studied, using large-scale consolidation test. It was found, based on the present study that the gain in strength is more on the top half of the clay layer, and also the shear strength of the sample ensuring radial drainage is slightly higher than that of the vertical drainage.Keywords: consolidation coefficient, degree of consolidation, PVDs, shear strength
Procedia PDF Downloads 23912907 Semantic Search Engine Based on Query Expansion with Google Ranking and Similarity Measures
Authors: Ahmad Shahin, Fadi Chakik, Walid Moudani
Abstract:
Our study is about elaborating a potential solution for a search engine that involves semantic technology to retrieve information and display it significantly. Semantic search engines are not used widely over the web as the majorities are still in Beta stage or under construction. Many problems face the current applications in semantic search, the major problem is to analyze and calculate the meaning of query in order to retrieve relevant information. Another problem is the ontology based index and its updates. Ranking results according to concept meaning and its relation with query is another challenge. In this paper, we are offering a light meta-engine (QESM) which uses Google search, and therefore Google’s index, with some adaptations to its returned results by adding multi-query expansion. The mission was to find a reliable ranking algorithm that involves semantics and uses concepts and meanings to rank results. At the beginning, the engine finds synonyms of each query term entered by the user based on a lexical database. Then, query expansion is applied to generate different semantically analogous sentences. These are generated randomly by combining the found synonyms and the original query terms. Our model suggests the use of semantic similarity measures between two sentences. Practically, we used this method to calculate semantic similarity between each query and the description of each page’s content generated by Google. The generated sentences are sent to Google engine one by one, and ranked again all together with the adapted ranking method (QESM). Finally, our system will place Google pages with higher similarities on the top of the results. We have conducted experimentations with 6 different queries. We have observed that most ranked results with QESM were altered with Google’s original generated pages. With our experimented queries, QESM generates frequently better accuracy than Google. In some worst cases, it behaves like Google.Keywords: semantic search engine, Google indexing, query expansion, similarity measures
Procedia PDF Downloads 42512906 Hypoglycemic Activity studies on Root Extracts of Sanseviera liberica Root in Streptozotocin-Induced Diabetic Rats
Authors: Omowunmi Amao
Abstract:
Sansevieria liberica belongs to the family Agavaceae (Ruscaceae or Dracaenaceae). They are widely distributed throughout the tropics. Literature review suggests that in Nigeria, the leaves and roots of Sansevieria liberica are used in traditional medicine for the treatment of asthma, abdominal pains, colic, diarrhea, eczema, gonorrhea, hemorrhoids, hypertension, monorrhagia, piles, sexual weakness, snake bites, and wounds of the foot. In this context, the standardized Methanolic extract of roots of Sansevieria liberica is hypothesized for the evaluation of the hypoglycemic activity. Material and Methods: Inbreed adult male sprague-Dawley albino rats were used in the experiment. The suspension of standardized Methanol extract (ME) of Sansevieria liberica was treated for hypoglycemic activity in oral glucose tolerance test (OGTT) method. The suspension of standardized Methanolic extract (ME) of Sansevieria liberica was also treated for hypoglycemic activity in streptozotocin-induced diabetic rats. Results: The Methanolic extract (ME) of Sanseviera liberica root (100 mg/kg, 200mg/kg, and 400 mg/kg) showed potential hypoglycemic activity in diabetic rats, and further in OGTT method. Furthermore, Methanolic extract of Sanseviera liberica root showed significant (P<0.05) increase in final body weight, total hemoglobin, insulin, albumin and high-density lipoprotein levels, however, decrease in fluid intake, glycosylated hemoglobin, urea, creatinine, total cholesterol, triglyceride and low-density lipoprotein levels. Additionally, it improved oxidative stress in terms of reducing lipid peroxidase and superoxide dismutase, and elevating catalase activity. Conclusions: These findings suggest that the Methanolic extract of Sanseviera liberica root was found to be potential hypoglycemic, and would be a promising candidate for the treatment of diabetes.Keywords: diabetes, Sanseviera liberica, hypoglycemic activity, diabetes and metabolism
Procedia PDF Downloads 36412905 Sustainable Technologies for Decommissioning of Nuclear Facilities
Authors: Ahmed Stifi, Sascha Gentes
Abstract:
The German nuclear industry, while implementing the German policy, believes that the journey towards the green-field, namely phasing out of nuclear energy, should be achieved through green techniques. The most important techniques required for the wide range of decommissioning activities are decontamination techniques, cutting techniques, radioactivity measuring techniques, remote control techniques, techniques for worker and environmental protection and techniques for treating, preconditioning and conditioning nuclear waste. Many decontamination techniques are used for removing contamination from metal, concrete or other surfaces like the scales inside pipes. As the pipeline system is one of the important components of nuclear power plants, the process of decontamination in tubing is of more significance. The development of energy sectors like oil sector, gas sector and nuclear sector, since the middle of 20th century, increased the pipeline industry and the research in the decontamination of tubing in each sector is found to serve each other. The extraction of natural products and material through the pipeline can result in scale formation. These scales can be radioactively contaminated through an accumulation process especially in the petrochemical industry when oil and gas are extracted from the underground reservoir. The radioactivity measured in these scales can be significantly high and pose a great threat to people and the environment. At present, the decontamination process involves using high pressure water jets with or without abrasive material and this technology produces a high amount of secondary waste. In order to overcome it, the research team within Karlsruhe Institute of Technology developed a new sustainable method to carry out the decontamination of tubing without producing any secondary waste. This method is based on vibration technique which removes scales and also does not require any auxiliary materials. The outcome of the research project proves that the vibration technique used for decontamination of tubing is environmental friendly in other words a sustainable technique.Keywords: sustainable technologies, decontamination, pipeline, nuclear industry
Procedia PDF Downloads 30312904 Degree of Approximation of Functions by Product Means
Authors: Hare Krishna Nigam
Abstract:
In this paper, for the first time, (E,q)(C,2) product summability method is introduced and two quite new results on degree of approximation of the function f belonging to Lip (alpha,r)class and W(L(r), xi(t)) class by (E,q)(C,2) product means of Fourier series, has been obtained.Keywords: Degree of approximation, (E, q)(C, 2) means, Fourier series, Lebesgue integral, Lip (alpha, r)class, W(L(r), xi(t))class of functions
Procedia PDF Downloads 51812903 Procedure to Optimize the Performance of Chemical Laser Using the Genetic Algorithm Optimizations
Authors: Mohammedi Ferhate
Abstract:
This work presents details of the study of the entire flow inside the facility where the exothermic chemical reaction process in the chemical laser cavity is analyzed. In our paper we will describe the principles of chemical lasers where flow reversal is produced by chemical reactions. We explain the device for converting chemical potential energy laser energy. We see that the phenomenon thus has an explosive trend. Finally, the feasibility and effectiveness of the proposed method is demonstrated by computer simulationKeywords: genetic, lasers, nozzle, programming
Procedia PDF Downloads 9412902 Ethanol in Carbon Monoxide Intoxication: Focus on Delayed Neuropsychological Sequelae
Authors: Hyuk-Hoon Kim, Young Gi Min
Abstract:
Background: In carbon monoxide (CO) intoxication, the pathophysiology of delayed neurological sequelae (DNS) is very complex and remains poorly understood. And predicting whether patients who exhibit resolved acute symptoms have escaped or will experience DNS represents a very important clinical issue. Brain magnetic resonance (MR) imaging has been conducted to assess the severity of brain damage as an objective method to predict prognosis. And co-ingestion of a second poison in patients with intentional CO poisoning occurs in almost one-half of patients. Among patients with co-ingestions, 66% ingested ethanol. We assessed the effects of ethanol on neurologic sequelae prevalence in acute CO intoxication by means of abnormal lesion in brain MR. Method: This study was conducted retrospectively by collecting data for patients who visited an emergency medical center during a period of 5 years. The enrollment criteria were diagnosis of acute CO poisoning and the measurement of the serum ethanol level and history of taking a brain MR during admission period. Official readout data by radiologist are used to decide whether abnormal lesion is existed or not. The enrolled patients were divided into two groups: patients with abnormal lesion and without abnormal lesion in Brain MR. A standardized extraction using medical record was performed; Mann Whitney U test and logistic regression analysis were performed. Result: A total of 112 patients were enrolled, and 68 patients presented abnormal brain lesion on MR. The abnormal brain lesion group had lower serum ethanol level (mean, 20.14 vs 46.71 mg/dL) (p-value<0.001). In addition, univariate logistic regression analysis showed the serum ethanol level (OR, 0.99; 95% CI, 0.98 -1.00) was independently associated with the development of abnormal lesion in brain MR. Conclusion: Ethanol could have neuroprotective effect in acute CO intoxication by sedative effect in stressful situation and mitigative effect in neuro-inflammatory reaction.Keywords: carbon monoxide, delayed neuropsychological sequelae, ethanol, intoxication, magnetic resonance
Procedia PDF Downloads 25212901 An Algebraic Geometric Imaging Approach for Automatic Dairy Cow Body Condition Scoring System
Authors: Thi Thi Zin, Pyke Tin, Ikuo Kobayashi, Yoichiro Horii
Abstract:
Today dairy farm experts and farmers have well recognized the importance of dairy cow Body Condition Score (BCS) since these scores can be used to optimize milk production, managing feeding system and as an indicator for abnormality in health even can be utilized to manage for having healthy calving times and process. In tradition, BCS measures are done by animal experts or trained technicians based on visual observations focusing on pin bones, pin, thurl and hook area, tail heads shapes, hook angles and short and long ribs. Since the traditional technique is very manual and subjective, the results can lead to different scores as well as not cost effective. Thus this paper proposes an algebraic geometric imaging approach for an automatic dairy cow BCS system. The proposed system consists of three functional modules. In the first module, significant landmarks or anatomical points from the cow image region are automatically extracted by using image processing techniques. To be specific, there are 23 anatomical points in the regions of ribs, hook bones, pin bone, thurl and tail head. These points are extracted by using block region based vertical and horizontal histogram methods. According to animal experts, the body condition scores depend mainly on the shape structure these regions. Therefore the second module will investigate some algebraic and geometric properties of the extracted anatomical points. Specifically, the second order polynomial regression is employed to a subset of anatomical points to produce the regression coefficients which are to be utilized as a part of feature vector in scoring process. In addition, the angles at thurl, pin, tail head and hook bone area are computed to extend the feature vector. Finally, in the third module, the extracted feature vectors are trained by using Markov Classification process to assign BCS for individual cows. Then the assigned BCS are revised by using multiple regression method to produce the final BCS score for dairy cows. In order to confirm the validity of proposed method, a monitoring video camera is set up at the milk rotary parlor to take top view images of cows. The proposed method extracts the key anatomical points and the corresponding feature vectors for each individual cows. Then the multiple regression calculator and Markov Chain Classification process are utilized to produce the estimated body condition score for each cow. The experimental results tested on 100 dairy cows from self-collected dataset and public bench mark dataset show very promising with accuracy of 98%.Keywords: algebraic geometric imaging approach, body condition score, Markov classification, polynomial regression
Procedia PDF Downloads 15812900 Thermosonic Devulcanization of Waste Ground Rubber Tires by Quaternary Ammonium-Based Ternary Deep Eutectic Solvents and the Effect of α-Hydrogen
Authors: Ricky Saputra, Rashmi Walvekar, Mohammad Khalid
Abstract:
Landfills, water contamination, and toxic gas emission are a few impacts faced by the environment due to the increasing number of αof waste rubber tires (WRT). In spite of such concerning issue, only minimal efforts are taken to reclaim or recycle these wastes as their products are generally not-profitable for companies. Unlike the typical reclamation process, devulcanization is a method to selectively cleave sulfidic bonds within vulcanizates to avoid polymeric scissions that compromise elastomer’s mechanical and tensile properties. The process also produces devulcanizates that are re-processable similar to virgin rubber. Often, a devulcanizing agent is needed. In the current study, novel and sustainable ammonium chloride-based ternary deep eutectic solvents (TDES), with a different number of α-hydrogens, were utilised to devulcanize ground rubber tire (GRT) as an effort to implement green chemistry to tackle such issue. 40-mesh GRT were soaked for 1 day with different TDESs and sonicated at 37-80 kHz for 60-120 mins and heated at 100-140oC for 30-90 mins. Devulcanizates were then filtered, dried, and evaluated based on the percentage of by means of Flory-Rehner calculation and swelling index. The result shows that an increasing number of α-Hs increases the degree of devulcanization, and the value achieved was around eighty-percent, thirty percent higher than the typical industrial-autoclave method. Resulting bondages of devulcanizates were also analysed by Fourier transform infrared spectrometer (FTIR), Horikx fitting, and thermogravimetric analyser (TGA). The earlier two confirms only sulfidic scissions were experienced by GRT through the treatment, while the latter proves the absence or negligibility of carbon-chains scission.Keywords: ammonium, sustainable, deep eutectic solvent, α-hydrogen, waste rubber tire
Procedia PDF Downloads 12712899 Measurement of CES Production Functions Considering Energy as an Input
Authors: Donglan Zha, Jiansong Si
Abstract:
Because of its flexibility, CES attracts much interest in economic growth and programming models, and the macroeconomics or micro-macro models. This paper focuses on the development, estimating methods of CES production function considering energy as an input. We leave for future research work of relaxing the assumption of constant returns to scale, the introduction of potential input factors, and the generalization method of the optimal nested form of multi-factor production functions.Keywords: bias of technical change, CES production function, elasticity of substitution, energy input
Procedia PDF Downloads 28212898 Analysis of Overall Thermo-Elastic Properties of Random Particulate Nanocomposites with Various Interphase Models
Authors: Lidiia Nazarenko, Henryk Stolarski, Holm Altenbach
Abstract:
In the paper, a (hierarchical) approach to analysis of thermo-elastic properties of random composites with interphases is outlined and illustrated. It is based on the statistical homogenization method – the method of conditional moments – combined with recently introduced notion of the energy-equivalent inhomogeneity which, in this paper, is extended to include thermal effects. After exposition of the general principles, the approach is applied in the investigation of the effective thermo-elastic properties of a material with randomly distributed nanoparticles. The basic idea of equivalent inhomogeneity is to replace the inhomogeneity and the surrounding it interphase by a single equivalent inhomogeneity of constant stiffness tensor and coefficient of thermal expansion, combining thermal and elastic properties of both. The equivalent inhomogeneity is then perfectly bonded to the matrix which allows to analyze composites with interphases using techniques devised for problems without interphases. From the mechanical viewpoint, definition of the equivalent inhomogeneity is based on Hill’s energy equivalence principle, applied to the problem consisting only of the original inhomogeneity and its interphase. It is more general than the definitions proposed in the past in that, conceptually and practically, it allows to consider inhomogeneities of various shapes and various models of interphases. This is illustrated considering spherical particles with two models of interphases, Gurtin-Murdoch material surface model and spring layer model. The resulting equivalent inhomogeneities are subsequently used to determine effective thermo-elastic properties of randomly distributed particulate composites. The effective stiffness tensor and coefficient of thermal extension of the material with so defined equivalent inhomogeneities are determined by the method of conditional moments. Closed-form expressions for the effective thermo-elastic parameters of a composite consisting of a matrix and randomly distributed spherical inhomogeneities are derived for the bulk and the shear moduli as well as for the coefficient of thermal expansion. Dependence of the effective parameters on the interphase properties is included in the resulting expressions, exhibiting analytically the nature of the size-effects in nanomaterials. As a numerical example, the epoxy matrix with randomly distributed spherical glass particles is investigated. The dependence of the effective bulk and shear moduli, as well as of the effective thermal expansion coefficient on the particle volume fraction (for different radii of nanoparticles) and on the radius of nanoparticle (for fixed volume fraction of nanoparticles) for different interphase models are compared to and discussed in the context of other theoretical predictions. Possible applications of the proposed approach to short-fiber composites with various types of interphases are discussed.Keywords: effective properties, energy equivalence, Gurtin-Murdoch surface model, interphase, random composites, spherical equivalent inhomogeneity, spring layer model
Procedia PDF Downloads 18512897 Passenger Preferences on Airline Check-In Methods: Traditional Counter Check-In Versus Common-Use Self-Service Kiosk
Authors: Cruz Queen Allysa Rose, Bautista Joymeeh Anne, Lantoria Kaye, Barretto Katya Louise
Abstract:
The study presents the preferences of passengers on the quality of service provided by the two airline check-in methods currently present in airports-traditional counter check-in and common-use self-service kiosks. Since a study has shown that airlines perceive self-service kiosks alone are sufficient enough to ensure adequate services and customer satisfaction, and in contrast, agents and passengers stated that it alone is not enough and that human interaction is essential. In reference with former studies that established opposing ideas about the choice of the more favorable airline check-in method to employ, it is the purpose of this study to present a recommendation that shall somehow fill-in the gap between the conflicting ideas by means of comparing the perceived quality of service through the RATER model. Furthermore, this study discusses the major competencies present in each method which are supported by the theories–FIRO Theory of Needs upholding the importance of inclusion, control and affection, and the Queueing Theory which points out the discipline of passengers and the length of the queue line as important factors affecting quality service. The findings of the study were based on the data gathered by the researchers from selected Thomasian third year and fourth year college students currently enrolled in the first semester of the academic year 2014-2015, who have already experienced both airline check-in methods through the implication of a stratified probability sampling. The statistical treatments applied in order to interpret the data were mean, frequency, standard deviation, t-test, logistic regression and chi-square test. The final point of the study revealed that there is a greater effect in passenger preference concerning the satisfaction experienced in common-use self-service kiosks in comparison with the application of the traditional counter check-in.Keywords: traditional counter check-in, common-use self-service Kiosks, airline check-in methods
Procedia PDF Downloads 40612896 E-Learning Platform for School Kids
Authors: Gihan Thilakarathna, Fernando Ishara, Rathnayake Yasith, Bandara A. M. R. Y.
Abstract:
E-learning is a crucial component of intelligent education. Even in the midst of a pandemic, E-learning is becoming increasingly important in the educational system. Several e-learning programs are accessible for students. Here, we decided to create an e-learning framework for children. We've found a few issues that teachers are having with their online classes. When there are numerous students in an online classroom, how does a teacher recognize a student's focus on academics and below-the-surface behaviors? Some kids are not paying attention in class, and others are napping. The teacher is unable to keep track of each and every student. Key challenge in e-learning is online exams. Because students can cheat easily during online exams. Hence there is need of exam proctoring is occurred. In here we propose an automated online exam cheating detection method using a web camera. The purpose of this project is to present an E-learning platform for math education and include games for kids as an alternative teaching method for math students. The game will be accessible via a web browser. The imagery in the game is drawn in a cartoonish style. This will help students learn math through games. Everything in this day and age is moving towards automation. However, automatic answer evaluation is only available for MCQ-based questions. As a result, the checker has a difficult time evaluating the theory solution. The current system requires more manpower and takes a long time to evaluate responses. It's also possible to mark two identical responses differently and receive two different grades. As a result, this application employs machine learning techniques to provide an automatic evaluation of subjective responses based on the keyword provided to the computer as student input, resulting in a fair distribution of marks. In addition, it will save time and manpower. We used deep learning, machine learning, image processing and natural language technologies to develop these research components.Keywords: math, education games, e-learning platform, artificial intelligence
Procedia PDF Downloads 156