Search results for: comprehensible input
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 2173

Search results for: comprehensible input

133 A Valid Professional Development Framework For Supporting Science Teachers In Relation To Inquiry-Based Curriculum Units

Authors: Fru Vitalis Akuma, Jenna Koenen

Abstract:

The science education community is increasingly calling for learning experiences that mirror the work of scientists. Although inquiry-based science education is aligned with these calls, the implementation of this strategy is a complex and daunting task for many teachers. Thus, policymakers and researchers have noted the need for continued teacher Professional Development (PD) in the enactment of inquiry-based science education, coupled with effective ways of reaching the goals of teacher PD. This is a complex problem for which educational design research is suitable. The purpose at this stage of our design research is to develop a generic PD framework that is valid as the blueprint of a PD program for supporting science teachers in relation to inquiry-based curriculum units. The seven components of the framework are the goal, learning theory, strategy, phases, support, motivation, and an instructional model. Based on a systematic review of the literature on effective (science) teacher PD, coupled with developer screening, we have generated a design principle per component of the PD framework. For example, as per the associated design principle, the goal of the framework is to provide science teachers with experiences in authentic inquiry, coupled with enhancing their competencies linked to the adoption, customization and design; then the classroom implementation and the revision of inquiry-based curriculum units. The seven design principles have allowed us to synthesize the PD framework, which, coupled with the design principles, are the preliminary outcomes of the current research. We are in the process of evaluating the content and construct validity of the framework, based on nine one-on-one interviews with experts in inquiry-based classroom and teacher learning. To this end, we have developed an interview protocol with the input of eight such experts in South Africa and Germany. Using the protocol, the expert appraisal of the PD framework will involve three experts from Germany, South Africa, and Cameroon, respectively. These countries, where we originate and/or work, provide a variety of inquiry-based science education contexts, making the countries suitable in the evaluation of the generic PD framework. Based on the evaluation, we will revise the framework and its seven design principles to arrive at the final outcomes of the current research. While the final content and construct a valid version of the framework will serve as an example of the needed ways through which effective inquiry-based science teacher PD may be achieved, the final design principles will be useful to researchers when transforming the framework for use in any specific educational context. For example, in our further research, we will transform the framework to one that is practical and effective in supporting inquiry-based practical work in resource-constrained physical sciences classrooms in South Africa. Researchers in other educational contexts may similarly consider the final framework and design principles in their work. Thus, our final outcomes will inform practice and research around the support of teachers to increase the incorporation of learning experiences that mirror the work of scientists in a worldwide manner.

Keywords: design principles, educational design research, evaluation, inquiry-based science education, professional development framework

Procedia PDF Downloads 129
132 Relationship between Thumb Length and Pointing Performance on Portable Terminal with Touch-Sensitive Screen

Authors: Takahiro Nishimura, Kouki Doi, Hiroshi Fujimoto

Abstract:

Touch-sensitive screens that serve as displays and input devices have been adopted in many portable terminals such as smartphones and personal media players, and the market of touch-sensitive screens has expanded greatly. One of the advantages of touch-sensitive screen is the flexibility in the graphical user interface (GUI) design, and it is imperative to design an appropriate GUI to realize an easy-to-use interface. Moreover, it is important to evaluate the relationship between pointing performance and GUI design. There is much knowledge regarding easy-to-use GUI designs for portable terminals with touch-sensitive screens, and most have focused on GUI design approaches for women or children with small hands. In contrast, GUI design approaches for users with large hands have not received sufficient attention. In this study, to obtain knowledge that contributes to the establishment of individualized easy-to-use GUI design guidelines, we conducted experiments to investigate the relationship between thumb length and pointing performance on portable terminals with touch-sensitive screens. In this study, fourteen college students who participated in the experiment were divided into two groups based on the length of their thumbs. Specifically, we categorized the participants into two groups, thumbs longer than 64.2 mm into L (Long) group, and thumbs longer than 57.4 mm but shorter than 64.2 mm into A (Average) group, based on Japanese anthropometric database. They took part in this study under the authorization of Waseda University’s ‘Ethics Review Committee on Research with Human Subjects’. We created an application for the experimental task and implemented it on the projected capacitive touch-sensitive screen portable terminal (iPod touch (4th generation)). The display size was 3.5 inch and 960 × 640 - pixel resolution at 326 ppi (pixels per inch). This terminal was selected as the experimental device, because of its wide use and market share. The operational procedure of the application is as follows. First, the participants placed their thumb on the start position. Then, one cross-shaped target in a 10 × 7 array of 70 positions appeared at random. The participants pointed the target with their thumb as accurately and as fast as possible. Then, they returned their thumb to the start position and waited. The operation ended when this procedure had been repeated until all 70 targets had each been pointed at once by the participants. We adopted the evaluation indices for absolute error, variable error, and pointing time to investigate pointing performance when using the portable terminal. The results showed that pointing performance varied with thumb length. In particular, on the lower right side of the screen, the performance of L group with long thumb was low. Further, we presented an approach for designing easy-to- use button GUI for users with long thumbs. The contributions of this study include revelation of the relationship between pointing performance and user’s thumb length when using a portable terminal in terms of accuracy, precision, and speed of pointing. We hope that these findings contribute to an easy-to-use GUI design for users with large hands.

Keywords: pointing performance, portable terminal, thumb length, touch-sensitive screen

Procedia PDF Downloads 140
131 Comparison of Sediment Rating Curve and Artificial Neural Network in Simulation of Suspended Sediment Load

Authors: Ahmad Saadiq, Neeraj Sahu

Abstract:

Sediment, which comprises of solid particles of mineral and organic material are transported by water. In river systems, the amount of sediment transported is controlled by both the transport capacity of the flow and the supply of sediment. The transport of sediment in rivers is important with respect to pollution, channel navigability, reservoir ageing, hydroelectric equipment longevity, fish habitat, river aesthetics and scientific interests. The sediment load transported in a river is a very complex hydrological phenomenon. Hence, sediment transport has attracted the attention of engineers from various aspects, and different methods have been used for its estimation. So, several experimental equations have been submitted by experts. Though the results of these methods have considerable differences with each other and with experimental observations, because the sediment measures have some limits, these equations can be used in estimating sediment load. In this present study, two black box models namely, an SRC (Sediment Rating Curve) and ANN (Artificial Neural Network) are used in the simulation of the suspended sediment load. The study is carried out for Seonath subbasin. Seonath is the biggest tributary of Mahanadi river, and it carries a vast amount of sediment. The data is collected for Jondhra hydrological observation station from India-WRIS (Water Resources Information System) and IMD (Indian Meteorological Department). These data include the discharge, sediment concentration and rainfall for 10 years. In this study, sediment load is estimated from the input parameters (discharge, rainfall, and past sediment) in various combination of simulations. A sediment rating curve used the water discharge to estimate the sediment concentration. This estimated sediment concentration is converted to sediment load. Likewise, for the application of these data in ANN, they are normalised first and then fed in various combinations to yield the sediment load. RMSE (root mean square error) and R² (coefficient of determination) between the observed load and the estimated load are used as evaluating criteria. For an ideal model, RMSE is zero and R² is 1. However, as the models used in this study are black box models, they don’t carry the exact representation of the factors which causes sedimentation. Hence, a model which gives the lowest RMSE and highest R² is the best model in this study. The lowest values of RMSE (based on normalised data) for sediment rating curve, feed forward back propagation, cascade forward back propagation and neural network fitting are 0.043425, 0.00679781, 0.0050089 and 0.0043727 respectively. The corresponding values of R² are 0.8258, 0.9941, 0.9968 and 0.9976. This implies that a neural network fitting model is superior to the other models used in this study. However, a drawback of neural network fitting is that it produces few negative estimates, which is not at all tolerable in the field of estimation of sediment load, and hence this model can’t be crowned as the best model among others, based on this study. A cascade forward back propagation produces results much closer to a neural network model and hence this model is the best model based on the present study.

Keywords: artificial neural network, Root mean squared error, sediment, sediment rating curve

Procedia PDF Downloads 300
130 Comparative Investigation of Two Non-Contact Prototype Designs Based on a Squeeze-Film Levitation Approach

Authors: A. Almurshedi, M. Atherton, C. Mares, T. Stolarski, M. Miyatake

Abstract:

Transportation and handling of delicate and lightweight objects is currently a significant issue in some industries. Two common contactless movement prototype designs, ultrasonic transducer design and vibrating plate design, are compared. Both designs are based on the method of squeeze-film levitation, and this study aims to identify the limitations, and challenges of each. The designs are evaluated in terms of levitation capabilities, and characteristics. To this end, theoretical and experimental explorations are made. It is demonstrated that the ultrasonic transducer prototype design is better suited to the terms of levitation capabilities. However, the design has some operating and mechanical designing difficulties. For making accurate industrial products in micro-fabrication and nanotechnology contexts, such as semiconductor silicon wafers, micro-components and integrated circuits, non-contact oil-free, ultra-precision and low wear transport along the production line is crucial for enabling. One of the designs (design A) is called the ultrasonic chuck, for which an ultrasonic transducer (Langevin, FBI 28452 HS) comprises the main part. Whereas the other (design B), is a vibrating plate design, which consists of a plain rectangular plate made of Aluminium firmly fastened at both ends. The size of the rectangular plate is 200x100x2 mm. In addition, four rounded piezoelectric actuators of size 28 mm diameter with 0.5 mm thickness are glued to the underside of the plate. The vibrating plate is clamped at both ends in the horizontal plane through a steel supporting structure. In addition, the dynamic of levitation using the designs (A and B) has been investigated based on the squeeze film levitation (SFL). The input apparatus that is used with designs consist of a sine wave signal generator connected to an amplifier type ENP-1-1U (Echo Electronics). The latter has to be utilised to magnify the sine wave voltage that is produced by the signal generator. The measurements of the maximum levitation for three different semiconductor wafers of weights 52, 70 and 88 [g] for design A are 240, 205 and 187 [um], respectively. Whereas the physical results show that the average separation distance for a disk of 5 [g] weight for design B reaches 70 [um]. By using the methodology of squeeze film levitation, it is possible to hold an object in a non-contact manner. The analyses of the investigation outcomes signify that the non-contact levitation of design A provides more improvement than design B. However, design A is more complicated than design B in terms of its manufacturing. In order to identify an adequate non-contact SFL design, a comparison between two common such designs has been adopted for the current investigation. Specifically, the study will involve making comparisons in terms of the following issues: floating component geometries and material type constraints; final created pressure distributions; dangerous interactions with the surrounding space; working environment constraints; and complication and compactness of the mechanical design. Considering all these matters is essential for proficiently distinguish the better SFL design.

Keywords: ANSYS, floating, piezoelectric, squeeze-film

Procedia PDF Downloads 124
129 Phenolic Composition of Wines from Cultivar Carménère during Aging with Inserts to Barrels

Authors: E. Obreque-Slier, P. Osorio-Umaña, G. Vidal-Acevedo, A. Peña-Neira, M. Medel-Marabolí

Abstract:

Sensory and nutraceutical characteristics of a wine are determined by different chemical compounds, such as organic acids, sugars, alcohols, polysaccharides, aromas, and polyphenols. The polyphenols correspond to secondary metabolites that are associated with the prevention of several pathologies, and those are responsible for color, aroma, bitterness, and astringency in wines. These compounds come from grapes and wood during aging in barrels, which correspond to the format of wood most widely used in wine production. However, the barrels is a high-cost input with a limited useful life (3-4 years). For this reason, some oenological products have been developed in order to renew the barrels and increase their useful life in some years. These formats are being used slowly because limited information exists about the effect on the wine chemical characteristics. The objective of the study was to evaluate the effect of different laubarrel renewal systems (staves and zigzag) on the polyphenolic characteristics of a Carménère wine (Vitis vinifera), an emblematic cultivar of Chile. For this, a completely randomized experimental design with 5 treatments and three replicates per treatment was used. The treatments were: new barrels (T0), used barrels during 4 years (T1), scraped used barrels (T2), used barrels with staves (T3) and used barrels with zigzag (T4). The study was performed for 12 months, and different spectrophotometric parameters (phenols, anthocyanins, and total tannins) and HPLC-DAD (low molecular weight phenols) were evaluated. The wood inputs were donated by Toneleria Nacional and corresponded to products from the same production batch. The total phenols content increased significantly after 40 days, while the total tannin concentration decreased gradually during the study. The anthocyanin concentration increased after 120 days of the assay in all treatments. Comparatively, it was observed that the wine of T2 presented the lowest values of these polyphenols, while the T0 and T4 presented the highest total phenol contents. Also, T1 presented the highest values of total tannins in relation to the rest of the treatments in some samples. The low molecular weight phenolic compounds identified by HPLC-DAD were 7 flavonoids (epigallocatechin, catechin, procyanidin gallate, epicatechin, quercetin, rutin and myricetin) and 14 non-flavonoids (gallic, protocatechuic, hydroxybenzoic, trans-cutaric, vanillinic, caffeic, syringic, p-coumaric and ellagic acids; tyrosol, vanillin, syringaldehyde, trans-resveratrol and cis-resveratrol). Tyrosol was the most abundant compound, whereas ellagic acid was the lowest in the samples. Comparatively, it was observed that the wines of T2 showed the lowest concentrations of flavonoid and non-flavonoid phenols during the study. In contrast, wines of T1, T3, and T4 presented the highest contents of non-flavonoid polyphenols. In summary, the use of barrel renovators (zig zag and staves) is an interesting alternative which would emulate the contribution of polyphenols from the barrels to the wine.

Keywords: barrels, oak wood aging, polyphenols, red wine

Procedia PDF Downloads 171
128 The Intensity of Root and Soil Respiration Is Significantly Determined by the Organic Matter and Moisture Content of the Soil

Authors: Zsolt Kotroczó, Katalin Juhos, Áron Béni, Gábor Várbíró, Tamás Kocsis, István Fekete

Abstract:

Soil organic matter plays an extremely important role in the functioning and regulation processes of ecosystems. It follows that the C content of organic matter in soil is one of the most important indicators of soil fertility. Part of the carbon stored in them is returned to the atmosphere during soil respiration. Climate change and inappropriate land use can accelerate these processes. Our work aimed to determine how soil CO2 emissions change over ten years as a result of organic matter manipulation treatments. With the help of this, we were able to examine not only the effects of the different organic matter intake but also the effects of the different microclimates that occur as a result of the treatments. We carried out our investigations in the area of the Síkfőkút DIRT (Detritus Input and Removal Treatment) Project. The research area is located in the southern, hilly landscape of the Bükk Mountains, northeast of Eger (Hungary). GPS coordinates of the project: 47°55′34′′ N and 20°26′ 29′′ E, altitude 320-340 m. The soil of the area is Luvisols. The 27-hectare protected forest area is now under the supervision of the Bükki National Park. The experimental plots in Síkfőkút were established in 2000. We established six litter manipulation treatments each with three 7×7 m replicate plots established under complete canopy cover. There were two types of detritus addition treatments (Double Wood and Double Litter). In three treatments, detritus inputs were removed: No Litter No Roots plots, No Inputs, and the Controls. After the establishment of the plots, during the drier periods, the NR and NI treatments showed the highest CO2 emissions. In the first few years, the effect of this process was evident, because due to the lack of living vegetation, the amount of evapotranspiration on the NR and NI plots was much lower, and transpiration practically ceased on these plots. In the wetter periods, the NL and NI treatments showed the lowest soil respiration values, which were significantly lower compared to the Co, DW, and DL treatments. Due to the lower organic matter content and the lack of surface litter cover, the water storage capacity of these soils was significantly limited, therefore we measured the lowest average moisture content among the treatments after ten years. Soil respiration is significantly influenced by temperature values. Furthermore, the supply of nutrients to the soil microorganisms is also a determining factor, which in this case is influenced by the litter production dictated by the treatments. In the case of dry soils with a moisture content of less than 20% in the initial period, litter removal treatments showed a strong correlation with soil moisture (r=0.74). In very dry soils, a smaller increase in moisture does not cause a significant increase in soil respiration, while it does in a slightly higher moisture range. In wet soils, the temperature is the main regulating factor, above a certain moisture limit, water displaces soil air from the soil pores, which inhibits aerobic decomposition processes, and so heterotrophic soil respiration also declines.

Keywords: soil biology, organic matter, nutrition, DIRT, soil respiration

Procedia PDF Downloads 42
127 Managing Human-Wildlife Conflicts Compensation Claims Data Collection and Payments Using a Scheme Administrator

Authors: Eric Mwenda, Shadrack Ngene

Abstract:

Human-wildlife conflicts (HWCs) are the main threat to conservation in Africa. This is because wildlife needs overlap with those of humans. In Kenya, about 70% of wildlife occurs outside protected areas. As a result, wildlife and human range overlap, causing HWCs. The HWCs in Kenya occur in the drylands adjacent to protected areas. The top five counties with the highest incidences of HWC are Taita Taveta, Narok, Lamu, Kajiado, and Laikipia. The common wildlife species responsible for HWCs are elephants, buffaloes, hyenas, hippos, leopards, baboons, monkeys, snakes, and crocodiles. To ensure individuals affected by the conflicts are compensated, Kenya has developed a model of HWC compensation claims data collection and payment. We collected data on HWC from all eight Kenya Wildlife Service (KWS) Conservation Areas from 2009 to 2019. Additional data was collected from stakeholders' consultative workshops held in the Conservation Areas and a literature review regarding payment of injuries and ongoing insurance schemes being practiced in areas. This was followed by the description of the claims administration process and calculation of the pricing of the compensation claims. We further developed a digital platform for data capture and processing of all reported conflict cases and payments. Our product recognized four categories of HWC (i.e., human death and injury, property damage, crop destruction, and livestock predation). Personal bodily injury and human death were provided based on the Continental Scale of Benefits. We proposed a maximum of Kenya Shillings (KES) 3,000,000 for death. Medical, pharmaceutical, and hospital expenses were capped at a maximum of KES 150,000, as well as funeral costs at KES 50,000. Pain and suffering were proposed to be paid for 12 months at the rate of KES 13,500 per month. Crop damage was to be based on farm input costs at a maximum of KES 150,000 per claim. Livestock predation leading to death was based on Tropical Livestock Unit (TLU), which is equivalent to KES 30,000, whick includes Cattle (1 TLU = KES 30,000), Camel (1.4 TLU = KES 42,000), Goat (0.15 TLU = 4,500), Sheep (0.15 TLU = 4,500), and Donkey (0.5 TLU = KES 15,000). Property destruction (buildings, outside structures and harvested crops) was capped at KES 150,000 per any one claim. We conclude that it is possible to use an administrator to collect data on HWC compensation claims and make payments using technology. The success of the new approach will depend on a piloting program. We recommended that a pilot scheme be initiated for eight months in Taita Taveta, Kajiado, Baringo, Laikipia, Narok, and Meru Counties. This will test the claims administration process as well as harmonize data collection methods. The results of this pilot will be crucial in adjusting the scheme before country-wide roll out.

Keywords: human-wildlife conflicts, compensation, human death and injury, crop destruction, predation, property destruction

Procedia PDF Downloads 27
126 Vieira Da Silva's Tiles at Universidade Federal Rural Do Rio de Janeiro: A Conservation and Restoration Project

Authors: Adriana Anselmo Oliveira

Abstract:

The present project showcases a tile work from the Franco-Portuguese artist Maria Helena Vieira da Silva (1908-1992). It is a set of 8 panels composed of figurative and geometric tiles, with extra tiles framing nearby doors and windows in a study room in the (UFRRJ, Universidade Federal Rural do Rio de Janeiro). The aforementioned work was created between 1942 and 1943, during the artist's 6 year exile in the Brazilian city. This one-of-a-kind tileset was designed and made by Vieira da Silva between 1942 and 1943. Over the years, several units were lost, which led to their replacement in the nineties. However, these replacements don't do justice to the original work of art. In 2007, a project was initiated to fully repair and maintain the set. Three panels are removed and restored, but the project is halted. To this day, the three fully restored panels remain in boxes. In 2016 a new restoration project is submitted by the (Faculdade de Belas Artes da Universidade de Lisboa) in collaboration with de (Fundacão Árpád Szenes-Vieira da Silva). There are many varied opinions on restoring and conserving older pieces of art, however, we have the moral duty to safeguard the original materials used by the artist along with the artists original vision and also to care for the future generations of students who will use the space in which the tile-work was inserted. Many tiles have been replaced by white tiles, tiles with a divergent colour pallet and technique, and in a few cases, the incorrect place or way around. These many factors make it increasingly difficult to maintain the artists original vision and destroy and chance of coherence within the artwork itself. The conservative technician cannot make new images to fill the empty spaces or mark the remaining images with their own creative input. with reliable photographic documentation that can provide us with the necessary vision to allow us to proceed with an accurate reconstruction, we have the obligation to proceed and return the piece of art to its true form, as in its current state, it is impossible to maintain its original glory. Using the information we have, we must find a way to differentiate the original tiles from the reconstructions in order to recreate and reclaim the original message from the artist. The objective of this project is to understand the significance of tiles in Vieira da Silva's art as well as the influence they had on the artist's pictorial language since the colour definition on tile work is vastly different from the painting process as the materials change during their merger. Another primary goal is to understand what the previous interventions achieved besides increasing the artworks durability. The main objective is to submit a proposal that can salvage the artist's visual intention and supports it for posteriority. In summary, this proposal goes further than the usual conservative interventions as it intends to recreate the original artistic worth, prioritising the aesthetics and keeping its soul alive.

Keywords: Vieira da Silva, tiles, conservation, restoration

Procedia PDF Downloads 128
125 Implementation of Deep Neural Networks for Pavement Condition Index Prediction

Authors: M. Sirhan, S. Bekhor, A. Sidess

Abstract:

In-service pavements deteriorate with time due to traffic wheel loads, environment, and climate conditions. Pavement deterioration leads to a reduction in their serviceability and structural behavior. Consequently, proper maintenance and rehabilitation (M&R) are necessary actions to keep the in-service pavement network at the desired level of serviceability. Due to resource and financial constraints, the pavement management system (PMS) prioritizes roads most in need of maintenance and rehabilitation action. It recommends a suitable action for each pavement based on the performance and surface condition of each road in the network. The pavement performance and condition are usually quantified and evaluated by different types of roughness-based and stress-based indices. Examples of such indices are Pavement Serviceability Index (PSI), Pavement Serviceability Ratio (PSR), Mean Panel Rating (MPR), Pavement Condition Rating (PCR), Ride Number (RN), Profile Index (PI), International Roughness Index (IRI), and Pavement Condition Index (PCI). PCI is commonly used in PMS as an indicator of the extent of the distresses on the pavement surface. PCI values range between 0 and 100; where 0 and 100 represent a highly deteriorated pavement and a newly constructed pavement, respectively. The PCI value is a function of distress type, severity, and density (measured as a percentage of the total pavement area). PCI is usually calculated iteratively using the 'Paver' program developed by the US Army Corps. The use of soft computing techniques, especially Artificial Neural Network (ANN), has become increasingly popular in the modeling of engineering problems. ANN techniques have successfully modeled the performance of the in-service pavements, due to its efficiency in predicting and solving non-linear relationships and dealing with an uncertain large amount of data. Typical regression models, which require a pre-defined relationship, can be replaced by ANN, which was found to be an appropriate tool for predicting the different pavement performance indices versus different factors as well. Subsequently, the objective of the presented study is to develop and train an ANN model that predicts the PCI values. The model’s input consists of percentage areas of 11 different damage types; alligator cracking, swelling, rutting, block cracking, longitudinal/transverse cracking, edge cracking, shoving, raveling, potholes, patching, and lane drop off, at three severity levels (low, medium, high) for each. The developed model was trained using 536,000 samples and tested on 134,000 samples. The samples were collected and prepared by The National Transport Infrastructure Company. The predicted results yielded satisfactory compliance with field measurements. The proposed model predicted PCI values with relatively low standard deviations, suggesting that it could be incorporated into the PMS for PCI determination. It is worth mentioning that the most influencing variables for PCI prediction are damages related to alligator cracking, swelling, rutting, and potholes.

Keywords: artificial neural networks, computer programming, pavement condition index, pavement management, performance prediction

Procedia PDF Downloads 108
124 Oil-price Volatility and Economic Prosperity in Nigeria: Empirical Evidence

Authors: Yohanna Panshak

Abstract:

The impact of macroeconomic instability on economic growth and prosperity has been at forefront in many discourses among researchers and policy makers and has generated a lot of controversies over the years. This has generated series of research efforts towards understanding the remote causes of this phenomenon; its nature, determinants and how it can be targeted and mitigated. While others have opined that the root cause of macroeconomic flux in Nigeria is attributed to Oil-Price volatility, others viewed the issue as resulting from some constellation of structural constraints both within and outside the shores of the country. Research works of scholars such as [Akpan (2009), Aliyu (2009), Olomola (2006), etc] argue that oil volatility can determine economic growth or has the potential of doing so. On the contrary, [Darby (1982), Cerralo (2005) etc] share the opinion that it can slow down growth. The earlier argument rest on the understanding that for a net balance of oil exporting economies, price upbeat directly increases real national income through higher export earnings, whereas, the latter allude to the case of net-oil importing countries (which experience price rises, increased input costs, reduced non-oil demand, low investment, fall in tax revenues and ultimately an increase in budget deficit which will further reduce welfare level). Therefore, assessing the precise impact of oil price volatility on virtually any economy is a function of whether it is an oil-exporting or importing nation. Research on oil price volatility and its outcome on the growth of the Nigerian economy are evolving and in a march towards resolving Nigeria’s macroeconomic instability as long as oil revenue still remain the mainstay and driver of socio-economic engineering. Recently, a major importer of Nigeria’s oil- United States made a historic breakthrough in more efficient source of energy for her economy with the capacity of serving significant part of the world. This undoubtedly suggests a threat to the exchange earnings of the country. The need to understand fluctuation in its major export commodity is critical. This paper leans on the Renaissance growth theory with greater focus on theoretical work of Lee (1998); a leading proponent of this school who makes a clear cut of difference between oil price changes and oil price volatility. Based on the above background, the research seeks to empirically examine the impact oil-price volatility on government expenditure using quarterly time series data spanning 1986:1 to 2014:4. Vector Auto Regression (VAR) econometric approach shall be used. The structural properties of the model shall be tested using Augmented Dickey-Fuller and Phillips-Perron. Relevant diagnostics tests of heteroscedasticity, serial correlation and normality shall also be carried out. Policy recommendation shall be offered on the empirical findings and believes it assist policy makers not only in Nigeria but the world-over.

Keywords: oil-price, volatility, prosperity, budget, expenditure

Procedia PDF Downloads 249
123 Microbial Fuel Cells: Performance and Applications

Authors: Andrea Pietrelli, Vincenzo Ferrara, Bruno Allard, Francois Buret, Irene Bavasso, Nicola Lovecchio, Francesca Costantini, Firas Khaled

Abstract:

This paper aims to show some applications of microbial fuel cells (MFCs), an energy harvesting technique, as clean power source to supply low power device for application like wireless sensor network (WSN) for environmental monitoring. Furthermore, MFC can be used directly as biosensor to analyse parameters like pH and temperature or arranged in form of cluster devices in order to use as small power plant. An MFC is a bioreactor that converts energy stored in chemical bonds of organic matter into electrical energy, through a series of reactions catalysed by microorganisms. We have developed a lab-scale terrestrial microbial fuel cell (TMFC), based on soil that acts as source of bacteria and flow of nutrient and a lab-scale waste water microbial fuel cell (WWMFC), where waste water acts as flow of nutrient and bacteria. We performed large series of tests to exploit the capability as biosensor. The pH value has strong influence on the open circuit voltage (OCV) delivered from TMFCs. We analyzed three condition: test A and B were filled with same soil but changing pH from 6 to 6.63, test C was prepared using a different soil with a pH value of 6.3. Experimental results clearly show how with higher pH value a higher OCV was produced; indeed reactors are influenced by different values of pH which increases the voltage in case of a higher pH value until the best pH value of 7 is achieved. The influence of pH on OCV of lab-scales WWMFC was analyzed at pH value of 6.5, 7, 7.2, 7.5 and 8. WWMFCs are influenced from temperature more than TMFCs. We tested the power performance of WWMFCs considering four imposed values of ambient temperature. Results show how power performance increase proportionally with higher temperature values, doubling the output power from 20° to 40°. The best value of power produced from our lab-scale TMFC was equal to 310 μW using peaty soil, at 1KΩ, corresponding to a current of 0.5 mA. A TMFC can supply proper energy to low power devices of a WSN by means of the design of three stages scheme of an energy management system, which adapts voltage level of TMFC to those required by a WSN node, as 3.3V. Using a commercial DC/DC boost converter, that needs an input voltage of 700 mV, the current source of 0.5 mA, charges a capacitor of 6.8 mF until it will have accumulated an amount of charge equal to 700 mV in a time of 10 s. The output stage includes an output switch that close the circuit after a time of 10s + 1.5ms because the converter can boost the voltage from 0.7V to 3.3V in 1.5 ms. Furthermore, we tested in form of clusters connected in series up to 20 WWMFCs, we have obtained a high voltage value as output, around 10V, but low current value. MFC can be considered a suitable clean energy source to be used to supply low power devices as a WSN node or to be used directly as biosensor.

Keywords: energy harvesting, low power electronics, microbial fuel cell, terrestrial microbial fuel cell, waste-water microbial fuel cell, wireless sensor network

Procedia PDF Downloads 184
122 Harnessing Emerging Creative Technology for Knowledge Discovery of Multiwavelenght Datasets

Authors: Basiru Amuneni

Abstract:

Astronomy is one domain with a rise in data. Traditional tools for data management have been employed in the quest for knowledge discovery. However, these traditional tools become limited in the face of big. One means of maximizing knowledge discovery for big data is the use of scientific visualisation. The aim of the work is to explore the possibilities offered by emerging creative technologies of Virtual Reality (VR) systems and game engines to visualize multiwavelength datasets. Game Engines are primarily used for developing video games, however their advanced graphics could be exploited for scientific visualization which provides a means to graphically illustrate scientific data to ease human comprehension. Modern astronomy is now in the era of multiwavelength data where a single galaxy for example, is captured by the telescope several times and at different electromagnetic wavelength to have a more comprehensive picture of the physical characteristics of the galaxy. Visualising this in an immersive environment would be more intuitive and natural for an observer. This work presents a standalone VR application that accesses galaxy FITS files. The application was built using the Unity Game Engine for the graphics underpinning and the OpenXR API for the VR infrastructure. The work used a methodology known as Design Science Research (DSR) which entails the act of ‘using design as a research method or technique’. The key stages of the galaxy modelling pipeline are FITS data preparation, Galaxy Modelling, Unity 3D Visualisation and VR Display. The FITS data format cannot be read by the Unity Game Engine directly. A DLL (CSHARPFITS) which provides a native support for reading and writing FITS files was used. The Galaxy modeller uses an approach that integrates cleaned FITS image pixels into the graphics pipeline of the Unity3d game Engine. The cleaned FITS images are then input to the galaxy modeller pipeline phase, which has a pre-processing script that extracts, pixel, galaxy world position, and colour maps the FITS image pixels. The user can visualise image galaxies in different light bands, control the blend of the image with similar images from different sources or fuse images for a holistic view. The framework will allow users to build tools to realise complex workflows for public outreach and possibly scientific work with increased scalability, near real time interactivity with ease of access. The application is presented in an immersive environment and can use all commercially available headset built on the OpenXR API. The user can select galaxies in the scene, teleport to the galaxy, pan, zoom in/out, and change colour gradients of the galaxy. The findings and design lessons learnt in the implementation of different use cases will contribute to the development and design of game-based visualisation tools in immersive environment by enabling informed decisions to be made.

Keywords: astronomy, visualisation, multiwavelenght dataset, virtual reality

Procedia PDF Downloads 64
121 Functionalization of Sanitary Pads with Probiotic Paste

Authors: O. Sauperl, L. Fras Zemljic

Abstract:

The textile industry is gaining increasing importance in the field of medical materials. Therefore, presented research is focused on textile materials for external (out-of-body) use. Such materials could be various hygienic textile products (diapers, tampons, sanitary napkins, incontinence products, etc.), protective textiles and various hospital linens (surgical covers, masks, gowns, cloths, bed linens, etc.) wound pillows, bandages, orthopedic socks, etc. Function of tampons and sanitary napkins is not only to provide protection during the menstrual cycle, but their function can be also to take care of physiological or pathological vaginal discharge. In general, women's intimate areas are against infection protected by a low pH value of the vaginal flora. High pH inhibits the development of harmful microorganisms, as it is difficult to be reproduced in an acidic environment. The normal vaginal flora in healthy women is highly colonized by lactobacilli. The lactic acid produced by these organisms maintains the constant acidity of the vagina. If the balance of natural protection breaks, infections can occur. In the market, there exist probiotic tampons as a medical product supplying the vagina with beneficial probiotic lactobacilli. But, many users have concerns about the use of tampons due to the possible dry-out of the vagina as well as the possible toxic shock syndrome, which is the reason that they use mainly sanitary napkins during the menstrual cycle. Functionalization of sanitary napkins with probiotics is, therefore, interesting in regard to maintain a healthy vaginal flora and to offer to users added value of the sanitary napkins in the sense of health- and environmentally-friendly products. For this reason, the presented research is oriented in functionalization of the sanitary napkins with the probiotic paste in order to activate the lactic acid bacteria presented in the core of the functionalized sanitary napkin at the time of the contact with the menstrual fluid. In this way, lactobacilli could penetrate into vagina and by maintaining healthy vaginal flora to reduce the risk of vaginal disorders. In regard to the targeted research problem, the influence of probiotic paste applied onto cotton hygienic napkins on selected properties was studied. The aim of the research was to determine whether the sanitary napkins with the applied probiotic paste may assure suitable vaginal pH to maintain a healthy vaginal flora during the use of this product. Together with this, sorption properties of probiotic functionalized sanitary napkins were evaluated and compared to the untreated one. The research itself was carried out on the basis of tracking and controlling the input parameters, currently defined by Slovenian producer (Tosama d.o.o.) as the most important. Successful functionalization of sanitary pads with the probiotic paste was confirmed by ATR-FTIR spectroscopy. Results of the methods used within the presented research show that the absorption of the pads treated with probiotic paste deteriorates compared to non-treated ones. The coating shows a 6-month stability. Functionalization of sanitary pads with probiotic paste is believed to have a commercial potential for lowering the probability of infection during the menstrual cycle.

Keywords: functionalization, probiotic paste, sanitary pads, textile materials

Procedia PDF Downloads 164
120 'You’re Not Alone': Peer Feedback Practices for Cross-Cultural Writing Classrooms and Centers

Authors: Cassandra Branham, Danielle Farrar

Abstract:

As writing instructors and writing center administrators at a large research university with a significant population of English language learners (ELLs), we are interested in how peer feedback pedagogy can be effectively translated for writing center purposes, as well as how various modes of peer feedback can enrich the learning experiences of L1 and L2 writers in these spaces. Although peer feedback is widely used in classrooms and centers, instructor, student, and researcher opinions vary in respect to its effectiveness. We argue that peer feedback - traditional and digital, synchronous and asynchronous - is an indispensable element for both classrooms and centers and emphasize that it should occur with both L1 and L2 students to further develop an array of reading and writing skills. We also believe that further understanding of the best practices of peer feedback in such cross-cultural spaces, like the classroom and center, can optimize the benefits of peer feedback. After a critical review of the literature, we implemented an embedded tutoring program in our university’s writing center in collaboration with its First-Year Composition (FYC) program and Language Institute. The embedded tutoring program matches a graduate writing consultant with L1 and L2 writers enrolled in controlled-matriculation composition courses where ELLs make up at least 50% of each class. Furthermore, this program is informed by what we argue to be some best practices of peer feedback for both classroom and center purposes, including expectation-based training through rubrics, modeling effective feedback, hybridizing traditional and digital modes of feedback, recognizing the significance the body in composition (what we call writer embodiment), and maximizing digital technologies to exploit extended cognition. After conducting surveys and follow-up interviews with students, instructors, and writing consultants in the embedded tutoring program, we found that not only did students see an increased value in peer feedback, but also instructors saw an improvement in both writing style and critical thinking skills. Our L2 participants noted improvements in language acquisition while our L1 students recognized a broadening of their worldviews. We believe that both L1 and L2 students developed self-efficacy and agency in their identities as writers because they gained confidence in their abilities to offer feedback, as well as in the legitimacy of feedback they received from peers. We also argue that these best practices situate novice writers as experts, as writers become a valued and integral part of the revision process with their own and their peers’ papers. Finally, the use of iPads in embedded tutoring recovered the importance of the body and its senses in writing; the highly sensory feedback from these multi-modal sessions that offer audio and visual input underscores the significant role both the body and mind play in compositional practices. After beginning with a brief review of the literature that sparked this research, this paper will discuss the embedded tutoring program in detail, report on the results of the pilot program, and will conclude with a discussion of the pedagogical implications that arise from this research for both classroom and center.

Keywords: English language learners, peer feedback, writing center, writing classroom

Procedia PDF Downloads 382
119 Test Rig Development for Up-to-Date Experimental Study of Multi-Stage Flash Distillation Process

Authors: Marek Vondra, Petr Bobák

Abstract:

Vacuum evaporation is a reliable and well-proven technology with a wide application range which is frequently used in food, chemical or pharmaceutical industries. Recently, numerous remarkable studies have been carried out to investigate utilization of this technology in the area of wastewater treatment. One of the most successful applications of vacuum evaporation principal is connected with seawater desalination. Since 1950’s, multi-stage flash distillation (MSF) has been the leading technology in this field and it is still irreplaceable in many respects, despite a rapid increase in cheaper reverse-osmosis-based installations in recent decades. MSF plants are conveniently operated in countries with a fluctuating seawater quality and at locations where a sufficient amount of waste heat is available. Nowadays, most of the MSF research is connected with alternative heat sources utilization and with hybridization, i.e. merging of different types of desalination technologies. Some of the studies are concerned with basic principles of the static flash phenomenon, but only few scientists have lately focused on the fundamentals of continuous multi-stage evaporation. Limited measurement possibilities at operating plants and insufficiently equipped experimental facilities may be the reasons. The aim of the presented study was to design, construct and test an up-to-date test rig with an advanced measurement system which will provide real time monitoring options of all the important operational parameters under various conditions. The whole system consists of a conventionally designed MSF unit with 8 evaporation chambers, versatile heating circuit for different kinds of feed water (e.g. seawater, waste water), sophisticated system for acquisition and real-time visualization of all the related quantities (temperature, pressure, flow rate, weight, conductivity, pH, water level, power input), access to a wide spectrum of operational media (salt, fresh and softened water, steam, natural gas, compressed air, electrical energy) and integrated transparent features which enable a direct visual control of selected physical mechanisms (water evaporation in chambers, water level right before brine and distillate pumps). Thanks to the adjustable process parameters, it is possible to operate the test unit at desired operational conditions. This allows researchers to carry out statistical design and analysis of experiments. Valuable results obtained in this manner could be further employed in simulations and process modeling. First experimental tests confirm correctness of the presented approach and promise interesting outputs in the future. The presented experimental apparatus enables flexible and efficient research of the whole MSF process.

Keywords: design of experiment, multi-stage flash distillation, test rig, vacuum evaporation

Procedia PDF Downloads 363
118 A Design Methodology and Tool to Support Ecodesign Implementation in Induction Hobs

Authors: Anna Costanza Russo, Daniele Landi, Michele Germani

Abstract:

Nowadays, the European Ecodesign Directive has emerged as a new approach to integrate environmental concerns into the product design and related processes. Ecodesign aims to minimize environmental impacts throughout the product life cycle, without compromising performances and costs. In addition, the recent Ecodesign Directives require products which are increasingly eco-friendly and eco-efficient, preserving high-performances. It is very important for producers measuring performances, for electric cooking ranges, hobs, ovens, and grills for household use, and a low power consumption of appliances represents a powerful selling point, also in terms of ecodesign requirements. The Ecodesign Directive provides a clear framework about the sustainable design of products and it has been extended in 2009 to all energy-related products, or products with an impact on energy consumption during the use. The European Regulation establishes measures of ecodesign of ovens, hobs, and kitchen hoods, and domestic use and energy efficiency of a product has a significant environmental aspect in the use phase which is the most impactful in the life cycle. It is important that the product parameters and performances are not affected by ecodesign requirements from a user’s point of view, and the benefits of reducing energy consumption in the use phase should offset the possible environmental impact in the production stage. Accurate measurements of cooking appliance performance are essential to help the industry to produce more energy efficient appliances. The development of ecodriven products requires ecoinnovation and ecodesign tools to support the sustainability improvement. The ecodesign tools should be practical and focused on specific ecoobjectives in order to be largely diffused. The main scope of this paper is the development, implementation, and testing of an innovative tool, which could be an improvement for the sustainable design of induction hobs. In particular, a prototypical software tool is developed in order to simulate the energy performances of the induction hobs. The tool is focused on a multiphysics model which is able to simulate the energy performances and the efficiency of induction hobs starting from the design data. The multiphysics model is composed by an electromagnetic simulation and a thermal simulation. The electromagnetic simulation is able to calculate the eddy current induced in the pot, which leads to the Joule heating of material. The thermal simulation is able to measure the energy consumption during the operational phase. The Joule heating caused from the eddy currents is the output of electromagnetic simulation and the input of thermal ones. The aims of the paper are the development of integrated tools and methodologies of virtual prototyping in the context of the ecodesign. This tool could be a revolutionary instrument in the field of industrial engineering and it gives consideration to the environmental aspects of product design and focus on the ecodesign of energy-related products, in order to achieve a reduced environmental impact.

Keywords: ecodesign, energy efficiency, induction hobs, virtual prototyping

Procedia PDF Downloads 236
117 Inputs and Outputs of Innovation Processes in the Colombian Services Sector

Authors: Álvaro Turriago-Hoyos

Abstract:

Most research tends to see innovation as an explanatory factor in achieving high levels of competitiveness and productivity. More recent studies have begun to analyze the determinants of innovation in the services sector as opposed to the much-discussed industrial sector of a country’s economy. This research paper focuses on the services sector in Colombia, one of Latin America’s fastest growing and biggest economies. Over the past decade, much of Colombia’s economic expansion has relied on commodity exports (mainly oil and coffee) whilst the industrial sector has performed relatively poorly. Such developments highlight the potential of the innovative role played by the services sector of the Colombian economy and its future growth prospects. This research paper analyzes the relationship between inputs, which at the same time are internal sources of innovation (such as R&D activities), and external sources that are improved by technology acquisition. The outputs are basically the four kinds of innovation that the OECD Oslo Manual recognizes: product, process, marketing and organizational innovations. The instrument used to measure this input-output relationship is based on Knowledge Production Function approaches. We run Probit models in order to identify the existing relationships between the above inputs and outputs, but also to identify spill-overs derived from interactions of the components of the value chain of the services firms analyzed: customers, suppliers, competitors, and complementary firms. Data are obtained from the Colombian National Administrative Department of Statistics for the period 2008 to 2013 published in the II and III Colombian National Innovation Survey. A short summary of the results obtained lead to conclude that firm size and a firm’s level of technological development turn out to be important discriminating factors for the description of the innovative process at the firm level. The model’s outcomes show a positive impact on the probability of introducing any kind of innovation both on R&D and Technology Acquisition investment. Also, cooperation agreements with customers, research institutes, competitors, and the suppliers are significant. Belonging to a particular industrial group is an important determinant but only to product and organizational innovation. It is possible to establish that Health Services, Education, Computer, Wholesale trade, and Financial Intermediation are the ISIC sectors, which report the highest number of frequencies of the considered set of firms. Those five sectors of the sixteen considered, in all cases, explained more than half of the total of all kinds of innovations. Product Innovation, which is followed by Marketing Innovation, gets the highest results. Displaying the same set of firms distinguishing by size, and belonging to high and low tech services sector shows that the larger the firms the larger a number of innovations, but also that always high-tech firms show a better innovation performance.

Keywords: Colombia, determinants of innovation, innovation, services sector

Procedia PDF Downloads 241
116 The Development of the Psychosomatic Nursing Model from an Evidence-Based Action Research on Proactive Mental Health Care for Medical Inpatients

Authors: Chia-Yi Wu, Jung-Chen Chang, Wen-Yu Hu, Ming-Been Lee

Abstract:

In nearly all physical health conditions, suicide risk is increased compared to healthy people even after adjustment for age, gender, mental health, and substance use diagnoses. In order to highlight the importance of suicide risk assessment for the inpatients and early identification and engagement for inpatients’ mental health problems, a study was designed aiming at developing a comprehensive psychosomatic nursing engagement (PSNE) model with standardized operation procedures informing how nurses communicate, assess, and engage with the inpatients with emotional distress. The purpose of the study was to promote the gatekeeping role of clinical nurses in performing brief assessment and interventions to detect depression and anxiety symptoms among the inpatients, particularly in non-psychiatric wards. The study will be carried out in a 2000-bed university hospital in Northern Taiwan in 2019. We will select a ward for trial and develop feasible procedures and in-job training course for the nurses to offer mental health care, which will also be validated through professional consensus meeting. The significance of the study includes the following three points: (1) The study targets at an important but less-researched area of PSNE model in the cultural background of Taiwan, where hospital service is highly accessible, but mental health and suicide risk assessment are hardly provided by non-psychiatric healthcare personnel. (2) The issue of PSNE could be efficient and cost-effective in the identification of suicide risks at an early stage to prevent inpatient suicide or to reduce future suicide risk by early treatment of mental illnesses among the high-risk group of hospitalized patients who are more than three-times lethal to suicide. (3) Utilizing a brief tool with its established APP ('The Five-item Brief Symptom Rating Scale, BSRS-5'), we will invent the standardized procedure of PSNE and referral steps in collaboration with the medical teams across the study hospital. New technological tools nested within nursing assessment/intervention will concurrently be invented to facilitate better care quality. The major outcome measurements will include tools for early identification of common mental distress and suicide risks, i.e., the BSRS-5, revised BSRS-5, and the 9-item Concise Mental Health Checklist (CMHC-9). The main purpose of using the CMHC-9 in clinical suicide risk assessment is mainly to provide care and build-up therapeutic relationship with the client, so it will also be used to nursing training highlighting the skills of supportive care. Through early identification of the inpatients’ depressive symptoms or other mental health care needs such as insomnia, anxiety, or suicide risk, the majority of the nursing clinicians would be able to engage in critical interventions that alleviate the inpatients’ suffering from mental health problems, given a feasible nursing input.

Keywords: mental health care, clinical outcome improvement, clinical nurses, suicide prevention, psychosomatic nursing

Procedia PDF Downloads 88
115 Enhancing Engineering Students Educational Experience: Studying Hydrostatic Pumps Association System in Fluid Mechanics Laboratories

Authors: Alexandre Daliberto Frugoli, Pedro Jose Gabriel Ferreira, Pedro Americo Frugoli, Lucio Leonardo, Thais Cavalheri Santos

Abstract:

Laboratory classes in Engineering courses are essential for students to be able to integrate theory with practical reality, by handling equipment and observing experiments. In the researches of physical phenomena, students can learn about the complexities of science. Over the past years, universities in developing countries have been reducing the course load of engineering courses, in accordance with cutting cost agendas. Quality education is the object of study for researchers and requires educators and educational administrators able to demonstrate that the institutions are able to provide great learning opportunities at reasonable costs. Didactic test benches are indispensable equipment in educational activities related to turbo hydraulic pumps and pumping facilities study, which have a high cost and require long class time due to measurements and equipment adjustment time. In order to overcome the aforementioned obstacles, aligned with the professional objectives of an engineer, GruPEFE - UNIP (Research Group in Physics Education for Engineering - Universidade Paulista) has developed a multi-purpose stand for the discipline of fluid mechanics which allows the study of velocity and flow meters, loads losses and pump association. In this work, results obtained by the association in series and in parallel of hydraulic pumps will be presented and discussed, mainly analyzing the repeatability of experimental procedures and their agreement with the theory. For the association in series two identical pumps were used, consisting of the connection of the discharge of a pump to the suction of the next one, allowing the fluid to receive the power of all machines in the association. The characteristic curve of the set is obtained from the curves of each of the pumps, by adding the heads corresponding to the same flow rates. The same pumps were associated in parallel. In this association, the discharge piping is common to the two machines together. The characteristic curve of the set was obtained by adding to each value of H (head height), the flow rates of each pump. For the tests, the input and output pressure of each pump were measured. For each set there were three sets of measurements, varying the flow rate in range from 6.0 to 8.5 m 3 / h. For the two associations, the results showed an excellent repeatability with variations of less than 10% between sets of measurements and also a good agreement with the theory. This variation agrees with the instrumental uncertainty. Thus, the results validate the use of the fluids bench designed for didactic purposes. As a future work, a digital acquisition system is being developed, using differential sensors of extremely low pressures (2 to 2000 Pa approximately) for the microcontroller Arduino.

Keywords: engineering education, fluid mechanics, hydrostatic pumps association, multi-purpose stand

Procedia PDF Downloads 201
114 Development and Preliminary Testing of the Dutch Version of the Program for the Education and Enrichment of Relational Skills

Authors: Sakinah Idris, Gabrine Jagersma, Bjorn Jaime Van Pelt, Kirstin Greaves-Lord

Abstract:

Background: The PEERS (Program for the Education and Enrichment of Relational Skills) intervention can be considered a well-established, evidence-based intervention in the USA. However, testing the efficacy of cultural adaptations of PEERS is still ongoing. More and more, the involvement of all stakeholders in the development and evaluation of interventions is acknowledged as crucial for the longer term implementation of interventions across settings. Therefore, in the current project, teens with ASD (Autism Spectrum Disorder), their neurotypical peers, parents, teachers, as well as clinicians were involved in the development and evaluation of the Dutch version of PEERS. Objectives: The current presentation covers (1) the formative phase and (2) the preliminary adaptation test phase of the cultural adaptation of evidence-based interventions. In the formative phase, we aim to describe the process of adaptation of the PEERS program to the Dutch culture and care system. In the preliminary adaptation phase, we will present results from the preliminary adaptation test among 32 adolescents with ASD. Methods: In phase 1, a group discussion on common vocabulary was conducted among 70 teenagers (and their teachers) from special and regular education aged 12-18 years old. This inventory concerned 14 key constructs from PEERS, e.g., areas of interests, locations for making friends, common peer groups and crowds inside and outside of school, activities with friends, commonly used ways for electronic communication, ways for handling disagreements, and common teasing comebacks. Also, 15 clinicians were involved in the translation and cultural adaptation process. The translation and cultural adaptation process were guided by the research team, and who included input and feedback from all stakeholders through an iterative feedback incorporation procedure. In phase 2, The parent-reported Social Responsiveness Scale (SRS), the Test of Adolescent Social Skills Knowledge (TASSK), and the Quality of Socialization Questionnaire (QSQ) were assessed pre- and post-intervention to evaluate potential treatment outcome. Results: The most striking cultural adaptation - reflecting the standpoints of all stakeholders - concerned the strategies for handling rumors and gossip, which were suggested to be taught using a similar approach as the teasing comebacks, more in line with ‘down-to-earth’ Dutch standards. The preliminary testing of this adapted version indicated that the adolescents with ASD significantly improved their social knowledge (TASSK; t₃₁ = -10.9, p < .01), social experience (QSQ-Parent; t₃₁ = -4.2, p < .01 and QSQ-Adolescent; t₃₂ = -3.8, p < .01), and in parent-reported social responsiveness (SRS; t₃₃ = 3.9, p < .01). In addition, subjective evaluations of teens with ASD, their parents and clinicians were positive. Conclusions: In order to further scrutinize the effectiveness of the Dutch version of the PEERS intervention, we recommended performing a larger scale randomized control trial (RCT) design, for which we provide several methodological considerations.

Keywords: cultural adaptation, PEERS, preliminary testing, translation

Procedia PDF Downloads 135
113 Economic Impacts of Nitrogen Fertilizer Use into Tropical Pastures for Beef Cattle in Brazil

Authors: Elieder P. Romanzini, Lutti M. Delevatti, Rhaony G. Leite, Ricardo A. Reis, Euclides B. Malheiros

Abstract:

Brazilian beef cattle production systems are an important profitability source for the national gross domestic product. The main characteristic of these systems is forage utilization as the exclusive feed source. Forage utilization had been causing on owners the false feeling of low production costs. However, this low cost is followed to low profit causing a lot times worst animal index what can result in activities changes or until land sold. Aiming to evaluate economic impacts into Brazilian beef cattle systems were evaluated four nitrogen fertilizer (N) application levels (0, 90, 180 and 270 kg per hectare [kg.ha-1]). Research was developed during 2015 into Forage Crops and Grasslands section of São Paulo State University, “Júlio de Mesquita Filho” (Unesp) (Jaboticabal, São Paulo, Brazil). Pastures were seeded with Brachiaria brizantha Stapf. ‘Marandu’ (Palisade grass) handled using continuous grazing system, with variable stocking rate, sward height maintained at 25 cm. The economic evaluation was developed in rearing e finishing phases. We evaluated the cash flows inside each phase on different N levels. Economic valuations were considering: cost-effective operating (CEO), cost-total operating (CTO), gross revenue (GR), operating profit (OP) and net income (NI), every measured in US$. Complementary analyses were developed, profitability was calculated by [OP/GR]. Pay back (measured in years) was calculated considering average capital stocktaking pondered by area in use (ACS) divided by [GR-CEO]. And the internal rate of return (IRR) was calculated by 100/(pay back). Input prices were prices during 2015 and were obtained from Anuário Brasileiro da Pecuária, Centro de Estudos Avançados em Economia Aplicada and quotation in the same region of animal production (northeast São Paulo State) during the period above mentioned. Values were calculated in US$ according exchange rate US$1.00 equal R$3.34. The CEO, CTO, GR, OP and NI per hectare for each N level were respectively US$1,919.66; US$2,048.47; US$2,905.72; US$857.25 and US$986.06 to 0 kg.ha-1; US$2,403.20; US$2,551.80; US$3,530.19; US$978.39 and US$1,126.99 to 90 kg.ha-1; US$3,180.42; US$3,364.81; US$4,985.03; US$1,620.23 and US$1,804.62 to 180 kg.ha-1andUS$3,709.14; US$3,915.15; US$5,554.95; US$1,639.80 and US$1,845.81 to 270 kg.ha-1. Relationship to another economic indexes, profitability, pay back and IRR, the results were respectively 29.50%, 6.44 and 15.54% to 0 kg.ha-1; 27.72%, 6.88 and 14.54% to 90 kg.ha-1; 32.50%, 4.08 and 24.50% to 180 kg.ha-1 and 29.52%, 3.42 and 29.27% to 270 kg.ha-1. Values previously presented in this evaluation allowing to affirm that the best result was obtained to N level 270 kg.ha-1. These results among all N levels evaluated could be explained by improve occurred on stocking rate caused by increase on N level. However, a crucial information about high N level application into pastures is the efficiency of N utilization (associated to environmental impacts) that normally decrease with the increase on N level. Hence, considering all situations (efficiency of N utilization and economic results) into tropical pastures used to beef cattle production could be recommended N level equal to 180kg.ha-1, which had better profitability and cause lesser environmental impacts, proved by other studies developed in the same area.

Keywords: Brachiaria brizantha, cost-total operating, gross revenue, profitability

Procedia PDF Downloads 142
112 Row Detection and Graph-Based Localization in Tree Nurseries Using a 3D LiDAR

Authors: Ionut Vintu, Stefan Laible, Ruth Schulz

Abstract:

Agricultural robotics has been developing steadily over recent years, with the goal of reducing and even eliminating pesticides used in crops and to increase productivity by taking over human labor. The majority of crops are arranged in rows. The first step towards autonomous robots, capable of driving in fields and performing crop-handling tasks, is for robots to robustly detect the rows of plants. Recent work done towards autonomous driving between plant rows offers big robotic platforms equipped with various expensive sensors as a solution to this problem. These platforms need to be driven over the rows of plants. This approach lacks flexibility and scalability when it comes to the height of plants or distance between rows. This paper proposes instead an algorithm that makes use of cheaper sensors and has a higher variability. The main application is in tree nurseries. Here, plant height can range from a few centimeters to a few meters. Moreover, trees are often removed, leading to gaps within the plant rows. The core idea is to combine row detection algorithms with graph-based localization methods as they are used in SLAM. Nodes in the graph represent the estimated pose of the robot, and the edges embed constraints between these poses or between the robot and certain landmarks. This setup aims to improve individual plant detection and deal with exception handling, like row gaps, which are falsely detected as an end of rows. Four methods were developed for detecting row structures in the fields, all using a point cloud acquired with a 3D LiDAR as an input. Comparing the field coverage and number of damaged plants, the method that uses a local map around the robot proved to perform the best, with 68% covered rows and 25% damaged plants. This method is further used and combined with a graph-based localization algorithm, which uses the local map features to estimate the robot’s position inside the greater field. Testing the upgraded algorithm in a variety of simulated fields shows that the additional information obtained from localization provides a boost in performance over methods that rely purely on perception to navigate. The final algorithm achieved a row coverage of 80% and an accuracy of 27% damaged plants. Future work would focus on achieving a perfect score of 100% covered rows and 0% damaged plants. The main challenges that the algorithm needs to overcome are fields where the height of the plants is too small for the plants to be detected and fields where it is hard to distinguish between individual plants when they are overlapping. The method was also tested on a real robot in a small field with artificial plants. The tests were performed using a small robot platform equipped with wheel encoders, an IMU and an FX10 3D LiDAR. Over ten runs, the system achieved 100% coverage and 0% damaged plants. The framework built within the scope of this work can be further used to integrate data from additional sensors, with the goal of achieving even better results.

Keywords: 3D LiDAR, agricultural robots, graph-based localization, row detection

Procedia PDF Downloads 106
111 Real-Time Neuroimaging for Rehabilitation of Stroke Patients

Authors: Gerhard Gritsch, Ana Skupch, Manfred Hartmann, Wolfgang Frühwirt, Hannes Perko, Dieter Grossegger, Tilmann Kluge

Abstract:

Rehabilitation of stroke patients is dominated by classical physiotherapy. Nowadays, a field of research is the application of neurofeedback techniques in order to help stroke patients to get rid of their motor impairments. Especially, if a certain limb is completely paralyzed, neurofeedback is often the last option to cure the patient. Certain exercises, like the imagination of the impaired motor function, have to be performed to stimulate the neuroplasticity of the brain, such that in the neighboring parts of the injured cortex the corresponding activity takes place. During the exercises, it is very important to keep the motivation of the patient at a high level. For this reason, the missing natural feedback due to a movement of the effected limb may be replaced by a synthetic feedback based on the motor-related brain function. To generate such a synthetic feedback a system is needed which measures, detects, localizes and visualizes the motor related µ-rhythm. Fast therapeutic success can only be achieved if the feedback features high specificity, comes in real-time and without large delay. We describe such an approach that offers a 3D visualization of µ-rhythms in real time with a delay of 500ms. This is accomplished by combining smart EEG preprocessing in the frequency domain with source localization techniques. The algorithm first selects the EEG channel featuring the most prominent rhythm in the alpha frequency band from a so-called motor channel set (C4, CZ, C3; CP6, CP4, CP2, CP1, CP3, CP5). If the amplitude in the alpha frequency band of this certain electrode exceeds a threshold, a µ-rhythm is detected. To prevent detection of a mixture of posterior alpha activity and µ-activity, the amplitudes in the alpha band outside the motor channel set are not allowed to be in the same range as the main channel. The EEG signal of the main channel is used as template for calculating the spatial distribution of the µ - rhythm over all electrodes. This spatial distribution is the input for a inverse method which provides the 3D distribution of the µ - activity within the brain which is visualized in 3D as color coded activity map. This approach mitigates the influence of lid artifacts on the localization performance. The first results of several healthy subjects show that the system is capable of detecting and localizing the rarely appearing µ-rhythm. In most cases the results match with findings from visual EEG analysis. Frequent eye-lid artifacts have no influence on the system performance. Furthermore, the system will be able to run in real-time. Due to the design of the frequency transformation the processing delay is 500ms. First results are promising and we plan to extend the test data set to further evaluate the performance of the system. The relevance of the system with respect to the therapy of stroke patients has to be shown in studies with real patients after CE certification of the system. This work was performed within the project ‘LiveSolo’ funded by the Austrian Research Promotion Agency (FFG) (project number: 853263).

Keywords: real-time EEG neuroimaging, neurofeedback, stroke, EEG–signal processing, rehabilitation

Procedia PDF Downloads 362
110 An Approach on Intelligent Tolerancing of Car Body Parts Based on Historical Measurement Data

Authors: Kai Warsoenke, Maik Mackiewicz

Abstract:

To achieve a high quality of assembled car body structures, tolerancing is used to ensure a geometric accuracy of the single car body parts. There are two main techniques to determine the required tolerances. The first is tolerance analysis which describes the influence of individually tolerated input values on a required target value. Second is tolerance synthesis to determine the location of individual tolerances to achieve a target value. Both techniques are based on classical statistical methods, which assume certain probability distributions. To ensure competitiveness in both saturated and dynamic markets, production processes in vehicle manufacturing must be flexible and efficient. The dimensional specifications selected for the individual body components and the resulting assemblies have a major influence of the quality of the process. For example, in the manufacturing of forming tools as operating equipment or in the higher level of car body assembly. As part of the metrological process monitoring, manufactured individual parts and assemblies are recorded and the measurement results are stored in databases. They serve as information for the temporary adjustment of the production processes and are interpreted by experts in order to derive suitable adjustments measures. In the production of forming tools, this means that time-consuming and costly changes of the tool surface have to be made, while in the body shop, uncertainties that are difficult to control result in cost-intensive rework. The stored measurement results are not used to intelligently design tolerances in future processes or to support temporary decisions based on real-world geometric data. They offer potential to extend the tolerancing methods through data analysis and machine learning models. The purpose of this paper is to examine real-world measurement data from individual car body components, as well as assemblies, in order to develop an approach for using the data in short-term actions and future projects. For this reason, the measurement data will be analyzed descriptively in the first step in order to characterize their behavior and to determine possible correlations. In the following, a database is created that is suitable for developing machine learning models. The objective is to create an intelligent way to determine the position and number of measurement points as well as the local tolerance range. For this a number of different model types are compared and evaluated. The models with the best result are used to optimize equally distributed measuring points on unknown car body part geometries and to assign tolerance ranges to them. The current results of this investigation are still in progress. However, there are areas of the car body parts which behave more sensitively compared to the overall part and indicate that intelligent tolerancing is useful here in order to design and control preceding and succeeding processes more efficiently.

Keywords: automotive production, machine learning, process optimization, smart tolerancing

Procedia PDF Downloads 89
109 Unveiling Drought Dynamics in the Cuneo District, Italy: A Machine Learning-Enhanced Hydrological Modelling Approach

Authors: Mohammadamin Hashemi, Mohammadreza Kashizadeh

Abstract:

Droughts pose a significant threat to sustainable water resource management, agriculture, and socioeconomic sectors, particularly in the field of climate change. This study investigates drought simulation using rainfall-runoff modelling in the Cuneo district, Italy, over the past 60-year period. The study leverages the TUW model, a lumped conceptual rainfall-runoff model with a semi-distributed operation capability. Similar in structure to the widely used Hydrologiska Byråns Vattenbalansavdelning (HBV) model, the TUW model operates on daily timesteps for input and output data specific to each catchment. It incorporates essential routines for snow accumulation and melting, soil moisture storage, and streamflow generation. Multiple catchments' discharge data within the Cuneo district form the basis for thorough model calibration employing the Kling-Gupta Efficiency (KGE) metric. A crucial metric for reliable drought analysis is one that can accurately represent low-flow events during drought periods. This ensures that the model provides a realistic picture of water availability during these critical times. Subsequent validation of monthly discharge simulations thoroughly evaluates overall model performance. Beyond model development, the investigation delves into drought analysis using the robust Standardized Runoff Index (SRI). This index allows for precise characterization of drought occurrences within the study area. A meticulous comparison of observed and simulated discharge data is conducted, with particular focus on low-flow events that characterize droughts. Additionally, the study explores the complex interplay between land characteristics (e.g., soil type, vegetation cover) and climate variables (e.g., precipitation, temperature) that influence the severity and duration of hydrological droughts. The study's findings demonstrate successful calibration of the TUW model across most catchments, achieving commendable model efficiency. Comparative analysis between simulated and observed discharge data reveals significant agreement, especially during critical low-flow periods. This agreement is further supported by the Pareto coefficient, a statistical measure of goodness-of-fit. The drought analysis provides critical insights into the duration, intensity, and severity of drought events within the Cuneo district. This newfound understanding of spatial and temporal drought dynamics offers valuable information for water resource management strategies and drought mitigation efforts. This research deepens our understanding of drought dynamics in the Cuneo region. Future research directions include refining hydrological modelling techniques and exploring future drought projections under various climate change scenarios.

Keywords: hydrologic extremes, hydrological drought, hydrological modelling, machine learning, rainfall-runoff modelling

Procedia PDF Downloads 17
108 The Effects of Adding Vibrotactile Feedback to Upper Limb Performance during Dual-Tasking and Response to Misleading Visual Feedback

Authors: Sigal Portnoy, Jason Friedman, Eitan Raveh

Abstract:

Introduction: Sensory substitution is possible due to the capacity of our brain to adapt to information transmitted by a synthetic receptor via an alternative sensory system. Practical sensory substitution systems are being developed in order to increase the functionality of individuals with sensory loss, e.g. amputees. For upper limb prosthetic-users the loss of tactile feedback compels them to allocate visual attention to their prosthesis. The effect of adding vibrotactile feedback (VTF) to the applied force has been studied, however its effect on the allocation if visual attention during dual-tasking and the response during misleading visual feedback have not been studied. We hypothesized that VTF will improve the performance and reduce visual attention during dual-task assignments in healthy individuals using a robotic hand and improve the performance in a standardized functional test, despite the presence of misleading visual feedback. Methods: For the dual-task paradigm, twenty healthy subjects were instructed to toggle two keyboard arrow keys with the left hand to retain a moving virtual car on a road on a screen. During the game, instructions for various activities, e.g. mix the sugar in the glass with a spoon, appeared on the screen. The subject performed these tasks with a robotic hand, attached to the right hand. The robotic hand was controlled by the activity of the flexors and extensors of the right wrist, recorded using surface EMG electrodes. Pressure sensors were attached at the tips of the robotic hand and induced VTF using vibrotactile actuators attached to the right arm of the subject. An eye-tracking system tracked to visual attention of the subject during the trials. The trials were repeated twice, with and without the VTF. Additionally, the subjects performed the modified box and blocks, hidden from eyesight, in a motion laboratory. A virtual presentation of a misleading visual feedback was be presented on a screen so that twice during the trial, the virtual block fell while the physical block was still held by the subject. Results: This is an ongoing study, which current results are detailed below. We are continuing these trials with transradial myoelectric prosthesis-users. In the healthy group, the VTF did not reduce the visual attention or improve performance during dual-tasking for the tasks that were typed transfer-to-target, e.g. place the eraser on the shelf. An improvement was observed for other tasks. For example, the average±standard deviation of time to complete the sugar-mixing task was 13.7±17.2s and 19.3±9.1s with and without the VTF, respectively. Also, the number of gaze shifts from the screen to the hand during this task were 15.5±23.7 and 20.0±11.6, with and without the VTF, respectively. The response of the subjects to the misleading visual feedback did not differ between the two conditions, i.e. with and without VTF. Conclusions: Our interim results suggest that the performance of certain activities of daily living may be improved by VTF. The substitution of visual sensory input by tactile feedback might require a long training period so that brain plasticity can occur and allow adaptation to the new condition.

Keywords: prosthetics, rehabilitation, sensory substitution, upper limb amputation

Procedia PDF Downloads 314
107 Reading and Writing Memories in Artificial and Human Reasoning

Authors: Ian O'Loughlin

Abstract:

Memory networks aim to integrate some of the recent successes in machine learning with a dynamic memory base that can be updated and deployed in artificial reasoning tasks. These models involve training networks to identify, update, and operate over stored elements in a large memory array in order, for example, to ably perform question and answer tasks parsing real-world and simulated discourses. This family of approaches still faces numerous challenges: the performance of these network models in simulated domains remains considerably better than in open, real-world domains, wide-context cues remain elusive in parsing words and sentences, and even moderately complex sentence structures remain problematic. This innovation, employing an array of stored and updatable ‘memory’ elements over which the system operates as it parses text input and develops responses to questions, is a compelling one for at least two reasons: first, it addresses one of the difficulties that standard machine learning techniques face, by providing a way to store a large bank of facts, offering a way forward for the kinds of long-term reasoning that, for example, recurrent neural networks trained on a corpus have difficulty performing. Second, the addition of a stored long-term memory component in artificial reasoning seems psychologically plausible; human reasoning appears replete with invocations of long-term memory, and the stored but dynamic elements in the arrays of memory networks are deeply reminiscent of the way that human memory is readily and often characterized. However, this apparent psychological plausibility is belied by a recent turn in the study of human memory in cognitive science. In recent years, the very notion that there is a stored element which enables remembering, however dynamic or reconstructive it may be, has come under deep suspicion. In the wake of constructive memory studies, amnesia and impairment studies, and studies of implicit memory—as well as following considerations from the cognitive neuroscience of memory and conceptual analyses from the philosophy of mind and cognitive science—researchers are now rejecting storage and retrieval, even in principle, and instead seeking and developing models of human memory wherein plasticity and dynamics are the rule rather than the exception. In these models, storage is entirely avoided by modeling memory using a recurrent neural network designed to fit a preconceived energy function that attains zero values only for desired memory patterns, so that these patterns are the sole stable equilibrium points in the attractor network. So although the array of long-term memory elements in memory networks seem psychologically appropriate for reasoning systems, they may actually be incurring difficulties that are theoretically analogous to those that older, storage-based models of human memory have demonstrated. The kind of emergent stability found in the attractor network models more closely fits our best understanding of human long-term memory than do the memory network arrays, despite appearances to the contrary.

Keywords: artificial reasoning, human memory, machine learning, neural networks

Procedia PDF Downloads 236
106 Valuing Cultural Ecosystem Services of Natural Treatment Systems Using Crowdsourced Data

Authors: Andrea Ghermandi

Abstract:

Natural treatment systems such as constructed wetlands and waste stabilization ponds are increasingly used to treat water and wastewater from a variety of sources, including stormwater and polluted surface water. The provision of ancillary benefits in the form of cultural ecosystem services makes these systems unique among water and wastewater treatment technologies and greatly contributes to determine their potential role in promoting sustainable water management practices. A quantitative analysis of these benefits, however, has been lacking in the literature. Here, a critical assessment of the recreational and educational benefits in natural treatment systems is provided, which combines observed public use from a survey of managers and operators with estimated public use as obtained using geotagged photos from social media as a proxy for visitation rates. Geographic Information Systems (GIS) are used to characterize the spatial boundaries of 273 natural treatment systems worldwide. Such boundaries are used as input for the Application Program Interfaces (APIs) of two popular photo-sharing websites (Flickr and Panoramio) in order to derive the number of photo-user-days, i.e., the number of yearly visits by individual photo users in each site. The adequateness and predictive power of four univariate calibration models using the crowdsourced data as a proxy for visitation are evaluated. A high correlation is found between photo-user-days and observed annual visitors (Pearson's r = 0.811; p-value < 0.001; N = 62). Standardized Major Axis (SMA) regression is found to outperform Ordinary Least Squares regression and count data models in terms of predictive power insofar as standard verification statistics – such as the root mean square error of prediction (RMSEP), the mean absolute error of prediction (MAEP), the reduction of error (RE), and the coefficient of efficiency (CE) – are concerned. The SMA regression model is used to estimate the intensity of public use in all 273 natural treatment systems. System type, influent water quality, and area are found to statistically affect public use, consistently with a priori expectations. Publicly available information regarding the home location of the sampled visitors is derived from their social media profiles and used to infer the distance they are willing to travel to visit the natural treatment systems in the database. Such information is analyzed using the travel cost method to derive monetary estimates of the recreational benefits of the investigated natural treatment systems. Overall, the findings confirm the opportunities arising from an integrated design and management of natural treatment systems, which combines the objectives of water quality enhancement and provision of cultural ecosystem services through public use in a multi-functional approach and compatibly with the need to protect public health.

Keywords: constructed wetlands, cultural ecosystem services, ecological engineering, waste stabilization ponds

Procedia PDF Downloads 159
105 Seismic Response Control of Multi-Span Bridge Using Magnetorheological Dampers

Authors: B. Neethu, Diptesh Das

Abstract:

The present study investigates the performance of a semi-active controller using magneto-rheological dampers (MR) for seismic response reduction of a multi-span bridge. The application of structural control to the structures during earthquake excitation involves numerous challenges such as proper formulation and selection of the control strategy, mathematical modeling of the system, uncertainty in system parameters and noisy measurements. These problems, however, need to be tackled in order to design and develop controllers which will efficiently perform in such complex systems. A control algorithm, which can accommodate un-certainty and imprecision compared to all the other algorithms mentioned so far, due to its inherent robustness and ability to cope with the parameter uncertainties and imprecisions, is the sliding mode algorithm. A sliding mode control algorithm is adopted in the present study due to its inherent stability and distinguished robustness to system parameter variation and external disturbances. In general a semi-active control scheme using an MR damper requires two nested controllers: (i) an overall system controller, which derives the control force required to be applied to the structure and (ii) an MR damper voltage controller which determines the voltage required to be supplied to the damper in order to generate the desired control force. In the present study a sliding mode algorithm is used to determine the desired optimal force. The function of the voltage controller is to command the damper to produce the desired force. The clipped optimal algorithm is used to find the command voltage supplied to the MR damper which is regulated by a semi active control law based on sliding mode algorithm. The main objective of the study is to propose a robust semi active control which can effectively control the responses of the bridge under real earthquake ground motions. Lumped mass model of the bridge is developed and time history analysis is carried out by solving the governing equations of motion in the state space form. The effectiveness of MR dampers is studied by analytical simulations by subjecting the bridge to real earthquake records. In this regard, it may also be noted that the performance of controllers depends, to a great extent, on the characteristics of the input ground motions. Therefore, in order to study the robustness of the controller in the present study, the performance of the controllers have been investigated for fourteen different earthquake ground motion records. The earthquakes are chosen in such a way that all possible characteristic variations can be accommodated. Out of these fourteen earthquakes, seven are near-field and seven are far-field. Also, these earthquakes are divided into different frequency contents, viz, low-frequency, medium-frequency, and high-frequency earthquakes. The responses of the controlled bridge are compared with the responses of the corresponding uncontrolled bridge (i.e., the bridge without any control devices). The results of the numerical study show that the sliding mode based semi-active control strategy can substantially reduce the seismic responses of the bridge showing a stable and robust performance for all the earthquakes.

Keywords: bridge, semi active control, sliding mode control, MR damper

Procedia PDF Downloads 107
104 A Comparative Assessment of Information Value, Fuzzy Expert System Models for Landslide Susceptibility Mapping of Dharamshala and Surrounding, Himachal Pradesh, India

Authors: Kumari Sweta, Ajanta Goswami, Abhilasha Dixit

Abstract:

Landslide is a geomorphic process that plays an essential role in the evolution of the hill-slope and long-term landscape evolution. But its abrupt nature and the associated catastrophic forces of the process can have undesirable socio-economic impacts, like substantial economic losses, fatalities, ecosystem, geomorphologic and infrastructure disturbances. The estimated fatality rate is approximately 1person /100 sq. Km and the average economic loss is more than 550 crores/year in the Himalayan belt due to landslides. This study presents a comparative performance of a statistical bivariate method and a machine learning technique for landslide susceptibility mapping in and around Dharamshala, Himachal Pradesh. The final produced landslide susceptibility maps (LSMs) with better accuracy could be used for land-use planning to prevent future losses. Dharamshala, a part of North-western Himalaya, is one of the fastest-growing tourism hubs with a total population of 30,764 according to the 2011 census and is amongst one of the hundred Indian cities to be developed as a smart city under PM’s Smart Cities Mission. A total of 209 landslide locations were identified in using high-resolution linear imaging self-scanning (LISS IV) data. The thematic maps of parameters influencing landslide occurrence were generated using remote sensing and other ancillary data in the GIS environment. The landslide causative parameters used in the study are slope angle, slope aspect, elevation, curvature, topographic wetness index, relative relief, distance from lineaments, land use land cover, and geology. LSMs were prepared using information value (Info Val), and Fuzzy Expert System (FES) models. Info Val is a statistical bivariate method, in which information values were calculated as the ratio of the landslide pixels per factor class (Si/Ni) to the total landslide pixel per parameter (S/N). Using this information values all parameters were reclassified and then summed in GIS to obtain the landslide susceptibility index (LSI) map. The FES method is a machine learning technique based on ‘mean and neighbour’ strategy for the construction of fuzzifier (input) and defuzzifier (output) membership function (MF) structure, and the FR method is used for formulating if-then rules. Two types of membership structures were utilized for membership function Bell-Gaussian (BG) and Trapezoidal-Triangular (TT). LSI for BG and TT were obtained applying membership function and if-then rules in MATLAB. The final LSMs were spatially and statistically validated. The validation results showed that in terms of accuracy, Info Val (83.4%) is better than BG (83.0%) and TT (82.6%), whereas, in terms of spatial distribution, BG is best. Hence, considering both statistical and spatial accuracy, BG is the most accurate one.

Keywords: bivariate statistical techniques, BG and TT membership structure, fuzzy expert system, information value method, machine learning technique

Procedia PDF Downloads 105