Search results for: cognitive performance test
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 8163

Search results for: cognitive performance test

63 Conflation Methodology Applied to Flood Recovery

Authors: E. L. Suarez, D. E. Meeroff, Y. Yong

Abstract:

Current flooding risk modeling focuses on resilience, defined as the probability of recovery from a severe flooding event. However, the long-term damage to property and well-being by nuisance flooding and its long-term effects on communities are not typically included in risk assessments. An approach was developed to address the probability of recovering from a severe flooding event combined with the probability of community performance during a nuisance event. A consolidated model, namely the conflation flooding recovery (&FR) model, evaluates risk-coping mitigation strategies for communities based on the recovery time from catastrophic events, such as hurricanes or extreme surges, and from everyday nuisance flooding events. The &FR model assesses the variation contribution of each independent input and generates a weighted output that favors the distribution with minimum variation. This approach is especially useful if the input distributions have dissimilar variances. The &FR is defined as a single distribution resulting from the product of the individual probability density functions. The resulting conflated distribution resides between the parent distributions, and it infers the recovery time required by a community to return to basic functions, such as power, utilities, transportation, and civil order, after a flooding event. The &FR model is more accurate than averaging individual observations before calculating the mean and variance or averaging the probabilities evaluated at the input values, which assigns the same weighted variation to each input distribution. The main disadvantage of these traditional methods is that the resulting measure of central tendency is exactly equal to the average of the input distribution’s means without the additional information provided by each individual distribution variance. When dealing with exponential distributions, such as resilience from severe flooding events and from nuisance flooding events, conflation results are equivalent to the weighted least squares method or best linear unbiased estimation. The combination of severe flooding risk with nuisance flooding improves flood risk management for highly populated coastal communities, such as in South Florida, USA, and provides a method to estimate community flood recovery time more accurately from two different sources, severe flooding events and nuisance flooding events.

Keywords: Community resilience, conflation, flood risk, nuisance flooding.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 67
62 Florida’s Groundwater and Surface Water System Reliability in Terms of Climate Change and Sea-Level Rise

Authors: Rahman Davtalab, Saba Ghotbi

Abstract:

Florida is one of the most vulnerable states to natural disasters among the 50 states of the USA. The state exposed by tropical storms, hurricanes, storm surge, landslide, etc. Besides the mentioned natural phenomena, global warming, sea-level rise, and other anthropogenic environmental changes make a very complicated and unpredictable system for decision-makers. In this study, we tried to highlight the effects of climate change and sea-level rise on surface water and groundwater systems for three different geographical locations in Florida; Main Canal of Jacksonville Beach in the northeast of Florida adjacent to the Atlantic Ocean, Grace Lake in central Florida, far away from surrounded coastal line, and Mc Dill in Florida and adjacent to Tampa Bay and Mexican Gulf. An integrated hydrologic and hydraulic model was developed and simulated for all three cases, including surface water, groundwater, or a combination of both. For the case study of Main Canal-Jacksonville Beach, the investigation showed that a 76 cm sea-level rise in time horizon 2060 could increase the flow velocity of the tide cycle for the main canal's outlet and headwater. This case also revealed how the sea level rise could change the tide duration, potentially affecting the coastal ecosystem. As expected, sea-level rise can raise the groundwater level. Therefore, for the Mc Dill case, the effect of groundwater rise on soil storage and the performance of stormwater retention ponds is investigated. The study showed that sea-level rise increased the pond’s seasonal high water up to 40 cm by time horizon 2060. The reliability of the retention pond is dropped from 99% for the current condition to 54% for the future. The results also proved that the retention pond could not retain and infiltrate the designed treatment volume within 72 hours, which is a significant indication of increasing pollutants in the future. Grace Lake case study investigates the effects of climate change on groundwater recharge. This study showed that using the dynamically downscaled data of the groundwater recharge can decline up to 24 % by the mid-21st century. 

Keywords: groundwater, surface water, Florida, retention pond, tide, sea-level rise

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 529
61 The Efficiency of Mechanization in Weed Control in Artificial Regeneration of Oriental Beech (Fagus orientalis Lipsky.)

Authors: Tuğrul Varol, Halil Barış Özel

Abstract:

In this study which has been conducted in Akçasu Forest Range District of Devrek Forest Directorate; 3 methods (weed control with labourer power, cover removal with Hitachi F20 Excavator, and weed control with agricultural equipment mounted on a Ferguson 240S agriculture tractor) were utilized in weed control efforts in regeneration of degraded oriental beech forests have been compared. In this respect, 3 methods have been compared by determining certain work hours and standard durations of unit areas (1 hectare). For this purpose, evaluating the tasks made with human and machine force from the aspects of duration, productivity and costs, it has been aimed to determine the most productive method in accordance with the actual ecological conditions of research field. Within the scope of the study, the time studies have been conducted for 3 methods used in weed control efforts. While carrying out those studies, the performed implementations have been evaluated by dividing them into business stages. Also, the actual data have been used while calculating the cost accounts. In those calculations, the latest formulas and equations which are also used in developed countries have been utilized. The variance of analysis (ANOVA) was used in order to determine whether there is any statistically significant difference among obtained results, and the Duncan test was used for grouping if there is significant difference. According to the measurements and findings carried out within the scope of this study, it has been found during living cover removal efforts in regeneration efforts in demolished oriental beech forests that the removal of weed layer in 1 hectare of field has taken 920 hours with labourer force, 15.1 hours with excavator and 60 hours with an equipment mounted on a tractor. On the other hand, it has been determined that the cost of removal of living cover in unit area (1 hectare) was 3220.00 TL for labourer power, 1250 TL for excavator and 1825 TL for equipment mounted on a tractor. According to the obtained results, it has been found that the utilization of excavator in weed control effort in regeneration of degraded oriental beech regions under actual ecological conditions of research field has been found to be more productive from both of aspects of duration and costs. These determinations carried out should be repeated in weed control efforts in degraded forest fields with different ecological conditions, it is compulsory for finding the most efficient weed control method. These findings will light the way of technical staff of forestry directorate in determination of the most effective and economic weed control method. Thus, the more actual data will be used while preparing the weed control budgets, and there will be significant contributions to national economy. Also the results of this and similar studies are very important for developing the policies for our forestry in short and long term.

Keywords: Artificial regeneration, weed control, oriental beech, productivity, mechanization, man power, cost analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1977
60 Predictive Factors of Exercise Behaviors of Junior High School Students in Chonburi Province

Authors: Tanida Julvanichpong

Abstract:

Exercise has been regarded as a necessary and important aspect to enhance physical performance and psychology health. Body weight statistics of students in junior high school students in Chonburi Province beyond a standard risk of obesity. Promoting exercise among Junior high school students in Chonburi Province, essential knowledge concerning factors influencing exercise is needed. Therefore, this study aims to (1) determine the levels of perceived exercise behavior, exercise behavior in the past, perceived barriers to exercise, perceived benefits of exercise, perceived self-efficacy to exercise, feelings associated with exercise behavior, influence of the family to exercise, influence of friends to exercise, and the perceived influence of the environment on exercise. (2) examine the predicting ability of each of the above factors while including personal factors (sex, educational level) for exercise behavior. Pender’s Health Promotion Model was used as a guide for the study. Sample included 652 students in junior high schools, Chonburi Provience. The samples were selected by Multi-Stage Random Sampling. Data Collection has been done by using self-administered questionnaires. Data were analyzed using descriptive statistics, Pearson’s product moment correlation coefficient, Eta, and stepwise multiple regression analysis. The research results showed that: 1. Perceived benefits of exercise, influence of teacher, influence of environmental, feelings associated with exercise behavior were at a high level. Influence of the family to exercise, exercise behavior, exercise behavior in the past, perceived self-efficacy to exercise and influence of friends were at a moderate level. Perceived barriers to exercise were at a low level. 2. Exercise behavior was positively significant related to perceived benefits of exercise, influence of the family to exercise, exercise behavior in the past, perceived self-efficacy to exercise, influence of friends, influence of teacher, influence of environmental and feelings associated with exercise behavior (p < .01, respectively) and was negatively significant related to educational level and perceived barriers to exercise (p < .01, respectively). Exercise behavior was significant related to sex (Eta = 0.243, p=.000). 3. Exercise behavior in the past, influence of the family to exercise significantly contributed 60.10 percent of the variance to the prediction of exercise behavior in male students (p < .01). Exercise behavior in the past, perceived self-efficacy to exercise, perceived barriers to exercise, and educational level significantly contributed 52.60 percent of the variance to the prediction of exercise behavior in female students (p < .01).

Keywords: Predictive factors, exercise behaviors, junior high school.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1151
59 Microstructure and Mechanical Characterization of Heat Treated Stir Cast Silica (Sea Sand) Reinforced 7XXX Al Alloy MMCs

Authors: S. S. Sharma, Jagannath K, P. R. Prabhu

Abstract:

Metal matrix composites consists of a metallic matrix combined with dispersed particulate phase as reinforcement. Aluminum alloys have been the primary material of choice for structural components of aircraft since about 1930. Well known performance characteristics, known fabrication costs, design experience, and established manufacturing methods and facilities, are just a few of the reasons for the continued confidence in 7XXX Al alloys that will ensure their use in significant quantities for the time to come. Particulate MMCs are of special interest owing to the low cost of their raw materials (primarily natural river sand here) and their ease of fabrication, making them suitable for applications requiring relatively high volume production. 7XXX Al alloys are precipitation hardenable and therefore amenable for thermomechanical treatment. Al–Zn alloys reinforced with particulate materials are used in aerospace industries in spite of the drawbacks of susceptibility to stress corrosion, poor wettability, poor weldability and poor fatigue resistance. The resistance offered by these particulates for the moving dislocations impart secondary hardening in turn contributes strain hardening. Cold deformation increases lattice defects, which in turn improves the properties of solution treated alloy. In view of this, six different Al–Zn–Mg alloy composites reinforced with silica (3 wt. % and 5 wt. %) are prepared by conventional semisolid synthesizing process. The cast alloys are solution treated and aged. The solution treated alloys are further severely cold rolled to enhance the properties. The hardness and strength values are analyzed and compared with silica free Al – Zn-Mg alloys. Precipitation hardening phenomena is accelerated due to the increased number of potential sites for precipitation. Higher peak hardness and lesser aging time are the characteristics of thermo mechanically treated samples. For obtaining maximum hardness, optimum number and volume of precipitate particles are required. The Al-5Zn-1Mg with 5% SiO2 alloy composite shows better result.

Keywords: Dislocation, hardness, matrix, thermomechanical, precipitation hardening, reinforcement.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1819
58 Achieving Design-Stage Elemental Cost Planning Accuracy: Case Study of New Zealand

Authors: Johnson Adafin, James O. B. Rotimi, Suzanne Wilkinson, Abimbola O. Windapo

Abstract:

An aspect of client expenditure management that requires attention is the level of accuracy achievable in design-stage elemental cost planning. This has been a major concern for construction clients and practitioners in New Zealand (NZ). Pre-tender estimating inaccuracies are significantly influenced by the level of risk information available to estimators. Proper cost planning activities should ensure the production of a project’s likely construction costs (initial and final), and subsequent cost control activities should prevent unpleasant consequences of cost overruns, disputes and project abandonment. If risks were properly identified and priced at the design stage, observed variance between design-stage elemental cost plans (ECPs) and final tender sums (FTS) (initial contract sums) could be reduced. This study investigates the variations between design-stage ECPs and FTS of construction projects, with a view to identifying risk factors that are responsible for the observed variance. Data were sourced through interviews, and risk factors were identified by using thematic analysis. Access was obtained to project files from the records of study participants (consultant quantity surveyors), and document analysis was employed in complementing the responses from the interviews. Study findings revealed the discrepancies between ECPs and FTS in the region of -14% and +16%. It is opined in this study that the identified risk factors were responsible for the variability observed. The values obtained from the analysis would enable greater accuracy in the forecast of FTS by Quantity Surveyors. Further, whilst inherent risks in construction project developments are observed globally, these findings have important ramifications for construction projects by expanding existing knowledge on what is needed for reasonable budgetary performance and successful delivery of construction projects. The findings contribute significantly to the study by providing quantitative confirmation to justify the theoretical conclusions generated in the literature from around the world. This therefore adds to and consolidates existing knowledge.

Keywords: Accuracy, design-stage, elemental cost plan, final tender sum, New Zealand.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1751
57 C-LNRD: A Cross-Layered Neighbor Route Discovery for Effective Packet Communication in Wireless Sensor Network

Authors: K. Kalaikumar, E. Baburaj

Abstract:

One of the problems to be addressed in wireless sensor networks is the issues related to cross layer communication. Cross layer architecture shares the information across the layer, ensuring Quality of Services (QoS). With this shared information, MAC protocol adapts effective functionality maintenance such as route selection on changeable sensor network environment. However, time slot assignment and neighbour route selection time duration for cross layer have not been carried out. The time varying physical layer communication over cross layer causes high traffic load in the sensor network. Though, the traffic load was reduced using cross layer optimization procedure, the computational cost is high. To improve communication efficacy in the sensor network, a self-determined time slot based Cross-Layered Neighbour Route Discovery (C-LNRD) method is presented in this paper. In the presented work, the initial process is to discover the route in the sensor network using Dynamic Source Routing based Medium Access Control (MAC) sub layers. This process considers MAC layer operation with dynamic route neighbour table discovery. Then, the discovered route path for packet communication employs Broad Route Distributed Time Slot Assignment method on Cross-Layered Sensor Network system. Broad Route means time slotting on varying length of the route paths. During packet communication in this sensor network, transmission of packets is adjusted over the different time with varying ranges for controlling the traffic rate. Finally, Rayleigh fading model is developed in C-LNRD to identify the performance of the sensor network communication structure. The main task of Rayleigh Fading is to measure the power level of each communication under MAC sub layer. The minimized power level helps to easily reduce the computational cost of packet communication in the sensor network. Experiments are conducted on factors such as power factor, on packet communication, neighbour route discovery time, and information (i.e., packet) propagation speed.

Keywords: Medium access control, neighbour route discovery, wireless sensor network, Rayleigh fading, distributed time slot assignment

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 740
56 Detecting Tomato Flowers in Greenhouses Using Computer Vision

Authors: Dor Oppenheim, Yael Edan, Guy Shani

Abstract:

This paper presents an image analysis algorithm to detect and count yellow tomato flowers in a greenhouse with uneven illumination conditions, complex growth conditions and different flower sizes. The algorithm is designed to be employed on a drone that flies in greenhouses to accomplish several tasks such as pollination and yield estimation. Detecting the flowers can provide useful information for the farmer, such as the number of flowers in a row, and the number of flowers that were pollinated since the last visit to the row. The developed algorithm is designed to handle the real world difficulties in a greenhouse which include varying lighting conditions, shadowing, and occlusion, while considering the computational limitations of the simple processor in the drone. The algorithm identifies flowers using an adaptive global threshold, segmentation over the HSV color space, and morphological cues. The adaptive threshold divides the images into darker and lighter images. Then, segmentation on the hue, saturation and volume is performed accordingly, and classification is done according to size and location of the flowers. 1069 images of greenhouse tomato flowers were acquired in a commercial greenhouse in Israel, using two different RGB Cameras – an LG G4 smartphone and a Canon PowerShot A590. The images were acquired from multiple angles and distances and were sampled manually at various periods along the day to obtain varying lighting conditions. Ground truth was created by manually tagging approximately 25,000 individual flowers in the images. Sensitivity analyses on the acquisition angle of the images, periods throughout the day, different cameras and thresholding types were performed. Precision, recall and their derived F1 score were calculated. Results indicate better performance for the view angle facing the flowers than any other angle. Acquiring images in the afternoon resulted with the best precision and recall results. Applying a global adaptive threshold improved the median F1 score by 3%. Results showed no difference between the two cameras used. Using hue values of 0.12-0.18 in the segmentation process provided the best results in precision and recall, and the best F1 score. The precision and recall average for all the images when using these values was 74% and 75% respectively with an F1 score of 0.73. Further analysis showed a 5% increase in precision and recall when analyzing images acquired in the afternoon and from the front viewpoint.

Keywords: Agricultural engineering, computer vision, image processing, flower detection.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2324
55 Creative Mapping Landuse and Human Activities: From the Inventories of Factories to the History of the City and Citizens

Authors: R. Tamborrino, F. Rinaudo

Abstract:

Digital technologies offer possibilities to effectively convert historical archives into instruments of knowledge able to provide a guide for the interpretation of historical phenomena. Digital conversion and management of those documents allow the possibility to add other sources in a unique and coherent model that permits the intersection of different data able to open new interpretations and understandings. Urban history uses, among other sources, the inventories that register human activities in a specific space (e.g. cadastres, censuses, etc.). The geographic localisation of that information inside cartographic supports allows for the comprehension and visualisation of specific relationships between different historical realities registering both the urban space and the peoples living there. These links that merge the different nature of data and documentation through a new organisation of the information can suggest a new interpretation of other related events. In all these kinds of analysis, the use of GIS platforms today represents the most appropriate answer. The design of the related databases is the key to realise the ad-hoc instrument to facilitate the analysis and the intersection of data of different origins. Moreover, GIS has become the digital platform where it is possible to add other kinds of data visualisation. This research deals with the industrial development of Turin at the beginning of the 20th century. A census of factories realized just prior to WWI provides the opportunity to test the potentialities of GIS platforms for the analysis of urban landscape modifications during the first industrial development of the town. The inventory includes data about location, activities, and people. GIS is shaped in a creative way linking different sources and digital systems aiming to create a new type of platform conceived as an interface integrating different kinds of data visualisation. The data processing allows linking this information to an urban space, and also visualising the growth of the city at that time. The sources, related to the urban landscape development in that period, are of a different nature. The emerging necessity to build, enlarge, modify and join different buildings to boost the industrial activities, according to their fast development, is recorded by different official permissions delivered by the municipality and now stored in the Historical Archive of the Municipality of Turin. Those documents, which are reports and drawings, contain numerous data on the buildings themselves, including the block where the plot is located, the district, and the people involved such as the owner, the investor, and the engineer or architect designing the industrial building. All these collected data offer the possibility to firstly re-build the process of change of the urban landscape by using GIS and 3D modelling technologies thanks to the access to the drawings (2D plans, sections and elevations) that show the previous and the planned situation. Furthermore, they access information for different queries of the linked dataset that could be useful for different research and targets such as economics, biographical, architectural, or demographical. By superimposing a layer of the present city, the past meets to the present-industrial heritage, and people meet urban history.

Keywords: Digital urban history, census, digitalisation, GIS, modelling, digital humanities.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1200
54 Review of the Model-Based Supply Chain Management Research in the Construction Industry

Authors: Aspasia Koutsokosta, Stefanos Katsavounis

Abstract:

This paper reviews the model-based qualitative and quantitative Operations Management research in the context of Construction Supply Chain Management (CSCM). Construction industry has been traditionally blamed for low productivity, cost and time overruns, waste, high fragmentation and adversarial relationships. The construction industry has been slower than other industries to employ the Supply Chain Management (SCM) concept and develop models that support the decision-making and planning. However the last decade there is a distinct shift from a project-based to a supply-based approach of construction management. CSCM comes up as a new promising management tool of construction operations and improves the performance of construction projects in terms of cost, time and quality. Modeling the Construction Supply Chain (CSC) offers the means to reap the benefits of SCM, make informed decisions and gain competitive advantage. Different modeling approaches and methodologies have been applied in the multi-disciplinary and heterogeneous research field of CSCM. The literature review reveals that a considerable percentage of the CSC modeling research accommodates conceptual or process models which present general management frameworks and do not relate to acknowledged soft Operations Research methods. We particularly focus on the model-based quantitative research and categorize the CSCM models depending on their scope, objectives, modeling approach, solution methods and software used. Although over the last few years there has been clearly an increase of research papers on quantitative CSC models, we identify that the relevant literature is very fragmented with limited applications of simulation, mathematical programming and simulation-based optimization. Most applications are project-specific or study only parts of the supply system. Thus, some complex interdependencies within construction are neglected and the implementation of the integrated supply chain management is hindered. We conclude this paper by giving future research directions and emphasizing the need to develop optimization models for integrated CSCM. We stress that CSC modeling needs a multi-dimensional, system-wide and long-term perspective. Finally, prior applications of SCM to other industries have to be taken into account in order to model CSCs, but not without translating the generic concepts to the context of construction industry.

Keywords: Construction supply chain management, modeling, operations research, optimization and simulation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2795
53 Production, Characterisation and Assessment of Biomixture Fuels for Compression Ignition Engine Application

Authors: K. Masera, A. K. Hossain

Abstract:

Hardly any neat biodiesel satisfies the European EN14214 standard for compression ignition engine application. To satisfy the EN14214 standard, various additives are doped into biodiesel; however, biodiesel additives might cause other problems such as increase in the particular emission and increased specific fuel consumption. In addition, the additives could be expensive. Considering the increasing level of greenhouse gas GHG emissions and fossil fuel depletion, it is forecasted that the use of biodiesel will be higher in the near future. Hence, the negative aspects of the biodiesel additives will likely to gain much more importance and need to be replaced with better solutions. This study aims to satisfy the European standard EN14214 by blending the biodiesels derived from sustainable feedstocks. Waste Cooking Oil (WCO) and Animal Fat Oil (AFO) are two sustainable feedstocks in the EU (including the UK) for producing biodiesels. In the first stage of the study, these oils were transesterified separately and neat biodiesels (W100 & A100) were produced. Secondly, the biodiesels were blended together in various ratios: 80% WCO biodiesel and 20% AFO biodiesel (W80A20), 60% WCO biodiesel and 40% AFO biodiesel (W60A40), 50% WCO biodiesel and 50% AFO biodiesel (W50A50), 30% WCO biodiesel and 70% AFO biodiesel (W30A70), 10% WCO biodiesel and 90% AFO biodiesel (W10A90). The prepared samples were analysed using Thermo Scientific Trace 1300 Gas Chromatograph and ISQ LT Mass Spectrometer (GC-MS). The GS-MS analysis gave Fatty Acid Methyl Ester (FAME) breakdowns of the fuel samples. It was found that total saturation degree of the samples was linearly increasing (from 15% for W100 to 54% for A100) as the percentage of the AFO biodiesel was increased. Furthermore, it was found that WCO biodiesel was mainly (82%) composed of polyunsaturated FAMEs. Cetane numbers, iodine numbers, calorific values, lower heating values and the densities (at 15 oC) of the samples were estimated by using the mass percentages data of the FAMEs. Besides, kinematic viscosities (at 40 °C and 20 °C), densities (at 15 °C), heating values and flash point temperatures of the biomixture samples were measured in the lab. It was found that estimated and measured characterisation results were comparable. The current study concluded that biomixture fuel samples W60A40 and W50A50 were perfectly satisfying the European EN 14214 norms without any need of additives. Investigation on engine performance, exhaust emission and combustion characteristics will be conducted to assess the full feasibility of the proposed biomixture fuels.

Keywords: Biodiesel, blending, characterisation, CI Engine.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 771
52 Solomon 300 OD (Betacyfluthrin+Imidacloprid): A Combi-Product for the Management of Insect-Pests of Chilli (Capsicum annum L.)

Authors: R. S. Giraddi, B. Thirupam Reddy, D. N. Kambrekar

Abstract:

Chilli (Capsicum annum L.) an important commercial vegetable crop is ravaged by a number of insect-pests during both vegetative and reproductive phase resulting into significant crop loss.Thrips, Scirtothripsdorsalis, mite, Polyphagotarsonemuslatus and whitefly, Bemisiatabaci are the key sap feeding insects, their infestation leads to leaf curl, stunted growth and yield loss.During flowering and fruit formation stage, gall midge fly, Asphondyliacapparis (Rubsaaman) infesting flower buds and young fruits andHelicoverpaarmigera (Hubner) feeding on matured green fruits are the important insect pests causing significant crop loss.The pest is known to infest both flower buds and young fruits resulting into malformation of flower buds and twisting of fruits.In order to manage these insect-pests a combi product consisting of imidacloprid and betacyfluthrin (Soloman 300 OD) was evaluated for its bio-efficacy, phytotoxicity and effect on predator activity.Imidacloprid, a systemic insecticide belonging to neo-nicotinoid group, is effective against insect pests such as aphids, whiteflies (sap feeders) and other insectsviz., termites and soil insects.Beta-Cyfluthrin is an insecticide of synthetic pyrethroid group which acts by contact action and ingestion. It acts on the insects' nervous system as sodium channel blocker consequently a disorder of the nervous system occurs leading finally to the death. The field experiments were taken up during 2015 and 2016 at the Main Agricultural Research Station of University of Agricultural Sciences, Dharwad, Karnataka, India.The trials were laid out in a Randomized Block Design (RBD) with three replications using popular land race of Byadagi crop variety.Results indicated that the product at 21.6 + 50.4% gai/ha (240 ml/ha) and 27.9 + 65% gai/ha (310 ml/ha) was found quite effective in controlling thrips (0.00 to 0.66 thrips per six leaves) as against the standard check insecticide recommended for thrips by the University of Agricultural Sciences, Dharwad wherein the density of thrips recorded was significantly higher (1.00 to 2.00 Nos./6 leaves). Similarly, the test insecticide was quite effective against other target insects, whiteflies, fruit borer and gall midge fly as indicated by lower insect population observed in the treatments as compared to standard insecticidal control. The predatory beetle activity was found to be normal in all experimental plots. Highest green fruit yield of 5100-5500 kg/ha was recorded in Soloman 300 OD applied crop at 310 ml/ha rate as compared to 4750 to 5050 kg/ha recorded in check. At present 6-8 sprays of insecticides are recommended for management of these insect-pests on the crop. If combi-products are used in pest management programmes, it is possible to reduce insecticide usages in crop ecosystem.

Keywords: Imidacloprid, Betacyfluthrin, gallmidge fly, thrips, chilli.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1780
51 Advanced Compound Coating for Delaying Corrosion of Fast-Dissolving Alloy in High Temperature and Corrosive Environment

Authors: Lei Zhao, Yi Song, Tim Dunne, Jiaxiang (Jason) Ren, Wenhan Yue, Lei Yang, Li Wen, Yu Liu

Abstract:

Fasting dissolving magnesium (DM) alloy technology has contributed significantly to the “Shale Revolution” in oil and gas industry. This application requires DM downhole tools dissolving initially at a slow rate, rapidly accelerating to a high rate after certain period of operation time (typically 8 h to 2 days), a contradicting requirement that can hardly be addressed by traditional Mg alloying or processing itself. Premature disintegration has been broadly reported in downhole DM tool from field trials. To address this issue, “temporary” thin polymers of various formulations are currently coated onto DM surface to delay its initial dissolving. Due to conveying parts, harsh downhole condition, and high dissolving rate of the base material, the current delay coatings relying on pure polymers are found to perform well only at low temperature (typical < 100 ℃) and parts without sharp edges or corners, as severe geometries prevent high quality thin film coatings from forming effectively. In this study, a coating technology combining Plasma Electrolytic Oxide (PEO) coatings with advanced thin film deposition has been developed, which can delay DM complex parts (with sharp corners) in corrosive fluid at 150 ℃ for over 2 days. Synergistic effects between porous hard PEO coating and chemical inert elastic-polymer sealing leads to its delaying dissolution improvement, and strong chemical/physical bonding between these two layers has been found to play essential role. Microstructure of this advanced coating and compatibility between PEO and various polymer selections has been thoroughly investigated and a model is also proposed to explain its delaying performance. This study could not only benefit oil and gas industry to unplug their High Temperature High Pressure (HTHP) unconventional resources inaccessible before, but also potentially provides a technical route for other industries (e.g., bio-medical, automobile, aerospace) where primer anti-corrosive protection on light Mg alloy is highly demanded.

Keywords: Dissolvable magnesium, coating, plasma electrolytic oxide, sealer.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 518
50 Towards End-To-End Disease Prediction from Raw Metagenomic Data

Authors: Maxence Queyrel, Edi Prifti, Alexandre Templier, Jean-Daniel Zucker

Abstract:

Analysis of the human microbiome using metagenomic sequencing data has demonstrated high ability in discriminating various human diseases. Raw metagenomic sequencing data require multiple complex and computationally heavy bioinformatics steps prior to data analysis. Such data contain millions of short sequences read from the fragmented DNA sequences and stored as fastq files. Conventional processing pipelines consist in multiple steps including quality control, filtering, alignment of sequences against genomic catalogs (genes, species, taxonomic levels, functional pathways, etc.). These pipelines are complex to use, time consuming and rely on a large number of parameters that often provide variability and impact the estimation of the microbiome elements. Training Deep Neural Networks directly from raw sequencing data is a promising approach to bypass some of the challenges associated with mainstream bioinformatics pipelines. Most of these methods use the concept of word and sentence embeddings that create a meaningful and numerical representation of DNA sequences, while extracting features and reducing the dimensionality of the data. In this paper we present an end-to-end approach that classifies patients into disease groups directly from raw metagenomic reads: metagenome2vec. This approach is composed of four steps (i) generating a vocabulary of k-mers and learning their numerical embeddings; (ii) learning DNA sequence (read) embeddings; (iii) identifying the genome from which the sequence is most likely to come and (iv) training a multiple instance learning classifier which predicts the phenotype based on the vector representation of the raw data. An attention mechanism is applied in the network so that the model can be interpreted, assigning a weight to the influence of the prediction for each genome. Using two public real-life data-sets as well a simulated one, we demonstrated that this original approach reaches high performance, comparable with the state-of-the-art methods applied directly on processed data though mainstream bioinformatics workflows. These results are encouraging for this proof of concept work. We believe that with further dedication, the DNN models have the potential to surpass mainstream bioinformatics workflows in disease classification tasks.

Keywords: Metagenomics, phenotype prediction, deep learning, embeddings, multiple instance learning.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 854
49 Obese and Overweight Women and Public Health Issues in Hillah City, Iraq

Authors: Amean A. Yasir, Zainab Kh. A. Al-Mahdi Al-Amean

Abstract:

In both developed and developing countries, obesity among women is increasing, but in different patterns and at very different speeds. It may have a negative effect on health, leading to reduced life expectancy and/or increased health problems. This research studied the age distribution among obese women, the types of overweight and obesity, and the extent of the problem of overweight/obesity and the obesity etiological factors among women in Hillah city in central Iraq. A total of 322 overweight and obese women were included in the study, those women were randomly selected. The Body Mass Index was used as indicator for overweight/ obesity. The incidence of overweight/obesity among age groups were estimated, the etiology factors included genetic, environmental, genetic/environmental and endocrine disease. The overweight and obese women were screened for incidence of infection and/or diseases. The study found that the prevalence of 322 overweight and obese women in Hillah city in central Iraq was 19.25% and 80.78%, respectively. The obese women types were recorded based on BMI and WHO classification as class-1 obesity (29.81%), class-2 obesity (24.22%) and class-3 obesity (26.70%), the result was discrepancy non-significant, P value < 0.05. The incidence of overweight in women was high among those aged 20-29 years (90.32%), 6.45% aged 30-39 years old and 3.22% among ≥ 60 years old, while the incidence of obesity was 20.38% for those in the age group 20-29 years, 17.30% were 30-39 years, 23.84% were 40-49 years, 16.92% were 50-59 years group and 21.53% were ≥ 60 years age group. These results confirm that the age can be considered as a significant factor for obesity types (P value < 0.0001). The result also showed that the both genetic factors and environmental factors were responsible for incidents of overweight or obesity (84.78%) p value < 0.0001. The results also recorded cases of different repeated infections (skin infection, recurrent UTI and influenza), cancer, gallstones, high blood pressure, type 2 diabetes, and infertility. Weight stigma and bias generally refers to negative attitudes; Obesity can affect quality of life, and the results of this study recorded depression among overweight or obese women. This can lead to sexual problems, shame and guilt, social isolation and reduced work performance. Overweight and Obesity are real problems among women of all age groups and is associated with the risk of diseases and infection and negatively affects quality of life. This result warrants further studies into the prevalence of obesity among women in Hillah City in central Iraq and the immune response of obese women.

Keywords: Obesity, overweight, Iraq, body mass index.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1242
48 Sensitivity Analysis of the Heat Exchanger Design in Net Power Oxy-Combustion Cycle for Carbon Capture

Authors: Hirbod Varasteh, Hamidreza Gohari Darabkhani

Abstract:

The global warming and its impact on climate change is one of main challenges for current century. Global warming is mainly due to the emission of greenhouse gases (GHG) and carbon dioxide (CO2) is known to be the major contributor to the GHG emission profile. Whilst the energy sector is the primary source for CO2 emission, Carbon Capture and Storage (CCS) are believed to be the solution for controlling this emission. Oxyfuel combustion (Oxy-combustion) is one of the major technologies for capturing CO2 from power plants. For gas turbines, several Oxy-combustion power cycles (Oxyturbine cycles) have been investigated by means of thermodynamic analysis. NetPower cycle is one of the leading oxyturbine power cycles with almost full carbon capture capability from a natural gas fired power plant. In this manuscript, sensitivity analysis of the heat exchanger design in NetPower cycle is completed by means of process modelling. The heat capacity variation and supercritical CO2 with gaseous admixtures are considered for multi-zone analysis with Aspen Plus software. It is found that the heat exchanger design has a major role to increase the efficiency of NetPower cycle. The pinch-point analysis is done to extract the composite and grand composite curve for the heat exchanger. In this paper, relationship between the cycle efficiency and the minimum approach temperature (∆Tmin) of the heat exchanger has also been evaluated.  Increase in ∆Tmin causes a decrease in the temperature of the recycle flue gases (RFG) and an overall decrease in the required power for the recycled gas compressor. The main challenge in the design of heat exchangers in power plants is a tradeoff between the capital and operational costs. To achieve lower ∆Tmin, larger size of heat exchanger is required. This means a higher capital cost but leading to a better heat recovery and lower operational cost. To achieve this, ∆Tmin is selected from the minimum point in the diagrams of capital and operational costs. This study provides an insight into the NetPower Oxy-combustion cycle’s performance analysis and operational condition based on its heat exchanger design.

Keywords: Carbon capture and storage, oxy-combustion, netpower cycle, oxyturbine power cycles, heat exchanger design, supercritical carbon dioxide, pinch point analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1651
47 Multi-Agent Searching Adaptation Using Levy Flight and Inferential Reasoning

Authors: Sagir M. Yusuf, Chris Baber

Abstract:

In this paper, we describe how to achieve knowledge understanding and prediction (Situation Awareness (SA)) for multiple-agents conducting searching activity using Bayesian inferential reasoning and learning. Bayesian Belief Network was used to monitor agents' knowledge about their environment, and cases are recorded for the network training using expectation-maximisation or gradient descent algorithm. The well trained network will be used for decision making and environmental situation prediction. Forest fire searching by multiple UAVs was the use case. UAVs are tasked to explore a forest and find a fire for urgent actions by the fire wardens. The paper focused on two problems: (i) effective agents’ path planning strategy and (ii) knowledge understanding and prediction (SA). The path planning problem by inspiring animal mode of foraging using Lévy distribution augmented with Bayesian reasoning was fully described in this paper. Results proof that the Lévy flight strategy performs better than the previous fixed-pattern (e.g., parallel sweeps) approaches in terms of energy and time utilisation. We also introduced a waypoint assessment strategy called k-previous waypoints assessment. It improves the performance of the ordinary levy flight by saving agent’s resources and mission time through redundant search avoidance. The agents (UAVs) are to report their mission knowledge at the central server for interpretation and prediction purposes. Bayesian reasoning and learning were used for the SA and results proof effectiveness in different environments scenario in terms of prediction and effective knowledge representation. The prediction accuracy was measured using learning error rate, logarithm loss, and Brier score and the result proves that little agents mission that can be used for prediction within the same or different environment. Finally, we described a situation-based knowledge visualization and prediction technique for heterogeneous multi-UAV mission. While this paper proves linkage of Bayesian reasoning and learning with SA and effective searching strategy, future works is focusing on simplifying the architecture.

Keywords: Lèvy flight, situation awareness, multi-agent system, multi-robot coordination, autonomous system, swarm intelligence.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 482
46 Malaria Prone Zones of West Bengal: A Spatio-Temporal Scenario

Authors: Meghna Maiti, Utpal Roy

Abstract:

In India, till today, malaria is considered to be one of the significant infectious diseases. Most of the cases regional geographical factors are the principal elements to let the places a unique identity. The incidence and intensity of infectious diseases are quite common and affect different places differently across the nation. The present study aims to identify spatial clusters of hot spots and cold spots of malaria incidence and their seasonal variation during the three periods of 2012-2014, 2015-2017 and 2018-20 in the state of West Bengal in India. As malaria is a vector-borne disease, numbers of positive test results are to be reported by the laboratories to the Department of Health, West Bengal (through the National Vector Borne Disease Control Programme). Data on block-wise monthly malaria positive cases are collected from Health Management Information System (HMIS), Ministry of Health and Family Welfare, Government of India. Moran’s I statistic is performed to assess the spatial autocorrelation of malaria incidence. The spatial statistical analysis mainly Local Indicators of Spatial Autocorrelation (LISA) cluster and Local Geary Cluster are applied to find the spatial clusters of hot spots and cold spots and seasonal variability of malaria incidence over the three periods. The result indicates that the spatial distribution of malaria is clustered during each of the three periods of 2012-2014, 2015-2017 and 2018-20. The analysis shows that in all the cases, high-high clusters are primarily concentrated in the western (Purulia, Paschim Medinipur districts), central (Maldah, Murshidabad districts) and the northern parts (Jalpaiguri, Kochbihar districts) and low-low clusters are found in the lower Gangetic plain (central-south) mainly and northern parts of West Bengal during the stipulated period. Apart from this seasonal variability inter-year variation is also visible. The results from different methods of this study indicate significant variation in the spatial distribution of malaria incidence in West Bengal and high incidence clusters are primarily persistently concentrated over the western part during 2012-2020 along with a strong seasonal pattern with a peak in rainy and autumn. By applying the different techniques in identifying the different degrees of incidence zones of malaria across West Bengal, some specific pockets or malaria hotspots are marked and identified where the incidence rates are quite harmonious over the different periods. From this analysis, it is clear that malaria is not a disease that is distributed uniformly across the state; some specific pockets are more prone to be affected in particular seasons of each year. Disease ecology and spatial patterns must be the factors in explaining the real factors for the higher incidence of this issue within those affected districts. The further study mainly by applying empirical approach is needed for discerning the strong relationship between communicable disease and other associated affecting factors.

Keywords: Malaria, infectious diseases, spatial statistics, spatial autocorrelation, LISA.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 450
45 Evaluation of Buckwheat Genotypes to Different Planting Geometries and Fertility Levels in Northern Transition Zone of Karnataka

Authors: U. K. Hulihalli, Shantveerayya

Abstract:

Buckwheat (Fagopyrum esculentum Moench) is an annual crop belongs to family Poligonaceae. The cultivated buckwheat species are notable for their exceptional nutritive values. It is an important source of carbohydrates, fibre, macro, and microelements such as K, Ca, Mg, Na and Mn, Zn, Se, and Cu. It also contains rutin, flavonoids, riboflavin, pyridoxine and many amino acids which have beneficial effects on human health, including lowering both blood lipid and sugar levels. Rutin, quercetin and some other polyphenols are potent carcinogens against colon and other cancers. Buckwheat has significant nutritive value and plenty of uses. Cultivation of buckwheat in Sothern part of India is very meager. Hence, a study was planned with an objective to know the performance of buckwheat genotypes to different planting geometries and fertility levels. The field experiment was conducted at Main Agriculture Research Station, University of Agriculture Sciences, Dharwad, India, during 2017 Kharif. The experiment was laid-out in split-plot design with three replications having three planting geometries as main plots, two genotypes as sub plots and three fertility levels as sub-sub plot treatments. The soil of the experimental site was vertisol. The standard procedures are followed to record the observations. The planting geometry of 30*10 cm was recorded significantly higher seed yield (893 kg/ha⁻¹), stover yield (1507 kg ha⁻¹), clusters plant⁻¹ (7.4), seeds clusters⁻¹ (7.9) and 1000 seed weight (26.1 g) as compared to 40*10 cm and 20*10 cm planting geometries. Between the genotypes, significantly higher seed yield (943 kg ha⁻¹) and harvest index (45.1) was observed with genotype IC-79147 as compared to PRB-1 genotype (687 kg ha⁻¹ and 34.2, respectively). However, the genotype PRB-1 recorded significantly higher stover yield (1344 kg ha⁻¹) as compared to genotype IC-79147 (1173 kg ha⁻¹). The genotype IC-79147 was recorded significantly higher clusters plant⁻¹ (7.1), seeds clusters⁻¹ (7.9) and 1000 seed weight (24.5 g) as compared PRB-1 (5.4, 5.8 and 22.3 g, respectively). Among the fertility levels tried, the fertility level of 60:30 NP kg ha⁻¹ recorded significantly higher seed yield (845 kg ha-1) and stover yield (1359 kg ha⁻¹) as compared to 40:20 NP kg ha-1 (808 and 1259 kg ha⁻¹ respectively) and 20:10 NP kg ha-1 (793 and 1144 kg ha⁻¹ respectively). Within the treatment combinations, IC 79147 genotype having 30*10 cm planting geometry with 60:30 NP kg ha⁻¹ recorded significantly higher seed yield (1070 kg ha⁻¹), clusters plant⁻¹ (10.3), seeds clusters⁻¹ (9.9) and 1000 seed weight (27.3 g) compared to other treatment combinations.

Keywords: Buckwheat, fertility levels, genotypes, geometry, polyphenols, rutin.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 793
44 Prospects of Iraq’s Maritime Openness and Their Effect on Its Economy

Authors: Mohanad Hammad

Abstract:

Port institutions serve as a link connecting the land areas that receive the goods and the areas from where ships sail. These areas hold great significance for the conversion of goods into commodities of economic value, capable of meeting the needs of the society. Development of ports constitutes a fundamental component of the comprehensive economic development process. Recognizing this fact, developing countries have always resorted to this infrastructural element to resolve the numerous problems they face, taking into account its contribution to the reformation of their economic conditions. Iraqi ports have played a major role in boosting the commercial movement in Iraq, as they are the starting point of its oil exports and a key constituent in fulfilling the consumer and production needs of the various economic sectors of Iraq. With the Gulf wars and the economic blockade, Iraqi ports have continued to deteriorate and become unable to perform their functions as first-generation ports, prompting Iraq to use the ports of neighboring countries such as Jordan's Aqaba commercial port. Meanwhile, Iraqi ports face strong competition from the ports of neighboring countries, which have achieved progress and advancement as opposed to the declining performance and efficiency of Iraqi ports. The great developments in the economic conditions of Iraq lay a too great burden on the Iraqi maritime transport and ports, which require development in order to be able to meet the challenges arising from the fierce international and regional competition in the markets. Therefore, it is necessary to find appropriate solutions in support of the role that can be played by Iraqi ports in serving Iraq's foreign trade transported by sea and in keeping up with the development of foreign trade. Thus, this research aims at tackling the current situation of the Iraqi ports and their commercial activity and studying the problems and obstacles they face. The research also studies the future prospects of these ports, the potentials of maritime openness to Iraq under the fierce competition of neighboring ports, and the possibility of enhancing Iraqi ports’ competitiveness. Among the results produced by this research is the future scenario it proposes for Iraqi ports, mainly represented in the establishment of Al-Faw Port, which will contribute to a greater openness of maritime transport in Iraq, and the rehabilitation and expansion of existing ports. This research seeks to develop solutions to Iraq ports so that they can be repositioned as a vital means of promoting economic development.

Keywords: Transport, port, regional openness, development.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 642
43 AI-Based Techniques for Online Social Media Network Sentiment Analysis: A Methodical Review

Authors: A. M. John-Otumu, M. M. Rahman, O. C. Nwokonkwo, M. C. Onuoha

Abstract:

Online social media networks have long served as a primary arena for group conversations, gossip, text-based information sharing and distribution. The use of natural language processing techniques for text classification and unbiased decision making has not been far-fetched. Proper classification of these textual information in a given context has also been very difficult. As a result, a systematic review was conducted from previous literature on sentiment classification and AI-based techniques. The study was done in order to gain a better understanding of the process of designing and developing a robust and more accurate sentiment classifier that could correctly classify social media textual information of a given context between hate speech and inverted compliments with a high level of accuracy using the knowledge gain from the evaluation of different artificial intelligence techniques reviewed. The study evaluated over 250 articles from digital sources like ACM digital library, Google Scholar, and IEEE Xplore; and whittled down the number of research to 52 articles. Findings revealed that deep learning approaches such as Convolutional Neural Network (CNN), Recurrent Neural Network (RNN), Bidirectional Encoder Representations from Transformer (BERT), and Long Short-Term Memory (LSTM) outperformed various machine learning techniques in terms of performance accuracy. A large dataset is also required to develop a robust sentiment classifier. Results also revealed that data can be obtained from places like Twitter, movie reviews, Kaggle, Stanford Sentiment Treebank (SST), and SemEval Task4 based on the required domain. The hybrid deep learning techniques like CNN+LSTM, CNN+ Gated Recurrent Unit (GRU), CNN+BERT outperformed single deep learning techniques and machine learning techniques. Python programming language outperformed Java programming language in terms of development simplicity and AI-based library functionalities. Finally, the study recommended the findings obtained for building robust sentiment classifier in the future.

Keywords: Artificial Intelligence, Natural Language Processing, Sentiment Analysis, Social Network, Text.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 514
42 Dynamic Simulation of IC Engine Bearings for Fault Detection and Wear Prediction

Authors: M. D. Haneef, R. B. Randall, Z. Peng

Abstract:

Journal bearings used in IC engines are prone to premature failures and are likely to fail earlier than the rated life due to highly impulsive and unstable operating conditions and frequent starts/stops. Vibration signature extraction and wear debris analysis techniques are prevalent in industry for condition monitoring of rotary machinery. However, both techniques involve a great deal of technical expertise, time, and cost. Limited literature is available on the application of these techniques for fault detection in reciprocating machinery, due to the complex nature of impact forces that confounds the extraction of fault signals for vibration-based analysis and wear prediction. In present study, a simulation model was developed to investigate the bearing wear behaviour, resulting because of different operating conditions, to complement the vibration analysis. In current simulation, the dynamics of the engine was established first, based on which the hydrodynamic journal bearing forces were evaluated by numerical solution of the Reynold’s equation. In addition, the essential outputs of interest in this study, critical to determine wear rates are the tangential velocity and oil film thickness between the journals and bearing sleeve, which if not maintained appropriately, have a detrimental effect on the bearing performance. Archard’s wear prediction model was used in the simulation to calculate the wear rate of bearings with specific location information as all determinative parameters were obtained with reference to crank rotation. Oil film thickness obtained from the model was used as a criterion to determine if the lubrication is sufficient to prevent contact between the journal and bearing thus causing accelerated wear. A limiting value of 1 μm was used as the minimum oil film thickness needed to prevent contact. The increased wear rate with growing severity of operating conditions is analogous and comparable to the rise in amplitude of the squared envelope of the referenced vibration signals. Thus on one hand, the developed model demonstrated its capability to explain wear behaviour and on the other hand it also helps to establish a co-relation between wear based and vibration based analysis. Therefore, the model provides a cost effective and quick approach to predict the impending wear in IC engine bearings under various operating conditions.

Keywords: Condition monitoring, IC engine, journal bearings, vibration analysis, wear prediction.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2248
41 Extremism among College and High School Students in Moscow: Diagnostics Features

Authors: Puzanova Zhanna Vasilyevna, Larina Tatiana Igorevna, Tertyshnikova Anastasia Gennadyevna

Abstract:

In this day and age, extremism in various forms of its manifestation is a real threat to the world community, the national security of a state and its territorial integrity, as well as to the constitutional rights and freedoms of citizens. Extremism, as it is known, in general terms described as a commitment to extreme views and actions, radically denying the existing social norms and rules. Supporters of extremism in the ideological and political struggles often adopt methods and means of psychological warfare, appeal not to reason and logical arguments, but to emotions and instincts of the people, to prejudices, biases, and a variety of mythological designs. They are dissatisfied with the established order and aim at increasing this dissatisfaction among the masses. Youth extremism holds a specific place among the existing forms and types of extremism. In this context in 2015, we conducted a survey among Moscow college and high school students. The aim of this study was to determine how great or small is the difference in understanding and attitudes towards extremism manifestations, inclination and readiness to take part in extremist activities and what causes this predisposition, if it exists. We performed multivariate analysis and found the Russian college and high school students' opinion about the extremism and terrorism situation in our country and also their cognition on these topics. Among other things, we showed, that the level of aggressiveness of young people were not above the average for the whole population. The survey was conducted using the questionnaire method. The sample included college and high school students in Moscow (642 and 382, respectively) by method of random selection. The questionnaire was developed by specialists of RUDN University Sociological Laboratory and included both original questions (projective questions, the technique of incomplete sentences), and the standard test Dayhoff S. to determine the level of internal aggressiveness. It is also used as an experiment, the technique of study option using of FACS and SPAFF to determine the psychotypes and determination of non-verbal manifestations of emotions. The study confirmed the hypothesis that in respondents’ opinion, the level of aggression is higher today than a few years ago. Differences were found in the understanding of and respect for such social phenomena as extremism, terrorism, and their danger and appeal for the two age groups of young people. Theory of psychotypes, SPAFF (specific affect cording system) and FACS (facial action cording system) are considered as additional techniques for the diagnosis of a tendency to extreme views. Thus, it is established that diagnostics of acceptance of extreme views among young people is possible thanks to simultaneous use of knowledge from the different fields of socio-humanistic sciences. The results of the research can be used in a comparative context with other countries and as a starting point for further research in the field, taking into account its extreme relevance.

Keywords: Extremism, youth extremism, diagnostics of extremist manifestations, forecast of behavior, Sociological polls, theory of psychotypes, FACS, SPAFF.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1787
40 Piezoelectric Bimorph Harvester Based on Different Lead Zirconate Titanate Materials to Enhance Energy Collection

Authors: Irene Perez-Alfaro, Nieves Murillo, Carlos Bernal, Daniel Gil-Hernandez

Abstract:

Nowadays, the increasing applicability of internet of things (IoT) systems has changed the way that the world around is perceived. The massive interconnection of systems by means of sensing, processing and communication, allows multitude of data to be at our fingertips. In this way, countless advances have been made in different fields such as personal care, predictive maintenance in industry, quality control in production processes, security, and in everything imaginable. However, all these electronic systems have in common the need to be electrically powered. In this context, batteries and wires are the most commonly used solutions, but they are not a definitive solution in some applications, because of the attainability, the serviceability, or the performance requirements. Therefore, the need arises to look for other types of solutions based on energy harvesting and long-life electronics. Energy Harvesting can be defined as the action of capturing energy from the environment and store it for an instantaneous use or later use. Among the materials capable of harvesting energy from the environment, such as thermoelectrics, electromagnetics, photovoltaics or triboelectrics, the most suitable is the piezoelectric material. The phenomenon of piezoelectricity is one of the most powerful sources for energy harvesting, ranging from a few micro wats to hundreds of wats, depending on certain factors such as material type, geometry, excitation frequency, mechanical and electrical configurations, among others. In this research work, an exhaustive study is carried out on how different types of piezoelectric materials and electrical configurations influence the maximum power that a bimorph harvester is able to extract from mechanical vibrations. A series of experiments has been carried out in which the manufactured bimorph specimens are excited under fixed inertial vibrational conditions. In addition, in order to evaluate the dependence of the maximum transferred power, different load resistors are tested. In this way, the pure active power that achieves the maximum power transfer can be approximated. In this paper, we present the design of low-cost energy harvesting solutions based on piezoelectric smart materials with tunable frequency. The results obtained show the differences in energy extraction between the PZT materials studied and their electrical configurations. The aim of this work is to gain a better understanding of the behavior of piezoelectric materials, and the design process of bimorph PZT harvesters to optimize environmental energy extraction.

Keywords: Bimorph harvesters, electrical impedance, energy harvesting, piezoelectric, smart material.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 400
39 Wetting Characterization of High Aspect Ratio Nanostructures by Gigahertz Acoustic Reflectometry

Authors: C. Virgilio, J. Carlier, P. Campistron, M. Toubal, P. Garnier, L. Broussous, V. Thomy, B. Nongaillard

Abstract:

Wetting efficiency of microstructures or nanostructures patterned on Si wafers is a real challenge in integrated circuits manufacturing. In fact, bad or non-uniform wetting during wet processes limits chemical reactions and can lead to non-complete etching or cleaning inside the patterns and device defectivity. This issue is more and more important with the transistors size shrinkage and concerns mainly high aspect ratio structures. Deep Trench Isolation (DTI) structures enabling pixels’ isolation in imaging devices are subject to this phenomenon. While low-frequency acoustic reflectometry principle is a well-known method for Non Destructive Test applications, we have recently shown that it is also well suited for nanostructures wetting characterization in a higher frequency range. In this paper, we present a high-frequency acoustic reflectometry characterization of DTI wetting through a confrontation of both experimental and modeling results. The acoustic method proposed is based on the evaluation of the reflection of a longitudinal acoustic wave generated by a 100 µm diameter ZnO piezoelectric transducer sputtered on the silicon wafer backside using MEMS technologies. The transducers have been fabricated to work at 5 GHz corresponding to a wavelength of 1.7 µm in silicon. The DTI studied structures, manufactured on the wafer frontside, are crossing trenches of 200 nm wide and 4 µm deep (aspect ratio of 20) etched into a Si wafer frontside. In that case, the acoustic signal reflection occurs at the bottom and at the top of the DTI enabling its characterization by monitoring the electrical reflection coefficient of the transducer. A Finite Difference Time Domain (FDTD) model has been developed to predict the behavior of the emitted wave. The model shows that the separation of the reflected echoes (top and bottom of the DTI) from different acoustic modes is possible at 5 Ghz. A good correspondence between experimental and theoretical signals is observed. The model enables the identification of the different acoustic modes. The evaluation of DTI wetting is then performed by focusing on the first reflected echo obtained through the reflection at Si bottom interface, where wetting efficiency is crucial. The reflection coefficient is measured with different water / ethanol mixtures (tunable surface tension) deposited on the wafer frontside. Two cases are studied: with and without PFTS hydrophobic treatment. In the untreated surface case, acoustic reflection coefficient values with water show that liquid imbibition is partial. In the treated surface case, the acoustic reflection is total with water (no liquid in DTI). The impalement of the liquid occurs for a specific surface tension but it is still partial for pure ethanol. DTI bottom shape and local pattern collapse of the trenches can explain these incomplete wetting phenomena. This high-frequency acoustic method sensitivity coupled with a FDTD propagative model thus enables the local determination of the wetting state of a liquid on real structures. Partial wetting states for non-hydrophobic surfaces or low surface tension liquids are then detectable with this method.

Keywords: Wetting, acoustic reflectometry, gigahertz, semiconductor.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1269
38 Capital Accumulation and Unemployment in Namibia, Nigeria, and South Africa

Authors: Abubakar Dikko

Abstract:

The research investigates the causes of unemployment in Namibia, Nigeria and South Africa and the role of Capital Accumulation in reducing the unemployment profile of these economies as proposed by the post-Keynesian economics. This is conducted through extensive review of literature on the NAIRU models and focused on the post-Keynesian view of unemployment within the NAIRU framework. The NAIRU (non-accelerating inflation rate of unemployment) model has become a dominant framework used in macroeconomic analysis of unemployment. The study views the post-Keynesian economics arguments that capital accumulation is a major determinant of unemployment. Unemployment remains the fundamental socio-economic challenge facing African economies. It has been a burden to citizens of those economies. Namibia, Nigeria, and South Africa are great African nations battling with high unemployment rates. The high unemployment rate in the country led the citizens to chase away foreigners in the country claiming that they have taken away their jobs. The study proposes there is a strong relationship between capital accumulation and unemployment in Namibia, Nigeria, and South Africa, and capital accumulation is responsible for high unemployment rates in these countries. For the economies to achieve steady state level of employment and satisfactory level of economic growth and development, there is need for capital accumulation to take place. The countries in the study have been selected after a critical research and investigations. They are selected based on the following criteria; African economies with high unemployment rates above 15% and have about 40% of their workforce unemployed. This level of unemployment is the critical level of unemployment in Africa as expressed by International Labour Organization (ILO). And finally, the African countries experience a slow growth in their Gross fixed capital formation. Adequate statistical measures have been employed using a time-series analysis in the study and the results revealed that capital accumulation is the main driver of unemployment performance in the chosen African countries. An increase in the accumulation of capital causes unemployment to reduce significantly. The results of the research work will be useful and relevant to federal governments and ministries, departments and agencies (MDAs) of Namibia, Nigeria and South Africa to resolve the issue of high and persistent unemployment rates in their economies which are great burden that slows growth and development of developing economies. Also, the result can be useful to World Bank, African Development Bank and International Labour Organization (ILO) in their further research and studies on how to tackle unemployment in developing and emerging economies.

Keywords: Capital accumulation, NAIRU, post-Keynesian economics, unemployment.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3223
37 Clean Sky 2 – Project PALACE: Aeration’s Experimental Sound Velocity Investigations for High-Speed Gerotor Simulations

Authors: Benoît Mary, Thibaut Gras, Gaëtan Fagot, Yvon Goth, Ilyes Mnassri-Cetim

Abstract:

A Gerotor pump is composed of an external and internal gear with conjugate cycloidal profiles. From suction to delivery ports, the fluid is transported inside cavities formed by teeth and driven by the shaft. From a geometric and conceptional side it is worth to note that the internal gear has one tooth less than the external one. Simcenter Amesim v.16 includes a new submodel for modelling the hydraulic Gerotor pumps behavior (THCDGP0). This submodel considers leakages between teeth tips using Poiseuille and Couette flows contributions. From the 3D CAD model of the studied pump, the “CAD import” tool takes out the main geometrical characteristics and the submodel THCDGP0 computes the evolution of each cavity volume and their relative position according to the suction or delivery areas. This module, based on international publications, presents robust results up to 6 000 rpm for pressure greater than atmospheric level. For higher rotational speeds or lower pressures, oil aeration and cavitation effects are significant and highly drop the pump’s performance. The liquid used in hydraulic systems always contains some gas, which is dissolved in the liquid at high pressure and tends to be released in a free form (i.e. undissolved as bubbles) when pressure drops. In addition to gas release and dissolution, the liquid itself may vaporize due to cavitation. To model the relative density of the equivalent fluid, modified Henry’s law is applied in Simcenter Amesim v.16 to predict the fraction of undissolved gas or vapor. Three parietal pressure sensors have been set up upstream from the pump to estimate the sound speed in the oil. Analytical models have been compared with the experimental sound speed to estimate the occluded gas content. Simcenter Amesim v.16 model was supplied by these previous analyses marks which have successfully improved the simulations results up to 14 000 rpm. This work provides a sound foundation for designing the next Gerotor pump generation reaching high rotation range more than 25 000 rpm. This improved module results will be compared to tests on this new pump demonstrator.

Keywords: Gerotor pump, high speed, simulations, aeronautic, aeration, cavitation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 524
36 Evaluation of Vitamin D Levels in Obese and Morbid Obese Children

Authors: Orkide Donma, Mustafa M. Donma

Abstract:

Obesity may lead to growing serious health problems throughout the world. Vitamin D appears to play a role in cardiovascular and metabolic health. Vitamin D deficiency may add to derangements in human metabolic systems, particularly those of children. Childhood obesity is associated with an increased risk of chronic and sophisticated diseases. The aim of this study is to investigate associations as well as possible differences related to parameters affected by obesity and their relations with vitamin D status in obese (OB) and morbid obese (MO) children. This study included a total of 78 children. Of them, 41 and 37 were OB and MO, respectively. WHO BMI-for age percentiles were used for the classification of obesity. The values above 99 percentile were defined as MO. Those between 95 and 99 percentiles were included into OB group. Anthropometric measurements were recorded. Basal metabolic rates (BMRs) were measured. Vitamin D status is determined by the measurement of 25-hydroxy cholecalciferol [25- hydroxyvitamin D3, 25(OH)D] using high-performance liquid chromatography. Vitamin D status was evaluated as deficient, insufficient and sufficient. Values < 20.0 ng/ml, values between 20-30 ng/ml and values > 30.0 ng/ml were defined as vitamin D deficient, insufficient and sufficient, respectively. Optimal 25(OH)D level was defined as ≥ 30 ng/ml. SPSSx statistical package program was used for the evaluation of the data. The statistical significance degree was accepted as p < 0.05. Mean ages did not differ between the groups. Significantly increased body mass index (BMI), waist circumference (C) and neck C as well as significantly decreased fasting blood glucose (FBG) and vitamin D values were observed in MO group (p < 0.05). In OB group, 37.5% of the children were vitamin D deficient, and in MO group the corresponding value was 53.6%. No difference between the groups in terms of lipid profile, systolic blood pressure (SBP), diastolic blood pressure (DBP) and insulin values was noted. There was a severe statistical significance between FBG values of the groups (p < 0.001). Important correlations between BMI, waist C, hip C, neck C and both SBP as well as DBP were found in OB group. In MO group, correlations only with SBP were obtained. In a similar manner, in OB group, correlations were detected between SBP-BMR and DBP-BMR. However, in MO children, BMR correlated only with SBP. The associations of vitamin D with anthropometric indices as well as some lipid parameters were defined. In OB group BMI, waist C, hip C and triglycerides (TRG) were negatively correlated with vitamin D concentrations whereas none of them were detected in MO group. Vitamin D deficiency may contribute to the complications associated with childhood obesity. Loss of correlations between obesity indices-DBP, vitamin D-TRG, as well as relatively lower FBG values, observed in MO group point out that the emergence of MetS components starts during obesity state just before the transition to morbid obesity. Aside from its deficiency state, associations of vitamin D with anthropometric measurements, blood pressures and TRG should also be evaluated before the development of morbid obesity.

Keywords: Children, morbid obesity, obesity, vitamin D.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 941
35 Surface Topography Assessment Techniques based on an In-process Monitoring Approach of Tool Wear and Cutting Force Signature

Authors: A. M. Alaskari, S. E. Oraby

Abstract:

The quality of a machined surface is becoming more and more important to justify the increasing demands of sophisticated component performance, longevity, and reliability. Usually, any machining operation leaves its own characteristic evidence on the machined surface in the form of finely spaced micro irregularities (surface roughness) left by the associated indeterministic characteristics of the different elements of the system: tool-machineworkpart- cutting parameters. However, one of the most influential sources in machining affecting surface roughness is the instantaneous state of tool edge. The main objective of the current work is to relate the in-process immeasurable cutting edge deformation and surface roughness to a more reliable easy-to-measure force signals using a robust non-linear time-dependent modeling regression techniques. Time-dependent modeling is beneficial when modern machining systems, such as adaptive control techniques are considered, where the state of the machined surface and the health of the cutting edge are monitored, assessed and controlled online using realtime information provided by the variability encountered in the measured force signals. Correlation between wear propagation and roughness variation is developed throughout the different edge lifetimes. The surface roughness is further evaluated in the light of the variation in both the static and the dynamic force signals. Consistent correlation is found between surface roughness variation and tool wear progress within its initial and constant regions. At the first few seconds of cutting, expected and well known trend of the effect of the cutting parameters is observed. Surface roughness is positively influenced by the level of the feed rate and negatively by the cutting speed. As cutting continues, roughness is affected, to different extents, by the rather localized wear modes either on the tool nose or on its flank areas. Moreover, it seems that roughness varies as wear attitude transfers from one mode to another and, in general, it is shown that it is improved as wear increases but with possible corresponding workpart dimensional inaccuracy. The dynamic force signals are found reasonably sensitive to simulate either the progressive or the random modes of tool edge deformation. While the frictional force components, feeding and radial, are found informative regarding progressive wear modes, the vertical (power) components is found more representative carrier to system instability resulting from the edge-s random deformation.

Keywords: Dynamic force signals, surface roughness (finish), tool wear and deformation, tool wear modes (nose, flank)

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1324
34 A Self Supervised Bi-directional Neural Network (BDSONN) Architecture for Object Extraction Guided by Beta Activation Function and Adaptive Fuzzy Context Sensitive Thresholding

Authors: Siddhartha Bhattacharyya, Paramartha Dutta, Ujjwal Maulik, Prashanta Kumar Nandi

Abstract:

A multilayer self organizing neural neural network (MLSONN) architecture for binary object extraction, guided by a beta activation function and characterized by backpropagation of errors estimated from the linear indices of fuzziness of the network output states, is discussed. Since the MLSONN architecture is designed to operate in a single point fixed/uniform thresholding scenario, it does not take into cognizance the heterogeneity of image information in the extraction process. The performance of the MLSONN architecture with representative values of the threshold parameters of the beta activation function employed is also studied. A three layer bidirectional self organizing neural network (BDSONN) architecture comprising fully connected neurons, for the extraction of objects from a noisy background and capable of incorporating the underlying image context heterogeneity through variable and adaptive thresholding, is proposed in this article. The input layer of the network architecture represents the fuzzy membership information of the image scene to be extracted. The second layer (the intermediate layer) and the final layer (the output layer) of the network architecture deal with the self supervised object extraction task by bi-directional propagation of the network states. Each layer except the output layer is connected to the next layer following a neighborhood based topology. The output layer neurons are in turn, connected to the intermediate layer following similar topology, thus forming a counter-propagating architecture with the intermediate layer. The novelty of the proposed architecture is that the assignment/updating of the inter-layer connection weights are done using the relative fuzzy membership values at the constituent neurons in the different network layers. Another interesting feature of the network lies in the fact that the processing capabilities of the intermediate and the output layer neurons are guided by a beta activation function, which uses image context sensitive adaptive thresholding arising out of the fuzzy cardinality estimates of the different network neighborhood fuzzy subsets, rather than resorting to fixed and single point thresholding. An application of the proposed architecture for object extraction is demonstrated using a synthetic and a real life image. The extraction efficiency of the proposed network architecture is evaluated by a proposed system transfer index characteristic of the network.

Keywords: Beta activation function, fuzzy cardinality, multilayer self organizing neural network, object extraction,

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1531