Search results for: high technology export
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 25860

Search results for: high technology export

3750 Identification of Blood Biomarkers Unveiling Early Alzheimer's Disease Diagnosis Through Single-Cell RNA Sequencing Data and Autoencoders

Authors: Hediyeh Talebi, Shokoofeh Ghiam, Changiz Eslahchi

Abstract:

Traditionally, Alzheimer’s disease research has focused on genes with significant fold changes, potentially neglecting subtle but biologically important alterations. Our study introduces an integrative approach that highlights genes crucial to underlying biological processes, regardless of their fold change magnitude. Alzheimer's Single-cell RNA-seq data related to the peripheral blood mononuclear cells (PBMC) was extracted from the Gene Expression Omnibus (GEO). After quality control, normalization, scaling, batch effect correction, and clustering, differentially expressed genes (DEGs) were identified with adjusted p-values less than 0.05. These DEGs were categorized based on cell-type, resulting in four datasets, each corresponding to a distinct cell type. To distinguish between cells from healthy individuals and those with Alzheimer's, an adversarial autoencoder with a classifier was employed. This allowed for the separation of healthy and diseased samples. To identify the most influential genes in this classification, the weight matrices in the network, which includes the encoder and classifier components, were multiplied, and focused on the top 20 genes. The analysis revealed that while some of these genes exhibit a high fold change, others do not. These genes, which may be overlooked by previous methods due to their low fold change, were shown to be significant in our study. The findings highlight the critical role of genes with subtle alterations in diagnosing Alzheimer's disease, a facet frequently overlooked by conventional methods. These genes demonstrate remarkable discriminatory power, underscoring the need to integrate biological relevance with statistical measures in gene prioritization. This integrative approach enhances our understanding of the molecular mechanisms in Alzheimer’s disease and provides a promising direction for identifying potential therapeutic targets.

Keywords: alzheimer's disease, single-cell RNA-seq, neural networks, blood biomarkers

Procedia PDF Downloads 66
3749 Adsorption of Chlorinated Pesticides in Drinking Water by Carbon Nanotubes

Authors: Hacer Sule Gonul, Vedat Uyak

Abstract:

Intensive use of pesticides in agricultural activity causes mixing of these compounds into water sources with surface flow. Especially after the 1970s, a number of limitations imposed on the use of chlorinated pesticides that have a carcinogenic risk potential and regulatory limit have been established. These chlorinated pesticides discharge to water resources, transport in the water and land environment and accumulation in the human body through the food chain raises serious health concerns. Carbon nanotubes (CNTs) have attracted considerable attention from on all because of their excellent mechanical, electrical, and environmental characteristics. Due to CNT particles' high degree of hydrophobic surfaces, these nanoparticles play critical role in the removal of water contaminants of natural organic matters, pesticides and phenolic compounds in water sources. Health concerns associated with chlorinated pesticides requires the removal of such contaminants from aquatic environment. Although the use of aldrin and atrazine was restricted in our country, repatriation of illegal entry and widespread use of such chemicals in agricultural areas cause increases for the concentration of these chemicals in the water supply. In this study, the compounds of chlorinated pesticides such as aldrin and atrazine compounds would be tried to eliminate from drinking water with carbon nanotube adsorption method. Within this study, 2 different types of CNT would be used including single-wall (SWCNT) and multi-wall (MWCNT) carbon nanotubes. Adsorption isotherms within the scope of work, the parameters affecting the adsorption of chlorinated pesticides in water are considered as pH, contact time, CNT type, CNT dose and initial concentration of pesticides. As a result, under conditions of neutral pH conditions with MWCNT respectively for atrazine and aldrin obtained adsorption capacity of determined as 2.24 µg/mg ve 3.84 µg/mg. On the other hand, the determined adsorption capacity rates for SWCNT for aldrin and atrazine has identified as 3.91 µg/mg ve 3.92 µg/mg. After all, each type of pesticide that provides superior performance in relieving SWCNT particles has emerged.

Keywords: pesticide, drinking water, carbon nanotube, adsorption

Procedia PDF Downloads 170
3748 A Nutritional Wellness Program for Overweight Health Care Providers in Hospital Setting: A Randomized Controlled Trial Pilot Study

Authors: Kim H. K. Choy, Oliva H. K. Chu, W. Y. Keung, B. Lim, Winnie P. Y. Tang

Abstract:

Background: The prevalence of workplace obesity is rising worldwide; therefore, the workplace is an ideal venue to implement weight control intervention. This pilot randomized controlled trial aimed to develop, implement, and evaluate a nutritional wellness program for obese health care providers working in a hospital. Methods: This hospital-based nutritional wellness program was an 8-week pilot randomized controlled trial for obese health care providers. The primary outcomes were body weight and body mass index (BMI). The secondary outcomes were serum fasting glucose, fasting cholesterol, triglyceride, high-density (HDL) and low-density (LDL) lipoprotein, body fat percentage, and body mass. Participants were randomly assigned to the intervention (n = 20) or control (n = 22) group. Participants in both groups received individual nutrition counselling and nutrition pamphlets, whereas only participants in the intervention group were given mobile phone text messages. Results: 42 participants completed the study. In comparison with the control group, the intervention group showed approximately 0.98 kg weight reduction after two months. Participants in intervention group also demonstrated clinically significant improvement in BMI, serum cholesterol level, and HDL level. There was no improvement of body fat percentage and body mass for both intervention and control groups. Conclusion: The nutritional wellness program for obese health care providers was feasible in hospital settings. Health care providers demonstrated short-term weight loss, decrease in serum fasting cholesterol level, and HDL level after completing the program.

Keywords: weight management, weight control, health care providers, hospital

Procedia PDF Downloads 243
3747 New Test Algorithm to Detect Acute and Chronic HIV Infection Using a 4th Generation Combo Test

Authors: Barun K. De

Abstract:

Acquired immunodeficiency syndrome (AIDS) is caused by two types of human immunodeficiency viruses, collectively designated HIV. HIV infection is spreading globally particularly in developing countries. Before an individual is diagnosed with HIV, the disease goes through different phases. First there is an acute early phase that is followed by an established or chronic phase. Subsequently, there is a latency period after which the individual becomes immunodeficient. It is in the acute phase that an individual is highly infectious due to a high viral load. Presently, HIV diagnosis involves use of tests that do not detect the acute phase infection during which both the viral RNA and p24 antigen are expressed. Instead, these less sensitive tests detect antibodies to viral antigens which are typically sero-converted later in the disease process following acute infection. These antibodies are detected in both asymptomatic HIV-infected individuals as well as AIDS patients. Studies indicate that early diagnosis and treatment of HIV infection can reduce medical costs, improve survival, and reduce spreading of infection to new uninfected partners. Newer 4th generation combination antigen/antibody tests are highly sensitive and specific for detection of acute and established HIV infection (HIV1 and HIV2) enabling immediate linkage to care. The CDC (Center of Disease Control, USA) recently recommended an algorithm involving three different tests to screen and diagnose acute and established infections of HIV-1 and HIV-2 in a general population. Initially a 4th generation combo test detects a viral antigen p24 and specific antibodies against HIV -1 and HIV-2 envelope proteins. If the test is positive it is followed by a second test known as a differentiation assay which detects antibodies against specific HIV-1 and HIV-2 envelope proteins confirming established infection of HIV-1 or HIV-2. However if it is negative then another test is performed that measures viral load confirming an acute HIV-1 infection. Screening results of a Phoenix area population detected 0.3% new HIV infections among which 32.4% were acute cases. Studies in the U.S. indicate that this algorithm effectively reduces HIV infection through immediate treatment and education following diagnosis.

Keywords: new algorithm, HIV, diagnosis, infection

Procedia PDF Downloads 410
3746 Predicting Radioactive Waste Glass Viscosity, Density and Dissolution with Machine Learning

Authors: Joseph Lillington, Tom Gout, Mike Harrison, Ian Farnan

Abstract:

The vitrification of high-level nuclear waste within borosilicate glass and its incorporation within a multi-barrier repository deep underground is widely accepted as the preferred disposal method. However, for this to happen, any safety case will require validation that the initially localized radionuclides will not be considerably released into the near/far-field. Therefore, accurate mechanistic models are necessary to predict glass dissolution, and these should be robust to a variety of incorporated waste species and leaching test conditions, particularly given substantial variations across international waste-streams. Here, machine learning is used to predict glass material properties (viscosity, density) and glass leaching model parameters from large-scale industrial data. A variety of different machine learning algorithms have been compared to assess performance. Density was predicted solely from composition, whereas viscosity additionally considered temperature. To predict suitable glass leaching model parameters, a large simulated dataset was created by coupling MATLAB and the chemical reactive-transport code HYTEC, considering the state-of-the-art GRAAL model (glass reactivity in allowance of the alteration layer). The trained models were then subsequently applied to the large-scale industrial, experimental data to identify potentially appropriate model parameters. Results indicate that ensemble methods can accurately predict viscosity as a function of temperature and composition across all three industrial datasets. Glass density prediction shows reliable learning performance with predictions primarily being within the experimental uncertainty of the test data. Furthermore, machine learning can predict glass dissolution model parameters behavior, demonstrating potential value in GRAAL model development and in assessing suitable model parameters for large-scale industrial glass dissolution data.

Keywords: machine learning, predictive modelling, pattern recognition, radioactive waste glass

Procedia PDF Downloads 115
3745 Bronchoscopy and Genexpert in the Diagnosis of Pulmonary Tuberculosis in the Indian Private Health Sector: A Short Case Series

Authors: J. J. Mathew

Abstract:

Pulmonary tuberculosis is highly prevalent in the Indian subcontinent. Most cases of pulmonary tuberculosis are diagnosed with sputum examinations and the vast majority of these are undertaken by the government run establishments. However, mycobacterial cultures are not routinely done, unless drug resistance is detected based on clinical response. Modern diagnostic tests like bronchoscopy and Genexpert are not routinely employed in the government institutions for the diagnosis of pulmonary tuberculosis, but have been accepted widely by good private institutions. The utility of these investigations in the private sector is not yet well recognized. This retrospective study aims to assess the usefulness of bronchoscopy and Genexpert in the diagnosis of pulmonary tuberculosis in quaternary care private hospital in India. 30 patients with respiratory symptoms raising the possibility of tuberculosis based on clinical and radiological features, but without any significant sputum production, were subject to bronchoscopy and BAL samples taken for microbiological studies, including Genexpert. 6 out of the 30 patients were found to be Genexpert positive and none of them showed Rifampicin resistance. All the 6 cases had upper zone predominant disease. One of the 6 cases of tuberculosis had another co-existent bacterial infection according to the routine culture studies. 6 other cases were proven to be due to other bacterial infections alone, 2 had a malignant diagnosis and the remaining cases were thought to be non-infective pathologies. The Genexpert results were made available within 48 hours in the 6 positive cases. All of them were commenced on standard anti-tuberculous regimen with excellent clinical response. The other infective cases were also managed successfully based on the drug susceptibilities. The study has shown the usefulness of these investigations as early intervention enabled diagnosis facilitating treatment and prevention of any clinical deterioration. The study lends support to early bronchoscopy and Genexpert testing in suspected cases of pulmonary tuberculosis without significant sputum production, in a high prevalence country which normally relies on sputum examination for the diagnosis of pulmonary tuberculosis.

Keywords: pulmonary, tuberculosis, bronchoscopy, genexpert

Procedia PDF Downloads 245
3744 Response of Yield and Morphological Characteristic of Rice Cultivars to Heat Stress at Different Growth Stages

Authors: Mohammad Taghi Karbalaei Aghamolki, Mohd Khanif Yusop, Fateh Chand Oad, Hamed Zakikhani, Hawa Zee Jaafar, Sharifh Kharidah, Mohamed Hanafi Musa, Shahram Soltani

Abstract:

The high temperatures during sensitive growth phases are changing rice morphology as well as influencing yield. In the glass house study, the treatments were: growing conditions [normal growing (32oC+2) and heat stress (38oC+2) day time and 22oC+2 night time], growth stages (booting, flowering and ripening) and four cultivars (Hovaze, Hashemi, Fajr, as exotic and MR219 as indigenous). The heat chamber was prepared covered with plastic, and automatic heater was adjusted at 38oC+2 (day) and 22oC+2 (night) for two weeks in every growth stages. Rice morphological and yield under the influence of heat stress during various growth stages showed taller plants in Hashsemi due to its tall character. The total tillers per hill were significantly higher in Fajr receiving heat stress during booting stage. In all growing conditions and growth stages, Hashemi recorded higher panicle exertion and flag leaf length. The flag leaf width in all situations was found higher in Hovaze. The total tillers per hill were more in Fajr, although heat stress was imposed during booting and flowering stages. The indigenous MR219 in all situations of growing conditions, growth stages recorded higher grain yield. However, its grain yield slightly decreased when heat stress was imposed during booting and flowering. Similar results were found in all other exotic cultivars recording to lower grain yield in the heat stress condition during booting and flowering. However, plants had no effect on heat stress during ripening stage.

Keywords: rice, growth, heat, temperature, stress, morphology, yield

Procedia PDF Downloads 276
3743 The Use of Additives to Prevent Fouling in Polyethylene and Polypropylene Gas and Slurry Phase Processes

Authors: L. Shafiq, A. Rigby

Abstract:

All polyethylene processes are highly exothermic, and the safe removal of the heat of reaction is a fundamental issue in the process design. In slurry and gas processes, the velocity of the polymer particles in the reactor and external coolers can be very high, and under certain conditions, this can lead to static charging of these particles. Such static charged polymer particles may start building up on the reactor wall, limiting heat transfer, and ultimately leading to severe reactor fouling and forced reactor shut down. Statsafe™ is an FDA approved anti-fouling additive currently used around the world for polyolefin production as an anti-fouling additive. The unique polymer chemistry aids static discharge, which prevents the build-up of charged polyolefin particles, which could lead to fouling. Statsafe™ is being used and trailed in gas, slurry, and a combination of these technologies around the world. We will share data to demonstrate how the use of Statsafe™ allows more stable operation at higher solids level by eliminating static, which would otherwise prevent closer packing of particles in the hydrocarbon slurry. Because static charge generation depends also on the concentration of polymer particles in the slurry, the maximum slurry concentration can be higher when using Statsafe™, leading to higher production rates. The elimination of fouling also leads to less downtime. Special focus will be made on the impact anti-static additives have on catalyst performance within the polymerization process and how this has been measured. Lab-scale studies have investigated the effect on the activity of Ziegler Natta catalysts when anti-static additives are used at various concentrations in gas and slurry, polyethylene and polypropylene processes. An in-depth gas phase study investigated the effect of additives on the final polyethylene properties such as particle size, morphology, fines, bulk density, melt flow index, gradient density, and melting point.

Keywords: anti-static additives, catalyst performance, FDA approved anti-fouling additive, polymerisation

Procedia PDF Downloads 203
3742 Design of an Active Compression System for Treating Vascular Disease Using a Series of Silicone Based Inflatable Mini Bladders

Authors: Gayani K. Nandasiri, Tilak Dias, William Hurley

Abstract:

Venous disease of human lower limb could range from minor asymptomatic incompetence of venous valves to chronic venous ulceration. The sheer prevalence of varicose veins and its associated significant costs of treating late complications such as chronic ulcers contribute to a higher burden on health care resources. In most of western countries with developed health care systems, treatment costs associated with Venous disease accounts for a considerable portion of their total health care budget, and it has become a high-cost burden to National Health Service (NHS), UK. The established gold standard of treatment for the venous disease is the graduated compression, where the pressure at the ankle being highest and decreasing towards the knee and thigh. Currently, medical practitioners use two main methods to treat venous disease; i.e. compression bandaging and compression stockings. Both these systems have their own disadvantages which lead to the current programme of research. The aim of the present study is to revolutionize the compression therapy by using a novel active compression system to deliver a controllable and more accurate pressure profiles using a series of inflatable mini bladders. Two types of commercially available silicones were tested for the application. The mini bladders were designed with a special fabrication procedure to provide required pressure profiles, and a series of experiments were conducted to characterise the mini bladders. The inflation/deflation heights of these mini bladders were investigated experimentally and using a finite element model (FEM), and the experimental data were compared to the results obtained from FEM simulations, which showed 70-80% agreement. Finally, the mini bladders were tested for its pressure transmittance characteristics, and the results showed a 70-80% of inlet air pressure transmitted onto the treated surface.

Keywords: finite element analysis, graduated compression, inflatable bladders, venous disease

Procedia PDF Downloads 185
3741 Spatial Differentiation of Elderly Care Facilities in Mountainous Cities: A Case Study of Chongqing

Authors: Xuan Zhao, Wen Jiang

Abstract:

In this study, a web crawler was used to collect POI sample data from 38 districts and counties of Chongqing in 2022, and ArcGIS was combined to coordinate and projection conversion and realize data visualization. Nuclear density analysis and spatial correlation analysis were used to explore the spatial distribution characteristics of elderly care facilities in Chongqing, and K mean cluster analysis was carried out with GeoDa to study the spatial concentration degree of elderly care resources in 38 districts and counties. Finally, the driving force of spatial differentiation of elderly care facilities in various districts and counties of Chongqing is studied by using the method of geographic detector. The results show that: (1) in terms of spatial distribution structure, the distribution of elderly care facilities in Chongqing is unbalanced, showing a distribution pattern of ‘large dispersion and small agglomeration’ and the asymmetric pattern of ‘west dense and east sparse, north dense and south sparse’ is prominent. (2) In terms of the spatial matching between elderly care resources and the elderly population, there is a weak coordination between the input of elderly care resources and the distribution of the elderly population at the county level in Chongqing. (3) The analysis of the results of the geographical detector shows that the single factor influence is mainly the number of elderly population, public financial revenue and district and county GDP. The high single factor influence is mainly caused by the elderly population, public financial income, and district and county GDP. The influence of each influence factor on the spatial distribution of elderly care facilities is not simply superimposed but has a nonlinear enhancement effect or double factor enhancement. It is necessary to strengthen the synergistic effect of two factors and promote the synergistic effect of multiple factors.

Keywords: aging, elderly care facilities, spatial differentiation, geographical detector, driving force analysis, Mountain city

Procedia PDF Downloads 38
3740 Authentication and Traceability of Meat Products from South Indian Market by Species-Specific Polymerase Chain Reaction

Authors: J. U. Santhosh Kumar, V. Krishna, Sebin Sebastian, G. S. Seethapathy, G. Ravikanth, R. Uma Shaanker

Abstract:

Food is one of the basic needs of human beings. It requires the normal function of the body part and a healthy growth. Recently, food adulteration increases day by day to increase the quantity and make more benefit. Animal source foods can provide a variety of micronutrients that are difficult to obtain in adequate quantities from plant source foods alone. Particularly in the meat industry, products from animals are susceptible targets for fraudulent labeling due to the economic profit that results from selling cheaper meat as meat from more profitable and desirable species. This work presents an overview of the main PCR-based techniques applied to date to verify the authenticity of beef meat and meat products from beef species. We were analyzed 25 market beef samples in South India. We examined PCR methods based on the sequence of the cytochrome b gene for source species identification. We found all sample were sold as beef meat as Bos Taurus. However, interestingly Male meats are more valuable high price compare to female meat, due to this reason most of the markets samples are susceptible. We were used sex determination gene of cattle like TSPY(Y-encoded, testis-specific protein TSPY is a Y-specific gene). TSPY homologs exist in several mammalian species, including humans, horses, and cattle. This gene is Y coded testis protein genes, which only amplify the male. We used multiple PCR products form species-specific “fingerprints” on gel electrophoresis, which may be useful for meat authentication. Amplicons were obtained only by the Cattle -specific PCR. We found 13 market meat samples sold as female beef samples. These results suggest that the species-specific PCR methods established in this study would be useful for simple and easy detection of adulteration of meat products.

Keywords: authentication, meat products, species-specific, TSPY

Procedia PDF Downloads 375
3739 Lithuanian Sign Language Literature: Metaphors at the Phonological Level

Authors: Anželika Teresė

Abstract:

In order to solve issues in sign language linguistics, address matters pertaining to maintaining high quality of sign language (SL) translation, contribute to dispelling misconceptions about SL and deaf people, and raise awareness and understanding of the deaf community heritage, this presentation discusses literature in Lithuanian Sign Language (LSL) and inherent metaphors that are created by using the phonological parameter –handshape, location, movement, palm orientation and nonmanual features. The study covered in this presentation is twofold, involving both the micro-level analysis of metaphors in terms of phonological parameters as a sub-lexical feature and the macro-level analysis of the poetic context. Cognitive theories underlie research of metaphors in sign language literature in a range of SL. The study follows this practice. The presentation covers the qualitative analysis of 34 pieces of LSL literature. The analysis employs ELAN software widely used in SL research. The target is to examine how specific types of each phonological parameter are used for the creation of metaphors in LSL literature and what metaphors are created. The results of the study show that LSL literature employs a range of metaphors created by using classifier signs and by modifying the established signs. The study also reveals that LSL literature tends to create reference metaphors indicating status and power. As the study shows, LSL poets metaphorically encode status by encoding another meaning in the same sign, which results in creating double metaphors. The metaphor of identity has been determined. Notably, the poetic context has revealed that the latter metaphor can also be identified as a metaphor for life. The study goes on to note that deaf poets create metaphors related to the importance of various phenomena significance of the lyrical subject. Notably, the study has allowed detecting locations, nonmanual features and etc., never mentioned in previous SL research as used for the creation of metaphors.

Keywords: Lithuanian sign language, sign language literature, sign language metaphor, metaphor at the phonological level, cognitive linguistics

Procedia PDF Downloads 136
3738 Low Pricing Strategy of Forest Products in Community Forestry Program: Subsidy to the Forest Users or Loss of Economy?

Authors: Laxuman Thakuri

Abstract:

Community-based forest management is often glorified as one of the best forest management alternatives in the developing countries like Nepal. It is also believed that the transfer of forest management authorities to local communities is decisive to take efficient decisions, maximize the forest benefits and improve the people’s livelihood. The community forestry of Nepal also aims to maximize the forest benefits; share them among the user households and improve their livelihood. However, how the local communities fix the price of forest products and local pricing made by the forest user groups affects to equitable forest benefits-sharing among the user households and their livelihood improvement objectives, the answer is largely silent among the researchers and policy-makers alike. This study examines local pricing system of forest products in the lowland community forestry and its effects on equitable benefit-sharing and livelihood improvement objectives. The study discovered that forest user groups fixed the price of forest products based on three criteria: i) costs incur in harvesting, ii) office operation costs, and iii) livelihood improvement costs through community development and income generating activities. Since user households have heterogeneous socio-economic conditions, the forest user groups have been applied low pricing strategy even for high-value forest products that the access of socio-economically worse-off households can be increased. However, the results of forest products distribution showed that as a result of low pricing strategy the access of socio-economically better-off households has been increasing at higher rate than worse-off and an inequality situation has been created. Similarly, the low pricing strategy is also found defective to livelihood improvement objectives. The study suggests for revising the forest products pricing system in community forest management and reforming the community forestry policy as well.

Keywords: community forestry, forest products pricing, equitable benefit-sharing, livelihood improvement, Nepal

Procedia PDF Downloads 299
3737 The Effects of Arginine, Glutamine and Threonine Supplementation in the Starting Phase on Subsequent Performance of Male Broile

Authors: Jalal Fazli Amiri, Mohammad Hossein Shahir, Mohammad Hossein Nemati, Afshin Heidarinia

Abstract:

The current study was performed to investigate the effects of arginine, threonine, and glutamine supplementation in excess of requirements in the starter period (17 days) on performance, intestinal morphology, and immune response of broilers. Four hundred and sixteen male day-old chicks were assigned in a 2×2×2 factorial arrangement to a completely randomized design with four replicates (13 birds per replicate ). Treatments were: a control group that received the basal diet, basal diet plus 1% glutamine, basal diet plus 0.2% threonine, basal diet plus 0.75 % arginine, and combination of these three amino acids (glutamine+arginine, glutamine+threonine, arginine+threonine and arginine+ glutamine+threonine). The effect of glutamine supplementation on feed intake was significant in week 4 (p < 0.05), week 6 (p < 0.001), and total feed intake (p < 0.05) and caused declined feed intake. No significant differences of glutamine addition were observed on intestinal morphology (villi height, crypt depth, villi height to crypt depth ratio, villi width). Threonine supplementation caused increased weight gain in week 2 (p < 0.001) and 3 and a decrease of total feed intake (p < 0.05). Duodenum and jejunum villi height, crypt depth, villi height to crypt depth ratio, villi width were not affected. The effect of arginine supplementation was the increase of breast percentage (p < 0.05) and a decrease of jejunum villi high (p < 0.05) and Jejunum crypt depth (p < 0.05). Supplementation of arginine, threonine, and glutamine had no significant effects on blood titer of antibodies against Newcastle disease, infectious bronchitis, avian influenza. Overall, it seems that the supplementation of arginine, threonine, and glutamine in excess of requirements in the starter period had no effect on performance in subsequent periods and intestinal morphology.

Keywords: intestinal morphology, immunity, broiler chickens, glutamine, arginine, threonine

Procedia PDF Downloads 137
3736 Evaluating Robustness of Conceptual Rainfall-runoff Models under Climate Variability in Northern Tunisia

Authors: H. Dakhlaoui, D. Ruelland, Y. Tramblay, Z. Bargaoui

Abstract:

To evaluate the impact of climate change on water resources at the catchment scale, not only future projections of climate are necessary but also robust rainfall-runoff models that are able to be fairly reliable under changing climate conditions. This study aims at assessing the robustness of three conceptual rainfall-runoff models (GR4j, HBV and IHACRES) on five basins in Northern Tunisia under long-term climate variability. Their robustness was evaluated according to a differential split sample test based on a climate classification of the observation period regarding simultaneously precipitation and temperature conditions. The studied catchments are situated in a region where climate change is likely to have significant impacts on runoff and they already suffer from scarcity of water resources. They cover the main hydrographical basins of Northern Tunisia (High Medjerda, Zouaraâ, Ichkeul and Cap bon), which produce the majority of surface water resources in Tunisia. The streamflow regime of the basins can be considered as natural since these basins are located upstream from storage-dams and in areas where withdrawals are negligible. A 30-year common period (1970‒2000) was considered to capture a large spread of hydro-climatic conditions. The calibration was based on the Kling-Gupta Efficiency (KGE) criterion, while the evaluation of model transferability is performed according to the Nash-Suttfliff efficiency criterion and volume error. The three hydrological models were shown to have similar behaviour under climate variability. Models prove a better ability to simulate the runoff pattern when transferred toward wetter periods compared to the case when transferred to drier periods. The limits of transferability are beyond -20% of precipitation and +1.5 °C of temperature in comparison with the calibration period. The deterioration of model robustness could in part be explained by the climate dependency of some parameters.

Keywords: rainfall-runoff modelling, hydro-climate variability, model robustness, uncertainty, Tunisia

Procedia PDF Downloads 292
3735 Transverse Momentum Dependent Factorization and Evolution for Spin Physics

Authors: Bipin Popat Sonawane

Abstract:

After 1988 Electron muon Collaboration (EMC) announcement of measurement of spin dependent structure function, it has been found that it has become a need to understand spin structure of a hadron. In the study of three-dimensional spin structure of a proton, we need to understand the foundation of quantum field theory in terms of electro-weak and strong theories using rigorous mathematical theories and models. In the process of understanding the inner dynamical stricture of proton we need understand the mathematical formalism in perturbative quantum chromodynamics (pQCD). In QCD processes like proton-proton collision at high energy we calculate cross section using conventional collinear factorization schemes. In this calculations, parton distribution functions (PDFs) and fragmentation function are used which provide the information about probability density of finding quarks and gluons ( partons) inside the proton and probability density of finding final hadronic state from initial partons. In transverse momentum dependent (TMD) PDFs and FFs, collectively called as TMDs, take an account for intrinsic transverse motion of partons. The TMD factorization in the calculation of cross sections provide a scheme of hadronic and partonic states in the given QCD process. In this study we review Transverse Momentum Dependent (TMD) factorization scheme using Collins-Soper-Sterman (CSS) Formalism. CSS formalism considers the transverse momentum dependence of the partons, in this formalism the cross section is written as a Fourier transform over a transverse position variable which has physical interpretation as impact parameter. Along with this we compare this formalism with improved CSS formalism. In this work we study the TMD evolution schemes and their comparison with other schemes. This would provide description in the process of measurement of transverse single spin asymmetry (TSSA) in hadro-production and electro-production of J/psi meson at RHIC, LHC, ILC energy scales. This would surely help us to understand J/psi production mechanism which is an appropriate test of QCD.

Keywords: QCD, PDF, TMD, CSS

Procedia PDF Downloads 69
3734 Study on the Wave Dissipation Performance of Double-Cylinder and Double-Plate Floating Breakwater

Authors: Liu Bijin

Abstract:

Floating breakwaters have several advantages, including being environmentally friendly, easy to construct, and cost-effective regardless of water depth. They have a broad range of applications in coastal engineering. However, they face significant challenges due to the unstable effect of wave dissipation, structural vulnerability, and high mooring system requirements. This paper investigates the wave dissipation performance of a floating breakwater structure. The structure consists of double cylinders, double vertical plates, and horizontal connecting plates. The investigation is carried out using physical model tests and numerical simulation methods based on STAR-CCM+. This paper discusses the impact of wave elements, relative vertical plate heights, and relative horizontal connecting plate widths on the wave dissipation performance of the double-cylinder, double-plate floating breakwater (DCDPFB). The study also analyses the changes in local vorticity and velocity fields around the DCDPFB to determine the optimal structural dimensions. The study found that the relative width of the horizontal connecting plate, the relative height of the vertical plate, and the size of the semi-cylinder are the key factors affecting the wave dissipation performance of the DCDPFB. The transmittance coefficient is minimally affected by the wave height and the depth of water entry. The local vortex and velocity field formed around the DCDPFB are important factors for dissipating wave energy. The test section of the DCDPFB, constructed according to the relative optimal structural dimensions, showed good wave dissipation performance during offshore prototype tests. The test section of DCDPFB, constructed with optimal structural dimensions, exhibits excellent wave dissipation performance in offshore prototype tests.

Keywords: floating breakwater, wave dissipation performance, transmittance coefficient, model test

Procedia PDF Downloads 56
3733 Mathematical Modeling of the AMCs Cross-Contamination Removal in the FOUPs: Finite Element Formulation and Application in FOUP’s Decontamination

Authors: N. Santatriniaina, J. Deseure, T. Q. Nguyen, H. Fontaine, C. Beitia, L. Rakotomanana

Abstract:

Nowadays, with the increasing of the wafer's size and the decreasing of critical size of integrated circuit manufacturing in modern high-tech, microelectronics industry needs a maximum attention to challenge the contamination control. The move to 300 mm is accompanied by the use of Front Opening Unified Pods for wafer and his storage. In these pods an airborne cross contamination may occur between wafers and the pods. A predictive approach using modeling and computational methods is very powerful method to understand and qualify the AMCs cross contamination processes. This work investigates the required numerical tools which are employed in order to study the AMCs cross-contamination transfer phenomena between wafers and FOUPs. Numerical optimization and finite element formulation in transient analysis were established. Analytical solution of one dimensional problem was developed and the calibration process of physical constants was performed. The least square distance between the model (analytical 1D solution) and the experimental data are minimized. The behavior of the AMCs intransient analysis was determined. The model framework preserves the classical forms of the diffusion and convection-diffusion equations and yields to consistent form of the Fick's law. The adsorption process and the surface roughness effect were also traduced as a boundary condition using the switch condition Dirichlet to Neumann and the interface condition. The methodology is applied, first using the optimization methods with analytical solution to define physical constants, and second using finite element method including adsorption kinetic and the switch of Dirichlet to Neumann condition.

Keywords: AMCs, FOUP, cross-contamination, adsorption, diffusion, numerical analysis, wafers, Dirichlet to Neumann, finite elements methods, Fick’s law, optimization

Procedia PDF Downloads 506
3732 An Infrared Inorganic Scintillating Detector Applied in Radiation Therapy

Authors: Sree Bash Chandra Debnath, Didier Tonneau, Carole Fauquet, Agnes Tallet, Julien Darreon

Abstract:

Purpose: Inorganic scintillating dosimetry is the most recent promising technique to solve several dosimetric issues and provide quality assurance in radiation therapy. Despite several advantages, the major issue of using scintillating detectors is the Cerenkov effect, typically induced in the visible emission range. In this context, the purpose of this research work is to evaluate the performance of a novel infrared inorganic scintillator detector (IR-ISD) in the radiation therapy treatment to ensure Cerenkov free signal and the best matches between the delivered and prescribed doses during treatment. Methods: A simple and small-scale infrared inorganic scintillating detector of 100 µm diameter with a sensitive scintillating volume of 2x10-6 mm3 was developed. A prototype of the dose verification system has been introduced based on PTIR1470/F (provided by Phosphor Technology®) material used in the proposed novel IR-ISD. The detector was tested on an Elekta LINAC system tuned at 6 MV/15MV and a brachytherapy source (Ir-192) used in the patient treatment protocol. The associated dose rate was measured in count rate (photons/s) using a highly sensitive photon counter (sensitivity ~20ph/s). Overall measurements were performed in IBATM water tank phantoms by following international Technical Reports series recommendations (TRS 381) for radiotherapy and TG43U1 recommendations for brachytherapy. The performance of the detector was tested through several dosimetric parameters such as PDD, beam profiling, Cerenkov measurement, dose linearity, dose rate linearity repeatability, and scintillator stability. Finally, a comparative study is also shown using a reference microdiamond dosimeter, Monte-Carlo (MC) simulation, and data from recent literature. Results: This study is highlighting the complete removal of the Cerenkov effect especially for small field radiation beam characterization. The detector provides an entire linear response with the dose in the 4cGy to 800 cGy range, independently of the field size selected from 5 x 5 cm² down to 0.5 x 0.5 cm². A perfect repeatability (0.2 % variation from average) with day-to-day reproducibility (0.3% variation) was observed. Measurements demonstrated that ISD has superlinear behavior with dose rate (R2=1) varying from 50 cGy/s to 1000 cGy/s. PDD profiles obtained in water present identical behavior with a build-up maximum depth dose at 15 mm for different small fields irradiation. A low dimension of 0.5 x 0.5 cm² field profiles have been characterized, and the field cross profile presents a Gaussian-like shape. The standard deviation (1σ) of the scintillating signal remains within 0.02% while having a very low convolution effect, thanks to lower sensitive volume. Finally, during brachytherapy, a comparison with MC simulations shows that considering energy dependency, measurement agrees within 0.8% till 0.2 cm source to detector distance. Conclusion: The proposed scintillating detector in this study shows no- Cerenkov radiation and efficient performance for several radiation therapy measurement parameters. Therefore, it is anticipated that the IR-ISD system can be promoted to validate with direct clinical investigations, such as appropriate dose verification and quality control in the Treatment Planning System (TPS).

Keywords: IR-Scintillating detector, dose measurement, micro-scintillators, Cerenkov effect

Procedia PDF Downloads 182
3731 Contrasting Infrastructure Sharing and Resource Substitution Synergies Business Models

Authors: Robin Molinier

Abstract:

Industrial symbiosis (I.S) rely on two modes of cooperation that are infrastructure sharing and resource substitution to obtain economic and environmental benefits. The former consists in the intensification of use of an asset while the latter is based on the use of waste, fatal energy (and utilities) as alternatives to standard inputs. Both modes, in fact, rely on the shift from a business-as-usual functioning towards an alternative production system structure so that in a business point of view the distinction is not clear. In order to investigate the way those cooperation modes can be distinguished, we consider the stakeholders' interplay in the business model structure regarding their resources and requirements. For infrastructure sharing (following economic engineering literature) the cost function of capacity induces economies of scale so that demand pooling reduces global expanses. Grassroot investment sizing decision and the ex-post pricing strongly depends on the design optimization phase for capacity sizing whereas ex-post operational cost sharing minimizing budgets are less dependent upon production rates. Value is then mainly design driven. For resource substitution, synergies value stems from availability and is at risk regarding both supplier and user load profiles and market prices of the standard input. Baseline input purchasing cost reduction is thus more driven by the operational phase of the symbiosis and must be analyzed within the whole sourcing policy (including diversification strategies and expensive back-up replacement). Moreover, while resource substitution involves a chain of intermediate processors to match quality requirements, the infrastructure model relies on a single operator whose competencies allow to produce non-rival goods. Transaction costs appear higher in resource substitution synergies due to the high level of customization which induces asset specificity, and non-homogeneity following transaction costs economics arguments.

Keywords: business model, capacity, sourcing, synergies

Procedia PDF Downloads 174
3730 Experimental Study of an Isobaric Expansion Heat Engine with Hydraulic Power Output for Conversion of Low-Grade-Heat to Electricity

Authors: Maxim Glushenkov, Alexander Kronberg

Abstract:

Isobaric expansion (IE) process is an alternative to conventional gas/vapor expansion accompanied by a pressure decrease typical of all state-of-the-art heat engines. The elimination of the expansion stage accompanied by useful work means that the most critical and expensive parts of ORC systems (turbine, screw expander, etc.) are also eliminated. In many cases, IE heat engines can be more efficient than conventional expansion machines. In addition, IE machines have a very simple, reliable, and inexpensive design. They can also perform all the known operations of existing heat engines and provide usable energy in a very convenient hydraulic or pneumatic form. This paper reports measurement made with the engine operating as a heat-to-shaft-power or electricity converter and a comparison of the experimental results to a thermodynamic model. Experiments were carried out at heat source temperature in the range 30–85 °C and heat sink temperature around 20 °C; refrigerant R134a was used as the engine working fluid. The pressure difference generated by the engine varied from 2.5 bar at the heat source temperature 40 °C to 23 bar at the heat source temperature 85 °C. Using a differential piston, the generated pressure was quadrupled to pump hydraulic oil through a hydraulic motor that generates shaft power and is connected to an alternator. At the frequency of about 0.5 Hz, the engine operates with useful powers up to 1 kW and an oil pumping flowrate of 7 L/min. Depending on the temperature of the heat source, the obtained efficiency was 3.5 – 6 %. This efficiency looks very high, considering such a low temperature difference (10 – 65 °C) and low power (< 1 kW). The engine’s observed performance is in good agreement with the predictions of the model. The results are very promising, showing that the engine is a simple and low-cost alternative to ORC plants and other known energy conversion systems, especially at low temperatures (< 100 °C) and low power range (< 500 kW) where other known technologies are not economic. Thus low-grade solar, geothermal energy, biomass combustion, and waste heat with a temperature above 30 °C can be involved into various energy conversion processes.

Keywords: isobaric expansion, low-grade heat, heat engine, renewable energy, waste heat recovery

Procedia PDF Downloads 226
3729 Expression of DNMT Enzymes-Regulated miRNAs Involving in Epigenetic Event of Tumor and Margin Tissues in Patients with Breast Cancer

Authors: Fatemeh Zeinali Sehrig

Abstract:

Background: miRNAs play an important role in the post-transcriptional regulation of genes, including genes involved in DNA methylation (DNMTs), and are also important regulators of oncogenic pathways. The study of microRNAs and DNMTs in breast cancer allows the development of targeted treatments and early detection of this cancer. Methods and Materials: Clinical Patients and Samples: Institutional guidelines, including ethical approval and informed consent, were followed by the Ethics Committee (Ethics code: IR.IAU.TABRIZ.REC.1401.063) of Tabriz Azad University, Tabriz, Iran. In this study, tissues of 100 patients with breast cancer and tissues of 100 healthy women were collected from Noor Nejat Hospital in Tabriz. The basic characteristics of the patients with breast cancer included: 1)Tumor grade(Grade 3 = 5%, Grade 2 = 87.5%, Grade 1 = 7.5%), 2)Lymph node(Yes = 87.5%, No = 12.5%), 3)Family cancer history(Yes = 47.5%, No = 41.3%, Unknown = 11.2%), 4) Abortion history(Yes = 36.2%).In silico methods (data gathering, process, and build networks): Gene Expression Omnibus (GEO), a high-throughput genomic database, was queried for miRNAs expression profiles in breast cancer. For Experimental protocol Tissue Processing, Total RNA isolation, complementary DNA(cDNA) synthesis, and quantitative real time PCR (QRT-PCR) analysis were performed. Results: In the present study, we found significant (p.value<0.05) changes in the expression level of miRNAs and DNMTs in patients with breast cancer. In bioinformatics studies, the GEO microarray data set, similar to qPCR results, showed a decreased expression of miRNAs and increased expression of DNMTs in breast cancer. Conclusion: According to the results of the present study, which showed a decrease in the expression of miRNAs and DNMTs in breast cancer, it can be said that these genes can be used as important diagnostic and therapeutic biomarkers in breast cancer.

Keywords: gene expression omnibus, microarray dataset, breast cancer, miRNA, DNMT (DNA methyltransferases)

Procedia PDF Downloads 34
3728 Ragging and Sludging Measurement in Membrane Bioreactors

Authors: Pompilia Buzatu, Hazim Qiblawey, Albert Odai, Jana Jamaleddin, Mustafa Nasser, Simon J. Judd

Abstract:

Membrane bioreactor (MBR) technology is challenged by the tendency for the membrane permeability to decrease due to ‘clogging’. Clogging includes ‘sludging’, the filling of the membrane channels with sludge solids, and ‘ragging’, the aggregation of short filaments to form long rag-like particles. Both sludging and ragging demand manual intervention to clear out the solids, which is time-consuming, labour-intensive and potentially damaging to the membranes. These factors impact on costs more significantly than membrane surface fouling which, unlike clogging, is largely mitigated by the chemical clean. However, practical evaluation of MBR clogging has thus far been limited. This paper presents the results of recent work attempting to quantify sludging and clogging based on simple bench-scale tests. Results from a novel ragging simulation trial indicated that rags can be formed within 24-36 hours from dispersed < 5 mm-long filaments at concentrations of 5-10 mg/L under gently agitated conditions. Rag formation occurred for both a cotton wool standard and samples taken from an operating municipal MBR, with between 15% and 75% of the added fibrous material forming a single rag. The extent of rag formation depended both on the material type or origin – lint from laundering operations forming zero rags – and the filament length. Sludging rates were quantified using a bespoke parallel-channel test cell representing the membrane channels of an immersed flat sheet MBR. Sludge samples were provided from two local MBRs, one treating municipal and the other industrial effluent. Bulk sludge properties measured comprised mixed liquor suspended solids (MLSS) concentration, capillary suction time (CST), particle size, soluble COD (sCOD) and rheology (apparent viscosity μₐ vs shear rate γ). The fouling and sludging propensity of the sludge was determined using the test cell, ‘fouling’ being quantified as the pressure incline rate against flux via the flux step test (for which clogging was absent) and sludging by photographing the channel and processing the image to determine the ratio of the clogged to unclogged regions. A substantial difference in rheological and fouling behaviour was evident between the two sludge sources, the industrial sludge having a higher viscosity but less shear-thinning than the municipal. Fouling, as manifested by the pressure increase Δp/Δt, as a function of flux from classic flux-step experiments (where no clogging was evident), was more rapid for the industrial sludge. Across all samples of both sludge origins the expected trend of increased fouling propensity with increased CST and sCOD was demonstrated, whereas no correlation was observed between clogging rate and these parameters. The relative contribution of fouling and clogging was appraised by adjusting the clogging propensity via increasing the MLSS both with and without a commensurate increase in the COD. Results indicated that whereas for the municipal sludge the fouling propensity was affected by the increased sCOD, there was no associated increased in the sludging propensity (or cake formation). The clogging rate actually decreased on increasing the MLSS. Against this, for the industrial sludge the clogging rate dramatically increased with solids concentration despite a decrease in the soluble COD. From this was surmised that sludging did not relate to fouling.

Keywords: clogging, membrane bioreactors, ragging, sludge

Procedia PDF Downloads 178
3727 Reasonable Adjustment for Students with Disabilities - Opportunities and Limits in Social Work Education

Authors: Bartelsen-Raemy Annabelle, Gerber Andrea

Abstract:

Objectives: The adoption of the UN Convention on the Rights of Persons with Disabilities has the effect that higher education institutions in Switzerland are called upon to promote inclusive university education. In this context, our School of Social Work aims to provide fair participation and the removal of barriers in our study programmes at bachelor’s and master’s levels. In 2015 we developed a concept of reasonable adjustments for students with disabilities and chronic illness as an instrument to provide equal opportunities for those students. We reviewed the implementation of this concept as part of our quality management process. Using a qualitative research design, we explored how affected students and lecturers experience the processes and measures taken and which barriers they still perceive. Methods: We captured subjective perspectives and experience of measures by conducting 15 problem-centred interviews with affected students and three experimental focus groups with lecturers. The data was processed using structured qualitative content analysis and summarised as key categories. Results: All respondents evaluated the concept of reasonable adjustment very positively and emphasised its importance for equal opportunities. Our analysis revealed differences in the usage and perception of both groups and showed that the students interviewed were a heterogeneous group with different needs. Overall, the students described the adjustments, in particular in relation to examinations and other assignments, as a great relief. The lecturers expressed high standards for their own teaching and supervision of students and, at the same time, wished for more support from the university. However, despite the positive evaluation by the lecturers, the limits of reasonable adjustment became evident. It is necessary to consider the limits of reasonable adjustments in terms of professional skills. Conclusion: Reasonable adjustments should, therefore, be seen as an element of an inclusive university culture that must be complemented by further measures. Taking this into account, we have planned further research as a basis for the development of a diversity and inclusion policy.

Keywords: opportunities and limits, reasonable adjustment, social work education, students with disabilities

Procedia PDF Downloads 132
3726 Transfer Function Model-Based Predictive Control for Nuclear Core Power Control in PUSPATI TRIGA Reactor

Authors: Mohd Sabri Minhat, Nurul Adilla Mohd Subha

Abstract:

The 1MWth PUSPATI TRIGA Reactor (RTP) in Malaysia Nuclear Agency has been operating more than 35 years. The existing core power control is using conventional controller known as Feedback Control Algorithm (FCA). It is technically challenging to keep the core power output always stable and operating within acceptable error bands for the safety demand of the RTP. Currently, the system could be considered unsatisfactory with power tracking performance, yet there is still significant room for improvement. Hence, a new design core power control is very important to improve the current performance in tracking and regulating reactor power by controlling the movement of control rods that suit the demand of highly sensitive of nuclear reactor power control. In this paper, the proposed Model Predictive Control (MPC) law was applied to control the core power. The model for core power control was based on mathematical models of the reactor core, MPC, and control rods selection algorithm. The mathematical models of the reactor core were based on point kinetics model, thermal hydraulic models, and reactivity models. The proposed MPC was presented in a transfer function model of the reactor core according to perturbations theory. The transfer function model-based predictive control (TFMPC) was developed to design the core power control with predictions based on a T-filter towards the real-time implementation of MPC on hardware. This paper introduces the sensitivity functions for TFMPC feedback loop to reduce the impact on the input actuation signal and demonstrates the behaviour of TFMPC in term of disturbance and noise rejections. The comparisons of both tracking and regulating performance between the conventional controller and TFMPC were made using MATLAB and analysed. In conclusion, the proposed TFMPC has satisfactory performance in tracking and regulating core power for controlling nuclear reactor with high reliability and safety.

Keywords: core power control, model predictive control, PUSPATI TRIGA reactor, TFMPC

Procedia PDF Downloads 241
3725 Dido: An Automatic Code Generation and Optimization Framework for Stencil Computations on Distributed Memory Architectures

Authors: Mariem Saied, Jens Gustedt, Gilles Muller

Abstract:

We present Dido, a source-to-source auto-generation and optimization framework for multi-dimensional stencil computations. It enables a large programmer community to easily and safely implement stencil codes on distributed-memory parallel architectures with Ordered Read-Write Locks (ORWL) as an execution and communication back-end. ORWL provides inter-task synchronization for data-oriented parallel and distributed computations. It has been proven to guarantee equity, liveness, and efficiency for a wide range of applications, particularly for iterative computations. Dido consists mainly of an implicitly parallel domain-specific language (DSL) implemented as a source-level transformer. It captures domain semantics at a high level of abstraction and generates parallel stencil code that leverages all ORWL features. The generated code is well-structured and lends itself to different possible optimizations. In this paper, we enhance Dido to handle both Jacobi and Gauss-Seidel grid traversals. We integrate temporal blocking to the Dido code generator in order to reduce the communication overhead and minimize data transfers. To increase data locality and improve intra-node data reuse, we coupled the code generation technique with the polyhedral parallelizer Pluto. The accuracy and portability of the generated code are guaranteed thanks to a parametrized solution. The combination of ORWL features, the code generation pattern and the suggested optimizations, make of Dido a powerful code generation framework for stencil computations in general, and for distributed-memory architectures in particular. We present a wide range of experiments over a number of stencil benchmarks.

Keywords: stencil computations, ordered read-write locks, domain-specific language, polyhedral model, experiments

Procedia PDF Downloads 127
3724 Root System Architecture Analysis of Sorghum Genotypes and Its Effect on Drought Adaptation

Authors: Hailemariam Solomon, Taye Tadesse, Daniel Nadew, Firezer Girma

Abstract:

Sorghum is an important crop in semi-arid regions and has shown resilience to drought stress. However, recurrent drought is affecting its productivity. Therefore, it is necessary to explore genes that contribute to drought stress adaptation to increase sorghum productivity. The aim of this study is to evaluate and determine the effect of root system traits, specifically root angle, on drought stress adaptation and grain yield performance in sorghum genotypes. A total of 428 sorghum genotypes from the Ethiopian breeding program were evaluated in three drought-stress environments. Field trials were conducted using a row-column design with three replications. Root system traits were phenotyped using a high-throughput phenotyping platform and analyzed using a row-column design with two replications. Data analysis was performed using R software and regression analysis. The study found significant variations in root system architecture among the sorghum genotypes. Non-stay-green genotypes had a grain yield ranging from 1.63 to 3.1 tons/ha, while stay-green genotypes had a grain yield ranging from 2.4 to 2.9 tons/ha. The analysis of root angle showed that non-stay-green genotypes had an angle ranging from 8.0 to 30.5 degrees, while stay-green genotypes had an angle ranging from 12.0 to 29.0 degrees. Improved varieties exhibited angles between 14.04 and 19.50 degrees. Positive and significant correlations were observed between leaf areas and shoot dry weight, as well as between leaf width and shoot dry weight. Negative correlations were observed between root angle and leaf area, as well as between root angle and root length. This research highlights the importance of root system architecture, particularly root angle traits, in enhancing grain yield production in drought-stressed conditions. It also establishes an association between root angle and grain yield traits for maximizing sorghum productivity.

Keywords: roor sysytem architecture, root angle, narrow root angle, wider root angle, drought

Procedia PDF Downloads 75
3723 Using of the Fractal Dimensions for the Analysis of Hyperkinetic Movements in the Parkinson's Disease

Authors: Sadegh Marzban, Mohamad Sobhan Sheikh Andalibi, Farnaz Ghassemi, Farzad Towhidkhah

Abstract:

Parkinson's disease (PD), which is characterized by the tremor at rest, rigidity, akinesia or bradykinesia and postural instability, affects the quality of life of involved individuals. The concept of a fractal is most often associated with irregular geometric objects that display self-similarity. Fractal dimension (FD) can be used to quantify the complexity and the self-similarity of an object such as tremor. In this work, we are aimed to propose a new method for evaluating hyperkinetic movements such as tremor, by using the FD and other correlated parameters in patients who are suffered from PD. In this study, we used 'the tremor data of Physionet'. The database consists of fourteen participants, diagnosed with PD including six patients with high amplitude tremor and eight patients with low amplitude. We tried to extract features from data, which can distinguish between patients before and after medication. We have selected fractal dimensions, including correlation dimension, box dimension, and information dimension. Lilliefors test has been used for normality test. Paired t-test or Wilcoxon signed rank test were also done to find differences between patients before and after medication, depending on whether the normality is detected or not. In addition, two-way ANOVA was used to investigate the possible association between the therapeutic effects and features extracted from the tremor. Just one of the extracted features showed significant differences between patients before and after medication. According to the results, correlation dimension was significantly different before and after the patient's medication (p=0.009). Also, two-way ANOVA demonstrates significant differences just in medication effect (p=0.033), and no significant differences were found between subject's differences (p=0.34) and interaction (p=0.97). The most striking result emerged from the data is that correlation dimension could quantify medication treatment based on tremor. This study has provided a technique to evaluate a non-linear measure for quantifying medication, nominally the correlation dimension. Furthermore, this study supports the idea that fractal dimension analysis yields additional information compared with conventional spectral measures in the detection of poor prognosis patients.

Keywords: correlation dimension, non-linear measure, Parkinson’s disease, tremor

Procedia PDF Downloads 244
3722 Impact of Contemporary Performance Measurement System and Organization Justice on Academic Staff Work Performance

Authors: Amizawati Mohd Amir, Ruhanita Maelah, Zaidi Mohd Noor

Abstract:

As part of the Malaysia Higher Institutions' Strategic Plan in promoting high-quality research and education, the Ministry of Higher Education has introduced various instrument to assess the universities performance. The aims are that university will produce more commercially-oriented research and continue to contribute in producing professional workforce for domestic and foreign needs. Yet the spirit of the success lies in the commitment of university particularly the academic staff to translate the vision into reality. For that reason, the element of fairness and justice in assessing individual academic staff performance is crucial to promote directly linked between university and individual work goals. Focusing on public research universities (RUs) in Malaysia, this study observes at the issue through the practice of university contemporary performance measurement system. Accordingly management control theory has conceptualized that contemporary performance measurement consisting of three dimension namely strategic, comprehensive and dynamic building upon equity theory, the relationships between contemporary performance measurement system and organizational justice and in turn the effect on academic staff work performance are tested based on online survey data administered on 365 academic staff from public RUs, which were analyzed using statistics analysis SPSS and Equation Structure Modeling. The findings validated the presence of strategic, comprehensive and dynamic in the contemporary performance measurement system. The empirical evidence also indicated that contemporary performance measure and procedural justice are significantly associated with work performance but not for distributive justice. Furthermore, procedural justice does mediate the relationship between contemporary performance measurement and academic staff work performance. Evidently, this study provides evidence on the importance of perceptions of justice towards influencing academic staff work performance. This finding may be a fruitful input in the setting up academic staff performance assessment policy.

Keywords: comprehensive, dynamic, distributive justice, contemporary performance measurement system, strategic, procedure justice, work performance

Procedia PDF Downloads 408
3721 Investigation of Mechanical Properties of Epoxy-Nanocomposite Reinforced with Copper Coated MWCNTs

Authors: M. Nazem Salimi, C. Abrinia, M. Baniassadi, M. Ehsani

Abstract:

Mechanical properties of epoxy based nanocomposites containing copper coated MWCNTs were investigated and a comparative study between nanocomposites containing functionalized MWCNTs and copper coated MWCNTs which are already functionalized was conducted. The MWCNTs was deposited with copper nanoparticles through electroless deposition process after accomplishment of "two-step" method as sensitization and activation procedures on oxidized MWCNTs. In addition, functionalization of MWCNTs was carried out through combination of two covalent and non-covalent funcionalization methods using HNO3 for acid solution of covalent treatment and Triton X100 as non-ionic surfactant of non-covalent treatment. The presence of functional groups and removal of impurities of MWCNTs were confirmed by FTIR and Raman spectroscopy, respectively. The layer of copper nanoparticles on the MWCNTs wall increasing its diameter was observed by SEM. Utilizing solution blending process, 0.1%, 0.5% and 1.5% wt loading of both copper coated MWCNTs and non-coated MWCNTs were used to prepare epoxy-based nanocomposites. The tensile, flexural and impact properties of nanocomposites were investigated. The results of tensile test demonstrated that nanocomposites containing copper coated MWCNTs exhibited brittle behavior compared to those reinforced with functionalized MWCNTs, whereas former one exhibited higher values of modulus than latter one for concentrations more than 0.4% wt. Presence of copper particles on MWCNTs surface decreased the tensile strength of nanocomposites. In comparison to pure epoxy, nanocomposites with treated-MWCNTs and Cu-MWCNTs loading of 0.1% wt showed an increase of 35% and 51.6% for flexural strength beside 20% and 30% increase in flexural modulus, respectively, whereas flexural properties of both naocomposites decreased with increasing of CNTs concentration. The results of impact strength of nanocomposites with Cu-CNTs demonstrated that impact properties decreased with increasing of filler content with a optimum value at 0.1% wt while in high concentrations impact properties of Cu-nanocomposites exhibited lower values than f-MWCNT nanocomposites.

Keywords: epoxyresin, nanocomposite, functionalization, copper, electroless deposition process, mechanical properties

Procedia PDF Downloads 367