Search results for: earth size
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 6621

Search results for: earth size

5781 Phyto-Assisted Synthesis of Magnesium Oxide Nanoparticles: Characterization and Applications

Authors: Surendra Kumar Gautam, Mahesh Dhungana

Abstract:

Magnesium oxide nanoparticles (MgO NPs) are less toxic to humans and the environment as compared to other metal oxide nanoparticles. Various conventional chemical and physical methods are used for synthesis whose toxicity level is high and highly expensive. As the best alternative, phyto-assisted synthesis has emerged, which uses extracts from plant parts for the synthesis of nanoparticles. Here, we report the synthesis of MgO nanoparticles with the assistance of beetroot extract and leaf extract of P. guajava and A. adenophora. The synthesized MgO NPs were characterized by X-ray diffraction (XRD), Fourier transforms infrared spectroscopy (FTIR), and UV-visible spectroscopy. X-ray analysis for the broadening of peaks was used to evaluate the crystallite size and lattice strain using Debye-Scherer and Williamson–Hall method. The results of crystallite size obtained by both methods are in close proximity. The crystallite size obtained by the Williamson-Hall method seems more accurate, with values being 8.1 nm and 13.2 nm for beetroot MgO NPs and P. guajava MgO NPs, respectively. The FT-IR spectroscopy revealed the dominance of chemical bonds as well as functional groups on MgO NPs surfaces. The UV-visible absorption spectra of MgO NPs were found to be 310 nm, 315 nm, and 315 nm for beetroot, P. guajava, and A. adenophora leaf extract, respectively. Among the three samples, beetroot-mediated MgO NPs were effective antibacterial against both gram-positive and Gram-negative bacteria. In addition, synthesized MgO NPs also show significant antioxidant efficacy against 1,1-diphenyl-2-picrylhydrazyl radical. Further, beetroot MgO NPs showed the highest photocatalytic activity of about 91% in comparison with other samples.

Keywords: MgO NPs, XRD, FTIR, antibacterial, antioxidant and photocatalytic activity

Procedia PDF Downloads 71
5780 Microbial Dark Matter Analysis Using 16S rRNA Gene Metagenomics Sequences

Authors: Hana Barak, Alex Sivan, Ariel Kushmaro

Abstract:

Microorganisms are the most diverse and abundant life forms on Earth and account for a large portion of the Earth’s biomass and biodiversity. To date though, our knowledge regarding microbial life is lacking, as it is based mainly on information from cultivated organisms. Indeed, microbiologists have borrowed from astrophysics and termed the ‘uncultured microbial majority’ as ‘microbial dark matter’. The realization of how diverse and unexplored microorganisms are, actually stems from recent advances in molecular biology, and in particular from novel methods for sequencing microbial small subunit ribosomal RNA genes directly from environmental samples termed next-generation sequencing (NGS). This has led us to use NGS that generates several gigabases of sequencing data in a single experimental run, to identify and classify environmental samples of microorganisms. In metagenomics sequencing analysis (both 16S and shotgun), sequences are compared to reference databases that contain only small part of the existing microorganisms and therefore their taxonomy assignment may reveal groups of unknown microorganisms or origins. These unknowns, or the ‘microbial sequences dark matter’, are usually ignored in spite of their great importance. The goal of this work was to develop an improved bioinformatics method that enables more complete analyses of the microbial communities in numerous environments. Therefore, NGS was used to identify previously unknown microorganisms from three different environments (industrials wastewater, Negev Desert’s rocks and water wells at the Arava valley). 16S rRNA gene metagenome analysis of the microorganisms from those three environments produce about ~4 million reads for 75 samples. Between 0.1-12% of the sequences in each sample were tagged as ‘Unassigned’. Employing relatively simple methodology for resequencing of original gDNA samples through Sanger or MiSeq Illumina with specific primers, this study demonstrates that the mysterious ‘Unassigned’ group apparently contains sequences of candidate phyla. Those unknown sequences can be located on a phylogenetic tree and thus provide a better understanding of the ‘sequences dark matter’ and its role in the research of microbial communities and diversity. Studying this ‘dark matter’ will extend the existing databases and could reveal the hidden potential of the ‘microbial dark matter’.

Keywords: bacteria, bioinformatics, dark matter, Next Generation Sequencing, unknown

Procedia PDF Downloads 239
5779 Phenology and Size in the Social Sweat Bee, Halictus ligatus, in an Urban Environment

Authors: Rachel A. Brant, Grace E. Kenny, Paige A. Muñiz, Gerardo R. Camilo

Abstract:

The social sweat bee, Halictus ligatus, has been documented to alter its phenology as a response to changes in temporal dynamics of resources. Furthermore, H. ligatus exhibits polyethism in natural environments as a consequence of the variation in resources. Yet, we do not know if or how H. ligatus responds to these variations in urban environments. As urban environments become much more widespread, and human population is expected to reach nine billion by 2050, it is crucial to distinguish how resources are allocated by bees in cities. We hypothesize that in urban regions, where floral availability varies with human activity, H. ligatus will exhibit polyethism in order to match the extremely localized spatial variability of resources. We predict that in an urban setting, where resources vary both spatially and temporally, the phenology of H. ligatus will alter in response to these fluctuations. This study was conducted in Saint Louis, Missouri, at fifteen sites each varying in size and management type (community garden, urban farm, prairie restoration). Bees were collected by hand netting from 2013-2016. Results suggest that the largest individuals, mostly gynes, occurred in lower income neighborhood community gardens in May and August. We used a model averaging procedure, based on information theoretical methods, to determine a best model for predicting bee size. Our results suggest that month and locality within the city are the best predictors of bee size. Halictus ligatus was observed to comply with the predictions of polyethism from 2013 to 2015. However, in 2016 there was an almost complete absence of the smallest worker castes. This is a significant deviation from what is expected under polyethism. This could be attributed to shifts in planting decisions, shifts in plant-pollinator matches, or local climatic conditions. Further research is needed to determine if this divergence from polyethism is a new strategy for the social sweat bee as climate continues to alter or a response to human dominated landscapes.

Keywords: polyethism, urban environment, phenology, social sweat bee

Procedia PDF Downloads 208
5778 Characterization of Agroforestry Systems in Burkina Faso Using an Earth Observation Data Cube

Authors: Dan Kanmegne

Abstract:

Africa will become the most populated continent by the end of the century, with around 4 billion inhabitants. Food security and climate changes will become continental issues since agricultural practices depend on climate but also contribute to global emissions and land degradation. Agroforestry has been identified as a cost-efficient and reliable strategy to address these two issues. It is defined as the integrated management of trees and crops/animals in the same land unit. Agroforestry provides benefits in terms of goods (fruits, medicine, wood, etc.) and services (windbreaks, fertility, etc.), and is acknowledged to have a great potential for carbon sequestration; therefore it can be integrated into reduction mechanisms of carbon emissions. Particularly in sub-Saharan Africa, the constraint stands in the lack of information about both areas under agroforestry and the characterization (composition, structure, and management) of each agroforestry system at the country level. This study describes and quantifies “what is where?”, earliest to the quantification of carbon stock in different systems. Remote sensing (RS) is the most efficient approach to map such a dynamic technology as agroforestry since it gives relatively adequate and consistent information over a large area at nearly no cost. RS data fulfill the good practice guidelines of the Intergovernmental Panel On Climate Change (IPCC) that is to be used in carbon estimation. Satellite data are getting more and more accessible, and the archives are growing exponentially. To retrieve useful information to support decision-making out of this large amount of data, satellite data needs to be organized so to ensure fast processing, quick accessibility, and ease of use. A new solution is a data cube, which can be understood as a multi-dimensional stack (space, time, data type) of spatially aligned pixels and used for efficient access and analysis. A data cube for Burkina Faso has been set up from the cooperation project between the international service provider WASCAL and Germany, which provides an accessible exploitation architecture of multi-temporal satellite data. The aim of this study is to map and characterize agroforestry systems using the Burkina Faso earth observation data cube. The approach in its initial stage is based on an unsupervised image classification of a normalized difference vegetation index (NDVI) time series from 2010 to 2018, to stratify the country based on the vegetation. Fifteen strata were identified, and four samples per location were randomly assigned to define the sampling units. For safety reasons, the northern part will not be part of the fieldwork. A total of 52 locations will be visited by the end of the dry season in February-March 2020. The field campaigns will consist of identifying and describing different agroforestry systems and qualitative interviews. A multi-temporal supervised image classification will be done with a random forest algorithm, and the field data will be used for both training the algorithm and accuracy assessment. The expected outputs are (i) map(s) of agroforestry dynamics, (ii) characteristics of different systems (main species, management, area, etc.); (iii) assessment report of Burkina Faso data cube.

Keywords: agroforestry systems, Burkina Faso, earth observation data cube, multi-temporal image classification

Procedia PDF Downloads 131
5777 Enhancing the Flotation of Fine and Ultrafine Pyrite Particles Using Electrolytically Generated Bubbles

Authors: Bogale Tadesse, Krutik Parikh, Ndagha Mkandawire, Boris Albijanic, Nimal Subasinghe

Abstract:

It is well established that the floatability and selectivity of mineral particles are highly dependent on the particle size. Generally, a particle size of 10 micron is considered as the critical size below which both flotation selectivity and recovery decline sharply. It is widely accepted that the majority of ultrafine particles, including highly liberated valuable minerals, will be lost in tailings during a conventional flotation process. This is highly undesirable particularly in the processing of finely disseminated complex and refractory ores where there is a requirement for fine grinding in order to liberate the valuable minerals. In addition, the continuing decline in ore grade worldwide necessitates intensive processing of low grade mineral deposits. Recent advances in comminution allow the economic grinding of particles down to 10 micron sizes to enhance the probability of liberating locked minerals from low grade ores. Thus, it is timely that the flotation of fine and ultrafine particles is improved in order to reduce the amount of valuable minerals lost as slimes. It is believed that the use of fine bubbles in flotation increases the bubble-particle collision efficiency and hence the flotation performance. Electroflotation, where bubbles are generated by the electrolytic breakdown of water to produce oxygen and hydrogen gases, leads to the formation of extremely finely dispersed gas bubbles with dimensions varying from 5 to 95 micron. The sizes of bubbles generated by this method are significantly smaller than those found in conventional flotation (> 600 micron). In this study, microbubbles generated by electrolysis of water were injected into a bench top flotation cell to assess the performance electroflotation in enhancing the flotation of fine and ultrafine pyrite particles of sizes ranging from 5 to 53 micron. The design of the cell and the results from optimization of the process variables such as current density, pH, percent solid and particle size will be presented at this conference.

Keywords: electroflotation, fine bubbles, pyrite, ultrafine particles

Procedia PDF Downloads 316
5776 Behavior of Square Reinforced-Concrete Columns Strenghtened with Carbon Fiber Reinforced Polymers (CFRP) under Concentric Loading

Authors: Dana Abed, Mu`Tasim Abdel-Jaber, Nasim Shatarat

Abstract:

This study aims at investigating the influence of cross-sectional size on axial compressive capacity of carbon fiber reinforced polymer (CFRP) wrapped square reinforced concrete short columns. Three sets of columns were built for this purpose: 200x200x1200 mm; 250x250x1500 mm and 300x300x1800 mm. Each set includes a control column and a strengthened column with one layer of CFRP sheets. All columns were tested under the effect of pure axial compression load. The results of the study show that using CFRP sheets resulted in capacity enhancement of 37%, 32% and 27% for the 200×200, 250×250, and 300×300 mm, respectively. The results of the experimental program demonstrated that the percentage of improvement in strength decreased by increasing the cross-sectional size of the column.

Keywords: CFRP, columns, concentric loading, cross-sectional

Procedia PDF Downloads 273
5775 H2 Permeation Properties of a Catalytic Membrane Reactor in Methane Steam Reforming Reaction

Authors: M. Amanipour, J. Towfighi, E. Ganji Babakhani, M. Heidari

Abstract:

Cylindrical alumina microfiltration membrane (GMITM Corporation, inside diameter=9 mm, outside diameter=13 mm, length= 50 mm) with an average pore size of 0.5 micrometer and porosity of about 0.35 was used as the support for membrane reactor. This support was soaked in boehmite sols, and the mean particle size was adjusted in the range of 50 to 500 nm by carefully controlling hydrolysis time, and calcined at 650 °C for two hours. This process was repeated with different boehmite solutions in order to achieve an intermediate layer with an average pore size of about 50 nm. The resulting substrate was then coated with a thin and dense layer of silica by counter current chemical vapour deposition (CVD) method. A boehmite sol with 10 wt.% of nickel which was prepared by a standard procedure was used to make the catalytic layer. BET, SEM, and XRD analysis were used to characterize this layer. The catalytic membrane reactor was placed in an experimental setup to evaluate the permeation and hydrogen separation performance for a steam reforming reaction. The setup consisted of a tubular module in which the membrane was fixed, and the reforming reaction occurred at the inner side of the membrane. Methane stream, diluted with nitrogen, and deionized water with a steam to carbon (S/C) ratio of 3.0 entered the reactor after the reactor was heated up to 500 °C with a specified rate of 2 °C/ min and the catalytic layer was reduced at presence of hydrogen for 2.5 hours. Nitrogen flow was used as sweep gas through the outer side of the reactor. Any liquid produced was trapped and separated at reactor exit by a cold trap, and the produced gases were analyzed by an on-line gas chromatograph (Agilent 7890A) to measure total CH4 conversion and H2 permeation. BET analysis indicated uniform size distribution for catalyst with average pore size of 280 nm and average surface area of 275 m2.g-1. Single-component permeation tests were carried out for hydrogen, methane, and carbon dioxide at temperature range of 500-800 °C, and the results showed almost the same permeance and hydrogen selectivity values for hydrogen as the composite membrane without catalytic layer. Performance of the catalytic membrane was evaluated by applying membranes as a membrane reactor for methane steam reforming reaction at gas hourly space velocity (GHSV) of 10,000 h−1 and 2 bar. CH4 conversion increased from 50% to 85% with increasing reaction temperature from 600 °C to 750 °C, which is sufficiently above equilibrium curve at reaction conditions, but slightly lower than membrane reactor with packed nickel catalytic bed because of its higher surface area compared to the catalytic layer.

Keywords: catalytic membrane, hydrogen, methane steam reforming, permeance

Procedia PDF Downloads 247
5774 Seam Slippage of Light Woven Fabrics with Regards to Sewing Parameters

Authors: Mona Shawky, Khaled M. Elsheikh, Heba M. Darwish, Eman Abd El Elsamea

Abstract:

Seams are the basic component in the structure of any apparel. The seam quality of the garment is a term that indicates both the aesthetic and functional performance of the garment. Seam slippage is one of the important properties that determine garment performance. Lightweight fabrics are preferred for their aesthetic properties. Since seam slippage is one of the most occurable faults for woven garments, in this study, a design of experiment of the following sewing parameters (three levels of needle size, three levels of stitch density, three levels of the seam allowance, two levels of sewing thread count, and two fabric types) was used to obtain the effect of the interaction between different sewing parameters on-seam slippage force. Two lightweight polyester woven fabrics with different constructions were used with lock stitch 301 to perform this study. Regression equations which can predict seam slippage force in both warp and weft directions were concluded. It was found that fabric type has a significant positive effect on seam slippage force in the warp direction, while it has a significant negative effect on seam slippage force on weft direction. Also, the interaction between needle size and stitch density has a significant positive effect on seam slippage force on warp direction, while the interaction between stitch density and seam allowance has a negative effect on seam slippage force in the weft direction.

Keywords: needle size, regression equation, seam allowance, seam slippage, stitch density

Procedia PDF Downloads 154
5773 Doxorubicin and Cyclosporine Loaded PLGA Nanoparticles to Combat Multidrug Resistance

Authors: Senthil Rajan Dharmalingam, Shamala Nadaraju, Srinivasan Ramamurthy

Abstract:

Doxorubicin is the most widely used anticancer drugs in chemotherapy treatment. However, problems related to the development of multidrug resistance (MDR) and acute cardiotoxicity have led researchers to investigate alternative forms of administering doxorubicin for cancer therapy. Several methods have been attempted to overcome MDR, including the co-administration of a chemosensitizer inhibiting the efflux caused by ATP binding cassette transporters with anticancer drugs, and the bypass of the efflux mechanism. Co encapsulation of doxorubicin (Dox) and cyclosporine A (CSA) into poly (DL-lactide-co-glycolide) nanoparticles was emulsification-solvent evaporation method using polyvinyl alcohol as emulsion stabilizers. The Dox-CSA loaded nanoparticles were evaluated for particle size, zeta potential and PDI by light scattering analysis and thermal characterizations by differential scanning calorimetry (DSC). Loading efficiency (LE %) and in-vitro dissolution samples were evaluated by developed and validated HPLC method. The optimum particle size obtained is 298.6.8±39.4 nm and polydispersity index (PDI) is 0.098±0.092. Zeta potential is found to be -29.9±4.23. Optimum pH to increase Dox LE% was found 7.1 which gave 42.5% and 58.9% increase of LE% for pH 6.6 and pH 8.6 compared respectively. LE% achieved for Dox is 0.07±0.01 % and CSA is 0.09±0.03%. Increased volume of PVA and weight of PLGA shows increase in size of nanoparticles. DSC thermograms showed shift in the melting peak for the nanoparticles compared to Dox and CSA indicating encapsulation of drugs. In conclusion, these preliminary studies showed the feasibility of PLGA nanoparticles to entrap Dox and CSA and require future in-vivo studies to be performed to establish its potential.

Keywords: doxorubicin, cyclosporine, PLGA, nanoparticles

Procedia PDF Downloads 451
5772 The Association between Saharran Dust and Emergency Department Admission and Hospitalization in Gaziantep, Turkey

Authors: Behcet Al, Mustafa Bogan, Mehmet Murat Oktay, Suat Zengin, Hasan Bayram

Abstract:

Objective: In the last two decades there is a strong scientific interest regarding the role of aerosols for the Earth’s climate and associated changes. Aerosol particles are very important to the Earth-atmosphere climate system playing a crucial role in cloud and precipitation processes, air quality and climate. Here, we evaluated the association between saharran dust and emergency department admission, hospitalization, and mortality. Method: The records of admission to emergency department of Gaziantep University and the dust stroms of 31 months were studied. Patients admitted to ED at dust strom with chronic obstructive lung disease (COLD), asthma bronchiale (AB), serebrovascular events (SVE), acute myocardial infarction (AMI), stabile and unstabile angina pectoris (SAAP andUSAP); and the days with and without dust stroms were included. The study was realized from March 2010 to October 2012. The admission of three days before strom (group 1), during strom days (group 2) and three days after strom (group 3) were determined. The mean level of dust PM10 particulate was calculated, and the results were compared. Results: 5864 patients with chronic obstructive lung disease, asthma bronchiale, serebrovascular events, acute myocardial infarction, stabile and unstabile angyina pectoris admitted during the days with and without dust stroms. 28 dust stroms ocurred during 31 months. The totaliy of stroms continiued 78 days. Of admissions, 35.5% (n=2075) were in group1, 29.8% (n=1746) in group 2, and 34.8% (n=2043) were in group 3. The mean of PM10 for groups (group 1, 2 and 3) were 78.53 mg/m3 (range 19–276) particulate, 108.7 mg/m3 (range 34–631) particulate, and 60.9 mg/m3 (range 17–160) particulate respectively. The mean admission per a day for groups were 24.86, 22.55, and 24.50 respectively. The mortality was 12 in group 1, 12 in group 2, and 17 in grou 3. The hospitalization ratio for groups were 0.24, 0.27, and 0.27 respectively. Conclusion: However, the mean level of PM10 particulate for groups 2 (in dust strom days) is significantly higher (p=0.001) than the days before (group 1) and after (group 3) dust stroms, the mean admissions/day, hostilalization and mortality related to deseases (COLD, AB, SVE, AMI, SAAP andUSA) for group 2 is lower than the group 1 and group 3.

Keywords: Saharran dust, PM10 particulate, emergency department admission, mortality

Procedia PDF Downloads 389
5771 Permeodynamic Particulate Matter Filtration for Improved Air Quality

Authors: Hamad M. Alnagran, Mohammed S. Imbabi

Abstract:

Particulate matter (PM) in the air we breathe is detrimental to health. Overcoming this problem has attracted interest and prompted research on the use of PM filtration in commercial buildings and homes to be carried out. The consensus is that tangible health benefits can result from the use of PM filters in most urban environments, to clean up the building’s fresh air supply and thereby reduce exposure of residents to airborne PM. The authors have investigated and are developing a new large-scale Permeodynamic Filtration Technology (PFT) capable of permanently filtering and removing airborne PMs from outdoor spaces, thus also benefiting internal spaces such as the interiors of buildings. Theoretical models were developed, and laboratory trials carried out to determine, and validate through measurement permeodynamic filtration efficiency and pressure drop as functions of PM particle size distributions. The conclusion is that PFT offers a potentially viable, cost effective end of pipe solution to the problem of airborne PM.

Keywords: air filtration, particulate matter, particle size distribution, permeodynamic

Procedia PDF Downloads 190
5770 Vibration Frequencies Analysis of Nanoporous Graphene Membrane

Authors: Haw-Long Lee, Win-Jin Chang, Yu-Ching Yang

Abstract:

In this study, we use the atomic-scale finite element method to investigate the vibrational behavior of the armchair- and zigzag-structured nanoporous graphene layers with different size under the SFSF and CFFF boundary conditions. The fundamental frequencies computed for the graphene layers without pore are compared with the results of previous studies. We observe very good correspondence of our results with that of the other studies in all the considered cases. For the armchair- and zigzag-structured nanoporous graphene layers under the SFSF and CFFF boundary conditions, the frequencies decrease as the size of the nanopore increase. When the positions of the pore are symmetric with respect to the center of the graphene, the frequency of the zigzag pore graphene is higher than that of the armchair one.

Keywords: atomic-scale finite element method, graphene, nanoporous, natural frequency

Procedia PDF Downloads 349
5769 An Improved Circulating Tumor Cells Analysis Method for Identifying Tumorous Blood Cells

Authors: Salvador Garcia Bernal, Chi Zheng, Keqi Zhang, Lei Mao

Abstract:

Circulating Tumor Cells (CTC) is used to detect tumoral cell metastases using blood samples of patients with cancer (lung, breast, etc.). Using an immunofluorescent method a three channel image (Red, Green, and Blue) are obtained. These set of images usually overpass the 11 x 30 M pixels in size. An aided tool is designed for imaging cell analysis to segmented and identify the tumorous cell based on the three markers signals. Our Method, it is cell-based (area and cell shape) considering each channel information and extracting and making decisions if it is a valid CTC. The system also gives information about number and size of tumor cells found in the sample. We present results in real-life samples achieving acceptable performance in identifying CTCs in short time.

Keywords: Circulating Tumor Cells (CTC), cell analysis, immunofluorescent, medical image analysis

Procedia PDF Downloads 203
5768 Identification of Flooding Attack (Zero Day Attack) at Application Layer Using Mathematical Model and Detection Using Correlations

Authors: Hamsini Pulugurtha, V.S. Lakshmi Jagadmaba Paluri

Abstract:

Distributed denial of service attack (DDoS) is one altogether the top-rated cyber threats presently. It runs down the victim server resources like a system of measurement and buffer size by obstructing the server to supply resources to legitimate shoppers. Throughout this text, we tend to tend to propose a mathematical model of DDoS attack; we discuss its relevancy to the choices like inter-arrival time or rate of arrival of the assault customers accessing the server. We tend to tend to further analyze the attack model in context to the exhausting system of measurement and buffer size of the victim server. The projected technique uses an associate in nursing unattended learning technique, self-organizing map, to make the clusters of identical choices. Lastly, the abstract applies mathematical correlation and so the standard likelihood distribution on the clusters and analyses their behaviors to look at a DDoS attack. These systems not exclusively interconnect very little devices exchanging personal data, but to boot essential infrastructures news standing of nuclear facilities. Although this interconnection brings many edges and blessings, it to boot creates new vulnerabilities and threats which might be conversant in mount attacks. In such sophisticated interconnected systems, the power to look at attacks as early as accomplishable is of paramount importance.

Keywords: application attack, bandwidth, buffer correlation, DDoS distribution flooding intrusion layer, normal prevention probability size

Procedia PDF Downloads 208
5767 Localized Treatment of Cutaneous Candidiasis through Cubosomes in vitro Evaluation

Authors: Aakanchha Jain, D. V. Kohli

Abstract:

Cubosomes are nanoparticles but instead of the solid particles, cubosomes are self-assembled liquid crystalline particles of certain surfactant with proper ratio of water with a microstructure that provides unique properties of practical interest. Cubosomes encapsulating Fluconazole were prepared by emulsification method and characterized for particle size, entrapment efficiency. The cubosomes prepared were 257.2±2.94 nm in size with drug entrapment efficiency of 66.2±2.69%. The optimized formulation characterized for shape and surface morphology by TEM and SEM analysis. SEM photograph showed the smooth surface of optimized cubosomes and TEM photograph revealed square somewhat circular intact shapes of cubosomes. MIC was determined by XTT based method and antifungal activity was determined in vitro. The cumulative percentage of Fnz from cubosomes permeated via dialysis membrane (MWCO 12-14 KD) showed a percent cumulative drug release of 76.86% while Fnz solution showed release up to 91.04% in 24 hours in PBS (pH 6.5)(p < 0.005).

Keywords: Candids albicans, cubosomes, fluconazole, topical delivery

Procedia PDF Downloads 288
5766 Impact of Audit Committee on Real Earnings Management: Cases of Netherlands

Authors: Sana Masmoudi Mardassi, Yosra Makni Fourati

Abstract:

Regulators highlight the importance of the Audit Committee (AC) as a key internal corporate governance mechanism. One of the most important roles of this committee is to oversee the financial reporting process. The purpose of this paper is to examine the link between the characteristics of an audit committee and the financial reporting quality by investigating whether the characteristics of audit committees are associated with improved financial reporting quality, especially the Real Earnings Management. In the current study, a panel data from 80 nonfinancial companies listed on the Amsterdam Stock Exchange during the period between 2010 and 2017 were used. To measure audit committee characteristics, four proxies have been used, specifically, audit committee independence, financial expertise, gender diversity and AC meetings. For this research, a linear regression model was used to identify the influence of a set of board characteristics of the audit committee on real earnings management after controlling for firm audit committee size, leverage, size, loss, growth and board size. This research provides empirical evidence of the association between audit committee independence, financial expertise, gender diversity and meetings and Real Earnings Management (REM) as a proxy of financial reporting quality. The study finds that independence and AC Gender diversity are strongly related to financial reporting quality. In fact, these two characteristics constrain REM. The results also suggest that AC- financial expertise reduces to some extent, the likelihood of engaging in REM. These conclusions provide support then to the audit committee requirement under the Dutch Corporate Governance Code rules regarding gender diversity and AC meetings.

Keywords: audit committee, financial expertise, independence, real earnings management

Procedia PDF Downloads 153
5765 Comparative Analysis of Yield before and after Access to Extension Services among Crop Farmers in Bauchi Local Government Area of Bauchi State, Nigeria

Authors: U. S. Babuga, A. H. Danwanka, A. Garba

Abstract:

The research was carried out to compare the yield of respondents before and after access to extension services on crop production technologies in the study area. Data were collected from the study area through questionnaires administered to seventy-five randomly selected respondents. Data were analyzed using descriptive statistics, t-test and regression models. The result disclosed that majority (97%) of the respondent attended one form of school or the other. The majority (78.67%) of the respondents had farm size ranging between 1-3 hectares. The majority of the respondent adopt improved variety of crops, plant spacing, herbicide, fertilizer application, land preparation, crop protection, crop processing and storage of farm produce. The result of the t-test between the yield of respondents before and after access to extension services shows that there was a significant (p<0.001) difference in yield before and after access to extension. It also indicated that farm size was significant (p<0.001) while household size, years of farming experience and extension contact were significant at (p<0.005). The major constraint to adoption of crop production technologies were shortage of extension agents, high cost of technology and lack of access to credit facility. The major pre-requisite for the improvement of extension service are employment of more extension agents or workers and adequate training. Adequate agricultural credit to farmers at low interest rates will enhance their adoption of crop production technologies.

Keywords: comparative, analysis, yield, access, extension

Procedia PDF Downloads 348
5764 Impact of Population Size on Symmetric Travelling Salesman Problem Efficiency

Authors: Wafa' Alsharafat, Suhila Farhan Abu-Owida

Abstract:

Genetic algorithm (GA) is a powerful evolutionary searching technique that is used successfully to solve and optimize problems in different research areas. Genetic Algorithm (GA) considered as one of optimization methods used to solve Travel salesman Problem (TSP). The feasibility of GA in finding a TSP solution is dependent on GA operators; encoding method, population size, termination criteria, in general. In specific, crossover and its probability play a significant role in finding possible solutions for Symmetric TSP (STSP). In addition, the crossover should be determined and enhanced in term reaching optimal or at least near optimal. In this paper, we spot the light on using a modified crossover method called modified sequential constructive crossover and its impact on reaching optimal solution. To justify the relevance of a parameter value in solving the TSP, a set comparative analysis conducted on different crossover methods values.

Keywords: genetic algorithm, crossover, mutation, TSP

Procedia PDF Downloads 209
5763 Catalytic and Non-Catalytic Pyrolysis of Walnut Shell Waste to Biofuel: Characterisation of Catalytic Biochar and Biooil

Authors: Saimatun Nisa

Abstract:

Walnut is an important export product from the Union Territory of Jammy and Kashmir. After extraction of the kernel, the walnut shell forms a solid waste that needs to be managed. Pyrolysis is one interesting option for the utilization of this walnut waste. In this study microwave pyrolysis reactor is used to convert the walnut shell biomass into its value-added products. Catalytic and non-catalytic conversion of walnut shell waste to oil, gas and char was evaluated using a Co-based catalyst. The catalyst was characterized using XPS and SEM analysis. Pyrolysis temperature, reaction time, particle size and sweeping gas (N₂) flow rate were set in the ranges of 400–600 °C, 40 min, <0.6mm to < 4.75mm and 300 ml min−1, respectively. The heating rate was fixed at 40 °C min−1. Maximum gas yield was obtained at 600 °C, 40 min, particle size range 1.18-2.36, 0.5 molar catalytic as 45.2%. The liquid product catalytic and non-catalytic was characterized by GC–MS analyses. In addition, the solid product was analyzed by means of FTIR & SEM.

Keywords: walnut shell, biooil, biochar, microwave pyrolysis

Procedia PDF Downloads 33
5762 Adsorption of Chromium Ions from Aqueous Solution by Carbon Adsorbent

Authors: S. Heydari, H. Sharififard, M. Nabavinia, H. Kiani, M. Parvizi

Abstract:

Rapid industrialization has led to increased disposal of heavy metals into the environment. Activated carbon adsorption has proven to be an effective process for the removal of trace metal contaminants from aqueous media. This paper was investigated chromium adsorption efficiency by commercial activated carbon. The sorption studied as a function of activated carbon particle size, dose of activated carbon and initial pH of solution. Adsorption tests for the effects of these factors were designed with Taguchi approach. According to the Taguchi parameter design methodology, L9 orthogonal array was used. Analysis of experimental results showed that the most influential factor was initial pH of solution. The optimum conditions for chromium adsorption by activated carbons were found to be as follows: Initial feed pH 6, adsorbent particle size 0.412 mm and activated carbon dose 6 g/l. Under these conditions, nearly %100 of chromium ions was adsorbed by activated carbon after 2 hours.

Keywords: chromium, adsorption, Taguchi method, activated carbon

Procedia PDF Downloads 388
5761 Two Component Source Apportionment Based on Absorption and Size Distribution Measurement

Authors: Tibor Ajtai, Noémi Utry, Máté Pintér, Gábor Szabó, Zoltán Bozóki

Abstract:

Beyond its climate and health related issues ambient light absorbing carbonaceous particulate matter (LAC) has also become a great scientific interest in terms of its regulations recently. It has been experimentally demonstrated in recent studies, that LAC is dominantly composed of traffic and wood burning aerosol particularly under wintertime urban conditions, when the photochemical and biological activities are negligible. Several methods have been introduced to quantitatively apportion aerosol fractions emitted by wood burning and traffic but most of them require costly and time consuming off-line chemical analysis. As opposed to chemical features, the microphysical properties of airborne particles such as optical absorption and size distribution can be easily measured on-line, with high accuracy and sensitivity, especially under highly polluted urban conditions. Recently a new method has been proposed for the apportionment of wood burning and traffic aerosols based on the spectral dependence of their absorption quantified by the Aerosol Angström Exponent (AAE). In this approach the absorption coefficient is deduced from transmission measurement on a filter accumulated aerosol sample and the conversion factor between the measured optical absorption and the corresponding mass concentration (the specific absorption cross section) are determined by on-site chemical analysis. The recently developed multi-wavelength photoacoustic instruments provide novel, in-situ approach towards the reliable and quantitative characterization of carbonaceous particulate matter. Therefore, it also opens up novel possibilities on the source apportionment through the measurement of light absorption. In this study, we demonstrate an in-situ spectral characterization method of the ambient carbon fraction based on light absorption and size distribution measurements using our state-of-the-art multi-wavelength photoacoustic instrument (4λ-PAS) and Single Mobility Particle Sizer (SMPS) The carbonaceous particulate selective source apportionment study was performed for ambient particulate matter in the city center of Szeged, Hungary where the dominance of traffic and wood burning aerosol has been experimentally demonstrated earlier. The proposed model is based on the parallel, in-situ measurement of optical absorption and size distribution. AAEff and AAEwb were deduced from the measured data using the defined correlation between the AOC(1064nm)/AOC(266nm) and N100/N20 ratios. σff(λ) and σwb(λ) were determined with the help of the independently measured temporal mass concentrations in the PM1 mode. Furthermore, the proposed optical source apportionment is based on the assumption that the light absorbing fraction of PM is exclusively related to traffic and wood burning. This assumption is indirectly confirmed here by the fact that the measured size distribution is composed of two unimodal size distributions identified to correspond to traffic and wood burning aerosols. The method offers the possibility of replacing laborious chemical analysis with simple in-situ measurement of aerosol size distribution data. The results by the proposed novel optical absorption based source apportionment method prove its applicability whenever measurements are performed at an urban site where traffic and wood burning are the dominant carbonaceous sources of emission.

Keywords: absorption, size distribution, source apportionment, wood burning, traffic aerosol

Procedia PDF Downloads 216
5760 Granule Morphology of Zirconia Powder with Solid Content on Two-Fluid Spray Drying

Authors: Hyeongdo Jeong, Jong Kook Lee

Abstract:

Granule morphology and microstructure were affected by slurry viscosity, chemical composition, particle size and spray drying process. In this study, we investigated granule morphology of zirconia powder with solid content on two-fluid spray drying. Zirconia granules after spray drying show sphere-like shapes with a diameter of 40-70 μm at low solid contents (30 or 40 wt%) and specific surface area of 5.1-5.6 m²/g. But a donut-like shape with a few cracks were observed on zirconia granules prepared from the slurry of high solid content (50 wt %), green compacts after cold isostatic pressing under the pressure of 200 MPa have the density of 2.1-2.2 g/cm³ and homogeneous fracture surface by complete destruction of granules. After the sintering at 1500 °C for 2 h, all specimens have relative density of 96.2-98.3 %. With increasing a solid content from 30 to 50 wt%, grain size increased from 0.3 to 0.6 μm, but relative density was inversely decreased from 98.3 to 96.2 %.

Keywords: zirconia, solid content, granulation, spray drying

Procedia PDF Downloads 205
5759 Listening to Circles, Playing Lights: A Study of Cross-Modal Perception in Music

Authors: Roni Granot, Erica Polini

Abstract:

Music is often described in terms of non-auditory adjectives such as a rising melody, a bright sound, or a zigzagged contour. Such cross modal associations have been studied with simple isolated musical parameters, but only rarely in rich musical contexts. The current study probes cross sensory associations with polarity based dimensions by means of pairings of 10 adjectives: blunt-sharp, relaxed-tense, heavy-light, low (in space)-high, low (pitch)-high, big-small, hard-soft, active-passive, bright-dark, sad-happy. 30 participants (randomly assigned to one of two groups) were asked to rate one of 27 short saxophone improvisations on a 1 to 6 scale where 1 and six correspond to the opposite pole of each dimension. The 27 improvisations included three exemplars for each of three dimensions (size, brightness, sharpness), played by three different players. Here we focus on the question of whether ratings of scales corresponding with the musical dimension were consistently rated as such (e.g. music improvised to represent a white circle rated as bright in contrast with music improvised to represent a dark circle rated as dark). Overall the average scores by dimension showed an upward trend in the equivalent verbal scale, with a low rating for small, bright and sharp musical improvisations and higher scores for large, dark and blunt improvisations. Friedman tests indicate a statistically significant difference for brightness (χ2 (2) = 19.704, p = .000) and sharpness dimensions (χ2 (2) = 15.750, p = .000), but not for size (χ2 (2) = 1.444, p = .486). Post hoc analysis with Wilcoxon signed-rank tests within the brightness dimension, show significant differences among all possible parings resulted in significant differences: the rankings of 'bright' and 'dark' (Z = -3.310, p = .001), of 'bright' and 'medium' (Z = -2.438, p = .015) and of 'dark' and 'medium' music (Z = -2.714, p = .007); but only differences between the extreme contrasts within the sharpness dimension : 'sharp' and 'blunt' music (Z = -3.147, p = .002) and between 'sharp' and 'medium' music rated on the sharpness scale (Z = - 3.054, p = .002), but not between 'medium' and 'blunt' music (Z = -.982, p = .326). In summary our study suggests a privileged link between music and the perceptual and semantic domain of brightness. In contrast, size seems to be very difficult to convey in music, whereas sharpness seems to be mapped onto the two extremes (sharp vs. blunt) rather than continuously. This is nicely reflected in the musical literature in titles and texts which stress the association between music and concepts of light or darkness rather than sharpness or size.

Keywords: audiovisual, brightness, cross-modal perception, cross-sensory correspondences, size, visual angularity

Procedia PDF Downloads 205
5758 Investigation of the Effect of Grid Size on External Store Separation Trajectory Using CFD

Authors: Alaa A. Osman, Amgad M. Bayoumy Aly, Ismail El baialy, Osama E. Abdellatif, Essam E. Khallil

Abstract:

In this paper, a numerical simulation of a finned store separating from a wing-pylon configuration has been studied and validated. A dynamic unstructured tetrahedral mesh approach is accomplished by using three grid sizes to numerically solving the discretized three dimensional, inviscid and compressible Navier-stokes equations. The method used for computations of separation of an external store assuming quasi-steady flow condition. Computations of quasi-steady flow have been directly coupled to a six degree-of-freedom (6DOF) rigid-body motion code to generate store trajectories. The pressure coefficients at four different angular cuts and time histories of various trajectory parameters during the store separation are compared for every grid size with published experimental data.

Keywords: CFD modelling, transonic store separation, quasi-steady flow, moving-body trajectories

Procedia PDF Downloads 379
5757 Effect of Electromagnetic Field on Capacitive Deionization Performance

Authors: Alibi Kilybay, Emad Alhseinat, Ibrahim Mustafa, Abdulfahim Arangadi, Pei Shui, Faisal Almarzooqi

Abstract:

In this work, the electromagnetic field has been used for improving the performance of the capacitive deionization process. The effect of electromagnetic fields on the efficiency of the capacitive deionization (CDI) process was investigated experimentally. The results showed that treating the feed stream of the CDI process using an electromagnetic field can enhance the electrosorption capacity from 20% up to 70%. The effect of the degree of time of exposure, concentration, and type of ions have been examined. The electromagnetic field enhanced the salt adsorption capacity (SAC) of the Ca²⁺ ions by 70%, while the SAC enhanced 20% to the Na⁺ ions. It is hypnotized that the electrometric field affects the hydration shell around the ions and thus reduces their effective size and enhances the mass transfer. This reduction in ion effective size and increase in mass transfer enhanced the electrosorption capacity and kinetics of the CDI process.

Keywords: capacitive deionization, desalination, electromagnetic treatment, water treatment

Procedia PDF Downloads 246
5756 Rapid and Long-term Alien Language Analysis - Forming Frameworks for the Interpretation of Alien Communication for More Intelligent Life

Authors: Samiksha Raviraja, Junaid Arif

Abstract:

One of the most important abilities in species is the ability to communicate. This paper proposes steps to take when and if aliens came in contact with humans, and how humans would communicate with them. The situation would be a time-sensitive scenario, meaning that communication is at the utmost importance if such an event were to happen. First, humans would need to establish mutual peace by conveying that there is no threat to the alien race. Second, the aliens would need to acknowledge this understanding and reciprocate. This would be extremely difficult to do regardless of their intelligence level unless they are very human-like and have similarities to our way of communicating. The first step towards understanding their mind is to analyze their level of intelligence - Level 1-Low intelligence, Level 2-Human-like intelligence or Level 3-Advanced or High Intelligence. These three levels go hand in hand with the Kardashev scale. Further, the Barrow scale will also be used to categorize alien species in hopes of developing a common universal language to communicate in. This paper will delve into how the level of intelligence can be used toward achieving communication with aliens by predicting various possible scenarios and outcomes by proposing an intensive categorization system. This can be achieved by studying their Emotional and Intelligence Quotient (along with technological and scientific knowledge/intelligence). The limitations and capabilities of their intelligence must also be studied. By observing how they respond and react (expressions and senses) to different kinds of scenarios, items and people, the data will help enable good categorisation. It can be hypothesised that the more human-like aliens are or can relate to humans, the more likely it is that communication is possible. Depending on the situation, either human can teach aliens a human language, or humans can learn an alien language, or both races work together to develop a mutual understanding or mode of communication. There are three possible ways of contact. Aliens visit Earth, or humans discover aliens while on space exploration or through technology in the form of signals. A much rarer case would be humans and aliens running into each other during a space expedition of their own. The first two possibilities allow a more in-depth analysis of the alien life and enhanced results compared. The importance of finding a method of talking with aliens is important in order to not only protect Earth and humans but rather for the advancement of Science through the shared knowledge between the two species.

Keywords: intelligence, Kardashev scale, Barrow scale, alien civilizations, emotional and intelligence quotient

Procedia PDF Downloads 57
5755 Continuous Improvement Programme as a Strategy for Technological Innovation in Developing Nations. Nigeria as a Case Study

Authors: Sefiu Adebowale Adewumi

Abstract:

Continuous improvement programme (CIP) adopts an approach to improve organizational performance with small incremental steps over time. In this approach, it is not the size of each step that is important, but the likelihood that the improvements will be ongoing. Many companies in developing nations are now complementing continuous improvement with innovation, which is the successful exploitation of new ideas. Focus area of CIP in the organization was in relation to the size of the organizations and also in relation to the generic classification of these organizations. Product quality was prevalent in the manufacturing industry while manpower training and retraining and marketing strategy were emphasized for improvement to be made in the service, transport and supply industries. However, focus on innovation in raw materials, process and methods are needed because these are the critical factors that influence product quality in the manufacturing industries.

Keywords: continuous improvement programme, developing countries, generic classfications, technological innovation

Procedia PDF Downloads 172
5754 Tumor Size and Lymph Node Metastasis Detection in Colon Cancer Patients Using MR Images

Authors: Mohammadreza Hedyehzadeh, Mahdi Yousefi

Abstract:

Colon cancer is one of the most common cancer, which predicted to increase its prevalence due to the bad eating habits of peoples. Nowadays, due to the busyness of people, the use of fast foods is increasing, and therefore, diagnosis of this disease and its treatment are of particular importance. To determine the best treatment approach for each specific colon cancer patients, the oncologist should be known the stage of the tumor. The most common method to determine the tumor stage is TNM staging system. In this system, M indicates the presence of metastasis, N indicates the extent of spread to the lymph nodes, and T indicates the size of the tumor. It is clear that in order to determine all three of these parameters, an imaging method must be used, and the gold standard imaging protocols for this purpose are CT and PET/CT. In CT imaging, due to the use of X-rays, the risk of cancer and the absorbed dose of the patient is high, while in the PET/CT method, there is a lack of access to the device due to its high cost. Therefore, in this study, we aimed to estimate the tumor size and the extent of its spread to the lymph nodes using MR images. More than 1300 MR images collected from the TCIA portal, and in the first step (pre-processing), histogram equalization to improve image qualities and resizing to get the same image size was done. Two expert radiologists, which work more than 21 years on colon cancer cases, segmented the images and extracted the tumor region from the images. The next step is feature extraction from segmented images and then classify the data into three classes: T0N0، T3N1 و T3N2. In this article, the VGG-16 convolutional neural network has been used to perform both of the above-mentioned tasks, i.e., feature extraction and classification. This network has 13 convolution layers for feature extraction and three fully connected layers with the softmax activation function for classification. In order to validate the proposed method, the 10-fold cross validation method used in such a way that the data was randomly divided into three parts: training (70% of data), validation (10% of data) and the rest for testing. It is repeated 10 times, each time, the accuracy, sensitivity and specificity of the model are calculated and the average of ten repetitions is reported as the result. The accuracy, specificity and sensitivity of the proposed method for testing dataset was 89/09%, 95/8% and 96/4%. Compared to previous studies, using a safe imaging technique (MRI) and non-use of predefined hand-crafted imaging features to determine the stage of colon cancer patients are some of the study advantages.

Keywords: colon cancer, VGG-16, magnetic resonance imaging, tumor size, lymph node metastasis

Procedia PDF Downloads 45
5753 A Cost Effective Approach to Develop Mid-Size Enterprise Software Adopted the Waterfall Model

Authors: Mohammad Nehal Hasnine, Md Kamrul Hasan Chayon, Md Mobasswer Rahman

Abstract:

Organizational tendencies towards computer-based information processing have been observed noticeably in the third-world countries. Many enterprises are taking major initiatives towards computerized working environment because of massive benefits of computer-based information processing. However, designing and developing information resource management software for small and mid-size enterprises under budget costs and strict deadline is always challenging for software engineers. Therefore, we introduced an approach to design mid-size enterprise software by using the Waterfall model, which is one of the SDLC (Software Development Life Cycles), in a cost effective way. To fulfill research objectives, in this study, we developed mid-sized enterprise software named “BSK Management System” that assists enterprise software clients with information resource management and perform complex organizational tasks. Waterfall model phases have been applied to ensure that all functions, user requirements, strategic goals, and objectives are met. In addition, Rich Picture, Structured English, and Data Dictionary have been implemented and investigated properly in engineering manner. Furthermore, an assessment survey with 20 participants has been conducted to investigate the usability and performance of the proposed software. The survey results indicated that our system featured simple interfaces, easy operation and maintenance, quick processing, and reliable and accurate transactions.

Keywords: end-user application development, enterprise software design, information resource management, usability

Procedia PDF Downloads 424
5752 Separating Landform from Noise in High-Resolution Digital Elevation Models through Scale-Adaptive Window-Based Regression

Authors: Anne M. Denton, Rahul Gomes, David W. Franzen

Abstract:

High-resolution elevation data are becoming increasingly available, but typical approaches for computing topographic features, like slope and curvature, still assume small sliding windows, for example, of size 3x3. That means that the digital elevation model (DEM) has to be resampled to the scale of the landform features that are of interest. Any higher resolution is lost in this resampling. When the topographic features are computed through regression that is performed at the resolution of the original data, the accuracy can be much higher, and the reported result can be adjusted to the length scale that is relevant locally. Slope and variance are calculated for overlapping windows, meaning that one regression result is computed per raster point. The number of window centers per area is the same for the output as for the original DEM. Slope and variance are computed by performing regression on the points in the surrounding window. Such an approach is computationally feasible because of the additive nature of regression parameters and variance. Any doubling of window size in each direction only takes a single pass over the data, corresponding to a logarithmic scaling of the resulting algorithm as a function of the window size. Slope and variance are stored for each aggregation step, allowing the reported slope to be selected to minimize variance. The approach thereby adjusts the effective window size to the landform features that are characteristic to the area within the DEM. Starting with a window size of 2x2, each iteration aggregates 2x2 non-overlapping windows from the previous iteration. Regression results are stored for each iteration, and the slope at minimal variance is reported in the final result. As such, the reported slope is adjusted to the length scale that is characteristic of the landform locally. The length scale itself and the variance at that length scale are also visualized to aid in interpreting the results for slope. The relevant length scale is taken to be half of the window size of the window over which the minimum variance was achieved. The resulting process was evaluated for 1-meter DEM data and for artificial data that was constructed to have defined length scales and added noise. A comparison with ESRI ArcMap was performed and showed the potential of the proposed algorithm. The resolution of the resulting output is much higher and the slope and aspect much less affected by noise. Additionally, the algorithm adjusts to the scale of interest within the region of the image. These benefits are gained without additional computational cost in comparison with resampling the DEM and computing the slope over 3x3 images in ESRI ArcMap for each resolution. In summary, the proposed approach extracts slope and aspect of DEMs at the lengths scales that are characteristic locally. The result is of higher resolution and less affected by noise than existing techniques.

Keywords: high resolution digital elevation models, multi-scale analysis, slope calculation, window-based regression

Procedia PDF Downloads 119