Search results for: drag coefficients of banner
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 1121

Search results for: drag coefficients of banner

191 Experimental Investigation of Nucleate Pool Boiling Heat Transfer on Laser-Structured Copper Surfaces of Different Patterns

Authors: Luvindran Sugumaran, Mohd Nashrul Mohd Zubir, Kazi Md Salim Newaz, Tuan Zaharinie Tuan Zahari, Suazlan Mt Aznam, Aiman Mohd Halil

Abstract:

With reference to Energy Roadmap 2050, the minimization of greenhouse gas emissions and the enhancement of energy efficiency are the two key factors that could facilitate a radical change in the world's energy infrastructure. However, the energy demands of electronic devices skyrocketed with the advent of the digital age. Currently, the two-phase cooling technique based on phase change pool boiling heat transfer has received a lot of attention because of its potential to fully utilize the latent heat of the fluid and produce a highly effective heat dissipation capacity while keeping the equipment's operating temperature within an acceptable range. There are numerous strategies available for the alteration of heating surfaces, but finding the best, simplest, and most dependable one remains a challenge. Lately, surface texturing via laser ablation has been used in a variety of investigations, demonstrating its significant potential for enhancing the pool boiling heat transfer performance. In this research, the nucleate pool boiling heat transfer performance of laser-structured copper surfaces of different patterns was investigated. The bare copper surface serves as a reference to compare the performance of laser-structured surfaces. It was observed that the heat transfer coefficients were increased with the increase of surface area ratio and the ratio of the peak-to-valley height of the microstructure. Laser machined grain structure produced extra nucleation sites, which ultimately caused the improved pool boiling performance. Due to an increase in nucleation site density and surface area, the enhanced nucleate boiling served as the primary heat transfer mechanism. The pool boiling performance of the laser-structured copper surfaces is superior to the bare copper surface in all aspects.

Keywords: heat transfer coefficient, laser structuring, micro structured surface, pool boiling

Procedia PDF Downloads 47
190 Experimental Investigation of Nucleate Pool Boiling Heat Transfer on Laser-Structured Copper Surfaces of Different Patterns

Authors: Luvindran Sugumaran, Mohd Nashrul Mohd Zubir, Kazi Md Salim Newaz, Tuan Zaharinie Tuan Zahari, Suazlan Mt Aznam, Aiman Mohd Halil

Abstract:

With reference to Energy Roadmap 2050, the minimization of greenhouse gas emissions, and the enhancement of energy efficiency are the two key factors that could facilitate a radical change in the world's energy infrastructure. However, the energy demands of electronic devices skyrocketed with the advent of the digital age. Currently, the two-phase cooling technique based on phase change pool boiling heat transfer has received a lot of attention because of its potential to fully utilize the latent heat of the fluid and produce a highly effective heat dissipation capacity while keeping the equipment's operating temperature within an acceptable range. There are numerous strategies available for the alteration of heating surfaces, but to find the best, simplest, and most dependable one remains a challenge. Lately, surface texturing via laser ablation has been used in a variety of investigations, demonstrating its significant potential for enhancing the pool boiling heat transfer performance. In this research, the nucleate pool boiling heat transfer performance of laser-structured copper surfaces of different patterns was investigated. The bare copper surface serves as a reference to compare the performance of laser-structured surfaces. It was observed that the heat transfer coefficients were increased with the increase of surface area ratio and the ratio of the peak-to-valley height of the microstructure. Laser machined grain structure produced extra nucleation sites, which ultimately caused the improved pool boiling performance. Due to an increase in nucleation site density and surface area, the enhanced nucleate boiling served as the primary heat transfer mechanism. The pool boiling performance of the laser-structured copper surfaces is superior to the bare copper surface in all aspects.

Keywords: heat transfer coefficient, laser structuring, micro structured surface, pool boiling

Procedia PDF Downloads 50
189 Experimental Investigation of Nucleate Pool Boiling Heat Transfer on Laser-Structured Copper Surfaces of Different Patterns

Authors: Luvindran Sugumaran, Mohd Nashrul Mohd Zubir, Kazi Md. Salim Newaz, Tuan Zaharinie Tuan Zahari, Suazlan Mt Aznam, Aiman Mohd Halil

Abstract:

With reference to Energy Roadmap 2050, the minimization of greenhouse gas emissions and the enhancement of energy efficiency are the two key factors that could facilitate a radical change in the world's energy infrastructure. However, the energy demands of electronic devices skyrocketed with the advent of the digital age. Currently, the two-phase cooling technique based on phase change pool boiling heat transfer has received a lot of attention because of its potential to fully utilize the latent heat of the fluid and produce a highly effective heat dissipation capacity while keeping the equipment's operating temperature within an acceptable range. There are numerous strategies available for the alteration of heating surfaces, but to find the best, simplest, and most dependable one remains a challenge. Lately, surface texturing via laser ablation has been used in a variety of investigations, demonstrating its significant potential for enhancing the pool boiling heat transfer performance. In this research, the nucleate pool boiling heat transfer performance of laser-structured copper surfaces of different patterns was investigated. The bare copper surface serves as a reference to compare the performance of laser-structured surfaces. It was observed that the heat transfer coefficients were increased with the increase of surface area ratio and the ratio of the peak-to-valley height of the microstructure. Laser-machined grain structure produced extra nucleation sites, which ultimately caused the improved pool boiling performance. Due to an increase in nucleation site density and surface area, the enhanced nucleate boiling served as the primary heat transfer mechanism. The pool boiling performance of the laser-structured copper surfaces is superior to the bare copper surface in all aspects.

Keywords: heat transfer coefficient, laser structuring, micro structured surface, pool boiling

Procedia PDF Downloads 55
188 Environmental and Socioeconomic Determinants of Climate Change Resilience in Rural Nigeria: Empirical Evidence towards Resilience Building

Authors: Ignatius Madu

Abstract:

The study aims at assessing the environmental and socioeconomic determinants of climate change resilience in rural Nigeria. This is necessary because researches and development efforts on building climate change resilience of rural areas in developing countries are usually made without the knowledge of the impacts of the inherent rural characteristics that determine resilient capacities of the households. This has, in many cases, led to costly mistakes, delayed responses, inaccurate outcomes, and other difficulties. Consequently, this assessment becomes crucial not only to policymakers and people living in risk-prone environments in rural areas but also to fill the research gap. To achieve the aim, secondary data were obtained from the Annual Abstract of Statistics 2017, LSMS-Integrated Surveys on Agriculture and General Household Survey Panel 2015/2016, and National Agriculture Sample Survey (NASS), 2010/2011.Resilience was calculated by weighting and adding the adaptive, absorptive and anticipatory measures of households variables aggregated at state levels and then regressed against rural environmental and socioeconomic characteristics influencing it. From the regression, the coefficients of the variables were used to compute the impacts of the variables using the Stochastic Regression of Impacts on Population, Affluence and Technology (STIRPAT) Model. The results showed that the northern States are generally low in resilient indices and are impacted less by the development indicators. The major determining factors are percentage of non-poor, environmental protection, road transport development, landholding, agricultural input, population density, dependency ratio (inverse), household asserts, education and maternal care. The paper concludes that any effort to a successful resilient building in rural areas of the country should first address these key factors that enhance rural development and wellbeing since it is better to take action before shocks take place.

Keywords: climate change resilience; spatial impacts; STIRPAT model; Nigeria

Procedia PDF Downloads 122
187 Validation Study of Radial Aircraft Engine Model

Authors: Lukasz Grabowski, Tytus Tulwin, Michal Geca, P. Karpinski

Abstract:

This paper presents the radial aircraft engine model which has been created in AVL Boost software. This model is a one-dimensional physical model of the engine, which enables us to investigate the impact of an ignition system design on engine performance (power, torque, fuel consumption). In addition, this model allows research under variable environmental conditions to reflect varied flight conditions (altitude, humidity, cruising speed). Before the simulation research the identifying parameters and validating of model were studied. In order to verify the feasibility to take off power of gasoline radial aircraft engine model, some validation study was carried out. The first stage of the identification was completed with reference to the technical documentation provided by manufacturer of engine and the experiments on the test stand of the real engine. The second stage involved a comparison of simulation results with the results of the engine stand tests performed on a WSK ’PZL-Kalisz’. The engine was loaded by a propeller in a special test bench. Identifying the model parameters referred to a comparison of the test results to the simulation in terms of: pressure behind the throttles, pressure in the inlet pipe, and time course for pressure in the first inlet pipe, power, and specific fuel consumption. Accordingly, the required coefficients and error of simulation calculation relative to the real-object experiments were determined. Obtained the time course for pressure and its value is compatible with the experimental results. Additionally the engine power and specific fuel consumption tends to be significantly compatible with the bench tests. The mapping error does not exceed 1.5%, which verifies positively the model of combustion and allows us to predict engine performance if the process of combustion will be modified. The next conducted tests verified completely model. The maximum mapping error for the pressure behind the throttles and the inlet pipe pressure is 4 %, which proves the model of the inlet duct in the engine with the charging compressor to be correct.

Keywords: 1D-model, aircraft engine, performance, validation

Procedia PDF Downloads 309
186 Dynamic Web-Based 2D Medical Image Visualization and Processing Software

Authors: Abdelhalim. N. Mohammed, Mohammed. Y. Esmail

Abstract:

In the course of recent decades, medical imaging has been dominated by the use of costly film media for review and archival of medical investigation, however due to developments in networks technologies and common acceptance of a standard digital imaging and communication in medicine (DICOM) another approach in light of World Wide Web was produced. Web technologies successfully used in telemedicine applications, the combination of web technologies together with DICOM used to design a web-based and open source DICOM viewer. The Web server allowance to inquiry and recovery of images and the images viewed/manipulated inside a Web browser without need for any preinstalling software. The dynamic site page for medical images visualization and processing created by using JavaScript and HTML5 advancements. The XAMPP ‘apache server’ is used to create a local web server for testing and deployment of the dynamic site. The web-based viewer connected to multiples devices through local area network (LAN) to distribute the images inside healthcare facilities. The system offers a few focal points over ordinary picture archiving and communication systems (PACS): easy to introduce, maintain and independently platforms that allow images to display and manipulated efficiently, the system also user-friendly and easy to integrate with an existing system that have already been making use of web technologies. The wavelet-based image compression technique on which 2-D discrete wavelet transform used to decompose the image then wavelet coefficients are transmitted by entropy encoding after threshold to decrease transmission time, stockpiling cost and capacity. The performance of compression was estimated by using images quality metrics such as mean square error ‘MSE’, peak signal to noise ratio ‘PSNR’ and compression ratio ‘CR’ that achieved (83.86%) when ‘coif3’ wavelet filter is used.

Keywords: DICOM, discrete wavelet transform, PACS, HIS, LAN

Procedia PDF Downloads 138
185 Comparison of Bismuth-Based Nanoparticles as Radiosensitization Agents for Radiotherapy

Authors: Merfat Algethami, Anton Blencowe, Bryce Feltis, Stephen Best, Moshi Geso

Abstract:

Nano-materials with high atomic number atoms have been demonstrated to enhance the effective radiation dose and thus potentially could improve therapeutic efficacy in radiotherapy. The optimal nanoparticulate agents require high X-ray absorption coefficients, low toxicity, and should be cost effective. The focus of our research is the development of a nanoparticle therapeutic agent that can be used in radiotherapy to provide optimal enhancement of the radiation effects on the target. In this study, we used bismuth (Bi) nanoparticles coated with starch and bismuth sulphide nanoparticles (Bi2S3) coated with polyvinylpyrrolidone (PVP). These NPs are of low toxicity and are one of the least expensive heavy metal-based nanoparticles. The aims of this study were to synthesise Bi2S3 and Bi NPs, and examine their cytotoxicity to human lung adenocarcinoma epithelial cells (A549). The dose enhancing effects of NPs on A549 cells were examined at both KV and MV energies. The preliminary results revealed that bismuth based nanoparticles show increased radio-sensitisation of cells, displaying dose enhancement with KV X-ray energies and to a lesser degree for the MV energies. We also observed that Bi NPs generated a greater dose enhancement effect than Bi2S3 NPs in irradiated A549 cells. The maximum Dose Enhancement Factor (DEF) was obtained at lower energy KV range when cells treated with Bi NPs (1.5) compared to the DEF of 1.2 when cells treated with Bi2S3NPs. Less radiation dose enhancement was observed when using high energy MV beam with higher DEF value of Bi NPs treatment (1.26) as compared to 1.06 DEF value with Bi2S3 NPs. The greater dose enhancement was achieved at KV energy range, due the effect of the photoelectric effect which is the dominant process of interaction of X-ray. The cytotoxic effect of Bi NPs on enhancing the X-ray dose was higher due to the higher amount of elemental Bismuth present in Bi NPs compared to Bi2S3 NPs. The results suggest that Bismuth based NPs can be considered as valuable dose enhancing agents when used in clinical applications.

Keywords: A549 lung cancer cells, Bi2S3 nanoparticles, dose enhancement effect, radio-sensitising agents

Procedia PDF Downloads 243
184 Exploring Regularity Results in the Context of Extremely Degenerate Elliptic Equations

Authors: Zahid Ullah, Atlas Khan

Abstract:

This research endeavors to explore the regularity properties associated with a specific class of equations, namely extremely degenerate elliptic equations. These equations hold significance in understanding complex physical systems like porous media flow, with applications spanning various branches of mathematics. The focus is on unraveling and analyzing regularity results to gain insights into the smoothness of solutions for these highly degenerate equations. Elliptic equations, fundamental in expressing and understanding diverse physical phenomena through partial differential equations (PDEs), are particularly adept at modeling steady-state and equilibrium behaviors. However, within the realm of elliptic equations, the subset of extremely degenerate cases presents a level of complexity that challenges traditional analytical methods, necessitating a deeper exploration of mathematical theory. While elliptic equations are celebrated for their versatility in capturing smooth and continuous behaviors across different disciplines, the introduction of degeneracy adds a layer of intricacy. Extremely degenerate elliptic equations are characterized by coefficients approaching singular behavior, posing non-trivial challenges in establishing classical solutions. Still, the exploration of extremely degenerate cases remains uncharted territory, requiring a profound understanding of mathematical structures and their implications. The motivation behind this research lies in addressing gaps in the current understanding of regularity properties within solutions to extremely degenerate elliptic equations. The study of extreme degeneracy is prompted by its prevalence in real-world applications, where physical phenomena often exhibit characteristics defying conventional mathematical modeling. Whether examining porous media flow or highly anisotropic materials, comprehending the regularity of solutions becomes crucial. Through this research, the aim is to contribute not only to the theoretical foundations of mathematics but also to the practical applicability of mathematical models in diverse scientific fields.

Keywords: elliptic equations, extremely degenerate, regularity results, partial differential equations, mathematical modeling, porous media flow

Procedia PDF Downloads 28
183 Ill-Posed Inverse Problems in Molecular Imaging

Authors: Ranadhir Roy

Abstract:

Inverse problems arise in medical (molecular) imaging. These problems are characterized by large in three dimensions, and by the diffusion equation which models the physical phenomena within the media. The inverse problems are posed as a nonlinear optimization where the unknown parameters are found by minimizing the difference between the predicted data and the measured data. To obtain a unique and stable solution to an ill-posed inverse problem, a priori information must be used. Mathematical conditions to obtain stable solutions are established in Tikhonov’s regularization method, where the a priori information is introduced via a stabilizing functional, which may be designed to incorporate some relevant information of an inverse problem. Effective determination of the Tikhonov regularization parameter requires knowledge of the true solution, or in the case of optical imaging, the true image. Yet, in, clinically-based imaging, true image is not known. To alleviate these difficulties we have applied the penalty/modified barrier function (PMBF) method instead of Tikhonov regularization technique to make the inverse problems well-posed. Unlike the Tikhonov regularization method, the constrained optimization technique, which is based on simple bounds of the optical parameter properties of the tissue, can easily be implemented in the PMBF method. Imposing the constraints on the optical properties of the tissue explicitly restricts solution sets and can restore uniqueness. Like the Tikhonov regularization method, the PMBF method limits the size of the condition number of the Hessian matrix of the given objective function. The accuracy and the rapid convergence of the PMBF method require a good initial guess of the Lagrange multipliers. To obtain the initial guess of the multipliers, we use a least square unconstrained minimization problem. Three-dimensional images of fluorescence absorption coefficients and lifetimes were reconstructed from contact and noncontact experimentally measured data.

Keywords: constrained minimization, ill-conditioned inverse problems, Tikhonov regularization method, penalty modified barrier function method

Procedia PDF Downloads 248
182 Diagnostic Clinical Skills in Cardiology: Improving Learning and Performance with Hybrid Simulation, Scripted Histories, Wearable Technology, and Quantitative Grading – The Assimilate Excellence Study

Authors: Daly M. J, Condron C, Mulhall C, Eppich W, O'Neill J.

Abstract:

Introduction: In contemporary clinical cardiology, comprehensive and holistic bedside evaluation including accurate cardiac auscultation is in decline despite having positive effects on patients and their outcomes. Methods: Scripted histories and scoring checklists for three clinical scenarios in cardiology were co-created and refined through iterative consensus by a panel of clinical experts; these were then paired with recordings of auscultatory findings from three actual patients with known valvular heart disease. A wearable vest with embedded pressure-sensitive panel speakers was developed to transmit these recordings when examined at the standard auscultation points. RCSI medical students volunteered for a series of three formative long case examinations in cardiology (LC1 – LC3) using this hybrid simulation. Participants were randomised into two groups: Group 1 received individual teaching from an expert trainer between LC1 and LC2; Group 2 received the same intervention between LC2 and LC3. Each participant’s long case examination performance was recorded and blindly scored by two peer participants and two RCSI examiners. Results: Sixty-eight participants were included in the study (age 27.6 ± 0.1 years; 74% female) and randomised into two groups; there were no significant differences in baseline characteristics between groups. Overall, the median total faculty examiner score was 39.8% (35.8 – 44.6%) in LC1 and increased to 63.3% (56.9 – 66.4%) in LC3, with those in Group 1 showing a greater improvement in LC2 total score than that observed in Group 2 (p < .001). Using the novel checklist, intraclass correlation coefficients (ICC) were excellent between examiners in all cases: ICC .994 – .997 (p < .001); correlation between peers and examiners improved in LC2 following peer grading of LC1 performances: ICC .857 – .867 (p < .001). Conclusion: Hybrid simulation and quantitative grading improve learning, standardisation of assessment, and direct comparisons of both performance and acumen in clinical cardiology.

Keywords: cardiology, clinical skills, long case examination, hybrid simulation, checklist

Procedia PDF Downloads 84
181 Robustness of the Deep Chroma Extractor and Locally-Normalized Quarter Tone Filters in Automatic Chord Estimation under Reverberant Conditions

Authors: Luis Alvarado, Victor Poblete, Isaac Gonzalez, Yetzabeth Gonzalez

Abstract:

In MIREX 2016 (http://www.music-ir.org/mirex), the deep neural network (DNN)-Deep Chroma Extractor, proposed by Korzeniowski and Wiedmer, reached the highest score in an audio chord recognition task. In the present paper, this tool is assessed under acoustic reverberant environments and distinct source-microphone distances. The evaluation dataset comprises The Beatles and Queen datasets. These datasets are sequentially re-recorded with a single microphone in a real reverberant chamber at four reverberation times (0 -anechoic-, 1, 2, and 3 s, approximately), as well as four source-microphone distances (32, 64, 128, and 256 cm). It is expected that the performance of the trained DNN will dramatically decrease under these acoustic conditions with signals degraded by room reverberation and distance to the source. Recently, the effect of the bio-inspired Locally-Normalized Cepstral Coefficients (LNCC), has been assessed in a text independent speaker verification task using speech signals degraded by additive noise at different signal-to-noise ratios with variations of recording distance, and it has also been assessed under reverberant conditions with variations of recording distance. LNCC showed a performance so high as the state-of-the-art Mel Frequency Cepstral Coefficient filters. Based on these results, this paper proposes a variation of locally-normalized triangular filters called Locally-Normalized Quarter Tone (LNQT) filters. By using the LNQT spectrogram, robustness improvements of the trained Deep Chroma Extractor are expected, compared with classical triangular filters, and thus compensating the music signal degradation improving the accuracy of the chord recognition system.

Keywords: chord recognition, deep neural networks, feature extraction, music information retrieval

Procedia PDF Downloads 196
180 Developing a Framework for Open Source Software Adoption in a Higher Education Institution in Uganda. A case of Kyambogo University

Authors: Kafeero Frank

Abstract:

This study aimed at developing a frame work for open source software adoption in an institution of higher learning in Uganda, with the case of KIU as a study area. There were mainly four research questions based on; individual staff interaction with open source software forum, perceived FOSS characteristics, organizational characteristics and external characteristics as factors that affect open source software adoption. The researcher used causal-correlation research design to study effects of these variables on open source software adoption. A quantitative approach was used in this study with self-administered questionnaire on a purposively and randomly sampled sample of university ICT staff. Resultant data was analyzed using means, correlation coefficients and multivariate multiple regression analysis as statistical tools. The study reveals that individual staff interaction with open source software forum and perceived FOSS characteristics were the primary factors that significantly affect FOSS adoption while organizational and external factors were secondary with no significant effect but significant correlation to open source software adoption. It was concluded that for effective open source software adoption to occur there must be more effort on primary factors with subsequent reinforcement of secondary factors to fulfill the primary factors and adoption of open source software. Lastly recommendations were made in line with conclusions for coming up with Kyambogo University frame work for open source software adoption in institutions of higher learning. Areas of further research recommended include; Stakeholders’ analysis of open source software adoption in Uganda; Challenges and way forward. Evaluation of Kyambogo University frame work for open source software adoption in institutions of higher learning. Framework development for cloud computing adoption in Ugandan universities. Framework for FOSS development in Uganda IT industry

Keywords: open source software., organisational characteristics, external characteristics, cloud computing adoption

Procedia PDF Downloads 39
179 Working Memory Capacity and Motivation in Japanese English as a Foreign Language Learners' Speaking Skills

Authors: Akiko Kondo

Abstract:

Although the effects of working memory capacity on second/foreign language speaking skills have been researched in depth, few studies have focused on Japanese English as a foreign language (EFL) learners as compared to other languages (Indo-European languages), and the sample sizes of the relevant Japanese studies have been relatively small. Furthermore, comparing the effects of working memory capacity and motivation which is another kind of frequently researched individual factor on L2 speaking skills would add to the scholarly literature in the field of second language acquisition research. Therefore, the purposes of this study were to investigate whether working memory capacity and motivation have significant relationships with Japanese EFL learners’ speaking skills and to investigate the degree to which working memory capacity and motivation contribute to their English speaking skills. One-hundred and ten Japanese EFL students aged 18 to 26 years participated in this study. All of them are native Japanese speakers and have learned English as s foreign language for 6 to 15. They completed the Versant English speaking test, which has been widely used to measure non-native speakers’ English speaking skills, two types of working memory tests (the L1-based backward digit span test and the L1-based listening span test), and the language learning motivation survey. The researcher designed the working memory tests and the motivation survey. To investigate the relationship between the variables (English speaking skills, working memory capacity, and language learning motivation), a correlation analysis was conducted, which showed that L2 speaking test scores were significantly related to both working memory capacity and language learning motivation, although the correlation coefficients were weak. Furthermore, a multiple regression analysis was performed, with L2 speaking skills as the dependent variable and working memory capacity and language learning motivation as the independent variables. The results showed that working memory capacity and motivation significantly explained the variance in L2 speaking skills and that the L2 motivation had slightly larger effects on the L2 speaking skills than the working memory capacity. Although this study includes several limitations, the results could contribute to the generalization of the effects of individual differences, such as working memory and motivation on L2 learning, in the literature.

Keywords: individual differences, motivation, speaking skills, working memory

Procedia PDF Downloads 140
178 Factors Contributing to Farmers’ Attitude Towards Climate Adaptation Farming Practices: A Farm Level Study in Bangladesh

Authors: Md Rezaul Karim, Farha Taznin

Abstract:

The purpose of this study was to assess and describe the individual and household characteristics of farmers, to measure the attitude of farmers towards climate adaptation farming practices and to explore the individual and household factors contributing in predicting their attitude towards climate adaptation farming practices. Data were collected through personal interviews using a pre-tested interview schedule. The data collection was done at Biral Upazila under Dinajpur district in Bangladesh from 1st November to 15 December 2018. Besides descriptive statistical parameters, Pearson’s Product Moment Correlation Coefficient (r), multiple regression and step-wise multiple regression analysis were used for the statistical analysis. Findings indicated that the highest proportion (77.6 percent) of the farmers had moderately favorable attitudes, followed by only 11.2 percent with highly favorable attitudes and 11.2 percent with slightly favorable attitudes towards climate adaptation farming practices. According to the computed correlation coefficients (r), among the 10 selected factors, five of them, such as education of household head, farm size, annual household income, organizational participation, and information access by extension services, had a significant relationship with the attitude of farmers towards climate-smart practices. The step-wise multiple regression results showed that two characteristics as education of household head and information access by extension services, contributed 26.2% and 5.1%, respectively, in predicting farmers' attitudes towards climate adaptation farming practices. In addition, more than two-thirds of farmers cited their opinion to the problems in response to ‘price of vermi species is high and it is not easily available’ as 1st ranked problem, followed by ‘lack of information for innovative climate-smart technologies’. This study suggests that policy implications are necessary to promote extension education and information services and overcome the obstacles to climate adaptation farming practices. It further recommends that research study should be conducted in diverse contexts of nationally or globally.

Keywords: factors, attitude, climate adaptation, farming practices, Bangladesh

Procedia PDF Downloads 57
177 Discerning Divergent Nodes in Social Networks

Authors: Mehran Asadi, Afrand Agah

Abstract:

In data mining, partitioning is used as a fundamental tool for classification. With the help of partitioning, we study the structure of data, which allows us to envision decision rules, which can be applied to classification trees. In this research, we used online social network dataset and all of its attributes (e.g., Node features, labels, etc.) to determine what constitutes an above average chance of being a divergent node. We used the R statistical computing language to conduct the analyses in this report. The data were found on the UC Irvine Machine Learning Repository. This research introduces the basic concepts of classification in online social networks. In this work, we utilize overfitting and describe different approaches for evaluation and performance comparison of different classification methods. In classification, the main objective is to categorize different items and assign them into different groups based on their properties and similarities. In data mining, recursive partitioning is being utilized to probe the structure of a data set, which allow us to envision decision rules and apply them to classify data into several groups. Estimating densities is hard, especially in high dimensions, with limited data. Of course, we do not know the densities, but we could estimate them using classical techniques. First, we calculated the correlation matrix of the dataset to see if any predictors are highly correlated with one another. By calculating the correlation coefficients for the predictor variables, we see that density is strongly correlated with transitivity. We initialized a data frame to easily compare the quality of the result classification methods and utilized decision trees (with k-fold cross validation to prune the tree). The method performed on this dataset is decision trees. Decision tree is a non-parametric classification method, which uses a set of rules to predict that each observation belongs to the most commonly occurring class label of the training data. Our method aggregates many decision trees to create an optimized model that is not susceptible to overfitting. When using a decision tree, however, it is important to use cross-validation to prune the tree in order to narrow it down to the most important variables.

Keywords: online social networks, data mining, social cloud computing, interaction and collaboration

Procedia PDF Downloads 119
176 Synthesis and Characterization of PH Sensitive Hydrogel and Its Application in Controlled Drug Release of Tramadol

Authors: Naima Bouslah, Leila Bounabi, Farid Ouazib, Nabila Haddadine

Abstract:

Conventional release dosage forms are known to provide an immediate release of the drug. Controlling the rate of drug release from polymeric matrices is very important for a number of applications, particularly in the pharmaceutical area. Hydrogels are polymers in three-dimensional network arrangement, which can absorb and retain large amounts of water without dissolution. They have been frequently used to develop controlled released formulations for oral administration because they can extend the duration of drug release and thus reduce dose to be administrated improving patient compliance. Tramadol is an opioid pain medication used to treat moderate to moderately severe pain. When taken as an immediate-release oral formulation, the onset of pain relief usually occurs within about an hour. In the present work, we synthesized pH-responsive hydrogels of (hydroxyl ethyl methacrylate-co-acrylic acid), (HEMA-AA) for control drug delivery of tramadol in the gastro-intestinal tractus. The hydrogels with different acrylic acid content, were synthesized by free radical polymerization and characterized by FTIR spectroscopy, X ray diffraction analysis (XRD), differential scanning calorimetry (DSC) and thermo gravimetric analysis (TGA). FTIR spectroscopy has shown specific hydrogen bonding interactions between the carbonyl groups of the hydrogels and hydroxyl groups of tramadol. Both the XRD and DSC studies revealed that the introduction of tramadol in the hydrogel network induced the amorphization of the drug. The swelling behaviour, absorptive kinetics and the release kinetics of tramadol in simulated gastric fluid (pH 1.2) and in simulated intestinal fluid (pH 7.4) were also investigated. The hydrogels exhibited pH-responsive behavior in the swelling study. The (HEMA-AA) hydrogel swelling was much higher in pH =7.4 medium. The tramadol release was significantly increased when pH of the medium was changed from simulated gastric fluid (pH 1.2) to simulated intestinal fluid (pH 7.4). Using suitable mathematical models, the apparent diffusional coefficients and the corresponding kinetic parameters have been calculated.

Keywords: biopolymres, drug delivery, hydrogels, tramadol

Procedia PDF Downloads 332
175 Thermal Stability and Electrical Conductivity of Ca₅Mg₄₋ₓMₓ(VO₄)₆ (0 ≤ x ≤ 4) where M = Zn, Ni Measured by Impedance Spectroscopy

Authors: Anna S. Tolkacheva, Sergey N. Shkerin, Kirill G. Zemlyanoi, Olga G. Reznitskikh, Pavel D. Khavlyuk

Abstract:

Calcium oxovanadates with garnet related structure are multifunctional oxides in various fields like photoluminescence, microwave dielectrics, and magneto-dielectrics. For example, vanadate garnets are self-luminescent compounds. They attract attention as RE-free broadband excitation and emission phosphors and are candidate materials for UV-based white light-emitting diodes (WLEDs). Ca₅M₄(VO₄)₆ (M = Mg, Zn, Co, Ni, Mn) compounds are also considered promising for application in microwave devices as substrate materials. However, the relation between their structure, composition and physical/chemical properties remains unclear. Given the above-listed observations, goals of this study are to synthesise Ca₅M₄(VO₄)₆ (M = Mg, Zn, Ni) and to study their thermal and electrical properties. Solid solutions Ca₅Mg₄₋ₓMₓ(VO₄)₆ (0 ≤ x ≤ 4) where M is Zn and Ni have been synthesized by sol-gel method. The single-phase character of the final products was checked by powder X-ray diffraction on a Rigaku D/MAX-2200 X-ray diffractometer using Cu Kα radiation in the 2θ range from 15° to 70°. The dependence of thermal properties on chemical composition of solid solutions was studied using simultaneous thermal analyses (DSC and TG). Thermal analyses were conducted in a Netzch simultaneous analyser STA 449C Jupiter, in Ar atmosphere, in temperature range from 25 to 1100°C heat rate was 10 K·min⁻¹. Coefficients of thermal expansion (CTE) were obtained by dilatometry measurements in air up to 800°C using a Netzsch 402PC dilatometer; heat rate was 1 K·min⁻¹. Impedance spectra were obtained via the two-probe technique with an impedance meter Parstat 2273 in air up to 700°C with the variation of pH₂O from 0.04 to 3.35 kPa. Cation deficiency in Ca and Mg sublattice under the substitution of MgO with ZnO up to 1/6 was observed using Rietveld refinement of the crystal structure. Melting point was found to decrease with x changing from 0 to 4 in Ca₅Mg₄₋ₓMₓ(VO₄)₆ where M is Zn and Ni. It was observed that electrical conductivity does not depend on air humidity. The reported study was funded by the RFBR Grant No. 17–03–01280. Sample attestation was carried out in the Shared Access Centers at the IHTE UB RAS.

Keywords: garnet structure, electrical conductivity, thermal expansion, thermal properties

Procedia PDF Downloads 133
174 Validation of the Female Sexual Function Index and the Female Sexual Distress Scale-Desire/Arousal/Orgasm in Chinese Women

Authors: Lan Luo, Jingjing Huang, Huafang Li

Abstract:

Introduction: Distressing low sexual desire is common in China, while the lack of reliable and valid instruments to evaluate symptoms of hypoactive sexual desire disorder (HSDD) impedes related research and clinical services. Aim: This study aimed to validate the reliability and validity of the Female Sexual Function Index (FSFI) and the Female Sexual Distress Scale-Desire/Arousal/Orgasm (FSDS-DAO) in Chinese female HSDD patients. Methods: We administered FSFI and FSDS-DAO in a convenient sample of Chinese adult women. Participants were diagnosed by a psychiatrist according to the Diagnostic and Statistical Manual of Mental Disorders, 4th Edition, Text Revision (DSM-IV-TR). Results: We had a valid analysis sample of 279 Chinese women, of which 107 were HSDD patients. The Cronbach's α of FSFI and FSDS-DAO were 0.947 and 0.956, respectively, and the intraclass correlation coefficients of which were 0.86 and 0.89, respectively (the interval was 13-15 days). The correlation coefficient between the Revised Adult Attachment Scale (RAAS) and FSFI (or FSDS-DAO) did not exceed 0.4; the area under the receiver operating characteristic (ROC) curve was 0. 83 when combined FSFI-d (the desire domain of FSFI) and FSDS-DAO to diagnose HSDD, which was significantly different from that of using these scales individually. FSFI-d of less than 2.7 (1.2-6) and FSDS-DAO of no less than 15 (0-60) (Sensitivity 65%, Specificity 83%), or FSFI-d of no more than 3.0 (1.2-6) and FSDS-DAO of no less than 14 (0-60) (Sensitivity 74%, Specificity 77%) can be used as cutoff scores in clinical research or outpatient screening. Clinical implications: FSFI (including FSFI-d) and FSDS-DAO are suitable for the screening and evaluation of Chinese female HSDD patients of childbearing age. Strengths and limitations: Strengths include a thorough validation of FSFI and FSDS-DAO and the exploration of the cutoff score combing FSFI-d and FSDS-DAO. Limitations include a small convenience sample and the requirement of being sexually active for HSDD patients. Conclusion: FSFI (including FSFI-d) and FSDS-DAO have good internal consistency, test-retest reliability, construct validity, and criterion validity in Chinese female HSDD patients of childbearing age.

Keywords: sexual desire, sexual distress, hypoactive sexual desire disorder, scale

Procedia PDF Downloads 49
173 Surface Motion of Anisotropic Half Space Containing an Anisotropic Inclusion under SH Wave

Authors: Yuanda Ma, Zhiyong Zhang, Zailin Yang, Guanxixi Jiang

Abstract:

Anisotropy is very common in underground media, such as rock, sand, and soil. Hence, the dynamic response of anisotropy medium under elastic waves is significantly different from the isotropic one. Moreover, underground heterogeneities and structures, such as pipelines, cylinders, or tunnels, are usually made by composite materials, leading to the anisotropy of these heterogeneities and structures. Both the anisotropy of the underground medium and the heterogeneities have an effect on the surface motion of the ground. Aiming at providing theoretical references for earthquake engineering and seismology, the surface motion of anisotropic half-space with a cylindrical anisotropic inclusion embedded under the SH wave is investigated in this work. Considering the anisotropy of the underground medium, the governing equation with three elastic parameters of SH wave propagation is introduced. Then, based on the complex function method and multipolar coordinates system, the governing equation in the complex plane is obtained. With the help of a pair of transformation, the governing equation is transformed into a standard form. By means of the same methods, the governing equation of SH wave propagation in the cylindrical inclusion with another three elastic parameters is normalized as well. Subsequently, the scattering wave in the half-space and the standing wave in the inclusion is deduced. Different incident wave angle and anisotropy are considered to obtain the reflected wave. Then the unknown coefficients in scattering wave and standing wave are solved by utilizing the continuous condition at the boundary of the inclusion. Through truncating finite terms of the scattering wave and standing wave, the equation of boundary conditions can be calculated by programs. After verifying the convergence and the precision of the calculation, the validity of the calculation is verified by degrading the model of the problem as well. Some parameters which influence the surface displacement of the half-space is considered: dimensionless wave number, dimensionless depth of the inclusion, anisotropic parameters, wave number ratio, shear modulus ratio. Finally, surface displacement amplitude of the half space with different parameters is calculated and discussed.

Keywords: anisotropy, complex function method, sh wave, surface displacement amplitude

Procedia PDF Downloads 96
172 How Social Support, Interaction with Clients and Work-Family Conflict Contribute to Mental Well-Being for Employees in the Human Service System

Authors: Uwe C. Fischer

Abstract:

Mental health and well-being for employees working in the human service system are getting more and more important given the increasing rate of absenteeism at work. Besides individual capacities, social and community factors seem to be important in the working setting. Starting from a demand resource framework including the classical demand control aspects, social support systems, specific demands and resources of the client work, and work-family conflict were considered in the present study. We state hypothetically, that these factors have a meaningful association with the mental quality of life of employees working in the field of social, educational and health sectors. 1140 employees, working in human service organizations (education, youth care, nursing etc.) were asked for strains and resources at work (selected scales from Salutogenetic Subjective Work Assessment SALSA and own new scales for client work), work-family conflict, and mental quality of life from the German Short Form Health Survey. Considering the complex influences of the variables, we conducted a multiple hierarchical regression analysis. One third of the whole variance of the mental quality of life can be declared by the different variables of the model. When the variables concerning social influences were included in the hierarchical regression, the influence of work related control resource decreased. Excessive workload, work-family conflict, social support by supervisors, co-workers and other persons outside work, as well as strains and resources associated with client work had significant regression coefficients. Conclusions: Social support systems are crucial in the social, educational and health related service sector, regarding the influence on mental well-being. Especially the work-family conflict focuses on the importance of the work-life balance. Also the specific strains and resources of the client work, measured with new constructed scales, showed great impact on mental health. Therefore occupational health promotion should focus more on the social factors within and outside the working place.

Keywords: client interaction, human service system, mental health, social support, work-family conflict

Procedia PDF Downloads 414
171 Small-Group Case-Based Teaching: Effects on Student Achievement, Critical Thinking, and Attitude toward Chemistry

Authors: Reynante E. Autida, Maria Ana T. Quimbo

Abstract:

The chemistry education curriculum provides an excellent avenue where students learn the principles and concepts in chemistry and at the same time, as a central science, better understand related fields. However, the teaching approach used by teachers affects student learning. Cased-based teaching (CBT) is one of the various forms of inductive method. The teacher starts with specifics then proceeds to the general principles. The students’ role in inductive learning shifts from being passive in the traditional approach to being active in learning. In this paper, the effects of Small-Group Case-Based Teaching (SGCBT) on college chemistry students’ achievement, critical thinking, and attitude toward chemistry including the relationships between each of these variables were determined. A quasi-experimental counterbalanced design with pre-post control group was used to determine the effects of SGCBT on Engineering students of four intact classes (two treatment groups and two control groups) in one of the State Universities in Mindanao. The independent variables are the type of teaching approach (SGCBT versus pure lecture-discussion teaching or PLDT) while the dependent variables are chemistry achievement (exam scores) and scores in critical thinking and chemistry attitude. Both Analysis of Covariance (ANCOVA) and t-tests (within and between groups and gain scores) were used to compare the effects of SGCBT versus PLDT on students’ chemistry achievement, critical thinking, and attitude toward chemistry, while Pearson product-moment correlation coefficients were calculated to determine the relationships between each of the variables. Results show that the use of SGCBT fosters positive attitude toward chemistry and provides some indications as well on improved chemistry achievement of students compared with PLDT. Meanwhile, the effects of PLDT and SGCBT on critical thinking are comparable. Furthermore, correlational analysis and focus group interviews indicate that the use of SGCBT not only supports development of positive attitude towards chemistry but also improves chemistry achievement of students. Implications are provided in view of the recent findings on SGCBT and topics for further research are presented as well.

Keywords: case-based teaching, small-group learning, chemistry cases, chemistry achievement, critical thinking, chemistry attitude

Procedia PDF Downloads 269
170 Teaching the Temperature Dependence of Electrical Resistance of Materials through Arduino Investigation

Authors: Vinit Srivastava, Abhay Singh Thakur, Shivam Dubey, Rahul Vaish, Bharat Singh Rajpurohit

Abstract:

This study examines the problem of students' poor comprehension of the thermal dependence of resistance by investigating this idea using an evidence-based inquiry approach. It suggests a practical exercise to improve secondary school students' comprehension of how materials' resistance to temperature changes. The suggested exercise uses an Arduino and Peltier device to test the resistance of aluminum and graphite at various temperatures. The study attempts to close the knowledge gap between the theoretical and practical facets of the subject, which students frequently find difficult to grasp. With the help of a variety of resistors made of various materials and pencils of varying grades, the Arduino experiment investigates the resistance of a metallic conductor (aluminum) and a semiconductor (graphite) at various temperatures. The purpose of the research is to clarify for students the relationship between temperature and resistance and to emphasize the importance of resistor material choice and measurement methods in obtaining precise and stable resistance values over dynamic temperature variations. The findings show that while the resistance of graphite decreases with temperature, the resistance of metallic conductors rises with temperature. The results also show that as softer lead pencils or pencils of a lower quality are used, the resistance values of the resistors drop. In addition, resistors showed greater stability at lower temperatures when their temperature coefficients of resistance (TCR) were smaller. Overall, the results of this article show that the suggested experiment is a useful and practical method for teaching students about resistance's relationship to temperature. It emphasizes how crucial it is to take into account the resistor material selection and the resistance measurement technique when designing and picking out resistors for various uses. The results of the study are anticipated to guide the creation of more efficient teaching methods to close the gap between science education's theoretical and practical components.

Keywords: electrical resistance, temperature dependence, science education, inquiry-based activity, resistor stability

Procedia PDF Downloads 50
169 Efficiency of Secondary Schools by ICT Intervention in Sylhet Division of Bangladesh

Authors: Azizul Baten, Kamrul Hossain, Abdullah-Al-Zabir

Abstract:

The objective of this study is to develop an appropriate stochastic frontier secondary schools efficiency model by ICT Intervention and to examine the impact of ICT challenges on secondary schools efficiency in the Sylhet division in Bangladesh using stochastic frontier analysis. The Translog stochastic frontier model was found an appropriate than the Cobb-Douglas model in secondary schools efficiency by ICT Intervention. Based on the results of the Cobb-Douglas model, it is found that the coefficient of the number of teachers, the number of students, and teaching ability had a positive effect on increasing the level of efficiency. It indicated that these are related to technical efficiency. In the case of inefficiency effects for both Cobb-Douglas and Translog models, the coefficient of the ICT lab decreased secondary school inefficiency, but the online class in school was found to increase the level of inefficiency. The coefficients of teacher’s preference for ICT tools like multimedia projectors played a contributor role in decreasing the secondary school inefficiency in the Sylhet division of Bangladesh. The interaction effects of the number of teachers and the classrooms, and the number of students and the number of classrooms, the number of students and teaching ability, and the classrooms and teaching ability of the teachers were recorded with the positive values and these have a positive impact on increasing the secondary school efficiency. The overall mean efficiency of urban secondary schools was found at 84.66% for the Translog model, while it was 83.63% for the Cobb-Douglas model. The overall mean efficiency of rural secondary schools was found at 80.98% for the Translog model, while it was 81.24% for the Cobb-Douglas model. So, the urban secondary schools performed better than the rural secondary schools in the Sylhet division. It is observed from the results of the Tobit model that the teacher-student ratio had a positive influence on secondary school efficiency. The teaching experiences of those who have 1 to 5 years and 10 years above, MPO type school, conventional teaching method have had a negative and significant influence on secondary school efficiency. The estimated value of σ-square (0.0625) was different from Zero, indicating a good fit. The value of γ (0.9872) was recorded as positive and it can be interpreted as follows: 98.72 percent of random variation around in secondary school outcomes due to inefficiency.

Keywords: efficiency, secondary schools, ICT, stochastic frontier analysis

Procedia PDF Downloads 117
168 Detection of Phoneme [S] Mispronounciation for Sigmatism Diagnosis in Adults

Authors: Michal Krecichwost, Zauzanna Miodonska, Pawel Badura

Abstract:

The diagnosis of sigmatism is mostly based on the observation of articulatory organs. It is, however, not always possible to precisely observe the vocal apparatus, in particular in the oral cavity of the patient. Speech processing can allow to objectify the therapy and simplify the verification of its progress. In the described study the methodology for classification of incorrectly pronounced phoneme [s] is proposed. The recordings come from adults. They were registered with the speech recorder at the sampling rate of 44.1 kHz and the resolution of 16 bit. The database of pathological and normative speech has been collected for the study including reference assessments provided by the speech therapy experts. Ten adult subjects were asked to simulate a certain type of stigmatism under the speech therapy expert supervision. In the recordings, the analyzed phone [s] was surrounded by vowels, viz: ASA, ESE, ISI, SPA, USU, YSY. Thirteen MFCC (mel-frequency cepstral coefficients) and RMS (root mean square) values are calculated within each frame being a part of the analyzed phoneme. Additionally, 3 fricative formants along with corresponding amplitudes are determined for the entire segment. In order to aggregate the information within the segment, the average value of each MFCC coefficient is calculated. All features of other types are aggregated by means of their 75th percentile. The proposed method of features aggregation reduces the size of the feature vector used in the classification. Binary SVM (support vector machine) classifier is employed at the phoneme recognition stage. The first group consists of pathological phones, while the other of the normative ones. The proposed feature vector yields classification sensitivity and specificity measures above 90% level in case of individual logo phones. The employment of a fricative formants-based information improves the sole-MFCC classification results average of 5 percentage points. The study shows that the employment of specific parameters for the selected phones improves the efficiency of pathology detection referred to the traditional methods of speech signal parameterization.

Keywords: computer-aided pronunciation evaluation, sibilants, sigmatism diagnosis, speech processing

Procedia PDF Downloads 256
167 Kinematic Analysis of the Calf Raise Test Using a Mobile iOS Application: Validation of the Calf Raise Application

Authors: Ma. Roxanne Fernandez, Josie Athens, Balsalobre-Fernandez, Masayoshi Kubo, Kim Hébert-Losier

Abstract:

Objectives: The calf raise test (CRT) is used in rehabilitation and sports medicine to evaluate calf muscle function. For testing, individuals stand on one leg and go up on their toes and back down to volitional fatigue. The newly developed Calf Raise application (CRapp) for iOS uses computer-vision algorithms enabling objective measurement of CRT outcomes. We aimed to validate the CRapp by examining its concurrent validity and agreement levels against laboratory-based equipment and establishing its intra- and inter-rater reliability. Methods: CRT outcomes (i.e., repetitions, positive work, total height, peak height, fatigue index, and peak power) were assessed in thirteen healthy individuals (6 males, 7 females) on three occasions and both legs using the CRapp, 3D motion capture, and force plate technologies simultaneously. Data were extracted from two markers: one placed immediately below the lateral malleolus and another on the heel. Concurrent validity and agreement measures were determined using intraclass correlation coefficients (ICC₃,ₖ), typical errors expressed as coefficient of variations (CV), and Bland-Altman methods to assess biases and precision. Reliability was assessed using ICC3,1 and CV values. Results: Validity of CRapp outcomes was good to excellent across measures for both markers (mean ICC ≥0.878), with precision plots showing good agreement and precision. CV ranged from 0% (repetitions) to 33.3% (fatigue index) and were, on average better for the lateral malleolus marker. Additionally, inter- and intra-rater reliability were excellent (mean ICC ≥0.949, CV ≤5.6%). Conclusion: These results confirm the CRapp is valid and reliable within and between users for measuring CRT outcomes in healthy adults. The CRapp provides a tool to objectivise CRT outcomes in research and practice, aligning with recent advances in mobile technologies and their increased use in healthcare.

Keywords: calf raise test, mobile application, validity, reliability

Procedia PDF Downloads 141
166 Financing the Welfare State in the United States: The Recent American Economic and Ideological Challenges

Authors: Rafat Fazeli, Reza Fazeli

Abstract:

This paper focuses on the study of the welfare state and social wage in the leading liberal economy of the United States. The welfare state acquired a broad acceptance as a major socioeconomic achievement of the liberal democracy in the Western industrialized countries during the postwar boom period. The modern and modified vision of capitalist democracy offered, on the one hand, the possibility of high growth rate and, on the other hand, the possibility of continued progression of a comprehensive system of social support for a wider population. The economic crises of the 1970s, provided the ground for a great shift in economic policy and ideology in several Western countries, most notably the United States and the United Kingdom (and to a lesser extent Canada under Prime Minister Brian Mulroney). In the 1980s, the free market oriented reforms undertaken under Reagan and Thatcher greatly affected the economic outlook not only of the United States and the United Kingdom, but of the whole Western world. The movement which was behind this shift in policy is often called neo-conservatism. The neoconservatives blamed the transfer programs for the decline in economic performance during the 1970s and argued that cuts in spending were required to go back to the golden age of full employment. The agenda for both Reagan and Thatcher administrations was rolling back the welfare state, and their budgets included a wide range of cuts for social programs. The question is how successful were Reagan and Thatcher’s efforts to achieve retrenchment? The paper involves an empirical study concerning the distributive role of the welfare state in the two countries. Other studies have often concentrated on the redistributive effect of fiscal policy on different income brackets. This study examines the net benefit/ burden position of the working population with respect to state expenditures and taxes in the postwar period. This measurement will enable us to find out whether the working population has received a net gain (or net social wage). This study will discuss how the expansion of social expenditures and the trend of the ‘net social wage’ can be linked to distinct forms of economic and social organizations. This study provides an empirical foundation for analyzing the growing significance of ‘social wage’ or the collectivization of consumption and the share of social or collective consumption in total consumption of the working population in the recent decades. The paper addresses three other major questions. The first question is whether the expansion of social expenditures has posed any drag on capital accumulation and economic growth. The findings of this study provide an analytical foundation to evaluate the neoconservative claim that the welfare state is itself the source of economic stagnation that leads to the crisis of the welfare state. The second question is whether the increasing ideological challenges from the right and the competitive pressures of globalization have led to retrenchment of the American welfare states in the recent decades. The third question is how social policies have performed in the presence of the rising inequalities in the recent decades.

Keywords: the welfare state, social wage, The United States, limits to growth

Procedia PDF Downloads 185
165 Lifting Body Concepts for Unmanned Fixed-Wing Transport Aircrafts

Authors: Anand R. Nair, Markus Trenker

Abstract:

Lifting body concepts were conceived as early as 1917 and patented by Roy Scroggs. It was an idea of using the fuselage as a lift producing body with no or small wings. Many of these designs were developed and even flight tested between 1920’s to 1970’s, but it was not pursued further for commercial flight as at lower airspeeds, such a configuration was incapable to produce sufficient lift for the entire aircraft. The concept presented in this contribution is combining the lifting body design along with a fixed wing to maximise the lift produced by the aircraft. Conventional aircraft fuselages are designed to be aerodynamically efficient, which is to minimise the drag; however, these fuselages produce very minimal or negligible lift. For the design of an unmanned fixed wing transport aircraft, many of the restrictions which are present for commercial aircraft in terms of fuselage design can be excluded, such as windows for the passengers/pilots, cabin-environment systems, emergency exits, and pressurization systems. This gives new flexibility to design fuselages which are unconventionally shaped to contribute to the lift of the aircraft. The two lifting body concepts presented in this contribution are targeting different applications: For a fast cargo delivery drone, the fuselage is based on a scaled airfoil shape with a cargo capacity of 500 kg for euro pallets. The aircraft has a span of 14 m and reaches 1500 km at a cruising speed of 90 m/s. The aircraft could also easily be adapted to accommodate pilot and passengers with modifications to the internal structures, but pressurization is not included as the service ceiling envisioned for this type of aircraft is limited to 10,000 ft. The next concept to be investigated is called a multi-purpose drone, which incorporates a different type of lifting body and is a much more versatile aircraft as it will have a VTOL capability. The aircraft will have a wingspan of approximately 6 m and flight speeds of 60 m/s within the same service ceiling as the fast cargo delivery drone. The multi-purpose drone can be easily adapted for various applications such as firefighting, agricultural purposes, surveillance, and even passenger transport. Lifting body designs are not a new concept, but their effectiveness in terms of cargo transportation has not been widely investigated. Due to their enhanced lift producing capability, lifting body designs enable the reduction of the wing area and the overall weight of the aircraft. This will, in turn, reduce the thrust requirement and ultimately the fuel consumption. The various designs proposed in this contribution will be based on the general aviation category of aircrafts and will be focussed on unmanned methods of operation. These unmanned fixed-wing transport drones will feature appropriate cargo loading/unloading concepts which can accommodate large size cargo for efficient time management and ease of operation. The various designs will be compared in performance to their conventional counterpart to understand their benefits/shortcomings in terms of design, performance, complexity, and ease of operation. The majority of the performance analysis will be carried out using industry relevant standards in computational fluid dynamics software packages.

Keywords: lifting body concept, computational fluid dynamics, unmanned fixed-wing aircraft, cargo drone

Procedia PDF Downloads 192
164 Two-Level Separation of High Air Conditioner Consumers and Demand Response Potential Estimation Based on Set Point Change

Authors: Mehdi Naserian, Mohammad Jooshaki, Mahmud Fotuhi-Firuzabad, Mohammad Hossein Mohammadi Sanjani, Ashknaz Oraee

Abstract:

In recent years, the development of communication infrastructure and smart meters have facilitated the utilization of demand-side resources which can enhance stability and economic efficiency of power systems. Direct load control programs can play an important role in the utilization of demand-side resources in the residential sector. However, investments required for installing control equipment can be a limiting factor in the development of such demand response programs. Thus, selection of consumers with higher potentials is crucial to the success of a direct load control program. Heating, ventilation, and air conditioning (HVAC) systems, which due to the heat capacity of buildings feature relatively high flexibility, make up a major part of household consumption. Considering that the consumption of HVAC systems depends highly on the ambient temperature and bearing in mind the high investments required for control systems enabling direct load control demand response programs, in this paper, a recent solution is presented to uncover consumers with high air conditioner demand among large number of consumers and to measure the demand response potential of such consumers. This can pave the way for estimating the investments needed for the implementation of direct load control programs for residential HVAC systems and for estimating the demand response potentials in a distribution system. In doing so, we first cluster consumers into several groups based on the correlation coefficients between hourly consumption data and hourly temperature data using K-means algorithm. Then, by applying a recent algorithm to the hourly consumption and temperature data, consumers with high air conditioner consumption are identified. Finally, demand response potential of such consumers is estimated based on the equivalent desired temperature setpoint changes.

Keywords: communication infrastructure, smart meters, power systems, HVAC system, residential HVAC systems

Procedia PDF Downloads 35
163 Application of Transportation Models for Analysing Future Intercity and Intracity Travel Patterns in Kuwait

Authors: Srikanth Pandurangi, Basheer Mohammed, Nezar Al Sayegh

Abstract:

In order to meet the increasing demand for housing care for Kuwaiti citizens, the government authorities in Kuwait are undertaking a series of projects in the form of new large cities, outside the current urban area. Al Mutlaa City located to the north-west of the Kuwait Metropolitan Area is one such project out of the 15 planned new cities. The city accommodates a wide variety of residential developments, employment opportunities, commercial, recreational, health care and institutional uses. This paper examines the application of comprehensive transportation demand modeling works undertaken in VISUM platform to understand the future intracity and intercity travel distribution patterns in Kuwait. The scope of models developed varied in levels of detail: strategic model update, sub-area models representing future demand of Al Mutlaa City, sub-area models built to estimate the demand in the residential neighborhoods of the city. This paper aims at offering model update framework that facilitates easy integration between sub-area models and strategic national models for unified traffic forecasts. This paper presents the transportation demand modeling results utilized in informing the planning of multi-modal transportation system for Al Mutlaa City. This paper also presents the household survey data collection efforts undertaken using GPS devices (first time in Kuwait) and notebook computer based digital survey forms for interviewing representative sample of citizens and residents. The survey results formed the basis of estimating trip generation rates and trip distribution coefficients used in the strategic base year model calibration and validation process.

Keywords: innovative methods in transportation data collection, integrated public transportation system, traffic forecasts, transportation modeling, travel behavior

Procedia PDF Downloads 190
162 An Association Model to Correlate the Experimentally Determined Mixture Solubilities of Methyl 10-Undecenoate with Methyl Ricinoleate in Supercritical Carbon Dioxide

Authors: V. Mani Rathnam, Giridhar Madras

Abstract:

Fossil fuels are depleting rapidly as the demand for energy, and its allied chemicals are continuously increasing in the modern world. Therefore, sustainable renewable energy sources based on non-edible oils are being explored as a viable option as they do not compete with the food commodities. Oils such as castor oil are rich in fatty acids and thus can be used for the synthesis of biodiesel, bio-lubricants, and many other fine industrial chemicals. There are several processes available for the synthesis of different chemicals obtained from the castor oil. One such process is the transesterification of castor oil, which results in a mixture of fatty acid methyl esters. The main products in the above reaction are methyl ricinoleate and methyl 10-undecenoate. To separate these compounds, supercritical carbon dioxide (SCCO₂) was used as a green solvent. SCCO₂ was chosen as a solvent due to its easy availability, non-toxic, non-flammable, and low cost. In order to design any separation process, the preliminary requirement is the solubility or phase equilibrium data. Therefore, the solubility of a mixture of methyl ricinoleate with methyl 10-undecenoate in SCCO₂ was determined in the present study. The temperature and pressure range selected for the investigation were T = 313 K to 333 K and P = 10 MPa to 18 MPa. It was observed that the solubility (mol·mol⁻¹) of methyl 10-undecenoate varied from 2.44 x 10⁻³ to 8.42 x 10⁻³ whereas it varied from 0.203 x 10⁻³ to 6.28 x 10⁻³ for methyl ricinoleate within the chosen operating conditions. These solubilities followed a retrograde behavior (characterized by the decrease in the solubility values with the increase in temperature) throughout the range of investigated operating conditions. An association theory model, coupled with regular solution theory for activity coefficients, was developed in the present study. The deviation from the experimental data using this model can be quantified using the average absolute relative deviation (AARD). The AARD% for the present compounds is 4.69 and 8.08 for methyl 10-undecenoate and methyl ricinoleate, respectively in a mixture of methyl ricinoleate and methyl 10-undecenoate. The maximum solubility enhancement of 32% was observed for the methyl ricinoleate in a mixture of methyl ricinoleate and methyl 10-undecenoate. The highest selectivity of SCCO₂ was observed to be 12 for methyl 10-undecenoate in a mixture of methyl ricinoleate and methyl 10-undecenoate.

Keywords: association theory, liquid mixtures, solubilities, supercritical carbon dioxide

Procedia PDF Downloads 108